身安不如心安,屋宽不如心宽 。

社区日报 第101期 (2017-11-15)

1、Elastic stack 6.0正式亮相
https://www.elastic.co/blog/el ... eased
2、vts数据处理搜索引擎选型 
http://t.cn/RjfQx01 
3、如何监控、调优golang应用程序且看下文分晓 
http://t.cn/RlA4biz 
4、elasticsearch性能分析123 
http://t.cn/RjfEjN3 
5、只等你来 | Elastic Meetup 广州交流会 
https://elasticsearch.cn/article/364 
 
编辑:wt 
归档:https://elasticsearch.cn/article/377
订阅:https://tinyletter.com/elastic-daily
 
继续阅读 »
1、Elastic stack 6.0正式亮相
https://www.elastic.co/blog/el ... eased
2、vts数据处理搜索引擎选型 
http://t.cn/RjfQx01 
3、如何监控、调优golang应用程序且看下文分晓 
http://t.cn/RlA4biz 
4、elasticsearch性能分析123 
http://t.cn/RjfEjN3 
5、只等你来 | Elastic Meetup 广州交流会 
https://elasticsearch.cn/article/364 
 
编辑:wt 
归档:https://elasticsearch.cn/article/377
订阅:https://tinyletter.com/elastic-daily
  收起阅读 »

spark elasticsearch 异常: PowerSet: Too many elements to create a power set 40


org.elasticsearch.hadoop.EsHadoopIllegalArgumentException: Too many elements to create a power set 54
at org.elasticsearch.hadoop.util.Assert.isTrue(Assert.java:50)
at org.elasticsearch.hadoop.rest.ShardSorter$PowerSet.(ShardSorter.java:218)
at org.elasticsearch.hadoop.rest.ShardSorter.powerList(ShardSorter.java:202)
at org.elasticsearch.hadoop.rest.ShardSorter.checkCombo(ShardSorter.java:89)
at org.elasticsearch.hadoop.rest.ShardSorter.find(ShardSorter.java:85)
at org.elasticsearch.hadoop.rest.RestRepository.doGetReadTargetShards(RestRepository.java:352)
at org.elasticsearch.hadoop.rest.RestRepository.getReadTargetShards(RestRepository.java:295)
at org.elasticsearch.hadoop.rest.RestService.findPartitions(RestService.java:253)
at org.elasticsearch.spark.rdd.AbstractEsRDD.esPartitions$lzycompute(AbstractEsRDD.scala:61)
at org.elasticsearch.spark.rdd.AbstractEsRDD.esPartitions(AbstractEsRDD.scala:60)
at org.elasticsearch.spark.rdd.AbstractEsRDD.getPartitions(AbstractEsRDD.scala:27)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)
at scala.Option.getOrElse(Option.scala:120)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:237)
at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)Users with the same issue
 
解释:
Read failure when index/alias spread among 32 or more nodes
 
https://github.com/elastic/ela ... s/737
继续阅读 »

org.elasticsearch.hadoop.EsHadoopIllegalArgumentException: Too many elements to create a power set 54
at org.elasticsearch.hadoop.util.Assert.isTrue(Assert.java:50)
at org.elasticsearch.hadoop.rest.ShardSorter$PowerSet.(ShardSorter.java:218)
at org.elasticsearch.hadoop.rest.ShardSorter.powerList(ShardSorter.java:202)
at org.elasticsearch.hadoop.rest.ShardSorter.checkCombo(ShardSorter.java:89)
at org.elasticsearch.hadoop.rest.ShardSorter.find(ShardSorter.java:85)
at org.elasticsearch.hadoop.rest.RestRepository.doGetReadTargetShards(RestRepository.java:352)
at org.elasticsearch.hadoop.rest.RestRepository.getReadTargetShards(RestRepository.java:295)
at org.elasticsearch.hadoop.rest.RestService.findPartitions(RestService.java:253)
at org.elasticsearch.spark.rdd.AbstractEsRDD.esPartitions$lzycompute(AbstractEsRDD.scala:61)
at org.elasticsearch.spark.rdd.AbstractEsRDD.esPartitions(AbstractEsRDD.scala:60)
at org.elasticsearch.spark.rdd.AbstractEsRDD.getPartitions(AbstractEsRDD.scala:27)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)
at scala.Option.getOrElse(Option.scala:120)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:237)
at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)Users with the same issue
 
解释:
Read failure when index/alias spread among 32 or more nodes
 
https://github.com/elastic/ela ... s/737 收起阅读 »

写在 社区 日报100期时 —— 相信社区的力量

community

我一直有看湾区日报的习惯,后来看到 Golang 中国社区的 astaxie 也在做 GoCN 每日新闻 的事情。日报对社区来讲是一件持续输出的事情,对个人来讲是一件持续输入的事情,于是我决定在 Elastic 社区也做这么一件事情,在2017年7月30日我发布了 Elastic日报 第1期

当天和 medcl 聊过后,他建议发动社区的力量来做,这样才能保证日报做好做久。于是我们开始在社区里面招募日报编辑,很快便有很多同学响应,接着 Elastic日报编辑部成立。到今天,我们一共有8位社区编辑,其中7位负责每周固定一天的日报,另一位负责审稿和公众号文章发布。他们分别是:

  • 江水
  • 金桥
  • bsll
  • 至尊宝
  • 叮咚光军
  • laoyang360
  • cyberdak
  • 陶文

感谢社区编辑们的付出,我们一同做了一件了不起的事情——持续100天的知识输出。如果有同学把这100天的日报内容都看完吃透,那它的Elastic 技术水准肯定提升了不止1个档次。

现在想来,如果是我一个人做这件事情,恐怕日报不会超过30期。个人的力量是有限的,而社区的力量是无限的。每天看到编辑们精挑细选的文章,我都会诧异 Elastic 相关的优秀文章可真是多啊!

相信社区的力量,让我们期待Elastic日报200期、300期甚至1000期的到来!

继续阅读 »

community

我一直有看湾区日报的习惯,后来看到 Golang 中国社区的 astaxie 也在做 GoCN 每日新闻 的事情。日报对社区来讲是一件持续输出的事情,对个人来讲是一件持续输入的事情,于是我决定在 Elastic 社区也做这么一件事情,在2017年7月30日我发布了 Elastic日报 第1期

当天和 medcl 聊过后,他建议发动社区的力量来做,这样才能保证日报做好做久。于是我们开始在社区里面招募日报编辑,很快便有很多同学响应,接着 Elastic日报编辑部成立。到今天,我们一共有8位社区编辑,其中7位负责每周固定一天的日报,另一位负责审稿和公众号文章发布。他们分别是:

  • 江水
  • 金桥
  • bsll
  • 至尊宝
  • 叮咚光军
  • laoyang360
  • cyberdak
  • 陶文

感谢社区编辑们的付出,我们一同做了一件了不起的事情——持续100天的知识输出。如果有同学把这100天的日报内容都看完吃透,那它的Elastic 技术水准肯定提升了不止1个档次。

现在想来,如果是我一个人做这件事情,恐怕日报不会超过30期。个人的力量是有限的,而社区的力量是无限的。每天看到编辑们精挑细选的文章,我都会诧异 Elastic 相关的优秀文章可真是多啊!

相信社区的力量,让我们期待Elastic日报200期、300期甚至1000期的到来!

收起阅读 »

社区日报 第100期 (2017-11-14)

1.使用Logstash将email索引至Elasticsearch。
http://t.cn/RjyNLge 
2.Elasticsearch选主流程详细分析。
http://t.cn/RjyNPLT 
3.手把手教你如何使用ES提高WordPress的搜索速度。
http://t.cn/RjbD1QK 
4.只等你来 | Elastic Meetup 广州交流会
https://elasticsearch.cn/article/364 


编辑:叮咚光军
归档:https://elasticsearch.cn/article/374 
订阅:https://tinyletter.com/elastic-daily
继续阅读 »
1.使用Logstash将email索引至Elasticsearch。
http://t.cn/RjyNLge 
2.Elasticsearch选主流程详细分析。
http://t.cn/RjyNPLT 
3.手把手教你如何使用ES提高WordPress的搜索速度。
http://t.cn/RjbD1QK 
4.只等你来 | Elastic Meetup 广州交流会
https://elasticsearch.cn/article/364 


编辑:叮咚光军
归档:https://elasticsearch.cn/article/374 
订阅:https://tinyletter.com/elastic-daily 收起阅读 »

从es2.3到5.6的迁移实践

config/elasticsearch.yml

# ======================== Elasticsearch Configuration =========================
#
# NOTE: Elasticsearch comes with reasonable defaults for most settings.
#       Before you set out to tweak and tune the configuration, make sure you
#       understand what are you trying to accomplish and the consequences.
#
# The primary way of configuring a node is via this file. This template lists
# the most important settings you may want to configure for a production cluster.
#
# Please consult the documentation for further information on configuration options:
# https://www.elastic.co/guide/e ... .html
#
# ---------------------------------- Cluster -----------------------------------
#
# Use a descriptive name for your cluster:
#
#cluster.name: my-application
cluster.name: es5_dev
#
# ------------------------------------ Node ------------------------------------
#
# Use a descriptive name for the node:
#
#node.name: node-1
node.name: es5-node03
#
# Add custom attributes to the node:
#
#node.attr.rack: r1
#
# ----------------------------------- Paths ------------------------------------
#
# Path to directory where to store the data (separate multiple locations by comma):
#
#path.data: /path/to/data
#
# Path to log files:
#
#path.logs: /path/to/logs
#
# ----------------------------------- Memory -----------------------------------
#
# Lock the memory on startup:
#
#bootstrap.memory_lock: true
bootstrap.memory_lock: true
#
# Make sure that the heap size is set to about half the memory available
# on the system and that the owner of the process is allowed to use this
# limit.
#
# Elasticsearch performs poorly when the system is swapping the memory.
#
# ---------------------------------- Network -----------------------------------
#
# Set the bind address to a specific IP (IPv4 or IPv6):
#
#network.host: 192.168.0.1
network.host: ["127.0.0.1","10.204.12.33"]
http.port: 9201
transport.tcp.port: 9301
#http.host: 127.0.0.1
#http.enabled: false
#
# Set a custom port for HTTP:
#
#http.port: 9200
#
# For more information, consult the network module documentation.
#
# --------------------------------- Discovery ----------------------------------
#
# Pass an initial list of hosts to perform discovery when new node is started:
# The default list of hosts is ["127.0.0.1", "[::1]"]
#
#discovery.zen.ping.unicast.hosts: ["host1", "host2"]
node.master: true
node.data: true
discovery.zen.minimum_master_nodes: 1
discovery.zen.ping.unicast.hosts:
   - 10.204.12.31:9301
   - 10.204.12.32:9301
   - 10.204.12.33:9301
#
# Prevent the "split brain" by configuring the majority of nodes (total number of master-eligible nodes / 2 + 1):
#
#discovery.zen.minimum_master_nodes: 3
#
# For more information, consult the zen discovery module documentation.
#
# ---------------------------------- Gateway -----------------------------------
#
# Block initial recovery after a full cluster restart until N nodes are started:
#
#gateway.recover_after_nodes: 3
#
# For more information, consult the gateway module documentation.
#
# ---------------------------------- Various -----------------------------------
#
# Require explicit names when deleting indices:
#
#action.destructive_requires_name: true
indices.requests.cache.size: 5%
config/jvm.options

## JVM configuration
################################################################
## IMPORTANT: JVM heap size
################################################################
##
## You should always set the min and max JVM heap
## size to the same value. For example, to set
## the heap to 4 GB, set:
##
## -Xms4g
## -Xmx4g
##
## See https://www.elastic.co/guide/e ... .html
## for more information
##
################################################################
# Xms represents the initial size of total heap space
# Xmx represents the maximum size of total heap space
-Xms2g
-Xmx2g
################################################################
## Expert settings
################################################################
##
## All settings below this section are considered
## expert settings. Don't tamper with them unless
## you understand what you are doing
##
################################################################
## GC configuration
-XX:+UseConcMarkSweepGC
-XX:CMSInitiatingOccupancyFraction=75
-XX:+UseCMSInitiatingOccupancyOnly
## optimizations
# pre-touch memory pages used by the JVM during initialization
-XX:+AlwaysPreTouch
## basic
# force the server VM (remove on 32-bit client JVMs)
-server
# explicitly set the stack size (reduce to 320k on 32-bit client JVMs)
-Xss1m
# set to headless, just in case
-Djava.awt.headless=true
# ensure UTF-8 encoding by default (e.g. filenames)
-Dfile.encoding=UTF-8
# use our provided JNA always versus the system one
-Djna.nosys=true
# use old-style file permissions on JDK9
-Djdk.io.permissionsUseCanonicalPath=true
# flags to configure Netty
-Dio.netty.noUnsafe=true
-Dio.netty.noKeySetOptimization=true
-Dio.netty.recycler.maxCapacityPerThread=0
# log4j 2
-Dlog4j.shutdownHookEnabled=false
-Dlog4j2.disable.jmx=true
-Dlog4j.skipJansi=true
## heap dumps
# generate a heap dump when an allocation from the Java heap fails
# heap dumps are created in the working directory of the JVM
-XX:+HeapDumpOnOutOfMemoryError
# specify an alternative path for heap dumps
# ensure the directory exists and has sufficient space
#-XX:HeapDumpPath=${heap.dump.path}
## GC logging
#-XX:+PrintGCDetails
#-XX:+PrintGCTimeStamps
#-XX:+PrintGCDateStamps
#-XX:+PrintClassHistogram
#-XX:+PrintTenuringDistribution
#-XX:+PrintGCApplicationStoppedTime
# log GC status to a file with time stamps
# ensure the directory exists
#-Xloggc:${loggc}
# By default, the GC log file will not rotate.
# By uncommenting the lines below, the GC log file
# will be rotated every 128MB at most 32 times.
#-XX:+UseGCLogFileRotation
#-XX:NumberOfGCLogFiles=32
#-XX:GCLogFileSize=128M
# Elasticsearch 5.0.0 will throw an exception on unquoted field names in JSON.
# If documents were already indexed with unquoted fields in a previous version
# of Elasticsearch, some operations may throw errors.
#
# WARNING: This option will be removed in Elasticsearch 6.0.0 and is provided
# only for migration purposes.
#-Delasticsearch.json.allow_unquoted_field_names=true



安装ik分词器


bin/elasticsearch-plugin install https://github.com/medcl/elast ... 1.zip




./bin/elasticsearch-plugin install https://github.com/medcl/elast ... 3.zip



配置ik远程扩展词典用于热词更新 elasticsearch-5.6.3/config/analysis-ik/IKAnalyzer.cfg.xml

<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE properties SYSTEM "http://java.sun.com/dtd/properties.dtd&quot;&gt;
<properties>
       <comment>IK Analyzer 扩展配置</comment>
       <!--用户可以在这里配置自己的扩展字典 -->
       <entry key="ext_dict"></entry>
        <!--用户可以在这里配置自己的扩展停止词字典-->
       <entry key="ext_stopwords"></entry>
       <!--用户可以在这里配置远程扩展字典 -->
       <entry key="remote_ext_dict">http://distribute.search.leju. ... gt%3B
       <!--用户可以在这里配置远程扩展停止词字典-->
       <!-- <entry key="remote_ext_stopwords">words_location</entry> -->
</properties>
安装拼音分词器

cd elasticsearch-5.5.1/plugins
wget https://github.com/medcl/elast ... 5.5.1
unzip v5.5.1




打包部署其他节点时,先清理data目录




集群监控可以利用head的chrome插件




数据迁移

迁移工具是自己写的elasticbak,目前更新了5.6.3驱动。github链接:https://github.com/jiashiwen/elasticbak

数据备份

java -jar elasticbak-2.3.3.jar \
--exp \
--cluster lejuesdev \
--host 10.204.12.31 \
--filesize 1000 \
--backupdir ./esbackupset \
--backupindexes "*" \
--threads 4
由于版本field的变化需要手工重建索引,这里举个例子,主要是2.x版本的string需要改为text。2.x版本我们通过index参数指定该字段是否被索引("index": "no")以及是否通过分词器分词("index": "not_analyzed")。在5.X版本里index只用来制定是否创建索引,如果需要整个字段不过分词器创建索引,需要通过keyword字段完成。  

 

curl -XPUT "http://10.204.12.31:9201/house_geo&quot; -H 'Content-Type: application/json' -d'
{
 "mappings": {
   "house": {
     "dynamic": "strict",
     "_all": {
       "enabled": false
     },
     "properties": {
       "_category": {
         "type": "keyword",
                 "store": true
       },
       "_content": {
         "type": "text",
         "store": true,
         "analyzer": "ik_max_word",
         "search_analyzer": "ik_smart"
       },
       "_deleted": {
         "type": "boolean",
         "store": true
       },
       "_doccreatetime": {
         "type": "date",
         "store": true,
         "format": "strict_date_optional_time||epoch_millis||yyyy/MM/dd HH:mm:ss||yyyy/MM/dd"
       },
       "_docupdatetime": {
         "type": "date",
         "store": true,
         "format": "strict_date_optional_time||epoch_millis||yyyy/MM/dd HH:mm:ss||yyyy/MM/dd"
       },
       "_flags": {
         "type": "text",
         "store": true,
         "analyzer": "whitespace"
       },
       "_hits": {
         "type": "text"
       },
       "_location": {
         "type": "geo_point"
       },
       "_multi": {
         "properties": {
           "_location": {
             "type": "geo_point"
           }
         }
       },
       "_origin": {
         "type": "object",
         "enabled": false
       },
       "_scope": {
         "type": "keyword",
         "store": true
       },
       "_tags": {
         "type": "text",
         "boost": 10,
         "store": true,
         "term_vector": "with_positions_offsets",
         "analyzer": "ik_max_word",
         "search_analyzer": "ik_smart"
       },
       "_title": {
         "type": "text",
         "store": true,
         "analyzer": "ik_max_word",
         "search_analyzer": "ik_smart"
       },
       "_uniqid": {
         "type": "keyword",
         "store": true
       },
       "_uniqsign": {
         "type": "keyword",
         "store": true
       },
       "_url": {
         "type": "text",
         "index": false,
         "store": true
       },
       "location": {
         "type": "geo_point"
       }
     }
   }
 },
 "settings": {
   "index": {
     "number_of_shards": "3",
     "requests": {
       "cache": {
         "enable": "true"
       }
     },
     "analysis": {
       "filter": {
         "my_synonym": {
           "type": "synonym",
           "synonyms_path": "analysis-ik/custom/synonym.dic"
         }
       },
       "analyzer": {
         "searchanalyzer": {
           "filter": "my_synonym",
           "type": "custom",
           "tokenizer": "ik_smart"
         },
         "indexanalyzer": {
           "filter": "my_synonym",
           "type": "custom",
           "tokenizer": "ik_max_word"
         }
       }
     },
     "number_of_replicas": "1"
   }
 }
}'
利用新版elasticbak导入索引数据


java -jar elasticbak-5.6.3.jar \
--imp \
--cluster es5_dev \
--host 10.204.12.31 \
--port 9301 \
--restoreindex house_geo \
--restoretype dataonly \
--backupset esbackupset/house_geo \
--threads 4
 
继续阅读 »
config/elasticsearch.yml

# ======================== Elasticsearch Configuration =========================
#
# NOTE: Elasticsearch comes with reasonable defaults for most settings.
#       Before you set out to tweak and tune the configuration, make sure you
#       understand what are you trying to accomplish and the consequences.
#
# The primary way of configuring a node is via this file. This template lists
# the most important settings you may want to configure for a production cluster.
#
# Please consult the documentation for further information on configuration options:
# https://www.elastic.co/guide/e ... .html
#
# ---------------------------------- Cluster -----------------------------------
#
# Use a descriptive name for your cluster:
#
#cluster.name: my-application
cluster.name: es5_dev
#
# ------------------------------------ Node ------------------------------------
#
# Use a descriptive name for the node:
#
#node.name: node-1
node.name: es5-node03
#
# Add custom attributes to the node:
#
#node.attr.rack: r1
#
# ----------------------------------- Paths ------------------------------------
#
# Path to directory where to store the data (separate multiple locations by comma):
#
#path.data: /path/to/data
#
# Path to log files:
#
#path.logs: /path/to/logs
#
# ----------------------------------- Memory -----------------------------------
#
# Lock the memory on startup:
#
#bootstrap.memory_lock: true
bootstrap.memory_lock: true
#
# Make sure that the heap size is set to about half the memory available
# on the system and that the owner of the process is allowed to use this
# limit.
#
# Elasticsearch performs poorly when the system is swapping the memory.
#
# ---------------------------------- Network -----------------------------------
#
# Set the bind address to a specific IP (IPv4 or IPv6):
#
#network.host: 192.168.0.1
network.host: ["127.0.0.1","10.204.12.33"]
http.port: 9201
transport.tcp.port: 9301
#http.host: 127.0.0.1
#http.enabled: false
#
# Set a custom port for HTTP:
#
#http.port: 9200
#
# For more information, consult the network module documentation.
#
# --------------------------------- Discovery ----------------------------------
#
# Pass an initial list of hosts to perform discovery when new node is started:
# The default list of hosts is ["127.0.0.1", "[::1]"]
#
#discovery.zen.ping.unicast.hosts: ["host1", "host2"]
node.master: true
node.data: true
discovery.zen.minimum_master_nodes: 1
discovery.zen.ping.unicast.hosts:
   - 10.204.12.31:9301
   - 10.204.12.32:9301
   - 10.204.12.33:9301
#
# Prevent the "split brain" by configuring the majority of nodes (total number of master-eligible nodes / 2 + 1):
#
#discovery.zen.minimum_master_nodes: 3
#
# For more information, consult the zen discovery module documentation.
#
# ---------------------------------- Gateway -----------------------------------
#
# Block initial recovery after a full cluster restart until N nodes are started:
#
#gateway.recover_after_nodes: 3
#
# For more information, consult the gateway module documentation.
#
# ---------------------------------- Various -----------------------------------
#
# Require explicit names when deleting indices:
#
#action.destructive_requires_name: true
indices.requests.cache.size: 5%
config/jvm.options

## JVM configuration
################################################################
## IMPORTANT: JVM heap size
################################################################
##
## You should always set the min and max JVM heap
## size to the same value. For example, to set
## the heap to 4 GB, set:
##
## -Xms4g
## -Xmx4g
##
## See https://www.elastic.co/guide/e ... .html
## for more information
##
################################################################
# Xms represents the initial size of total heap space
# Xmx represents the maximum size of total heap space
-Xms2g
-Xmx2g
################################################################
## Expert settings
################################################################
##
## All settings below this section are considered
## expert settings. Don't tamper with them unless
## you understand what you are doing
##
################################################################
## GC configuration
-XX:+UseConcMarkSweepGC
-XX:CMSInitiatingOccupancyFraction=75
-XX:+UseCMSInitiatingOccupancyOnly
## optimizations
# pre-touch memory pages used by the JVM during initialization
-XX:+AlwaysPreTouch
## basic
# force the server VM (remove on 32-bit client JVMs)
-server
# explicitly set the stack size (reduce to 320k on 32-bit client JVMs)
-Xss1m
# set to headless, just in case
-Djava.awt.headless=true
# ensure UTF-8 encoding by default (e.g. filenames)
-Dfile.encoding=UTF-8
# use our provided JNA always versus the system one
-Djna.nosys=true
# use old-style file permissions on JDK9
-Djdk.io.permissionsUseCanonicalPath=true
# flags to configure Netty
-Dio.netty.noUnsafe=true
-Dio.netty.noKeySetOptimization=true
-Dio.netty.recycler.maxCapacityPerThread=0
# log4j 2
-Dlog4j.shutdownHookEnabled=false
-Dlog4j2.disable.jmx=true
-Dlog4j.skipJansi=true
## heap dumps
# generate a heap dump when an allocation from the Java heap fails
# heap dumps are created in the working directory of the JVM
-XX:+HeapDumpOnOutOfMemoryError
# specify an alternative path for heap dumps
# ensure the directory exists and has sufficient space
#-XX:HeapDumpPath=${heap.dump.path}
## GC logging
#-XX:+PrintGCDetails
#-XX:+PrintGCTimeStamps
#-XX:+PrintGCDateStamps
#-XX:+PrintClassHistogram
#-XX:+PrintTenuringDistribution
#-XX:+PrintGCApplicationStoppedTime
# log GC status to a file with time stamps
# ensure the directory exists
#-Xloggc:${loggc}
# By default, the GC log file will not rotate.
# By uncommenting the lines below, the GC log file
# will be rotated every 128MB at most 32 times.
#-XX:+UseGCLogFileRotation
#-XX:NumberOfGCLogFiles=32
#-XX:GCLogFileSize=128M
# Elasticsearch 5.0.0 will throw an exception on unquoted field names in JSON.
# If documents were already indexed with unquoted fields in a previous version
# of Elasticsearch, some operations may throw errors.
#
# WARNING: This option will be removed in Elasticsearch 6.0.0 and is provided
# only for migration purposes.
#-Delasticsearch.json.allow_unquoted_field_names=true



安装ik分词器


bin/elasticsearch-plugin install https://github.com/medcl/elast ... 1.zip




./bin/elasticsearch-plugin install https://github.com/medcl/elast ... 3.zip



配置ik远程扩展词典用于热词更新 elasticsearch-5.6.3/config/analysis-ik/IKAnalyzer.cfg.xml

<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE properties SYSTEM "http://java.sun.com/dtd/properties.dtd&quot;&gt;
<properties>
       <comment>IK Analyzer 扩展配置</comment>
       <!--用户可以在这里配置自己的扩展字典 -->
       <entry key="ext_dict"></entry>
        <!--用户可以在这里配置自己的扩展停止词字典-->
       <entry key="ext_stopwords"></entry>
       <!--用户可以在这里配置远程扩展字典 -->
       <entry key="remote_ext_dict">http://distribute.search.leju. ... gt%3B
       <!--用户可以在这里配置远程扩展停止词字典-->
       <!-- <entry key="remote_ext_stopwords">words_location</entry> -->
</properties>
安装拼音分词器

cd elasticsearch-5.5.1/plugins
wget https://github.com/medcl/elast ... 5.5.1
unzip v5.5.1




打包部署其他节点时,先清理data目录




集群监控可以利用head的chrome插件




数据迁移

迁移工具是自己写的elasticbak,目前更新了5.6.3驱动。github链接:https://github.com/jiashiwen/elasticbak

数据备份

java -jar elasticbak-2.3.3.jar \
--exp \
--cluster lejuesdev \
--host 10.204.12.31 \
--filesize 1000 \
--backupdir ./esbackupset \
--backupindexes "*" \
--threads 4
由于版本field的变化需要手工重建索引,这里举个例子,主要是2.x版本的string需要改为text。2.x版本我们通过index参数指定该字段是否被索引("index": "no")以及是否通过分词器分词("index": "not_analyzed")。在5.X版本里index只用来制定是否创建索引,如果需要整个字段不过分词器创建索引,需要通过keyword字段完成。  

 

curl -XPUT "http://10.204.12.31:9201/house_geo&quot; -H 'Content-Type: application/json' -d'
{
 "mappings": {
   "house": {
     "dynamic": "strict",
     "_all": {
       "enabled": false
     },
     "properties": {
       "_category": {
         "type": "keyword",
                 "store": true
       },
       "_content": {
         "type": "text",
         "store": true,
         "analyzer": "ik_max_word",
         "search_analyzer": "ik_smart"
       },
       "_deleted": {
         "type": "boolean",
         "store": true
       },
       "_doccreatetime": {
         "type": "date",
         "store": true,
         "format": "strict_date_optional_time||epoch_millis||yyyy/MM/dd HH:mm:ss||yyyy/MM/dd"
       },
       "_docupdatetime": {
         "type": "date",
         "store": true,
         "format": "strict_date_optional_time||epoch_millis||yyyy/MM/dd HH:mm:ss||yyyy/MM/dd"
       },
       "_flags": {
         "type": "text",
         "store": true,
         "analyzer": "whitespace"
       },
       "_hits": {
         "type": "text"
       },
       "_location": {
         "type": "geo_point"
       },
       "_multi": {
         "properties": {
           "_location": {
             "type": "geo_point"
           }
         }
       },
       "_origin": {
         "type": "object",
         "enabled": false
       },
       "_scope": {
         "type": "keyword",
         "store": true
       },
       "_tags": {
         "type": "text",
         "boost": 10,
         "store": true,
         "term_vector": "with_positions_offsets",
         "analyzer": "ik_max_word",
         "search_analyzer": "ik_smart"
       },
       "_title": {
         "type": "text",
         "store": true,
         "analyzer": "ik_max_word",
         "search_analyzer": "ik_smart"
       },
       "_uniqid": {
         "type": "keyword",
         "store": true
       },
       "_uniqsign": {
         "type": "keyword",
         "store": true
       },
       "_url": {
         "type": "text",
         "index": false,
         "store": true
       },
       "location": {
         "type": "geo_point"
       }
     }
   }
 },
 "settings": {
   "index": {
     "number_of_shards": "3",
     "requests": {
       "cache": {
         "enable": "true"
       }
     },
     "analysis": {
       "filter": {
         "my_synonym": {
           "type": "synonym",
           "synonyms_path": "analysis-ik/custom/synonym.dic"
         }
       },
       "analyzer": {
         "searchanalyzer": {
           "filter": "my_synonym",
           "type": "custom",
           "tokenizer": "ik_smart"
         },
         "indexanalyzer": {
           "filter": "my_synonym",
           "type": "custom",
           "tokenizer": "ik_max_word"
         }
       }
     },
     "number_of_replicas": "1"
   }
 }
}'
利用新版elasticbak导入索引数据


java -jar elasticbak-5.6.3.jar \
--imp \
--cluster es5_dev \
--host 10.204.12.31 \
--port 9301 \
--restoreindex house_geo \
--restoretype dataonly \
--backupset esbackupset/house_geo \
--threads 4
  收起阅读 »

社区日报 第99期 (2017-11-13)

1、使用于商业智能的复杂模型建模教程
http://t.cn/Rj2uLh9
2、logstash配置文件的vscode插件,从其编辑配置文件不再发愁。
http://t.cn/Rj21ncE
3、elk告警插件sentinl。随着版本的更新,目前已经可以媲美x-pack的reporter以及watcher。
http://t.cn/Rj216Ef
4、只等你来 | Elastic Meetup 广州交流会
https://elasticsearch.cn/article/364 

编辑:cyberdak
归档:https://elasticsearch.cn/article/372
订阅:https://tinyletter.com/elastic-daily
继续阅读 »
1、使用于商业智能的复杂模型建模教程
http://t.cn/Rj2uLh9
2、logstash配置文件的vscode插件,从其编辑配置文件不再发愁。
http://t.cn/Rj21ncE
3、elk告警插件sentinl。随着版本的更新,目前已经可以媲美x-pack的reporter以及watcher。
http://t.cn/Rj216Ef
4、只等你来 | Elastic Meetup 广州交流会
https://elasticsearch.cn/article/364 

编辑:cyberdak
归档:https://elasticsearch.cn/article/372
订阅:https://tinyletter.com/elastic-daily 收起阅读 »

社区日报 第98期 (2017-11-12)

1. (自备梯子)推荐三个数据库相关(MySQL,MongoDB,Elasticsearch)的课程。
http://t.cn/RjPvlq1
2. 将 ELASTICSEARCH 写入速度优化到极限
http://t.cn/RWs8yvS
3. 零点之战!探访阿里巴巴8大技术专家,提前揭秘2017双11关键技术。
http://t.cn/RjPPzGc
4. 只等你来 | Elastic Meetup 广州交流会
https://elasticsearch.cn/article/364

编辑:至尊宝
归档:https://elasticsearch.cn/article/371
订阅:https://tinyletter.com/elastic-daily
继续阅读 »
1. (自备梯子)推荐三个数据库相关(MySQL,MongoDB,Elasticsearch)的课程。
http://t.cn/RjPvlq1
2. 将 ELASTICSEARCH 写入速度优化到极限
http://t.cn/RWs8yvS
3. 零点之战!探访阿里巴巴8大技术专家,提前揭秘2017双11关键技术。
http://t.cn/RjPPzGc
4. 只等你来 | Elastic Meetup 广州交流会
https://elasticsearch.cn/article/364

编辑:至尊宝
归档:https://elasticsearch.cn/article/371
订阅:https://tinyletter.com/elastic-daily 收起阅读 »

三步上手 esrally 完成 elasticsearch 压测任务

原文链接

距离上一篇 esrally 教程过去快2个月了,这期间不停有同学来询问使用中遇到的问题,尤其由于其测试数据存储在国外 aws 上,导致下载极慢。为了让大家快速上手使用 esrally,我 build 了一个可用的 docker 镜像,然后将 13GB 的测试数据拉取到国内的存储上,通过百度网盘的方式分享给大家。大家只要按照下面简单的几步操作就可以顺畅地使用 esrally 来进行相关测试了。

操作步骤

废话不多说,先上菜!

  1. 拉取镜像
    docker pull rockybean/esrally
  2. 下载数据文件 链接:http://pan.baidu.com/s/1eSrjZgA 密码:aagl
  3. 进入下载后的文件夹 rally_track,执行如下命令开始测试
    docker run -it -v $(PWD):/root/track rockybean/esrally esrally race --track-path=/root/track/logging --offline --pipeline=benchmark-only --target-hosts=192.168.1.105:9200

打完收工!

几点说明

数据文件介绍

esrally 自带的测试数据即为 rally_track 文件夹中的内容,主要包括:

  • Geonames(geonames): for evaluating the performance of structured data.
  • Geopoint(geopoint): for evaluating the performance of geo queries.
  • Percolator(percolator): for evaluating the performance of percolation queries.
  • PMC(pmc): for evaluating the performance of full text search.
  • NYC taxis(nyc_taxis): for evaluating the performance for highly structured data.
  • Nested(nested): for evaluating the performance for nested documents.
  • Logging(logging): for evaluating the performance of (Web) server logs.
  • noaa(noaa): for evaluating the performance of range fields.

可以根据自己的需要下载对应的测试数据,不必下载全部,保证对应文件夹下载完全即可。

命令解释

docker 相关

docker run -it rockybean/esrally esrally 为执行的 esrally 命令,-v $(PWD):/root/track是将 rally_docker 文件夹映射到 docker 容器中,$(PWD)是获取当前目录的意思,所以在此前要 cd 到 rally_docker 目录,当然你写全路径也是没有问题的。

esrally 的 docker 镜像比较简单,可以参看 github 项目介绍

esrally 相关

该镜像是通过自定义 track 的方式来加载数据,所以命令行中用到 --track=/root/track/logging 的命令行参数。注意这里的 /root/track 即上面我们绑定到容器的目录,更换 logging 为其他的数据集名称即可加载其他的测试数据。

该容器只支持测试第三方 es 集群,即 --pipeline=benchmark-only 模式。这应该也是最常见的压测需求了。

愉快地去玩耍吧!

继续阅读 »

原文链接

距离上一篇 esrally 教程过去快2个月了,这期间不停有同学来询问使用中遇到的问题,尤其由于其测试数据存储在国外 aws 上,导致下载极慢。为了让大家快速上手使用 esrally,我 build 了一个可用的 docker 镜像,然后将 13GB 的测试数据拉取到国内的存储上,通过百度网盘的方式分享给大家。大家只要按照下面简单的几步操作就可以顺畅地使用 esrally 来进行相关测试了。

操作步骤

废话不多说,先上菜!

  1. 拉取镜像
    docker pull rockybean/esrally
  2. 下载数据文件 链接:http://pan.baidu.com/s/1eSrjZgA 密码:aagl
  3. 进入下载后的文件夹 rally_track,执行如下命令开始测试
    docker run -it -v $(PWD):/root/track rockybean/esrally esrally race --track-path=/root/track/logging --offline --pipeline=benchmark-only --target-hosts=192.168.1.105:9200

打完收工!

几点说明

数据文件介绍

esrally 自带的测试数据即为 rally_track 文件夹中的内容,主要包括:

  • Geonames(geonames): for evaluating the performance of structured data.
  • Geopoint(geopoint): for evaluating the performance of geo queries.
  • Percolator(percolator): for evaluating the performance of percolation queries.
  • PMC(pmc): for evaluating the performance of full text search.
  • NYC taxis(nyc_taxis): for evaluating the performance for highly structured data.
  • Nested(nested): for evaluating the performance for nested documents.
  • Logging(logging): for evaluating the performance of (Web) server logs.
  • noaa(noaa): for evaluating the performance of range fields.

可以根据自己的需要下载对应的测试数据,不必下载全部,保证对应文件夹下载完全即可。

命令解释

docker 相关

docker run -it rockybean/esrally esrally 为执行的 esrally 命令,-v $(PWD):/root/track是将 rally_docker 文件夹映射到 docker 容器中,$(PWD)是获取当前目录的意思,所以在此前要 cd 到 rally_docker 目录,当然你写全路径也是没有问题的。

esrally 的 docker 镜像比较简单,可以参看 github 项目介绍

esrally 相关

该镜像是通过自定义 track 的方式来加载数据,所以命令行中用到 --track=/root/track/logging 的命令行参数。注意这里的 /root/track 即上面我们绑定到容器的目录,更换 logging 为其他的数据集名称即可加载其他的测试数据。

该容器只支持测试第三方 es 集群,即 --pipeline=benchmark-only 模式。这应该也是最常见的压测需求了。

愉快地去玩耍吧!

收起阅读 »

sense不能用了改用kibana吧

elasticsearch的dsl开发工具sense被google下架了,kibana console是很好的替代品。但是,我们的es集群前些日子因为应付安全检查改为https+basic auth方式(详细配置过程见本人博文:http://blog.csdn.net/jiashiwen ... 14374),kibana需要进行若干配置才能工作。另外老系统中还有elasticsearch2.3.3遗留,需要kibana4.5.1+sense。




一、elasticsearch5.5.2+kibana5.5.2

1.下载与elasticsearch版本号一致的kibana安装包,笔者目前开发环境5.5.2,对应kibana版本也为5.5.2(最新的5.6版本会报不兼容错误,不能运行)。




2.配置config/kibana.yml文件,主要配置项如下 
# The URL of the Elasticsearch instance to use for all your queries.
#elasticsearch.url: "http://localhost:9200&quot;
elasticsearch.url: "https://192.168.1.1:9281/&quot;


# If your Elasticsearch is protected with basic authentication, these settings provide
# the username and password that the Kibana server uses to perform maintenance on the Kibana
# index at startup. Your Kibana users still need to authenticate with Elasticsearch, which
# is proxied through the Kibana server.
#elasticsearch.username: "user"
#elasticsearch.password: "pass"
elasticsearch.username: "admin"
elasticsearch.password: "admin"


# Optional settings that provide the paths to the PEM-format SSL certificate and key files.
# These files validate that your Elasticsearch backend uses the same key files.
#elasticsearch.ssl.certificate: /path/to/your/client.crt
#elasticsearch.ssl.key: /path/to/your/client.key
elasticsearch.ssl.certificate: /home/develop/kibana-5.6.3-linux-x86_64/config/crts/eshttp.crt
elasticsearch.ssl.key: /home/develop/kibana-5.6.3-linux-x86_64/config/crts/eshttp.key


# To disregard the validity of SSL certificates, change this setting's value to 'none'.
#elasticsearch.ssl.verificationMode: full
elasticsearch.ssl.verificationMode: none各项配置看文件内说明,写的很清楚,这里就不翻译了,其中最重要的是这两样elasticsearch.ssl.certificate和elasticsearch.ssl.key,一定要与服务端保持一致。由于证书是自己生成的,校验项elasticsearch.ssl.verificationMode的值需要改为none。




启动kibana后,通过http://localhose:5601访问即可
继续阅读 »
elasticsearch的dsl开发工具sense被google下架了,kibana console是很好的替代品。但是,我们的es集群前些日子因为应付安全检查改为https+basic auth方式(详细配置过程见本人博文:http://blog.csdn.net/jiashiwen ... 14374),kibana需要进行若干配置才能工作。另外老系统中还有elasticsearch2.3.3遗留,需要kibana4.5.1+sense。




一、elasticsearch5.5.2+kibana5.5.2

1.下载与elasticsearch版本号一致的kibana安装包,笔者目前开发环境5.5.2,对应kibana版本也为5.5.2(最新的5.6版本会报不兼容错误,不能运行)。




2.配置config/kibana.yml文件,主要配置项如下 
# The URL of the Elasticsearch instance to use for all your queries.
#elasticsearch.url: "http://localhost:9200&quot;
elasticsearch.url: "https://192.168.1.1:9281/&quot;


# If your Elasticsearch is protected with basic authentication, these settings provide
# the username and password that the Kibana server uses to perform maintenance on the Kibana
# index at startup. Your Kibana users still need to authenticate with Elasticsearch, which
# is proxied through the Kibana server.
#elasticsearch.username: "user"
#elasticsearch.password: "pass"
elasticsearch.username: "admin"
elasticsearch.password: "admin"


# Optional settings that provide the paths to the PEM-format SSL certificate and key files.
# These files validate that your Elasticsearch backend uses the same key files.
#elasticsearch.ssl.certificate: /path/to/your/client.crt
#elasticsearch.ssl.key: /path/to/your/client.key
elasticsearch.ssl.certificate: /home/develop/kibana-5.6.3-linux-x86_64/config/crts/eshttp.crt
elasticsearch.ssl.key: /home/develop/kibana-5.6.3-linux-x86_64/config/crts/eshttp.key


# To disregard the validity of SSL certificates, change this setting's value to 'none'.
#elasticsearch.ssl.verificationMode: full
elasticsearch.ssl.verificationMode: none各项配置看文件内说明,写的很清楚,这里就不翻译了,其中最重要的是这两样elasticsearch.ssl.certificate和elasticsearch.ssl.key,一定要与服务端保持一致。由于证书是自己生成的,校验项elasticsearch.ssl.verificationMode的值需要改为none。




启动kibana后,通过http://localhose:5601访问即可 收起阅读 »

社区日报 第97期 (2017-11-11)

  1. sense为什么不能用了,看看ES官方怎么说? http://t.cn/RlB3B62

  2. 使用allocation API快速定位分片分配问题 http://t.cn/RlrzTsD

  3. ES6.0有关防止硬盘被填满的改进 http://t.cn/RlrU3Nr

  4. 喜大普奔,ES社区支持Markdown编辑器了 https://elasticsearch.cn/article/366

  5. Elastic 收购网站搜索 SaaS 服务领导者 Swiftype http://t.cn/Rl3a4P2

  6. 只等你来 | Elastic Meetup 广州交流会 https://elasticsearch.cn/article/364
继续阅读 »
  1. sense为什么不能用了,看看ES官方怎么说? http://t.cn/RlB3B62

  2. 使用allocation API快速定位分片分配问题 http://t.cn/RlrzTsD

  3. ES6.0有关防止硬盘被填满的改进 http://t.cn/RlrU3Nr

  4. 喜大普奔,ES社区支持Markdown编辑器了 https://elasticsearch.cn/article/366

  5. Elastic 收购网站搜索 SaaS 服务领导者 Swiftype http://t.cn/Rl3a4P2

  6. 只等你来 | Elastic Meetup 广州交流会 https://elasticsearch.cn/article/364
收起阅读 »

Elastic 招聘技术支持工程师,坐标北京

DG5LrOuXUAEgq4Q.jpg

立即申请

Support Engineer - Mandarin Speaking

Location: Beijing, China

Department: Support

Responsibilities

  • Ensuring customer issues are resolved within our committed service level agreements.
  • Maintain strong relationships with our customers for the delivery of support.
  • Have a mindset of continuous improvement, in terms of efficiency of support processes and customer satisfaction.

Experience

  • Demonstrable experience in of support in technology businesses
  • Experience working across multi-cultural and geographically distributed teams

Key Skills

  • Strong verbal and written communication skills in both Mandarin and English.
  • Customer orientated focus.
  • Team player, ability to work in a fast pace environment with a positive and adaptable approach.
  • Knowledge of databases or search technologies a plus.
  • Demonstrated strong technical understanding of software products.

Additional Information

  • Competitive pay and benefits
  • Stock options
  • Catered lunches, snacks, and beverages in most offices
  • An environment in which you can balance great work with a great life
  • Passionate people building great products
  • Employees with a wide variety of interests
  • Distributed-first company with employees in over 30 countries, spread across 18 time zones, and speaking over 30 languages!

立即申请

Elastic is an Equal Employment employer committed to the principles of equal employment opportunity and affirmative action for all applicants and employees. Qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender perception or identity, national origin, age, marital status, protected veteran status, or disability status or any other basis protected by federal, state or local law, ordinance or regulation. Elastic also makes reasonable accommodations for disabled employees consistent with applicable law.

About Elastic

Elastic is the world's leading software provider for making structured and unstructured data usable in real time for search, logging, security, and analytics use cases. Founded in 2012 by the people behind the Elasticsearch, Kibana, Beats, and Logstash open source projects, Elastic's global community has more than 80,000 members across 45 countries, and since its initial release. Elastic's products have achieved more than 100 million cumulative downloads. Today thousands of organizations, including Cisco, eBay, Dell, Goldman Sachs, Groupon, HP, Microsoft, Netflix, The New York Times, Uber, Verizon, Yelp, and Wikipedia, use the Elastic Stack, X-Pack, and Elastic Cloud to power mission-critical systems that drive new revenue opportunities and massive cost savings. Elastic is backed by more than $104 million in funding from Benchmark Capital, Index Ventures, and NEA; has headquarters in Amsterdam, the Netherlands, and Mountain View, California; and has over 500 employees in more than 30 countries around the world.

Our Philosophy

We’re always on the search for amazing people, people who have deep passion for technology and are masters at their craft. We build highly sophisticated distributed systems and we don’t take our technology lightly. In Elasticsearch, you’ll have the opportunity to work in a vibrant young company next to some of the smartest and highly skilled technologists the industry has to offer. We’re looking for great team players, yet we also promote independence and ownership. We’re hackers… but of the good kind. The kind that innovates and creates cutting edge products that eventually translates to a lot of happy, smiling faces.

LifeAtElastic

DEFMWTGV0AAOxBT.jpg

DFml2iQXcAAI4Dk.jpg

DFR1IpGVoAAyYfH.jpg

DHciIhpUMAAwk7W.jpg

DI-WzDWXoAAx8Z0.jpg

DKSTN1SVAAAoVgd.jpg

DJtbgW6UEAAmKu_.jpg

DIpqETTXoAAktBP.jpg

DKv1Gu1UMAAwJal.jpg

有兴趣加入我们么?点击申请

继续阅读 »

DG5LrOuXUAEgq4Q.jpg

立即申请

Support Engineer - Mandarin Speaking

Location: Beijing, China

Department: Support

Responsibilities

  • Ensuring customer issues are resolved within our committed service level agreements.
  • Maintain strong relationships with our customers for the delivery of support.
  • Have a mindset of continuous improvement, in terms of efficiency of support processes and customer satisfaction.

Experience

  • Demonstrable experience in of support in technology businesses
  • Experience working across multi-cultural and geographically distributed teams

Key Skills

  • Strong verbal and written communication skills in both Mandarin and English.
  • Customer orientated focus.
  • Team player, ability to work in a fast pace environment with a positive and adaptable approach.
  • Knowledge of databases or search technologies a plus.
  • Demonstrated strong technical understanding of software products.

Additional Information

  • Competitive pay and benefits
  • Stock options
  • Catered lunches, snacks, and beverages in most offices
  • An environment in which you can balance great work with a great life
  • Passionate people building great products
  • Employees with a wide variety of interests
  • Distributed-first company with employees in over 30 countries, spread across 18 time zones, and speaking over 30 languages!

立即申请

Elastic is an Equal Employment employer committed to the principles of equal employment opportunity and affirmative action for all applicants and employees. Qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender perception or identity, national origin, age, marital status, protected veteran status, or disability status or any other basis protected by federal, state or local law, ordinance or regulation. Elastic also makes reasonable accommodations for disabled employees consistent with applicable law.

About Elastic

Elastic is the world's leading software provider for making structured and unstructured data usable in real time for search, logging, security, and analytics use cases. Founded in 2012 by the people behind the Elasticsearch, Kibana, Beats, and Logstash open source projects, Elastic's global community has more than 80,000 members across 45 countries, and since its initial release. Elastic's products have achieved more than 100 million cumulative downloads. Today thousands of organizations, including Cisco, eBay, Dell, Goldman Sachs, Groupon, HP, Microsoft, Netflix, The New York Times, Uber, Verizon, Yelp, and Wikipedia, use the Elastic Stack, X-Pack, and Elastic Cloud to power mission-critical systems that drive new revenue opportunities and massive cost savings. Elastic is backed by more than $104 million in funding from Benchmark Capital, Index Ventures, and NEA; has headquarters in Amsterdam, the Netherlands, and Mountain View, California; and has over 500 employees in more than 30 countries around the world.

Our Philosophy

We’re always on the search for amazing people, people who have deep passion for technology and are masters at their craft. We build highly sophisticated distributed systems and we don’t take our technology lightly. In Elasticsearch, you’ll have the opportunity to work in a vibrant young company next to some of the smartest and highly skilled technologists the industry has to offer. We’re looking for great team players, yet we also promote independence and ownership. We’re hackers… but of the good kind. The kind that innovates and creates cutting edge products that eventually translates to a lot of happy, smiling faces.

LifeAtElastic

DEFMWTGV0AAOxBT.jpg

DFml2iQXcAAI4Dk.jpg

DFR1IpGVoAAyYfH.jpg

DHciIhpUMAAwk7W.jpg

DI-WzDWXoAAx8Z0.jpg

DKSTN1SVAAAoVgd.jpg

DJtbgW6UEAAmKu_.jpg

DIpqETTXoAAktBP.jpg

DKv1Gu1UMAAwJal.jpg

有兴趣加入我们么?点击申请

收起阅读 »

社区支持 Markdown 编辑器

为了改善大家的创作体验,提高大家的写作和分享热情!?,经过两天的不懈奋斗,终于把 Markdown 编辑器搬上来了。 目前只支持文章的发布,可以通过切换编辑器来选择 Markdown 编辑模式。 希望不要再以编辑器作为理由发只有链接的文章了。 ???????????

  • 支持 Github 风格的 Markdown 格式
  • 支持本站附件功能
  • 支持 emoj 符号
  • 支持自动的页面导航
  • 以前的文章可再次编辑,切换 Markdown 模式然后修改保存

如何使用?

  1. 点击【发起】,选择文章
  2. 切换绿色按钮,将编辑器切换到 Markdown,然后在文本框内输入 Markdown 格式的内容即可。

在线 Markdown 脚本编辑预览工具:https://elasticsearch.cn/static/js/editor/markdown/


以下为样式测试参考,忽略其意义。

1510307393004.jpg

----------- 常用格式-----------------


# 标题1
## 标题2
### 标题3
#### 标题4
##### 标题5
###### 标题6
超大标题   //等于号写于文字下方
===
标题      //同超大标题
---

`短代码`
_ 注:长代码块,用三个: `  _

> This is the first level of quoting.
>
> > This is nested blockquote.
>
> Back to the first level.

* Red
* Green
* Blue

- Red
- Green
- Blue

+ Red
+ Green
+ Blue

1. 这是第一个
1. 这是第二个
1. 这是第三个

* * *
***
*****
- - -
---

[markdown-syntax](http://daringfireball.net/projects/markdown/syntax)

[id]: http://example.com/  "Optional Title Here"
This is [an example][id] reference-style link.

*内容*
**内容**
_内容_
__内容__

![这是张外链图片](https://static-www.elastic.co/assets/bltbfcd44f1256d8c88/blog-swifttype-thumb.jpg?q=845)

<http://elastic.co/>

<info@elastic.o>

    四个空格
    一个tab

----------- 样式预览-----------------

标题1

标题2

标题3

标题4

标题5
标题6

超大标题 //等于号写于文字下方

标题 //同超大标题

短代码

This is the first level of quoting.

This is nested blockquote.

Back to the first level.

  • Red
  • Green
  • Blue

  • Red
  • Green
  • Blue

  • Red
  • Green
  • Blue
  1. 这是第一个
  2. 这是第二个
  3. 这是第三个





markdown-syntax

This is an example reference-style link.

内容 内容 内容 内容

这是张外链图片

http://elastic.co/

info@elastic.o

四个空格
一个tab

https://github.com/infinitbyte/gopa 的 README 内容


What a Spider!

GOPA, A Spider Written in Go.

Travis Go Report Card Coverage Status Join the chat at https://gitter.im/infinitbyte/gopa

Goal

  • Light weight, low footprint, memory requirement should < 100MB
  • Easy to deploy, no runtime or dependency required
  • Easy to use, no programming or scripts ability needed, out of box features

Screenshoot

What a Spider! GOPA Spider!

How to use

Setup

First of all, get it, two opinions: download the pre-built package or compile it yourself.

Download Pre Built Package

Go to Release or Snapshot page, download the right package for your platform.

Note: Darwin is for Mac

Compile The Package Manually

So far, we have:

gopa, the main program, a single binary.
config/, elasticsearch related scripts etc.
gopa.yml, main configuration for gopa.

Optional Config

By default, Gopa works well except indexing, if you want to use elasticsearch as indexing, follow these steps:

  • Create a index in elasticsearch with script config/gopa-index-mapping.sh

    Example
    curl -XPUT "http://localhost:9200/gopa-index" -H 'Content-Type: application/json' -d'
    {
    "mappings": {
    "doc": {
      "properties": {
        "host": {
            "type": "keyword",
            "ignore_above": 256
        },
        "snapshot": {
          "properties": {
            "bold": {
              "type": "text"
            },
            "url": {
              "type": "keyword",
              "ignore_above": 256
            },
            "content_type": {
              "type": "keyword",
              "ignore_above": 256
            },
            "file": {
              "type": "keyword",
              "ignore_above": 256
            },
            "h1": {
              "type": "text"
            },
            "h2": {
              "type": "text"
            },
            "h3": {
              "type": "text"
            },
            "h4": {
              "type": "text"
            },
            "hash": {
              "type": "keyword",
              "ignore_above": 256
            },
            "id": {
              "type": "keyword",
              "ignore_above": 256
            },
            "images": {
              "properties": {
                "external": {
                  "properties": {
                    "label": {
                      "type": "text"
                    },
                    "url": {
                      "type": "keyword",
                      "ignore_above": 256
                    }
                  }
                },
                "internal": {
                  "properties": {
                    "label": {
                      "type": "text"
                    },
                    "url": {
                      "type": "keyword",
                      "ignore_above": 256
                    }
                  }
                }
              }
            },
            "italic": {
              "type": "text"
            },
            "links": {
              "properties": {
                "external": {
                  "properties": {
                    "label": {
                      "type": "text"
                    },
                    "url": {
                      "type": "keyword",
                      "ignore_above": 256
                    }
                  }
                },
                "internal": {
                  "properties": {
                    "label": {
                      "type": "text"
                    },
                    "url": {
                      "type": "keyword",
                      "ignore_above": 256
                    }
                  }
                }
              }
            },
            "path": {
              "type": "keyword",
              "ignore_above": 256
            },
            "sim_hash": {
              "type": "keyword",
              "ignore_above": 256
            },
            "lang": {
              "type": "keyword",
              "ignore_above": 256
            },
            "size": {
              "type": "long"
            },
            "text": {
              "type": "text"
            },
            "title": {
              "type": "text",
              "fields": {
                "keyword": {
                  "type": "keyword"
                }
              }
            },
            "version": {
              "type": "long"
            }
          }
        },
        "task": {
          "properties": {
            "breadth": {
              "type": "long"
            },
            "created": {
              "type": "date"
            },
            "depth": {
              "type": "long"
            },
            "id": {
              "type": "keyword",
              "ignore_above": 256
            },
            "original_url": {
              "type": "keyword",
              "ignore_above": 256
            },
            "reference_url": {
              "type": "keyword",
              "ignore_above": 256
            },
            "schema": {
              "type": "keyword",
              "ignore_above": 256
            },
            "status": {
              "type": "integer"
            },
            "updated": {
              "type": "date"
            },
            "url": {
              "type": "keyword",
              "ignore_above": 256
            }
          }
        }
      }
    }
    }
    }'

Note: Elasticsearch version should > v5.0

  • Enable index module in gopa.yml, update the elasticsearch's setting:
    - module: index
    enabled: true
    ui:
      enabled: true
    elasticsearch:
      endpoint: http://dev:9200
      index_prefix: gopa-
      username: elastic
      password: changeme

Start

Gopa doesn't require any dependencies, simply run ./gopa to start the program.

Gopa can be run as daemon(Note: Only available on Linux and Mac):

Example

➜  gopa git:(master) ✗ ./bin/gopa --daemon
  ________ ________ __________  _____
 /  _____/ \_____  \\______   \/  _  \
/   \  ___  /   |   \|     ___/  /_\  \
\    \_\  \/    |    \    |  /    |    \
 \______  /\_______  /____|  \____|__  /
        \/         \/                \/
[gopa] 0.10.0_SNAPSHOT
///last commit: 99616a2, Fri Oct 20 14:04:54 2017 +0200, medcl, update version to 0.10.0 ///

[10-21 16:01:09] [INF] [instance.go:23] workspace: data/gopa/nodes/0
[gopa] started.

Also run ./gopa -h to get the full list of command line options.

Example

➜  gopa git:(master) ✗ ./bin/gopa -h
  ________ ________ __________  _____
 /  _____/ \_____  \\______   \/  _  \
/   \  ___  /   |   \|     ___/  /_\  \
\    \_\  \/    |    \    |  /    |    \
 \______  /\_______  /____|  \____|__  /
        \/         \/                \/
[gopa] 0.10.0_SNAPSHOT
///last commit: 99616a2, Fri Oct 20 14:04:54 2017 +0200, medcl, update version to 0.10.0 ///

Usage of ./bin/gopa:
  -config string
        the location of config file (default "gopa.yml")
  -cpuprofile string
        write cpu profile to this file
  -daemon
        run in background as daemon
  -debug
        run in debug mode, wi
  -log string
        the log level,options:trace,debug,info,warn,error (default "info")
  -log_path string
        the log path (default "log")
  -memprofile string
        write memory profile to this file
  -pidfile string
        pidfile path (only for daemon)
  -pprof string
        enable and setup pprof/expvar service, eg: localhost:6060 , the endpoint will be: http://localhost:6060/debug/pprof/ and http://localhost:6060/debug/vars

Stop

It's safety to press ctrl+c stop the current running Gopa, Gopa will handle the rest,saving the checkpoint, you may restore the job later,the world is still in your hand.

If you are running Gopa as daemon, you may stop it like this:

 kill -QUIT `pgrep gopa`

Configuration

UI

  • Search Console http://127.0.0.1:9001/
  • Admin Console http://127.0.0.1:9001/admin/

API

  • TBD

Contributing

You are sincerely and warmly welcomed to play with this project, from UI style to core features, or just a piece of document, welcome! let's make it better.

License

Released under the Apache License, Version 2.0 .

Also XSS Test

alert('XSS test');

继续阅读 »

为了改善大家的创作体验,提高大家的写作和分享热情!?,经过两天的不懈奋斗,终于把 Markdown 编辑器搬上来了。 目前只支持文章的发布,可以通过切换编辑器来选择 Markdown 编辑模式。 希望不要再以编辑器作为理由发只有链接的文章了。 ???????????

  • 支持 Github 风格的 Markdown 格式
  • 支持本站附件功能
  • 支持 emoj 符号
  • 支持自动的页面导航
  • 以前的文章可再次编辑,切换 Markdown 模式然后修改保存

如何使用?

  1. 点击【发起】,选择文章
  2. 切换绿色按钮,将编辑器切换到 Markdown,然后在文本框内输入 Markdown 格式的内容即可。

在线 Markdown 脚本编辑预览工具:https://elasticsearch.cn/static/js/editor/markdown/


以下为样式测试参考,忽略其意义。

1510307393004.jpg

----------- 常用格式-----------------


# 标题1
## 标题2
### 标题3
#### 标题4
##### 标题5
###### 标题6
超大标题   //等于号写于文字下方
===
标题      //同超大标题
---

`短代码`
_ 注:长代码块,用三个: `  _

> This is the first level of quoting.
>
> > This is nested blockquote.
>
> Back to the first level.

* Red
* Green
* Blue

- Red
- Green
- Blue

+ Red
+ Green
+ Blue

1. 这是第一个
1. 这是第二个
1. 这是第三个

* * *
***
*****
- - -
---

[markdown-syntax](http://daringfireball.net/projects/markdown/syntax)

[id]: http://example.com/  "Optional Title Here"
This is [an example][id] reference-style link.

*内容*
**内容**
_内容_
__内容__

![这是张外链图片](https://static-www.elastic.co/assets/bltbfcd44f1256d8c88/blog-swifttype-thumb.jpg?q=845)

<http://elastic.co/>

<info@elastic.o>

    四个空格
    一个tab

----------- 样式预览-----------------

标题1

标题2

标题3

标题4

标题5
标题6

超大标题 //等于号写于文字下方

标题 //同超大标题

短代码

This is the first level of quoting.

This is nested blockquote.

Back to the first level.

  • Red
  • Green
  • Blue

  • Red
  • Green
  • Blue

  • Red
  • Green
  • Blue
  1. 这是第一个
  2. 这是第二个
  3. 这是第三个





markdown-syntax

This is an example reference-style link.

内容 内容 内容 内容

这是张外链图片

http://elastic.co/

info@elastic.o

四个空格
一个tab

https://github.com/infinitbyte/gopa 的 README 内容


What a Spider!

GOPA, A Spider Written in Go.

Travis Go Report Card Coverage Status Join the chat at https://gitter.im/infinitbyte/gopa

Goal

  • Light weight, low footprint, memory requirement should < 100MB
  • Easy to deploy, no runtime or dependency required
  • Easy to use, no programming or scripts ability needed, out of box features

Screenshoot

What a Spider! GOPA Spider!

How to use

Setup

First of all, get it, two opinions: download the pre-built package or compile it yourself.

Download Pre Built Package

Go to Release or Snapshot page, download the right package for your platform.

Note: Darwin is for Mac

Compile The Package Manually

So far, we have:

gopa, the main program, a single binary.
config/, elasticsearch related scripts etc.
gopa.yml, main configuration for gopa.

Optional Config

By default, Gopa works well except indexing, if you want to use elasticsearch as indexing, follow these steps:

  • Create a index in elasticsearch with script config/gopa-index-mapping.sh

    Example
    curl -XPUT "http://localhost:9200/gopa-index" -H 'Content-Type: application/json' -d'
    {
    "mappings": {
    "doc": {
      "properties": {
        "host": {
            "type": "keyword",
            "ignore_above": 256
        },
        "snapshot": {
          "properties": {
            "bold": {
              "type": "text"
            },
            "url": {
              "type": "keyword",
              "ignore_above": 256
            },
            "content_type": {
              "type": "keyword",
              "ignore_above": 256
            },
            "file": {
              "type": "keyword",
              "ignore_above": 256
            },
            "h1": {
              "type": "text"
            },
            "h2": {
              "type": "text"
            },
            "h3": {
              "type": "text"
            },
            "h4": {
              "type": "text"
            },
            "hash": {
              "type": "keyword",
              "ignore_above": 256
            },
            "id": {
              "type": "keyword",
              "ignore_above": 256
            },
            "images": {
              "properties": {
                "external": {
                  "properties": {
                    "label": {
                      "type": "text"
                    },
                    "url": {
                      "type": "keyword",
                      "ignore_above": 256
                    }
                  }
                },
                "internal": {
                  "properties": {
                    "label": {
                      "type": "text"
                    },
                    "url": {
                      "type": "keyword",
                      "ignore_above": 256
                    }
                  }
                }
              }
            },
            "italic": {
              "type": "text"
            },
            "links": {
              "properties": {
                "external": {
                  "properties": {
                    "label": {
                      "type": "text"
                    },
                    "url": {
                      "type": "keyword",
                      "ignore_above": 256
                    }
                  }
                },
                "internal": {
                  "properties": {
                    "label": {
                      "type": "text"
                    },
                    "url": {
                      "type": "keyword",
                      "ignore_above": 256
                    }
                  }
                }
              }
            },
            "path": {
              "type": "keyword",
              "ignore_above": 256
            },
            "sim_hash": {
              "type": "keyword",
              "ignore_above": 256
            },
            "lang": {
              "type": "keyword",
              "ignore_above": 256
            },
            "size": {
              "type": "long"
            },
            "text": {
              "type": "text"
            },
            "title": {
              "type": "text",
              "fields": {
                "keyword": {
                  "type": "keyword"
                }
              }
            },
            "version": {
              "type": "long"
            }
          }
        },
        "task": {
          "properties": {
            "breadth": {
              "type": "long"
            },
            "created": {
              "type": "date"
            },
            "depth": {
              "type": "long"
            },
            "id": {
              "type": "keyword",
              "ignore_above": 256
            },
            "original_url": {
              "type": "keyword",
              "ignore_above": 256
            },
            "reference_url": {
              "type": "keyword",
              "ignore_above": 256
            },
            "schema": {
              "type": "keyword",
              "ignore_above": 256
            },
            "status": {
              "type": "integer"
            },
            "updated": {
              "type": "date"
            },
            "url": {
              "type": "keyword",
              "ignore_above": 256
            }
          }
        }
      }
    }
    }
    }'

Note: Elasticsearch version should > v5.0

  • Enable index module in gopa.yml, update the elasticsearch's setting:
    - module: index
    enabled: true
    ui:
      enabled: true
    elasticsearch:
      endpoint: http://dev:9200
      index_prefix: gopa-
      username: elastic
      password: changeme

Start

Gopa doesn't require any dependencies, simply run ./gopa to start the program.

Gopa can be run as daemon(Note: Only available on Linux and Mac):

Example

➜  gopa git:(master) ✗ ./bin/gopa --daemon
  ________ ________ __________  _____
 /  _____/ \_____  \\______   \/  _  \
/   \  ___  /   |   \|     ___/  /_\  \
\    \_\  \/    |    \    |  /    |    \
 \______  /\_______  /____|  \____|__  /
        \/         \/                \/
[gopa] 0.10.0_SNAPSHOT
///last commit: 99616a2, Fri Oct 20 14:04:54 2017 +0200, medcl, update version to 0.10.0 ///

[10-21 16:01:09] [INF] [instance.go:23] workspace: data/gopa/nodes/0
[gopa] started.

Also run ./gopa -h to get the full list of command line options.

Example

➜  gopa git:(master) ✗ ./bin/gopa -h
  ________ ________ __________  _____
 /  _____/ \_____  \\______   \/  _  \
/   \  ___  /   |   \|     ___/  /_\  \
\    \_\  \/    |    \    |  /    |    \
 \______  /\_______  /____|  \____|__  /
        \/         \/                \/
[gopa] 0.10.0_SNAPSHOT
///last commit: 99616a2, Fri Oct 20 14:04:54 2017 +0200, medcl, update version to 0.10.0 ///

Usage of ./bin/gopa:
  -config string
        the location of config file (default "gopa.yml")
  -cpuprofile string
        write cpu profile to this file
  -daemon
        run in background as daemon
  -debug
        run in debug mode, wi
  -log string
        the log level,options:trace,debug,info,warn,error (default "info")
  -log_path string
        the log path (default "log")
  -memprofile string
        write memory profile to this file
  -pidfile string
        pidfile path (only for daemon)
  -pprof string
        enable and setup pprof/expvar service, eg: localhost:6060 , the endpoint will be: http://localhost:6060/debug/pprof/ and http://localhost:6060/debug/vars

Stop

It's safety to press ctrl+c stop the current running Gopa, Gopa will handle the rest,saving the checkpoint, you may restore the job later,the world is still in your hand.

If you are running Gopa as daemon, you may stop it like this:

 kill -QUIT `pgrep gopa`

Configuration

UI

  • Search Console http://127.0.0.1:9001/
  • Admin Console http://127.0.0.1:9001/admin/

API

  • TBD

Contributing

You are sincerely and warmly welcomed to play with this project, from UI style to core features, or just a piece of document, welcome! let's make it better.

License

Released under the Apache License, Version 2.0 .

Also XSS Test

alert('XSS test');

收起阅读 »

社区日报 第96期 (2017-11-10)

1、源码解析 | elasticsearch 索引创建和更新
http://t.cn/RlHuOKx
2、业界良心 |  《Elasticsearch5.6.3 Java API 中文手册》
https://elasticsearch.cn/article/362
3、PPT | 基于 Mesos/Docker 的 Elasticsearch 容器化私有云
http://t.cn/RlHuTQR 
4、只等你来 | Elastic Meetup 广州交流会
https://elasticsearch.cn/article/364

编辑:laoyang360
归档:https://elasticsearch.cn/article/365
订阅:https://tinyletter.com/elastic-daily
 
继续阅读 »
1、源码解析 | elasticsearch 索引创建和更新
http://t.cn/RlHuOKx
2、业界良心 |  《Elasticsearch5.6.3 Java API 中文手册》
https://elasticsearch.cn/article/362
3、PPT | 基于 Mesos/Docker 的 Elasticsearch 容器化私有云
http://t.cn/RlHuTQR 
4、只等你来 | Elastic Meetup 广州交流会
https://elasticsearch.cn/article/364

编辑:laoyang360
归档:https://elasticsearch.cn/article/365
订阅:https://tinyletter.com/elastic-daily
  收起阅读 »

Elastic Meetup 广州交流会

Elastic Meetup 线下交流活动再次来到羊城广州,算是社区在广州的第二次线下聚会了,广州的小伙伴们,快快报名吧! 回顾去年的线下活动,可以点击这里:https://elasticsearch.cn/article/71​  

timg.jpeg

主办:

本次活动由 Elastic 与 网易游戏运维与基础架构部 联合举办。  

媒体:

本次活动由 IT大咖说 独家提供现场直播。  

时间:

2017.11.25​  下午2:00-5:00(1点半开始签到)  

地点:

广州市天河区科韵路16号广州信息港E栋网易大厦 一楼博学堂  

主题:

  1. 网易 - 杜鑫 - ELK在藏宝阁中的应用
  2. 酷狗 - 钟旺 - 基于ES的音乐搜索引擎 
  3. 阿里云 - 赵弘扬 - Elasticsearch在阿里云的实践分享
  4. 网易 - 林邦骏 - 网易ELK 系统综述
  5. 数说故事 - 吴文杰 - Data Warehouse with ElasticSearch in Datastory
  6. 闪电分享(5-10分钟,可现场报名) 

参会报名:

http://elasticsearch.mikecrm.com/O6o0yq3  

现场直播:

直播连接:http://www.itdks.com/eventlist/detail/1673

WechatIMG176.jpeg

主题介绍:

ELK在藏宝阁中的应用 

内容介绍:

 1. 藏宝阁项目介绍  主要介绍一下藏宝阁项目,让不熟悉藏宝阁的听众有一个基本的了解,熟悉应用的背景。

  1. ELK在藏宝阁中的应用(概述)  大致简要的阐述一下ELK在藏宝阁中哪些地方发挥了什么样的作用。

  2. ELK在藏宝阁推荐系统中的应用(重点)  较为详细的剖析一下ELK在推荐系统中的发挥的作用,具备的优势。

分享嘉宾:

讲师照片.jpg

杜鑫,网易藏宝阁工作室资深开发工程师,目前主要从事藏宝阁推荐业务相关的研发工作。  

网易ELK 系统综述

内容介绍:

  从架构以及功能两个角度去阐述网易的 ELK 平台,介绍系统内部各个组件及其管理方式。进而以用户的视角介绍平台中包含的自动化服务等功能,从管理员的视角去讨论组件的配置管理、资源调度回收等问题。

分享嘉宾:

lin.jpg

林邦骏,网易 GDC产品组资深运维工程师,主要负责内部 ELK 产品的运维、功能开发等工作。  

基于ES的音乐搜索引擎

内容介绍:

1、酷狗音乐搜索引擎架构变迁

2、构建音乐搜索引擎经验之谈 

分享嘉宾:

zhong.jpg

钟旺,酷狗后台开发工程师,从事JAVA、ES相关的开发工作。  

Data Warehouse with ElasticSearch in Datastory

内容介绍:

ES最多使用的场景是搜索和日志分析,然而ES强大的实时索引查询、全文检索和聚合能力也能成为数据仓库与OLAP场景的强力支持。

本次分享将为大家带来数说故事如何借助ES和Hadoop生态在不同的数据场景下构建起数据仓库能力。

分享嘉宾:

wu.jpg

吴文杰 ,数说故事平台架构团队 高级工程师,负责数说故事百亿级数据的存储查询及内部基础平台建设。  

Elasticsearch在阿里云的实践分享

内容介绍

介绍阿里云Elastiserach服务的技术架构和Xpack相关功能,并分享在云上环境搭建ELK的实践案例。

分享嘉宾

zhao.jpg

赵弘扬,阿里巴巴搜索产品专家,负责阿里云搜索产品规划和开发。


  深圳也在筹备中,可以提前报名!:https://elasticsearch.cn/article/261  


关于 Elastic Meetup

Elastic Meetup 由 Elastic 中文社区定期举办的线下交流活动,主要围绕 Elastic 的开源产品(Elasticsearch、Logstash、Kibana 和 Beats)及周边技术,探讨在搜索、数据实时分析、日志分析、安全等领域的实践与应用。

 

关于 Elastic

a998ed2d2cf1e319a841178bbc796fb8.jpg

Elastic 通过构建软件,让用户能够实时地、大规模地将数据用于搜索、日志和分析场景。Elastic 创立于 2012 年,相继开发了开源的 Elastic Stack(Elasticsearch、Kibana、Beats 和 Logstash)、X-Pack(商业功能)和 Elastic Cloud(托管服务)。截至目前,累计下载量超过 1.5 亿。Benchmark Capital、Index Ventures 和 NEA 为 Elastic 提供了超过 1 亿美元资金作为支持,Elastic 共有 600 多名员工,分布在 30 个国家/地区。有关更多信息,请访问 http://elastic.co/cn 。  

关于网易游戏运维与基础架构部

运基LOGO(红底白字).png

网易游戏运维与基础架构部, 主要负责网易游戏产品的可靠性保障以及基础设施的开发和部署,旨在:

  1. 专注为产品全生命周期提供可靠性保障服务,依托于大数据为运维提供决策
  2. 通过智能监控提高问题发现和解决能力,以自动化驱动低成本的业务管理
  3. 打造混合云方案,站在游戏业务角度驱动的TCO优化和运维智能化

关于IT大咖说

 

IT大咖说LOGO(知识分享平台).jpg

IT大咖说,IT垂直领域的大咖知识分享平台,践行“开源是一种态度”,通过线上线下开放模式分享行业TOP大咖干货,技术大会在线直播点播,在线活动直播平台。http://www.itdks.com 。  

再次感谢网易游戏运维与基础架构部和IT大咖说的大力支持!

继续阅读 »

Elastic Meetup 线下交流活动再次来到羊城广州,算是社区在广州的第二次线下聚会了,广州的小伙伴们,快快报名吧! 回顾去年的线下活动,可以点击这里:https://elasticsearch.cn/article/71​  

timg.jpeg

主办:

本次活动由 Elastic 与 网易游戏运维与基础架构部 联合举办。  

媒体:

本次活动由 IT大咖说 独家提供现场直播。  

时间:

2017.11.25​  下午2:00-5:00(1点半开始签到)  

地点:

广州市天河区科韵路16号广州信息港E栋网易大厦 一楼博学堂  

主题:

  1. 网易 - 杜鑫 - ELK在藏宝阁中的应用
  2. 酷狗 - 钟旺 - 基于ES的音乐搜索引擎 
  3. 阿里云 - 赵弘扬 - Elasticsearch在阿里云的实践分享
  4. 网易 - 林邦骏 - 网易ELK 系统综述
  5. 数说故事 - 吴文杰 - Data Warehouse with ElasticSearch in Datastory
  6. 闪电分享(5-10分钟,可现场报名) 

参会报名:

http://elasticsearch.mikecrm.com/O6o0yq3  

现场直播:

直播连接:http://www.itdks.com/eventlist/detail/1673

WechatIMG176.jpeg

主题介绍:

ELK在藏宝阁中的应用 

内容介绍:

 1. 藏宝阁项目介绍  主要介绍一下藏宝阁项目,让不熟悉藏宝阁的听众有一个基本的了解,熟悉应用的背景。

  1. ELK在藏宝阁中的应用(概述)  大致简要的阐述一下ELK在藏宝阁中哪些地方发挥了什么样的作用。

  2. ELK在藏宝阁推荐系统中的应用(重点)  较为详细的剖析一下ELK在推荐系统中的发挥的作用,具备的优势。

分享嘉宾:

讲师照片.jpg

杜鑫,网易藏宝阁工作室资深开发工程师,目前主要从事藏宝阁推荐业务相关的研发工作。  

网易ELK 系统综述

内容介绍:

  从架构以及功能两个角度去阐述网易的 ELK 平台,介绍系统内部各个组件及其管理方式。进而以用户的视角介绍平台中包含的自动化服务等功能,从管理员的视角去讨论组件的配置管理、资源调度回收等问题。

分享嘉宾:

lin.jpg

林邦骏,网易 GDC产品组资深运维工程师,主要负责内部 ELK 产品的运维、功能开发等工作。  

基于ES的音乐搜索引擎

内容介绍:

1、酷狗音乐搜索引擎架构变迁

2、构建音乐搜索引擎经验之谈 

分享嘉宾:

zhong.jpg

钟旺,酷狗后台开发工程师,从事JAVA、ES相关的开发工作。  

Data Warehouse with ElasticSearch in Datastory

内容介绍:

ES最多使用的场景是搜索和日志分析,然而ES强大的实时索引查询、全文检索和聚合能力也能成为数据仓库与OLAP场景的强力支持。

本次分享将为大家带来数说故事如何借助ES和Hadoop生态在不同的数据场景下构建起数据仓库能力。

分享嘉宾:

wu.jpg

吴文杰 ,数说故事平台架构团队 高级工程师,负责数说故事百亿级数据的存储查询及内部基础平台建设。  

Elasticsearch在阿里云的实践分享

内容介绍

介绍阿里云Elastiserach服务的技术架构和Xpack相关功能,并分享在云上环境搭建ELK的实践案例。

分享嘉宾

zhao.jpg

赵弘扬,阿里巴巴搜索产品专家,负责阿里云搜索产品规划和开发。


  深圳也在筹备中,可以提前报名!:https://elasticsearch.cn/article/261  


关于 Elastic Meetup

Elastic Meetup 由 Elastic 中文社区定期举办的线下交流活动,主要围绕 Elastic 的开源产品(Elasticsearch、Logstash、Kibana 和 Beats)及周边技术,探讨在搜索、数据实时分析、日志分析、安全等领域的实践与应用。

 

关于 Elastic

a998ed2d2cf1e319a841178bbc796fb8.jpg

Elastic 通过构建软件,让用户能够实时地、大规模地将数据用于搜索、日志和分析场景。Elastic 创立于 2012 年,相继开发了开源的 Elastic Stack(Elasticsearch、Kibana、Beats 和 Logstash)、X-Pack(商业功能)和 Elastic Cloud(托管服务)。截至目前,累计下载量超过 1.5 亿。Benchmark Capital、Index Ventures 和 NEA 为 Elastic 提供了超过 1 亿美元资金作为支持,Elastic 共有 600 多名员工,分布在 30 个国家/地区。有关更多信息,请访问 http://elastic.co/cn 。  

关于网易游戏运维与基础架构部

运基LOGO(红底白字).png

网易游戏运维与基础架构部, 主要负责网易游戏产品的可靠性保障以及基础设施的开发和部署,旨在:

  1. 专注为产品全生命周期提供可靠性保障服务,依托于大数据为运维提供决策
  2. 通过智能监控提高问题发现和解决能力,以自动化驱动低成本的业务管理
  3. 打造混合云方案,站在游戏业务角度驱动的TCO优化和运维智能化

关于IT大咖说

 

IT大咖说LOGO(知识分享平台).jpg

IT大咖说,IT垂直领域的大咖知识分享平台,践行“开源是一种态度”,通过线上线下开放模式分享行业TOP大咖干货,技术大会在线直播点播,在线活动直播平台。http://www.itdks.com 。  

再次感谢网易游戏运维与基础架构部和IT大咖说的大力支持!

收起阅读 »

社区日报 第95期 (2017-11-09)

1.Bulk异常引发的Elasticsearch内存泄漏
http://t.cn/RlY7tMh
2.Spring Boot 中使用 Java API 调用 Elasticsearch
http://t.cn/RljQNFJ
3.一个实时查看,搜索尾部日志事件的kibana插件
http://t.cn/RcXglR2
招聘:京东北京招聘ES高级工程师
https://elasticsearch.cn/article/358

编辑:金桥
归档:https://elasticsearch.cn/article/363
订阅:https://tinyletter.com/elastic-daily
继续阅读 »
1.Bulk异常引发的Elasticsearch内存泄漏
http://t.cn/RlY7tMh
2.Spring Boot 中使用 Java API 调用 Elasticsearch
http://t.cn/RljQNFJ
3.一个实时查看,搜索尾部日志事件的kibana插件
http://t.cn/RcXglR2
招聘:京东北京招聘ES高级工程师
https://elasticsearch.cn/article/358

编辑:金桥
归档:https://elasticsearch.cn/article/363
订阅:https://tinyletter.com/elastic-daily 收起阅读 »