首页 > 其他 > 详细

LEK-Introduction

时间:2015-06-23 17:30:56      阅读:273      评论:0      收藏:0      [点我收藏+]

LEK is a set of tools which can take data from any source and search, analyze, and visualize it in real time. There are three tools as below:

L- Logstash: collect , enrich , transport data

E- ElasticSearch: search , analyse data in real time

K- Kibana: explore , visualize your data

Install LEK is easy, download the related software, then extract them(tar -zxvf), cd bin, ./xxx, last, you can use them.

1.Logstash: Download, Installation, usage

  A Java runtime is required to run Logstash. We recommend running the latest version of Java. At a minimum, you need Java 7. 

wget https://download.elastic.co/logstash/logstash/logstash-1.5.0.tar.gz
tar -zxvf logstash-1.5.0.tar.gz
cd logstash-1.5.0/bin
./logstash -e ‘input { stdin { } } output { stdout { codec => rubydebug } }‘ //-f: use config file

you also can use ‘-f‘ to run config file:

./logstash -f  configFileName.conf

The configFileName.conf ‘s content like this:

input {
     stdin{}
}
output {
    elasticsearch { host => localhost }
    stdout { codec => rubydebug }
}

2.ElasticSearch: Download, Installation, usage

wget https://download.elastic.co/elasticsearch/elasticsearch/elasticsearch-1.6.0.tar.gz
tar -zxvf elasticsearch-1.6.0.tar.gz
cd elasticsearch-1.6.0/bin
./elasticsearch -d //running on backend

3.Kibana: Download, Installation, usage

wget https://download.elastic.co/kibana/kibana/kibana-4.1.0-linux-x64.tar.gz
tar -zxvf kibana-4.1.0-linux-x64.tar.gz
cd kibana-4.1.0-linux-x64/bin
./kibana

The running sequence of LEK :

First, run elasticsearch, then use ‘jps‘ to see whether it is running

$ ./elasticsearch -d
$ jps

技术分享

Second, you can run logstash, and connect it with elasticsearch

./logstash -e ‘input { stdin { } } output {   elasticsearch { host => localhost }  stdout { codec => rubydebug } }‘ //-f: use config file

Last, you can run kibana to search, view, and interact with data stored in Elasticsearch indices

./kibana

Note: Before you can start using Kibana, you need to tell it which Elasticsearch indices you want to explore. By default, Kibana connects to the Elasticsearch instance running on localhost.

To connect to a different Elasticsearch instance, modify the Elasticsearch URL in thekibana.yml configuration file and restart Kibana. For information about using Kibana with your production nodes, see 

     https://www.elastic.co/guide/en/kibana/current/production.html

The Final portal : Input your (IP:5601) on your brower

技术分享

The Logstash event processing pipeline has three stages: inputs → filters → outputs

Inputs

You use inputs to get data into Logstash. 

Filters

Filters are intermediary processing devices in the Logstash pipeline. You can combine filters with conditionals to perform an action on an event if it meets certain criteria. Some useful filters include:

  • grok: parse and structure arbitrary text. Grok is currently the best way in Logstash to parse unstructured log data into something structured and queryable. With 120 patterns built-in to Logstash, it’s more than likely you’ll find one that meets your needs!
  • mutate: perform general transformations on event fields. You can rename, remove, replace, and modify fields in your events.
  • drop: drop an event completely, for example, debug events.
  • clone: make a copy of an event, possibly adding or removing fields.
  • geoip: add information about geographical location of IP addresses (also displays amazing charts in Kibana!)

Outputs

Outputs are the final phase of the Logstash pipeline. An event can pass through multiple outputs, but once all output processing is complete, the event has finished its execution. 

For more info, this is link: https://www.elastic.co/guide/en/logstash/current/pipeline.html

 

LEK-Introduction

原文:http://www.cnblogs.com/zbhdsfdx-2015/p/4595661.html

(0)
(0)
   
举报
评论 一句话评论(0
关于我们 - 联系我们 - 留言反馈 - 联系我们:wmxa8@hotmail.com
© 2014 bubuko.com 版权所有
打开技术之扣,分享程序人生!