Skip to content

ksenji/kafka-connect-es

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 

Repository files navigation

kafka-connect-es

KafkaConnect (CopyCat) for writing data to ElasticSearch. The ElasticSearchSinkTask can be configured with the following configuration.


es.cluster=127.0.0.1:9300
es.cluster.name=elasticsearch
index=person-index
type=person
bulk.size=1200
action.type=index
es.converter=org.apache.kafka.connect.es.converter.impl.KeyIgnoringJsonConverter

es.converter is a org.apache.kafka.connect.es.converter.Converter that needs to be configured. This will take a SinkRecord object and serialize it in to JSON bytes that can be written to ElasticSearch.

If the data in Kafka is already in JSON format and if you ignore Key (or Key is null in Kafka) then you can use the org.apache.kafka.connect.es.converter.impl.KeyIgnoringJsonConverter that is available with this library.

There is org.apache.kafka.connect.es.converter.impl.KeyValueUnionJsonConverter Converter available which will combine both Key & Value and both need to be JSON data in Kafka.

If you have any other format in Kafka (for example Avro), you would have to code a Converter to convert a SinkRecord to JSON format.

This Sink takes care of fault tolerance. Only when all the records are successfully committed in ElasticSearch, it instructs KafkaConnect to procceed and commit offsets.

About

Kafka Connect for ElasticSearch Sink

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages