-
Notifications
You must be signed in to change notification settings - Fork 121
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Kafka Input Polling issue #265
Comments
Same problem here :( |
(stealing this response from @original-brownbear ) Kafka tracks the individual consumers in a consumer group (i.e. a number of LS instances) and tries to give each consumer one or more specific partitions of the data in the topic they’re consuming. Fixing the problem is easy by reducing the number of records per request that LS polls from the Kafka Broker in on request, reducing the number of Kafka input threads and/or increasing the relevant timeouts in the Kafka Consumer configuration. The number of records to pull in one request is set by the option |
Hi @andrewvc, thanks for sharing the detailed tuning steps. What changes you suggest in case of slower indexing rates than incoming messages ? After making these changes and running pipeline for couple of days, I am encountering heap issues on logstash indexer. Error are something like this,
|
@andrewvc any additional information on this? Also is there a way to have the plugin restart after it disconnects? (some sort of retry logic) |
Moved from elastic/logstash#9421
More Info:
With Kafka input and Elasticsearch output, logstash stops if elasticsearch is unreachable for a while.
The process is still running, but doesnt process new data.
The logstash is running in Kubernetes, so i have the problem that i need to manualy restart it each time ES was not reachable. IMHO it should end/stop the sevice or wait for the cluster to get up again.
Log output while ES is down (for some hours in my case):
Log output at the time es is up again:
The text was updated successfully, but these errors were encountered: