You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
INFO ElasticsearchDependenciesJob: Running Dependencies job for 2020-10-29T00:00Z, reading from hour-jaeger-span-2020-10-29 index, result storing to hour-jaeger-dependencies-2020-10-29
[Stage 0:> ERROR Executor: Exception in task 3.0 in stage 0.0 (TID 3)
org.elasticsearch.hadoop.rest.EsHadoopInvalidRequest: The number of slices [1217] is too large. It must be less than [1024]. This limit can be set by changing the [index.max_slices_per_scroll] index level setting.
spark-dependencies running,but yesterday spark-dependencies error,log info:
i find source is :https://github.com/jaegertracing/spark-dependencies/blob/master/jaeger-spark-dependencies-elasticsearch/src/main/java/io/jaegertracing/spark/dependencies/elastic/ElasticsearchDependenciesJob.java
run method i find 230 lines:
i read https://www.elastic.co/guide/en/elasticsearch/reference/current/paginate-search-results.html#scroll-search-results
i update the index.max_slices_per_scroll index setting, i use 2048 but log info:
what should I do ? how to run job ????
The text was updated successfully, but these errors were encountered: