{
"status" : "DOWN",
"zipkin" : {
"status" : "DOWN",
"details" : {
"ElasticsearchStorage{initialEndpoints=https://xyz.es.amazonaws.com, index=zipkin}" : {
"status" : "DOWN",
"details" : {
"error" : "java.lang.IllegalArgumentException: Elasticsearch versions 5-7.x are supported, was: 1.0"
}
}
}
}
}
Hello people !
We're going to use zipkin, and we're wondering which backend to choose between cassandra and elasticsearch. I couldn't find much information detailed for zipkin's usage, are there some guidelines somewhere to help us with this choice ? pros and cons, performances or whatever
cheers !
Hello, apologies if it's a stupid question, but we've instrumented a chain of services such as A -> POST -> B -> POST -> C
. My tracing json contains details for this chained call, but on the Zipking UI I can only see the calls from A -> B
.
Would be correct to expect that call B -> C
should be shown on the same zipkin screen, as a child node of service A?
/zipkin/traces/{traceId}
sudo docker run -d -p 9411:9411 \
-e ES_HOSTS=http://172.30.5.99:3002 \
-e ES_INDEX=tracing \
-e STORAGE_TYPE=elasticsearch \
--name=zipkin_test openzipkin/zipkin
The container does not start, it stays in "starting" mode in portainer, then it stops. are the env-var ES_HOSTS
and ES_INDEX
supported by default in the image?
Connection error: ('Unable to connect to any servers', {'127.0.0.1:9042': OperationTimedOut('errors=Timed out creating connection (5 seconds), last_host=None',)})
- JAVA_OPTS=-Darmeria.ssl.key-store=/zipkin/keystore.p12 -Darmeria.ssl.key-store-password=password -Darmeria.ssl.enabled=true -Darmeria.ports[0].port=9411 -Darmeria.ports[0].protocols[0]=https
The server starts up and the UI works with https, but zipkin logs Caused by: javax.net.ssl.SSLHandshakeException: error:1000009c:SSL routines:OPENSSL_internal:HTTP_REQUEST
every 5 seconds. Any ideas what I'm doing wrong?
22/01/05 10:26:59 INFO ElasticsearchDependenciesJob: Processing spans from zipkin:span-2022-01-05/span
22/01/05 10:26:59 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Exception in thread "main" java.lang.ExceptionInInitializerError
at org.elasticsearch.hadoop.util.Version.<clinit>(Version.java:66)
at org.elasticsearch.hadoop.rest.RestService.findPartitions(RestService.java:216)
at org.elasticsearch.spark.rdd.AbstractEsRDD.esPartitions$lzycompute(AbstractEsRDD.scala:79)
at org.elasticsearch.spark.rdd.AbstractEsRDD.esPartitions(AbstractEsRDD.scala:78)
at org.elasticsearch.spark.rdd.AbstractEsRDD.getPartitions(AbstractEsRDD.scala:48)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:273)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:269)
at scala.Option.getOrElse(Option.scala:121)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:269)
at org.apache.spark.Partitioner$$anonfun$4.apply(Partitioner.scala:78)
at org.apache.spark.Partitioner$$anonfun$4.apply(Partitioner.scala:78)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
at scala.collection.immutable.List.foreach(List.scala:392)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
at scala.collection.immutable.List.map(List.scala:296)
at org.apache.spark.Partitioner$.defaultPartitioner(Partitioner.scala:78)
at org.apache.spark.rdd.RDD$$anonfun$groupBy$1.apply(RDD.scala:714)
at org.apache.spark.rdd.RDD$$anonfun$groupBy$1.apply(RDD.scala:714)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:385)
at org.apache.spark.rdd.RDD.groupBy(RDD.scala:713)
at org.apache.spark.api.java.JavaRDDLike$class.groupBy(JavaRDDLike.scala:243)
at org.apache.spark.api.java.AbstractJavaRDDLike.groupBy(JavaRDDLike.scala:45)
at zipkin2.dependencies.elasticsearch.ElasticsearchDependenciesJob.run(ElasticsearchDependenciesJob.java:189)
at zipkin2.dependencies.elasticsearch.ElasticsearchDependenciesJob.run(ElasticsearchDependenciesJob.java:170)
at zipkin2.dependencies.ZipkinDependenciesJob.main(ZipkinDependenciesJob.java:79)
Caused by: java.lang.reflect.InaccessibleObjectException: Unable to make field protected byte[] java.io.ByteArrayInputStream.buf accessible: module java.base does not "opens java.io" to unnamed module @5274766b
at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:354)
at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:297)
at java.base/java.lang.reflect.Field.checkCanSetAccessible(Field.java:178)
at java.base/java.lang.reflect.Field.setAccessible(Field.java:172)
at org.elasticsearch.hadoop.util.ReflectionUtils.makeAccessible(ReflectionUtils.java:70)
at org.elasticsearch.hadoop.util.IOUtils.<clinit>(IOUtils.java:53)
... 28 more