@pa-vishal I guess I know what the problem is. A Queue listener is just a passive observer. It does not consume queue elements. When you have listeners registered on multiple clients then each listeners will be invoke for each item.
This is probably not what you want. Not only multiple listeners will be invoked for each item, but the items won't ever be consumer and the queue will leak memory.
You most likely do not want a queue listener, but a queue consumer. Please have a look at the reference manual how to consume queue: https://docs.hazelcast.org/docs/4.0.2/manual/html-single/index.html#taking-items-off-the-queue
@jerrinot
I checked your advice and it looks that it works. I mean almost because now we are facing with another problem.
When I set max size policy to PER_NODE and 15000 entries I observe that during requesting to our application:
and the limit is restricted so each instance never contains more than this.
For me it looks ok, I see in logs that entires are evicted.
But when i set USED_HEAP_SIZE on 100MB (does it mean 50MB per entries memory and 50MB per backup right?) I observe that when each instance reachs 50MB then eviction is started (mostly it happens after 10000 entries per instance). But when I request for example for another 1000000 entries and we have 1000000 putAsync operations, memory grows more than 50MB:
summing up the memory on each instance we still have the 100 MB but the distribution is different, only 16 MB for backup. Is this appropriate behavior?
@dawidjasiokwork
I'm glad the threshold is being respected when configured properly.
the primary/backup imbalance looks like a possible bug in eviction. for some reason backups entries are evicted more frequently then primaries.
please open a new ticket at https://github.com/hazelcast/hazelcast/issues/new and describe your scenario. thanks!
hi ,
Currently i am using client sever topology model with 5 java clients .All 5 java clients are web applications and am getting below error in logs.
Failed to execute java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask@6bb6a27
java.lang.NoClassDefFoundError: com/hazelcast/client/impl/protocol/util/BufferBuilder
could some one guide me to fix this.
and the other error is getting printed in logs continuously is as below.
Illegal access: this web application instance has been stopped already.
Could not load [com.hazelcast.client.impl.protocol.ClientMessage]. The following stack trace is thrown for debugging purposes as well as to attempt to terminate the thread which caused the illegal access.
java.lang.IllegalStateException: I
llegal access: this web application instance has been stopped already.
Could not load [com.hazelcast.client.impl.protocol.ClientMessage].
The following stack trace is thrown for debugging purposes as well as to attempt to terminate the thread which caused the illegal access.
3.12.8
Management Center image. With 4.0 release, the way Management Center connects to and communicates with Hazelcast clusters has been changed so we have modified related commands and configurations at our Helm chart as well. To start 3.12.8
Management Center, you need to use 2.10.0
as a chart version --> helm install <name> --set <parameters> --version 2.10.0 hazelcast/hazelcast
Hi all
I am getting
`[0.0.0.0]:8099 [dev] [4.0.2] While sending op error... op: com.hazelcast.internal.cluster.impl.operations.JoinMastershipClaimOp{serviceName='hz:core:clusterService', identityHash=188472785, partitionId=-1, replicaIndex=0, callId=5, invocationTime=1597066923255 (2020-08-10 19:12:03.255), waitTimeout=-1, callTimeout=60000}, error: java.lang.IllegalArgumentException: Target is this node! -> [0.0.0.0]:8099
java.lang.IllegalArgumentException: Target is this node! -> [0.0.0.0]:8099, response: ErrorResponse{callId=5, urgent=true, cause=java.lang.IllegalArgumentException: Target is this node! -> [0.0.0.0]:8099}`
// setting up the cluster
System.out.println( Arrays.asList(environment).toString());
JoinConfig join = config.getNetworkConfig().getJoin();
join.getMulticastConfig().setEnabled(false);
join.getAwsConfig().setEnabled(false);
join.getTcpIpConfig().setEnabled(true).setMembers(Arrays.asList(environment));
CacheSimpleConfig cacheConfig = new CacheSimpleConfig();
cacheConfig.setName("buckets");
config.addCacheConfig(cacheConfig);```