Hi Bela Ban,

We have been trying to use JGroups 3.6.8 with Infinispan 8.2.2., in this 
context we are observing an issue with the Replication Mode of Infinispan usage.
We see the below exception from Jgroups:
JGRP000029: failed sending message to <>:7800 (100 bytes): 
java.lang.NullPointerException, headers: TCPGOSSIP: [type=GET_MBRS_REQ, 
cluster=x-cluster], TP: [cluster_name=x-cluster]

The configurations used are highlighted below:

Infinispan config

<infinispan>

    <jgroups>

        <stack-file name="configurationFile" path="config/jgroups.xml"/>

    </jgroups>

    <cache-container>

        <transport cluster="x-cluster" stack="configurationFile" />

<replicated-cache name="transactional-type" mode="SYNC">

            <transaction mode="NON_XA"  locking="OPTIMISTIC" 
transaction-manager-lookup="org.infinispan.transaction.lookup.JBossStandaloneJTAManagerLookup"
 auto-commit="true"  />

<locking acquire-timeout="60000"/>

<expiration lifespan="43200000"/>

        </replicated-cache>

    </cache-container>

</infinispan>

 

Jgroups configuration

<!--

    TCP based stack, with flow control and message bundling. This is usually 
used when IP

    multicasting cannot be used in a network, e.g. because it is disabled 
(routers discard multicast).

    Note that TCP.bind_addr and TCPPING.initial_hosts should be set, possibly 
via system properties, e.g.

    -Djgroups.bind_addr=192.168.5.2 and 
-Djgroups.tcpping.initial_hosts=192.168.5.2[7800]".

    author: Bela Ban

-->

<config xmlns="urn:org:jgroups"

        xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance";

xsi:schemaLocation="urn:org:jgroups  
http://www.jgroups.org/schema/jgroups-3.6.xsd";>

 

  <TCP loopback="true"

bind_addr="${jgroups.tcp.address:127.0.0.1}"

bind_port="${jgroups.tcp.port:7800}"

recv_buf_size="${tcp.recv_buf_size:20M}"

send_buf_size="${tcp.send_buf_size:640K}"

discard_incompatible_packets="true"

max_bundle_size="64K"

max_bundle_timeout="30"

enable_bundling="true"

       use_send_queues="true"

sock_conn_timeout="300"

timer_type="new"

timer.min_threads="4"

timer.max_threads="10"

timer.keep_alive_time="3000"

timer.queue_max_size="500"

thread_pool.enabled="true"

thread_pool.min_threads="2"

thread_pool.max_threads="30"

thread_pool.keep_alive_time="60000"

thread_pool.queue_enabled="false"

thread_pool.queue_max_size="100"

thread_pool.rejection_policy="discard"

oob_thread_pool.enabled="true"

       oob_thread_pool.min_threads="2"

oob_thread_pool.max_threads="30"

oob_thread_pool.keep_alive_time="60000"

oob_thread_pool.queue_enabled="false"

oob_thread_pool.queue_max_size="100"

oob_thread_pool.rejection_policy="discard"/>

 

    <!-- <TCP_NIO -->

<!-- bind_port="7800" -->

<!-- bind_interface="${jgroups.tcp_nio.bind_interface:bond0}" -->

<!-- use_send_queues="true" -->

<!-- sock_conn_timeout="300" -->

<!--         reader_threads="3" -->

<!-- writer_threads="3" -->

<!-- processor_threads="0" -->

<!-- processor_minThreads="0" -->

<!-- processor_maxThreads="0" -->

<!-- processor_queueSize="100" -->

<!--         processor_keepAliveTime="9223372036854775807"/> -->

    <TCPGOSSIP initial_hosts="${jgroups.tcpgossip.initial_hosts}"/>

    <!--    <TCPPING async_discovery="true" 
initial_hosts="${jgroups.tcpping.initial_hosts}"

            port_range="2" timeout="3000" /> -->

    <MERGE2 max_interval="30000" min_interval="10000"/>

    <FD_SOCK/>

    <FD timeout="3000" max_tries="3"/>

    <VERIFY_SUSPECT timeout="1500"/>

    <pbcast.NAKACK

use_mcast_xmit="false"

retransmit_timeout="300,600,1200,2400,4800"

discard_delivered_msgs="false"/>

    <UNICAST2 timeout="300,600,1200"

stable_interval="5000"

max_bytes="1m"/>

    <pbcast.STABLE stability_delay="500" desired_avg_gossip="5000" 
max_bytes="1m"/>

    <pbcast.GMS print_local_addr="false" join_timeout="3000" 
view_bundling="true"/>

    <UFC max_credits="200k" min_threshold="0.20"/>

    <MFC max_credits="200k" min_threshold="0.20"/>

    <FRAG2 frag_size="60000"/>

    <RSVP timeout="60000" resend_interval="500" ack_on_delivery="false" />

</config>

Any help on this would be really great.
Kindly let us know if you would need any further information on this.

Regs,
Manohar.

_______________________________________________
infinispan-dev mailing list
infinispan-dev@lists.jboss.org
https://lists.jboss.org/mailman/listinfo/infinispan-dev

Reply via email to