Anytime you want to measure distributed processes, you should look into
logging, or sending the time data asynchronously to a persistent data store. I
haven’t measured joinWithCassandra but I’ve measured other parts of spark using
a Kafka topic to send execution times. I then consumed the times
I'm using the driver from Cassandra-Spark, I would like to know if there is
a way to measure the time which takes a " joinWithCassandraTable" to
execute easily inside a spark code.
This error comes from com.datastax.driver.core.HostConnectionPool#enqueue,
which is the client side pool. Cassandra can handle more requests, the
application needs to be fixed.
As per the java docs:
/**
* Indicates that a connection pool has run out of available connections.
*
* This
Hi,
I'm getting "Pool is Busy (limit is 256)", while connecting to a single node
cassandra cluster. The whole client side application is a 3rd-party lib
which i can't change it's source and its session builder is not using any
PoolingOptions.
Is there any config on cassandra side that could