Cassandra reads slow when add new node - cassandra

I'm using cassandra 3.11.0 on Ubuntu 16.04 (VM).
So with 2 nodes in same cluster using 0.8s for select * from emp.
But when i add new node in same cluster it using 2.0s
Keyspace
class: SimpleStrategy
replication_factor: 2
Table
CREATE TABLE emp (key int PRIMARY KEY,empname text);
Cassandra.yaml
Node1 and Node2 have a same config
autobootstrap: false
seed provider: "node1,node2"
num_tokens: 256
endpoint_snitch: GossipingPropertyFileSnitch
Node3 (new node) config
autobootstrap: true
seed provider: "node1,node2"
num_tokens: 256
endpoint_snitch: GossipingPropertyFileSnitch

Related

Best approach to remove cassandra-topology.properties file in running cluster nodes

There is 3 node cassandra cluster running and which is serving Production Traffic And in cassandra.yaml file "endpoint_snitch: GossipingPropertyFileSnitch" is configured but somehow we have forgot to remove file cassandra-topology.properties from cassandra conf directory. As per Cassandra documentation if you are using GossipingPropertyFileSnitch you should remove cassandra-topology.properties file.
Now As all three nodes are running and serving Production traffic So can I remove this file all three nodes or I have to remove this file after shutdown the nodes one by one.
Apache Cassandra Version is "3.11.2"
./bin/nodetool status
Datacenter: dc1
================
Status=Up/Down
|/ State=Normal/Leaving/Joining/Moving
-- Address Load Tokens Owns (effective) Host ID Rack
UN x.x.x.x1 409.39 GiB 256 62.9% cshdkd-6065-4813-ae53-sdh89hs98so RAC1
UN x.x.x.x2 546.33 GiB 256 67.8% jfdsdk-f18f-4d46-af95-33jw9yhfcsd RAC2
UN x.x.x.x3 594.73 GiB 256 69.3% 7s9skk-a27f-4875-a410-sdsiudw9eww RAC3
If the cluster is already migrated to GossippingPropertyFileSnitch, then you can safely remove that file without stopping the cluster nodes. See the item 7 in DSE 5.1 documentation (compatible with Cassandra 3.11)

Can't Connect to Cassandras default cluster (Test Cluster) Using Opscenter

My Error is as below
OpsCenter was not able to add the cluster: OpsCenter was unable to resolve the ip for Test_Cluster, please specify seed nodes using the rpc_address
My OS is CentOS 7
I install DSE 6
i found that datastax forbidden my ip

how to connect Mesos to existing dse cluster?

I have a dse cluster base on docker. Now I want to use Mesos to manage my dse cluster for running spark job. So I installed Mesos master, Mesos slave docker container. What should I do next to connect dse to mesos?

Solr: How to upgrade solrcloud 5.5.1 to 6.6.0

I want to upgrade the solrcloud to the newest version in CentOS.IS there have a step by step to do this process. and I don't want to loose my indexes also.
Lets answers your question by taking example. Assume that below is your Solr Cloud structure.
Shard1
node1 (20 docs) (5.5.1)
node2 (20 docs) (5.5.1)
Shard2
node3 (20 docs) (5.5.1)
node4 (20 docs) (5.5.1)
Now you will add a replica to both of the shards.
First you add two nodes to Zookeeper. But this both nodes will be of Solr 6.6.0.
bin/solr start -cloud -s /Users/XYZ/Downloads/solr-6.5.0/example/cloud/node6/solr -p 8957 -z localhost:9983
Add one replica to Both shards
localhost:8983/solr/admin/collections?action=ADDREPLICA&collection=collectionName&shard=shard1
localhost:8983/solr/admin/collections?action=ADDREPLICA&collection=collectionName&shard=shard2
After adding replicas your Solr Cloud Structure will look like :-
Shard1
node1 (20 docs) (5.5.1)
node2 (20 docs) (5.5.1)
node5 (20 docs) (6.6.0)
Shard2
node3 (20 docs) (5.5.1)
node4 (20 docs) (5.5.1)
node6 (20 docs) (6.6.0)
Remove node1, node2, node3 and node4. Your cloud structure will be in Solr 6.6.0.

Datastax Spark worker is always looking for master at 127.0.0.1

I am trying to bring up datastax cassandra in analytics mode by using "dse cassandra -k -s". I am using DSE 5.0 sandbox on a single node setup.
I have configured the spark-env.sh with SPARK_MASTER_IP as well as SPARK_LOCAL_IP to point to my LAN IP.
export SPARK_LOCAL_IP="172.40.9.79"
export SPARK_MASTER_HOST="172.40.9.79"
export SPARK_WORKER_HOST="172.40.9.79"
export SPARK_MASTER_IP="172.40.9.79"
All above variables are setup in spark-env.sh.
Despite these, the worker will not come up. It is always looking for a master at 127.0.0.1.This is the error i am seeing in /var/log/cassandra/system.log
WARN [worker-register-master-threadpool-8] 2016-10-04 08:02:45,832 SPARK-WORKER Logging.scala:91 - Failed to connect to master 127.0.0.1:7077
java.io.IOException: Failed to connect to /127.0.0.1:7077
Result from dse client-tool shows 127.0.0.1
$ dse client-tool -u cassandra -p cassandra spark master-address
spark://127.0.0.1:7077
However i am able to access the spark web UI from the LAN IP 172.40.9.79
Spark Web UI screenshot
Any help is greatly appreciated
Try add in file spark-defaults.conf this parameter: spark.master local[*] or spark.master 172.40.9.79. Maybe this solves your problem

Resources