I have previously setup cassandra using the datastax community edition and have tried to move to Enterprise 4.
I've tried installing via the optscenter web interface and had it 'Start Errored: Timed out waiting for Cassandra to start.' on all 4 nodes.
I've also tried the manual approach outlined on the site. In this case just as the other it launches the dse service 'successfully'. Output.log and system.logs show the classpath as the last entry and no errors in them at all.
Java: Jre-1.7.0_51
Os: centos 6.5 Final
Vagrant box: https://github.com/2creatives/vagrant-centos/releases/download/v6.5.1/centos65-x86_64-20131205.box
My suspicion would be that your VM does not have enough memory. If physical addressable memory is smaller than the MAX_HEAP_SIZE configured in resources/cassandra/conf/cassandra-env.sh jna will go bonkers. You want to have at least 4GB of memory, or change the value in cassandra-env.sh.
Related
I have tried connecting to a remote Cassandra from Windows 10 using the latest Simba-Datastax ODBC Driver (trial version). I was successful with Cassandra 2.1 (I connected to a Cassandra docker actually) but failed with Cassandra 3.0.15 and 3.11. I have installed the driver and I am able to see it in the Windows Data sources tool (64 bits), under the System DSN tab.
When I specify the host, port and keyspace of my Cassandra 3.0 docker (exactly the same values that work allright for me with the Cassandra 2.1 docker) and press the "Test..." button to launch the connectivity test, I am getting a strange error that "not even procol version 1 is available".
According to this web site, Simba says the driver is compatible with Cassandra 3.X. Could you think of any reason why this fails but 2.1 is successful? :-(
PS: I see other people complaining here but with a different error message (No hosts available for the control connection)
I fixed it! I think I was using a wrong version of the driver - I was using Datastax driver which apparently does not work for Cassandra 3.X. I have now downloaded the latest version of the ODBC driver from the Simba website (30-day trial version) and it is working :-)
The confusion came from the fact that I thought the Datastax driver and the Simba driver were the same as I read somewhere that "Simba and Datastax have partnered to develop a driver...".
Thank you very much Aaron anyway.
We're currently evaluating memsql and have two setups. One is running on CentOS 6.7, one on CentOS 7.1.
While using CentOS 7.1, after a system reboot the master has all services started, but the CentOS 6.7 variant does not and shows that the aggregator is offline. We had to run memsql-ops cluster-start found in MemSql leaf down on Single server Cluster. We're wondering if this is related to the init.d/systemctl diffs on the machines. Any reply appreciated!
Cheers,
µatthias
currently Ops only sets up a Sys-V style init script in /etc/init.d when it is installed by root. However, once Ops starts up correctly it should immediately check whether or not MemSQL is running. If it is not running but it should be Ops will start the cluster automatically. Can you confirm that you didn't run memsql-ops cluster-stop before shutting down the cluster? If you do that, when Ops comes back up it will not start the MemSQL cluster.
me and a partner have been trying to set up a simple multinode cluster on 4 servers. The servers are running Cent Os 6.5 and each had Cassandra 2.1.3 installed. We changed Cassandra.yaml as instructed in Datastax documentation, only to run nodetool status and find just the local instance listed as UN. We already opened the internode ports (7000,7001,7199) and nothing happens, the only anomaly I've found it's the fact that Host ID seems to repeat itself in each node. Any hints on what might be the problem?
So I have a Cassandra cluster of 6 nodes on my Ubuntu machines, now I have got another machine running Windows Server 2008. I have installed DataStax Apache Cassandra on this new Windows machine, and I want to be able to run all the CQL commands from Windows machine onto Ubuntu machines. So its like remote command execution.
I tried opening cqlsh in cmd using cqlsh with the IP of my one of the nodes and port like cqlsh 192.168.4.7 9160
But I can't seem to make it work. Also I don't want to add the new machine to my existing cluster Please suggest.
Provided version 3.1.1 is not supported by this server (supported: 2.0.0, 3.0.5)
any workaround u could suggest?
Basically, you have two options here. The harder one would be to upgrade your cluster (the tough, long-term solution). But there have been many improvements since 1.2.9 that you could take advantage of. Not to mention bugs fixed long ago that you may be running into.
The other, quicker option would be to install 1.2.9 on your Windows machine. Probably the easiest way to do this, would be to zip-up your Cassandra dir on Ubuntu (minus the data, commitlog, and saved caches dirs of course), copy it to your Windows machine, and expand it. Then the cqlsh versions would match-up, and you could solve your immediate problem.
I started with a Windows 7 computer, and set up an Ubuntu Linux virtual machine which I run using VirtualBox. The Cloudera Manager Free Edition version 4 has been executed, and I have been following the prompts on localhost:7180.
I am now stuck when the prompt asks me to "Specify hosts for your CDH cluster installation." Can I install all of the Hadoop components, as well as run them, in the linux virtual machine alone?
Please help point me in the right direction in which host I should specify.
Yes, you can run cdh in a linux virtual machine alone. You could do it using "standalone" or "pseudo distributed" modes. IMHO, the most effective method for doing it is to use the "pseudo distributed" mode.
In this case, there are multiple java-virtual-machines (JVM) running, so they simulated as they were a cluster with multiples nodes (each thread simulated to be a cluster node).
Cloudera has documented how to get deployed as "pseudo distributed":
https://www.cloudera.com/documentation/enterprise/5-6-x/topics/cdh_qs_cdh5_pseudo.html
Note: 3 ways for deploying cdh:
standalone: using a machine alone, with a unique jvm
pseudo-distributed: using a machine alone, but several jvm's, so
simulated to be a cluster
distributed: using a cluster, so several
nodes with different purposes (workers, namenode, etc).
you can specify hostname of your machine. it will install everything on your machine only.