I am trying to upgrade a cassandra single node installation from 1.2.10 to 2.0.1.
I am using still the old RandomPartitioner, I change all the parameters in the cassandra.yaml file. But when I tried to start the node I get this error:
java.lang.RuntimeException: Can't open incompatible SSTable! Current version jb
In the notes from the documentation I can not find anything related with the change in the SSTables, did anyone deal with this problem?
If you guys had previously upgraded from an earlier version to 1.2.10, did you guys perform a 'nodetool upgradesstables'? This would have upgraded sstables to prevent possible incompatibilities.
The upgrade procedure from 1.2.9 to 2.x does not seem to recommend running it, so those sstables should be compatible, but trying it again might be worth it.
http://www.datastax.com/documentation/cassandra/1.2/webhelp/#upgrade/upgradeC_c.html#task_ds_gqm_vzr_ck
Related
We are trying to upgrade Rundeck version from 3.0.27 to the latest version (3.4.7)
We got below questions,
What is the approach to upgrade from 3.0.xxxx to 3.4.7? Could we do a direct upgrade or otherwise, if we have to run through multiple intermediary versions, what are those?
Does upgrading from 3.0.xxx to 3.4.7 fix the log4j vulnerability?
Please help to clarify above.
Thanks
Due to the big gap between 3.0.X and 3.4.X (and the upcoming 4.0.X) the best way to upgrade your instance is to create a fresh 3.4/4.0 instance and then import your projects and keys (exported previously), as a tip, try this on a non-prod environment first.
Rundeck 3.0.x uses log4j 1, so, even if the vulnerability doesn't affect explicitly the log4j 1.x EOL was in 2015 and isn't supported anymore, so, it's important to move to the latest version (Rundeck 3.4.10 at this moment).
We have apache cassandra 3.11.0 installed i see new version 3.11.1 is installed can you guys point me to the document for the bugs fixed and if any new functionalities added in latest version
3.11.1 has been released on 2017.10.10. You can check the changes and the release notes.
Currently I have Spark version 1.6.2 installed.
I want to upgrade the Spark version to the newest 2.0.1. How do I do this without losing the existing configurations?
Any help would be appreciated.
If its maven or sbt application you simply change dependency version of spark and also migrate your code according to 2.0 so you will not lose you configurations. and for spark binary you can take backup of config folder.
There is no much change related to configuration, some method signatures are changed , major changes i observed was mapPartitions method signature and some changes to metrics/listener api, apart from new features.
I am trying to build a simple project with Spark+Cassandra for a SQL-analytics demo.
I need to use Cassandra v2.0.14 (can't upgrade it for now). I am unable to find the correct version of Spark and Spark-cassandra-connector. I referred to Datastax's git project at - https://github.com/datastax/spark-cassandra-connector, and I know that the Spark and Spark-cassandra-connector versions need to match and be compatible with Cassandra. Hence, would like anyone to help pointing out the exact versions for Spark, Spark-Cassandra-connector. I tried using v1.1.0 and v1.2.1 for both Spark and Spark-Cassandra-connector - but unable to build the spark-cassandra-connector jat jar with neither the supplied sbt (fails because the downloaded sbt-launch jar just contains a 404 not found html), nor my local sbt v0.13.8 (fails for compilation error for "import sbtassembly.Plugin.", "import AssemblyKeys.")
The connector works with Cassandra 2.0 and 2.1 but some features may also work fine with 2.2 and 3.0 (not officially supported yet) using the older Java driver 2.1. This is because C* Java driver supports a wide range of Cassandra versions. The newer driver works with older C* versions, but also the older driver versions work with newer C* versions, excluding new C* features.
However, there is a one minor caveat with using C* 2.0:
Since version 1.3.0, we dropped the thrift client from the connector. This move was to simplify connectivity code and make it easier to debug - debugging one type of connection should be easier than two. It either connects or not, no more surprises of a kind "it writes fine, but can't connect for reading". Unfortunately, not all of the thrift functionality was exposed by the native protocol in C* 2.0 nor in the system tables. Therefore, if you use C* prior to version 2.1.5, automatic split sizing won't work properly and you have to tell the connector the preferred number of splits. This is to be set in ReadConf object passed at the creation of the RDD.
As for the interface between the Connector and Spark, there is much less freedom. Spark APIs change quite often and you typically need a connector dedicated to the Spark version you use. See the version table in the README.
(fails because the downloaded sbt-launch jar just contains a 404 not found html)
This looks like an SBT problem, not a connector problem.
I just tried to do sbt clean assembly on all v1.2.5, v1.3.0, b1.4 and it worked fine.
if you can upgrade version of spark then you can connect with spark with cassandra .
put following maven dependency in pom file :-
cassandra-all
cassandra-core
cassandra-mapping
cassandra-thrift
cassandra-client
spark-cassandra-connector
spark-cassandra-connector-java
this will be work.
How can I upgrade the cassandra version and thrift version in Brisk?
Is there any tutorials kind of stuffs avalailble? I changed the build.properties file but on build, the version I mentioned couldn't be located at any repository.
Im trying to upgrade because of Cassandra Insertion Error
this issue im facing. I hope the comment mentioned there helps
What is the latest version of cassandra that brisk supports?
In support forums I see people mentioning, replace the jars. I don't understand where should I do that.
Thanks for Help in advance
Regards,
Tamil
Finally I found the way to upgrade Brisk's cassandra version.
I Downloaded apache-cassandra-0.8.6-bin from apache site.
Copied
lib/apache-cassandra-0.8.6.jar and lib/apache-cassandra-thrift-0.8.6.jar
to
brisk-1.0~beta2-bin/resources/cassandra/lib/
and removed the older version's cassandra and thrift jars
and just restarted brisk with ./brisk cassandra
and tried this
$ ./nodetool -h x.x.x.x version
ReleaseVersion: 0.8.6
So, I hope there won't any compatability issues, since in a forum a brisk dev mentioned tat brisk core works independently against cassandra core. But I'm yet to know the way to do it by building from src rather replacing jars in bin
Regards,
Tamil