Upgrade Arangodb on Mac OS - arangodb

What is the best procedure to use to upgrade from 2.1.2 to 2.2.0 on Mac OS?
I have tried to use:
$ brew upgrade arangodb
That results in a message that 2.1.2 is already installed.
Do I need to remove 2.1.2 and then install 2.2.0? If so what happens to existing databases?
thanks
geoff coleman

The brew formular is not yet updated. You can use
brew upgrade https://raw.githubusercontent.com/fceller/homebrew/master/Library/Formula/arangodb.rb
for the time being.
The database needs to be converted after upgrading. Please execute
/usr/local/opt/arangodb/sbin/arangod --log.file - --upgrade
This will upgrade your data files to 2.2.

Related

How to downgrade Thrift version on Ubuntu?

I am using Thrift version 0.14.0 on my Ubuntu (20.04.2 LTS).
However, I need to downgrade it to version 0.9.3 for working purposes.
I try to replace version 0.9.3, remove the current thrift version but it doesn't work.
Despite many ways to install the custom version, the current version is still 0.14.0 (thrift --version)
Can you give me the solution? Thank you so much.

Unable to install robotframework-ride in ubuntu

I am trying to install robotframwrok-ride in ubuntu18.04 but getting "could not find a version satisfies the requirement Pywin32" error.
please help me to get the ride installed.
That problem is fixed in the next release, which is 1.7.4.
You can install the Beta version by following the instructions in the release notes, https://github.com/robotframework/RIDE/blob/master/doc/releasenotes/ride-1.7.4.rst
You can also install the current development version, which includes more fixes.
pip install -U https://github.com/robotframework/RIDE/archive/master.zip
(edit)
The version 1.7.4 has been released on December the 24th, 2019.
You can install (after having installed wxPython) with:
sudo pip install robotframework-ride

How to install apache-spark 2.3.3 with homebrew on Mac

$brew install apache-spark only installs the most recent version of spark (2.4) and $brew search apache-spark does not give any other options. Is there a way to install an older version of spark with homebrew?
Type
brew tap eddies/spark-tap
Then install the specific version
brew install apache-spark#2.3.3
brew tap eddies/spark-tap
brew install apache-spark#2.3.2

How to update OpenCV 3.4.2 to OpenCV 4 or latest stable version?

I have installed OpenCV 3.4.2 successfully by following the tutorial given here:
https://www.pyimagesearch.com/2015/07/20/install-opencv-3-0-and-python-3-4-on-ubuntu/
Now, I would like to update to openCV 4 or the latest stable version.
Do I need to uninstall 3.4.2 first?
If so, how should I uninstall it.
I am afraid to creating another virtual environment and installing version 4 or master package from github by following the same steps might create conflicts. Please advice.
Working on Ubuntu 16.04 LTS, python 3.5
For Python interface, I guess you can try something like pip install opencv-python==4.0.0.21. Note, you might need to run pip3 install opencv-python==4.0.0.21 depending upon your pip version.

cqlsh not able to locate cassandra running on localhost [duplicate]

I've tried all the measures from this post and Cassandra doc.
I've tried running all the versions of Cassandra including the latest release 3.7 from tarball and Debian package, but I keep getting errors when I execute cqlsh.
Error:
Connection error: ('Unable to connect to any servers', {'127.0.0.1': TypeError('ref() does not take keyword arguments',)})
I had no problem running Cassandra before I upgraded my Linux Mint from 17.3 to 18.
I believe I installed all the necessary packages such as java 8 and python 2.7.12.
I think the problem exists in cassandra.yaml file since the default setting isn't working, but I'm not sure how to configure properly to get it running.
Any suggestions appreciated.
You are running into CASSANDRA-11850, where cqlsh breaks with Python 2.7.11+. This ticket has been marked as "Resolved" and a patch has been applied to Cassandra 3.9 which has not been released yet.
I believe I installed all the necessary packages such as java 8 and python 2.7.12.
In the interim (until 3.9 is released) you can roll back to Python 2.7.10, and cqlsh should work (not trivial). Otherwise, DataStax DevCenter should work with Cassandra 3.7.
Edit 20161020
Cassandra 3.9 was released a few weeks ago, and can now be downloaded.
refer https://issues.apache.org/jira/browse/CASSANDRA-11850
After setting environment variable
CQLSH_NO_BUNDLED=TRUE
it resolved.
I use windows 7, python 2.7.12, cassandra 3.7
Need to add following command
sudo apt install python-pip
pip install cassandra-driver
export CQLSH_NO_BUNDLED=true
This works for Ubuntu 16.04 in Amazon EC2:
sudo apt-get --no-install-recommends install python-cassandra python3-cassandra
CQLSH_NO_BUNDLED=TRUE cqlsh "$(ec2metadata --local-ipv4)"
Even if after rolling back to python 2.7.10, the issue persists. It means the python 2.7.10 is not set as the default python version.
Go to /usr/bin directory and check the different python versions available, say python2.7 corresponds to version 2.7.10 ( you can check it by running command python2.7 in your terminal and the python version will be mentioned in the first line of the Interpreter,try the same with all other python versions available in the folder to find the one which corresponds to version 2.7.10).
Now, use the following commands to make correct python version (python2.7 in my case) as default choice
update-alternatives --install /usr/bin/python python /usr/bin/python2.7
It may be possible as you have not installed cassandra-driver.
As I also faced the same problem and I solved using following such steps.
Try installing python pip then install cassandra-driver.
1.sudo apt install python-pip
2.pip install cassandra-driver

Resources