Cassandra is active (exited) after starting - cassandra

I am trying to get Cassandra to work on Linux Mint and I follow these instructions on Youtube (https://www.youtube.com/watch?v=pGhkX5z_vW8&list=PLalrWAGybpB-L1PGA-NfFu2uiWHEsdscD&index=4). They are the same like in the documentation:
JRE is openjdk version 1.8.0_242.
Cassandra is version 3.11.
My system is Linux Mint 19.3.
After starting Cassandra, I get this in my console:
There seems to be two problems, although I think if I fix one, I fix both:
(1) after "sudo service cassandra status" --> active (exited) should be active (running)
(2) after "nodetool status" --> Conncetion refused
Any idea? Thanks

I don't know, if this is a common problem or because of my ignorance. I could not make Cassandra work via Debian packages, but the INSTALLATION VIA BINARY TARBALL FILES worked fine on my machine. Here is the link to the official homepage:
https://cassandra.apache.org/doc/latest/getting_started/installing.html

You should check listen_address and rpc_address whether it is default or server IP if not you have to update the IP address then try nodetool status with host. It should work.

Related

Ubuntu 18.04 Server Beta 2 - restart / reboot not working

I have a fresh install of 18.04 SERVER installed on an Advantech SBC. 16.04 Server was working find but on 18.04 Server when I do a:
sudo shutdown -r now
The system starts the shutdown but stops with the last status line:
[ OK ] Stopped LVM2 metadata daemon.
I've tried some different bios configurations with power management but I can't seem to get it to restart. Only option at that point is to physically kill power and power up.
Anyone seen this? Any ideas on what to try?
On my SBC setting the South Bridge USB Configuration for Windows 8.x compatibility resolved this issue. Seems that the V4.15 kernel interactions with hardware are different enough to have caused this problem.

Run artifactory on browser

I installed artifactory on linux server. I want to know how to check whether it is properly installed.
Secondly, I want to use it's user interface on google chrome. From where I can get SERVER_DOMAIN and PORT number for it's url from installed files on server?
Thanks in advance.
Your question is a bit vague.
How did you installed it (ZIP, RPM, Docker, Deb...)?
Which database you used (default, Oracle, MySQL etc..)? You can see if the server is up and running by running the following REST API from the machine you have installed it on:
curl -iuadmin:password localhost:8081/artifactory/api/system/ping
Regarding the IP, this can be checked by running 'ifconfig' on the linux server and get the IP of that machine.
Artifactory by default is running on port 8081, unless you changed it.
It seems that you will need to follow the basic installation wiki:
https://www.jfrog.com/confluence/display/RTF/Installing+Artifactory

Datastax Enterprise Installation on Virtual Box CentOS

Can anyone please guide me step by step installation one by one for Datastax Enterprise Installation on Virtual Box CentOS .
I checked Datastax Documentation , but getting little bit confused in few steps and due to which I am not satisfied. Also checked other resources but not able to understand completely.
So Help me to know installation process one by one with all basis steps.
Thanks in advance .
You may have an easier time using OpsCenter's Lifecycle Manager to deploy DSE. (Disclaimer, I am a Lifecycle Manager dev so am biased.)
First you need to install OpsCenter in a separate VM or Centos box. If you're able to get through the Java install and yum repository setup parts of DSE setup, this won't be difficult: https://docs.datastax.com/en/opscenter/6.0/opsc/install/opscInstallRHEL_t.html
Then run an install job from LCM: https://docs.datastax.com/en/opscenter/6.0/opsc/LCM/opscLCMinstallJob.html Example the pre-requisite section of that page carefully. It will show you the things you need to do in LCM to get ready to run the job, it's all point-and-click, though.
The only pre-requisites on your target DSE machine are "python" (usually installed by default) and for the minute "which", though we'll be removing that dependency in an upcoming version.
Note at the end of this process, you'll need to provide cqlsh an IP address, username, and password to connect to the cluster... even when making a "local" connection from your DSE vm. For example: "cqlsh 192.168.1.100 -u cassandra -p the-password-you-chose-during-lcm-install"

Hadoop multi-node cluster manual installation over Ubuntu 14.04

I am a newcomer to Hadoop. For my College project we are given 4 VMs. I need to configure a multi-mode Hadoop cluster on this ( 1 master 3 slaves) and run my webapp on it. I would be using HBase in my project. Usually CentOS is used for installation and deployment of HDP, whereas I was given ubuntu. I cannot use Apache ambari plugin for installation as it is not supported in Ubuntu. I need to manually deploy them, Hence I tried looking out for tutorials.
I looked out for a tutorial to install HDP multinode clusters on ubuntu and found this [http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/]
But its too outdated (2010)
I have the official documentation here, but I am not able to follow it properly.
[http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.1-latest/bk_installing_manually_book/content/rpm-chap2-3.html] and I tried following them.
Could someone suggest me somelinks which are latest, a tutorial with decent amount of screenshots for installation of multinode clusters over Ubuntu 14.04 ( 12.04 is also fine).
Thanks a lot!!
The Michael Noll tutorial is too old, I think. I found this site:
https://www.digitalocean.com/community/tutorials/how-to-install-hadoop-on-ubuntu-13-10
I have a mini cluster (with 5 slaves and a master) in my University Lab. Ubuntu 12.04 and Hadoop 2.5.0 is there. Furthermore, I have a VM cluster in my laptop (2 slaves and a master) of Hadoop 1.2.1 on Ubuntu 12.04 too.
But I couldn't install Hadoop (any version) in Ubuntu 14.04. I don't remember the cause, but I think it was some problem with Java version (I don't check that).
I hope that help you!
I can across the same issue to install HDP 2.2 on Ubuntu 14.04, and found a solution.
I documented everything here: http://www.swiss-scalability.com/2014/12/install-hdp-22-on-ubuntu-1404-trusty.html
In a nutshell, the magic happens here:
sed -e "s/14.04/12.04/g" -i /etc/*-release
And the you can install or restart ambari-agent, it will be able to communicate with ambari-server.

Error - Apache Cassandra 2.0.5 & Datastax OpsCenter 4.1

I am using cassandra 2.0.5 on Centos 6.5 and OpsCenter 4 worked fine until i updated OpsCenter to version 4.1 . I access OpsCenter page, click on manage existing cluster and give the ip address of my node (127.0.0.1) and it gives me the following: "Error creating cluster: max() arg is an empty sequence".
Any clues ?
The bug is on 4.1.0, and is affecting those running Python 2.6. The complete fix for this is 4.1.1 (http://www.datastax.com/dev/blog/opscenter-4-1-1-now-available). To workaround this issue on 4.1.0, users should disable the auto-update feature, and manually re-populate the latest definitions. This will only need to be done once. This doesn't need to be done with 4.1.1, and that's the best fix. See the Known issues of the release notes (http://www.datastax.com/documentation/opscenter/4.1/opsc/release_notes/opscReleaseNotes410.html)
Add the following to opscenterd.conf to disable auto-update:
[definitions]
auto_update = False
Manually download the definition files
for tarball installs:
cd ./conf/definitions
for packages installs:
cd /etc/opscenter/definitions
Apply the latest definitions
curl https://opscenter.datastax.com/definitions/4.1.0/definition_files.tgz | tar xz
Restart opscenterd
I jus had today the same problem that you. I downloaded an older versions of opscenter (particulary version 4.0.2) from http://rpm.datastax.com/community/noarch/ and the error has gone.
I am also using the sam cassandra version and also on centos

Resources