Cassandra version 2.1.3 java.lang.AssertionError: attempted to delete non-existing file system-compaction_history - cassandra

I upgraded from Cassandra 2.1.2 to 2.1.3 following the upgrade instructions. However, I started to see this new java.lang.AssertionError. See log below.
Environment: Windows 7 (yes I konw...), Java 1.8.0_11
ERROR [NonPeriodicTasks:1] 2015-04-05 20:34:39,638 CassandraDaemon.java:167 - Exception in thread Thread[NonPeriodicTasks:1,5,main]
java.lang.AssertionError: attempted to delete non-existing file system-compaction_history-tmplink-ka-548-Data.db
at org.apache.cassandra.io.util.FileUtils.deleteWithConfirm(FileUtils.java:126) ~[apache-cassandra-2.1.3.jar:2.1.3]
at org.apache.cassandra.io.sstable.SSTableReader$Tidier$1.run(SSTableReader.java:2072) ~[apache-cassandra-2.1.3.jar:2.1.3]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) ~[na:1.8.0_11]
at java.util.concurrent.FutureTask.run(FutureTask.java:266) ~[na:1.8.0_11]
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) ~[na:1.8.0_11]
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) ~[na:1.8.0_11]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) ~[na:1.8.0_11]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [na:1.8.0_11]
at java.lang.Thread.run(Thread.java:745) [na:1.8.0_11]
Does anyone know what this error is about and how I can correct this? Or if there is already an open JIRA at Apache?

See open JIRA ticket at the link below on this issue. Basically this is potentially a known issue. bug has been filed.
https://issues.apache.org/jira/browse/CASSANDRA-9121

Related

Redhawk V3 - IDE Not displaying domain correctly

On a fresh install of Redhawk version 3 everything appears to be working and interacting with the domain works fine through Python. I appreciate the IDE has been deprecated but I am wondering if there is a fix to an issue when inspecting the domain. When connecting to a domain the IDE is producing the following error
Failed to fetch profile object from profile path: 'sca:///domain/DomainManager.dmd.xml?domain=REDHAWK_DEV&fs=IOR%3A000000000000001749444C3A43462F46696C654D616E616765723A312E300000000000010000000000000070010102000B00000031302E31312E322E3132000022CC00001C000000FF446F6D61696E4D616E61676572FEDD11376201003E5C00000000000200000000000000080000000100000000545441010000001C00000001000000010001000100000001000105090101000100000009010100#/'
The full stack indicates it is trying to access a file that doesn't exist (DomainManager.dmd.xml, which doesn't exist), full stack
org.eclipse.emf.ecore.resource.impl.ResourceSetImpl$1DiagnosticWrappedException: org.eclipse.core.runtime.CoreException: File cache error: /home/centos/workspace/.metadata/.plugins/gov.redhawk.sca.efs/fileCache/sdr_REDHAWK_DEV/dom/domain/DomainManager.dmd.xml at org.eclipse.emf.ecore.resource.impl.ResourceSetImpl.handleDemandLoadException(ResourceSetImpl.java:319) at org.eclipse.emf.ecore.resource.impl.ResourceSetImpl.demandLoadHelper(ResourceSetImpl.java:278) at org.eclipse.emf.ecore.resource.impl.ResourceSetImpl.getResource(ResourceSetImpl.java:406) at org.eclipse.emf.ecore.resource.impl.ResourceSetImpl.getEObject(ResourceSetImpl.java:220) at gov.redhawk.model.sca.ProfileObjectWrapper$Util.fetchProfileObject(ProfileObjectWrapper.java:225) at gov.redhawk.model.sca.impl.ScaPropertyContainerImpl.fetchProfileObject(ScaPropertyContainerImpl.java:379) at gov.redhawk.model.sca.impl.ScaPropertyContainerImpl.fetchPropertyDefinitions(ScaPropertyContainerImpl.java:708) at gov.redhawk.model.sca.impl.ScaPropertyContainerImpl.fetchProperties(ScaPropertyContainerImpl.java:611) at gov.redhawk.model.sca.impl.ScaDomainManagerImpl.fetchAttributes(ScaDomainManagerImpl.java:2187) at gov.redhawk.model.sca.impl.CorbaObjWrapperImpl.refresh(CorbaObjWrapperImpl.java:761) at gov.redhawk.sca.model.provider.refresh.internal.RefresherSwitch$1.refresh(RefresherSwitch.java:90) at gov.redhawk.sca.model.provider.refresh.internal.RefresherSwitch$1.refresh(RefresherSwitch.java:84) at gov.redhawk.sca.model.provider.refresh.internal.RefreshTasker.refresh(RefreshTasker.java:264) at gov.redhawk.sca.model.provider.refresh.internal.RefreshTasker$RefreshTask.run(RefreshTasker.java:214) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:750) Caused by: org.eclipse.core.runtime.CoreException: File cache error: /home/centos/workspace/.metadata/.plugins/gov.redhawk.sca.efs/fileCache/sdr_REDHAWK_DEV/dom/domain/DomainManager.dmd.xml at gov.redhawk.efs.sca.internal.cache.FileCache.downloadFile(FileCache.java:159) at gov.redhawk.efs.sca.internal.cache.FileCache.update(FileCache.java:74) at gov.redhawk.efs.sca.internal.cache.FileCache.openInputStream(FileCache.java:84) at gov.redhawk.efs.sca.internal.ScaFileStore.openInputStream(ScaFileStore.java:306) at sun.reflect.GeneratedMethodAccessor25.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.eclipse.emf.ecore.resource.impl.EFSURIHandlerImpl.createInputStream(EFSURIHandlerImpl.java:249) at org.eclipse.emf.ecore.resource.impl.ExtensibleURIConverterImpl.createInputStream(ExtensibleURIConverterImpl.java:360) at org.eclipse.emf.ecore.resource.impl.ResourceImpl.load(ResourceImpl.java:1314) at org.eclipse.emf.ecore.resource.impl.ResourceSetImpl.demandLoad(ResourceSetImpl.java:259) at org.eclipse.emf.ecore.resource.impl.ResourceSetImpl.demandLoadHelper(ResourceSetImpl.java:274) ... 19 more Caused by: java.util.concurrent.ExecutionException: org.omg.CORBA.MARSHAL: Server-side Exception: null vmcid: 0x41540000 minor code: 10 completed: No at mil.jpeojtrs.sca.util.ProtectedThreadExecutor.submit(ProtectedThreadExecutor.java:86) at gov.redhawk.efs.sca.internal.cache.FileCache.readProtected(FileCache.java:199) at gov.redhawk.efs.sca.internal.cache.FileCache.copyLarge(FileCache.java:187) at gov.redhawk.efs.sca.internal.cache.FileCache.downloadFile(FileCache.java:137) ... 30 more Caused by: org.omg.CORBA.MARSHAL: Server-side Exception: null vmcid: 0x41540000 minor code: 10 completed: No at sun.reflect.GeneratedConstructorAccessor26.newInstance(Unknown Source) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at org.jacorb.orb.SystemExceptionHelper.read(SystemExceptionHelper.java:222) at org.jacorb.orb.ReplyReceiver.getReply(ReplyReceiver.java:456) at org.jacorb.orb.Delegate._invoke_internal(Delegate.java:1419) at org.jacorb.orb.Delegate.invoke_internal(Delegate.java:1188) at org.jacorb.orb.Delegate.invoke(Delegate.java:1176) at org.omg.CORBA.portable.ObjectImpl._invoke(ObjectImpl.java:475) at CF._FileStub.read(_FileStub.java:66) at gov.redhawk.efs.sca.internal.ScaFileInputStream.read(ScaFileInputStream.java:84) at java.io.InputStream.read(InputStream.java:101) at gov.redhawk.efs.sca.internal.cache.FileCache$1.call(FileCache.java:203) at gov.redhawk.efs.sca.internal.cache.FileCache$1.call(FileCache.java:1) at mil.jpeojtrs.sca.util.ProtectedThreadExecutor.submit(ProtectedThreadExecutor.java:84) ... 33 more
Confirmed files exist on Redhawk V2 system, removed and re-added workspace directory
It seems that the issue was fixed in v3.0.1 - https://github.com/RedhawkSDR/redhawk/releases/tag/3.0.1.
The effort in REDHAWK 3.0.1 focused on:
Replace log4j with reload4j in the Core Framework and in the IDE.
Resolved bug preventing fei3-generator from generating python code
Resolved bug preventing IDE from connecting to a running redhawk-3.0.0 domain

Error when running Zeppelin to connect Neo4j & Spark

I'm trying to run : https://github.com/conker84/zep-neo4j-gc2k18
The 1st paragraph Add Neo4j-Spark-Connector dependency:
%spark.dep
z.reset()
z.load("neo4j-contrib:neo4j-spark-connector:2.2.1-M5")
but an error occurred as follow:
java.lang.NullPointerException
at org.sonatype.aether.impl.internal.DefaultRepositorySystem.resolveDependencies(DefaultRepositorySystem.java:352)
at org.apache.zeppelin.spark.dep.SparkDependencyContext.fetchArtifactWithDep(SparkDependencyContext.java:171)
at org.apache.zeppelin.spark.dep.SparkDependencyContext.fetch(SparkDependencyContext.java:121)
at org.apache.zeppelin.spark.DepInterpreter.interpret(DepInterpreter.java:247)
at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:103)
at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:633)
at org.apache.zeppelin.scheduler.Job.run(Job.java:188)
at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:140)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
What should I do to correct the error?
You are looking at an outdated version of Zeppelin integration. Nowadays Zeppelin supports querying Neo4j out of the box with https://zeppelin.apache.org/docs/0.8.0/interpreter/neo4j.html.
You can check out how I have done it in my repo: https://github.com/tomasonjo/zeppelin-graph-algo

Spark2 Zeppelin interpreter dependency results in Null pointer NullPointerException

I am trying to add dependency to Spark2 interpreter in Zeppelin as follows:
org.bdgenomics.adam:adam-core-spark2_2.11:0.23.0
There are no jars at localRepo location. So after I add this
dependency and run a simple command like println("hello world") I
am getting this NullPointerException:
java.lang.NullPointerException
at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:861)
at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493)
at org.apache.zeppelin.scheduler.Job.run(Job.java:175)
at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Now there are a lot of packages with jars in local-repo folder and
also under local-repo/2C4U48MY3_spark2 which is the location for
parameter zeppelin.interpreter.localRepo.
Here is zeppelin spark2 interpreter log: http://pasted.co/06cbc12a
Without this dependency, everything works fine.
I'm unable to say why I am getting this NPE. Can you help me please?

org.apache.thrift.protocol.TProtocolException: Bad version in readMessageBegin

I'm trying to connect cassandra with zeppelin. Every time I try to execute a query I get the following error. What should I do?
%cassandra
select * from rawtweet1
org.apache.thrift.protocol.TProtocolException: Bad version in readMessageBegin
at org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:223)
at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69)
at org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$Client.recv_createInterpreter(RemoteInterpreterService.java:184)
at org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$Client.createInterpreter(RemoteInterpreterService.java:168)
at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.init(RemoteInterpreter.java:172)
at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.getFormType(RemoteInterpreter.java:328)
at org.apache.zeppelin.interpreter.LazyOpenInterpreter.getFormType(LazyOpenInterpreter.java:105)
at org.apache.zeppelin.notebook.Paragraph.jobRun(Paragraph.java:260)
at org.apache.zeppelin.scheduler.Job.run(Job.java:176)
at org.apache.zeppelin.scheduler.RemoteScheduler$JobRunner.run(RemoteScheduler.java:328)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
The problem was in the configuration of the Cassandra interpreter. I had to change it back to the default settings and than everything worked.

Apache Zeppelin configuration with Spark

I have been trying to configure Apache Zeppeling with Spark 2.0. I managed to install them both on a linux os and I set the the spark on the 8080 port while zeppelin server on the 8082 port number.
In the zeppelin-env.sh file from zeppelin I set the SPARK_HOME variable to the location of the Spark folder.
However when I try to create a new node nothing compiles properly. From what it seems I didn't configure the interpreters as the interpreter tab is missing from in the home tab.
Any help would be much appreciated.
EDIT: E.I. when I am trying to run the zeppelin tutorial, the 'Load data into table' process I receive the following error:
java.lang.ClassNotFoundException:
org.apache.spark.repl.SparkCommandLine at
java.net.URLClassLoader.findClass(URLClassLoader.java:381) at
java.lang.ClassLoader.loadClass(ClassLoader.java:424) at
sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331) at
java.lang.ClassLoader.loadClass(ClassLoader.java:357) at
org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:400)
at
org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
at
org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:93)
at
org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:341)
at org.apache.zeppelin.scheduler.Job.run(Job.java:176) at
org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266) at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
I don't think it's possible to use spark 2.0 without building from
source, since some relatively big changes have happened with this release.
You can clone the zeppelin git repo and build using the spark 2.0 profile as mentioned in the readme on github https://github.com/apache/zeppelin.
I've tried it and it works.

Resources