I am using CQLSSTableWriter in an application which prepares sstables for bulk load into Cassandra 2.1.2 using sstableloader.
"Sometimes" I encounter this exception in the application's output:
14:23:58.694 [Thread-4] DEBUG o.apache.cassandra.io.util.FileUtils - Renaming /mnt/datasstables-generated/2006/05/sensordata/packets_area_cell1/sensordata-packets_area_cell1-tmp-ka-318-CompressionInfo.db to /mnt/datasstables-generated/2006/05/sensordata/packets_area_cell1/sensordata-packets_area_cell1-ka-318-CompressionInfo.db
14:23:58.695 [Thread-4] DEBUG o.apache.cassandra.io.util.FileUtils - Renaming /mnt/datasstables-generated/2006/05/sensordata/packets_area_cell1/sensordata-packets_area_cell1-tmp-ka-318-Data.db to /mnt/datasstables-generated/2006/05/sensordata/packets_area_cell1/sensordata-packets_area_cell1-ka-318-Data.db
Exception in thread "Thread-7" java.lang.ArrayIndexOutOfBoundsException: 132385
at org.apache.cassandra.db.ArrayBackedSortedColumns.append(ArrayBackedSortedColumns.java:196)
at org.apache.cassandra.db.ArrayBackedSortedColumns.sortCells(ArrayBackedSortedColumns.java:167)
at org.apache.cassandra.db.ArrayBackedSortedColumns.maybeSortCells(ArrayBackedSortedColumns.java:125)
at org.apache.cassandra.db.ArrayBackedSortedColumns.access$1100(ArrayBackedSortedColumns.java:44)
at org.apache.cassandra.db.ArrayBackedSortedColumns$CellCollection.iterator(ArrayBackedSortedColumns.java:622)
at org.apache.cassandra.db.ColumnFamily.iterator(ColumnFamily.java:476)
at org.apache.cassandra.db.ColumnFamily.getColumnStats(ColumnFamily.java:429)
at org.apache.cassandra.io.sstable.SSTableWriter.append(SSTableWriter.java:225)
at org.apache.cassandra.io.sstable.SSTableSimpleUnsortedWriter$DiskWriter.run(SSTableSimpleUnsortedWriter.java:215)
Exception in thread "main" java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:483)
at dk.company.commons.app.CliCommandList$1.execute(CliCommandList.java:50)
at dk.company.commons.app.CliCommandList.invoke(CliCommandList.java:80)
at dk.company.ais.store.Main.main(Main.java:33)
Caused by: java.lang.ArrayIndexOutOfBoundsException: 132385
at org.apache.cassandra.db.ArrayBackedSortedColumns.append(ArrayBackedSortedColumns.java:196)
at org.apache.cassandra.db.ArrayBackedSortedColumns.sortCells(ArrayBackedSortedColumns.java:167)
at org.apache.cassandra.db.ArrayBackedSortedColumns.maybeSortCells(ArrayBackedSortedColumns.java:125)
at org.apache.cassandra.db.ArrayBackedSortedColumns.access$1100(ArrayBackedSortedColumns.java:44)
at org.apache.cassandra.db.ArrayBackedSortedColumns$CellCollection.iterator(ArrayBackedSortedColumns.java:622)
at org.apache.cassandra.db.ColumnFamily.iterator(ColumnFamily.java:476)
at org.apache.cassandra.db.ColumnFamily.getColumnStats(ColumnFamily.java:429)
at org.apache.cassandra.io.sstable.SSTableWriter.append(SSTableWriter.java:225)
at org.apache.cassandra.io.sstable.SSTableSimpleUnsortedWriter$DiskWriter.run(SSTableSimpleUnsortedWriter.java:215)
14:24:45.763 [Thread-3] DEBUG o.apache.cassandra.io.util.FileUtils - Renaming /mnt/datasstables-generated/2006/05/sensordata/packets_id/sensordata-packets_id-tmp-ka-320-Index.db to /mnt/datasstables-generated/2006/05/sensordata/packets_id/sensordata-packets_id-ka-320-Index.db
14:24:45.763 [Thread-3] DEBUG o.apache.cassandra.io.util.FileUtils - Renaming /mnt/datasstables-generated/2006/05/sensordata/packets_id/sensordata-packets_id-tmp-ka-320-Filter.db to /mnt/datasstables-generated/2006/05/sensordata/packets_id/sensordata-packets_id-ka-320-Filter.db
14:24:45.765 [Thread-3] DEBUG o.apache.cassandra.io.util.FileUtils - Renaming /mnt/datasstables-generated/2006/05/sensordata/packets_id/sensordata-packets_id-tmp-ka-320-Digest.sha1 to /mnt/datasstables-generated/2006/05/sensordata/packets_id/sensordata-packets_id-ka-320-Digest.sha1
14:24:45.765 [Thread-3] DEBUG o.apache.cassandra.io.util.FileUtils - Renaming /mnt/datasstables-generated/2006/05/sensordata/packets_id/sensordata-packets_id-tmp-ka-320-Statistics.db to /mnt/datasstables-generated/2006/05/sensordata/packets_id/sensordata-packets_id-ka-320-Statistics.db
14:24:45.766 [Thread-3] DEBUG o.apache.cassandra.io.util.FileUtils - Renaming /mnt/datasstables-generated/2006/05/sensordata/packets_id/sensordata-packets_id-tmp-ka-320-TOC.txt to /mnt/datasstables-generated/2006/05/sensordata/packets_id/sensordata-packets_id-ka-320-TOC.txt
14:24:45.766 [Thread-3] DEBUG o.apache.cassandra.io.util.FileUtils - Renaming /mnt/datasstables-generated/2006/05/sensordata/packets_id/sensordata-packets_id-tmp-ka-320-CompressionInfo.db to /mnt/datasstables-generated/2006/05/sensordata/packets_id/sensordata-packets_id-ka-320-CompressionInfo.db
14:24:45.767 [Thread-3] DEBUG o.apache.cassandra.io.util.FileUtils - Renaming /mnt/datasstables-generated/2006/05/sensordata/packets_id/sensordata-packets_id-tmp-ka-320-Data.db to /mnt/datasstables-generated/2006/05/sensordata/packets_id/sensordata-packets_id-ka-320-Data.db
I remain yet to figure out whether "sometimes" is always at the same place in my input files; or whether the exception is triggered more unpredictably. The dataset is quite large - so it takes a while to reproduce. The example above occured after 9 hours of processing raw input from files.
Has someone else seen this - or knows how to resolve it?
UPDATE: I have filed a bug report:
https://issues.apache.org/jira/browse/CASSANDRA-8978.
Related
This question already has answers here:
java.lang.IllegalStateException:Could not find backup for factory javax.faces.application.ApplicationFactory
(5 answers)
Closed 3 years ago.
I have a application with JSF 2.3, Primefaces 6.3, I always deploy with Wildfly 15. But now I need deploy with Tomcat 8.5 or higher, but when I run "startup.bat", always not successfull. I already changes the jsf-api jar, impl.jar, and another things. Someone know who I Fix it?
Already changed some jars, alredy add listener on web.xml
com.sun.faces.config.ConfigurationException: CONFIGURATION FAILED! null
at com.sun.faces.config.ConfigManager.initialize(ConfigManager.java:377)
at com.sun.faces.config.ConfigureListener.contextInitialized(ConfigureListener.java:223)
at org.apache.catalina.core.StandardContext.listenerStart(StandardContext.java:4680)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5150)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:713)
at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:690)
at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:705)
at org.apache.catalina.startup.HostConfig.deployWAR(HostConfig.java:978)
at org.apache.catalina.startup.HostConfig$DeployWar.run(HostConfig.java:1849)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at org.apache.tomcat.util.threads.InlineExecutorService.execute(InlineExecutorService.java:75)
at java.base/java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:118)
at org.apache.catalina.startup.HostConfig.deployWARs(HostConfig.java:773)
at org.apache.catalina.startup.HostConfig.deployApps(HostConfig.java:427)
at org.apache.catalina.startup.HostConfig.start(HostConfig.java:1576)
at org.apache.catalina.startup.HostConfig.lifecycleEvent(HostConfig.java:309)
at org.apache.catalina.util.LifecycleBase.fireLifecycleEvent(LifecycleBase.java:123)
at org.apache.catalina.util.LifecycleBase.setStateInternal(LifecycleBase.java:423)
at org.apache.catalina.util.LifecycleBase.setState(LifecycleBase.java:366)
at org.apache.catalina.core.ContainerBase.startInternal(ContainerBase.java:929)
at org.apache.catalina.core.StandardHost.startInternal(StandardHost.java:841)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1377)
at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1367)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at org.apache.tomcat.util.threads.InlineExecutorService.execute(InlineExecutorService.java:75)
at java.base/java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:140)
at org.apache.catalina.core.ContainerBase.startInternal(ContainerBase.java:902)
at org.apache.catalina.core.StandardEngine.startInternal(StandardEngine.java:262)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.core.StandardService.startInternal(StandardService.java:421)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.core.StandardServer.startInternal(StandardServer.java:932)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.startup.Catalina.start(Catalina.java:633)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.apache.catalina.startup.Bootstrap.start(Bootstrap.java:344)
at org.apache.catalina.startup.Bootstrap.main(Bootstrap.java:475)
Caused by: java.lang.UnsupportedOperationException
at javax.faces.application.Application.addSearchKeywordResolver(Application.java:2055)
at org.primefaces.util.Jsf23Helper.addSearchKeywordResolvers(Jsf23Helper.java:27)
at org.primefaces.webapp.PostConstructApplicationEventListener.processEvent(PostConstructApplicationEventListener.java:48)
at javax.faces.event.SystemEvent.processListener(SystemEvent.java:147)
at com.sun.faces.application.ApplicationImpl.processListeners(ApplicationImpl.java:2169)
at com.sun.faces.application.ApplicationImpl.invokeListenersFor(ApplicationImpl.java:2145)
at com.sun.faces.application.ApplicationImpl.publishEvent(ApplicationImpl.java:296)
at com.sun.faces.config.ConfigManager.publishPostConfigEvent(ConfigManager.java:602)
at com.sun.faces.config.ConfigManager.initialize(ConfigManager.java:371)
... 42 more
04-Jun-2019 11:38:06.401 GRAVE [main] org.apache.catalina.core.StandardContext.startInternal One or more listeners failed to start. Full details will be found in the appropriate container log file
04-Jun-2019 11:38:06.651 ADVERTÊNCIA [main] org.apache.catalina.util.SessionIdGeneratorBase.createSecureRandom Creation of SecureRandom instance for session ID generation using [SHA1PRNG] took [242] milliseconds.
04-Jun-2019 11:38:06.651 GRAVE [main] org.apache.catalina.core.StandardContext.startInternal Context [/hstchipcentre-web] startup failed due to previous errors
04-Jun-2019 11:38:06.666 GRAVE [main] javax.faces.FactoryFinderInstance.getFactory O aplicativo não foi iniciado corretamente na inicialização. Não foi possível encontrar Fábrica: javax.faces.application.ApplicationFactory. Attempting to find backup.
04-Jun-2019 11:38:06.682 GRAVE [main] com.sun.faces.config.ConfigureListener.contextDestroyed Unexpected exception when attempting to tear down the Mojarra runtime
java.lang.IllegalStateException: Could not find backup for factory javax.faces.application.ApplicationFactory.
at javax.faces.FactoryFinderInstance.getFactory(FactoryFinderInstance.java:541)
at javax.faces.FactoryFinder.getFactory(FactoryFinder.java:292)
at com.sun.faces.config.InitFacesContext.getApplication(InitFacesContext.java:136)
at com.sun.faces.config.ConfigureListener.contextDestroyed(ConfigureListener.java:328)
at org.apache.catalina.core.StandardContext.listenerStop(StandardContext.java:4727)
at org.apache.catalina.core.StandardContext.stopInternal(StandardContext.java:5388)```
In the project I founded jsf-api jar, impl.jar duplicated. Problem solved.
I have download Pentaho 7 CE edition and I have not been able to make it work. I have noticed that there is System Error 0014 that pops up in the log.
I have also downloaded Java jre1.8.0_131 and I am using Windows 10 machine. The following is a snippet of the error log in the tomcat.
20:39:08,993 ERROR [OSGIBoot] Error starting Karaf
java.lang.RuntimeException: Could not resolve mvn:org.apache.felix/org.apache.felix.framework/4.2.1
at org.apache.karaf.main.util.SimpleMavenResolver.resolve(SimpleMavenResolver.java:59)
at org.apache.karaf.main.Main.createClassLoader(Main.java:309)
at org.apache.karaf.main.Main.launch(Main.java:229)
at org.pentaho.platform.osgi.KarafBoot$4.run(KarafBoot.java:239)
at java.lang.Thread.run(Unknown Source)
at org.pentaho.platform.osgi.KarafBoot.startup(KarafBoot.java:247)
at org.pentaho.platform.engine.core.system.PentahoSystem$2.call(PentahoSystem.java:443)
at org.pentaho.platform.engine.core.system.PentahoSystem$2.call(PentahoSystem.java:434)
at org.pentaho.platform.engine.core.system.PentahoSystem.runAsSystem(PentahoSystem.java:413)
at org.pentaho.platform.engine.core.system.PentahoSystem.notifySystemListenersOfStartup(PentahoSystem.java:434)
at org.pentaho.platform.engine.core.system.PentahoSystem.access$000(PentahoSystem.java:84)
at org.pentaho.platform.engine.core.system.PentahoSystem$1.call(PentahoSystem.java:365)
at org.pentaho.platform.engine.core.system.PentahoSystem$1.call(PentahoSystem.java:362)
at org.pentaho.platform.engine.core.system.PentahoSystem.runAsSystem(PentahoSystem.java:413)
at org.pentaho.platform.engine.core.system.PentahoSystem.notifySystemListenersOfStartup(PentahoSystem.java:362)
at org.pentaho.platform.engine.core.system.PentahoSystem.init(PentahoSystem.java:332)
at org.pentaho.platform.engine.core.system.PentahoSystem.init(PentahoSystem.java:228)
at org.pentaho.platform.web.http.context.SolutionContextListener.contextInitialized(SolutionContextListener.java:162)
at org.apache.catalina.core.StandardContext.listenerStart(StandardContext.java:4853)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5314)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:145)
at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:753)
at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:729)
at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:717)
at org.apache.catalina.startup.HostConfig.deployDirectory(HostConfig.java:1092)
at org.apache.catalina.startup.HostConfig$DeployDirectory.run(HostConfig.java:1834)
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.util.concurrent.FutureTask.run(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.lang.Thread.run(Unknown Source)
20:39:09,001 ERROR [Logger] Error: Pentaho
20:39:09,005 ERROR [Logger] misc-org.pentaho.platform.engine.core.system.PentahoSystem: org.pentaho.platform.api.engine.PentahoSystemException: PentahoSystem.ERROR_0014 - Error while trying to execute startup sequence for org.pentaho.platform.osgi.OSGIBoot
org.pentaho.platform.api.engine.PentahoSystemException: org.pentaho.platform.api.engine.PentahoSystemException: PentahoSystem.ERROR_0014 - Error while trying to execute startup sequence for org.pentaho.platform.osgi.OSGIBoot
at org.pentaho.platform.engine.core.system.PentahoSystem.notifySystemListenersOfStartup(PentahoSystem.java:370)
at org.pentaho.platform.engine.core.system.PentahoSystem.init(PentahoSystem.java:332)
at org.pentaho.platform.engine.core.system.PentahoSystem.init(PentahoSystem.java:228)
at org.pentaho.platform.web.http.context.SolutionContextListener.contextInitialized(SolutionContextListener.java:162)
at org.apache.catalina.core.StandardContext.listenerStart(StandardContext.java:4853)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5314)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:145)
at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:753)
at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:729)
at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:717)
at org.apache.catalina.startup.HostConfig.deployDirectory(HostConfig.java:1092)
at org.apache.catalina.startup.HostConfig$DeployDirectory.run(HostConfig.java:1834)
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.util.concurrent.FutureTask.run(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.lang.Thread.run(Unknown Source)
Caused by: org.pentaho.platform.api.engine.PentahoSystemException: PentahoSystem.ERROR_0014 - Error while trying to execute startup sequence for org.pentaho.platform.osgi.OSGIBoot
at org.pentaho.platform.engine.core.system.PentahoSystem$2.call(PentahoSystem.java:452)
at org.pentaho.platform.engine.core.system.PentahoSystem$2.call(PentahoSystem.java:434)
at org.pentaho.platform.engine.core.system.PentahoSystem.runAsSystem(PentahoSystem.java:413)
at org.pentaho.platform.engine.core.system.PentahoSystem.notifySystemListenersOfStartup(PentahoSystem.java:434)
at org.pentaho.platform.engine.core.system.PentahoSystem.access$000(PentahoSystem.java:84)
at org.pentaho.platform.engine.core.system.PentahoSystem$1.call(PentahoSystem.java:365)
at org.pentaho.platform.engine.core.system.PentahoSystem$1.call(PentahoSystem.java:362)
at org.pentaho.platform.engine.core.system.PentahoSystem.runAsSystem(PentahoSystem.java:413)
at org.pentaho.platform.engine.core.system.PentahoSystem.notifySystemListenersOfStartup(PentahoSystem.java:362)
... 16 more
Caused by: org.pentaho.platform.api.engine.PentahoSystemException: PentahoSystem.ERROR_0014 - Error while trying to execute startup sequence for org.pentaho.platform.osgi.OSGIBoot
at org.pentaho.platform.engine.core.system.PentahoSystem$2.call(PentahoSystem.java:444)
... 24 more
20:39:09,009 ERROR [Logger] Error end:
20:39:09,009 INFO [PeriodicStatusLogger] Warning, one or more errors occurred during the initialization process.
Pentaho BI Platform server failed to properly initialize. The system will not be available for requests. (pentaho-platform-core 7.1.0.0-12) Fully Qualified Server Url = http://localhost:8080/pentaho/, Solution Path = C:\Program Files\Pentaho_BI\pentaho-server-ce-7.1.0.0-12\pentaho-server\pentaho-solutions
10-Jul-2018 20:39:29.970 INFO [localhost-startStop-1] org.apache.catalina.startup.HostConfig.deployDirectory Deployment of web application directory C:\Program Files\Pentaho_BI\pentaho-server-ce-7.1.0.0-12\pentaho-server\tomcat\webapps\pentaho has finished in 295,164 ms
10-Jul-2018 20:39:29.970 INFO [localhost-startStop-1] org.apache.catalina.startup.HostConfig.deployDirectory Deploying web application directory C:\Program Files\Pentaho_BI\pentaho-server-ce-7.1.0.0-12\pentaho-server\tomcat\webapps\pentaho-style
10-Jul-2018 20:39:30.894 INFO [localhost-startStop-1] org.apache.jasper.servlet.TldScanner.scanJars At least one JAR was scanned for TLDs yet contained no TLDs. Enable debug logging for this logger for a complete list of JARs that were scanned but no TLDs were found in them. Skipping unneeded JARs during scanning can improve startup time and JSP compilation time.
10-Jul-2018 20:39:30.910 INFO [localhost-startStop-1] org.apache.catalina.startup.HostConfig.deployDirectory Deployment of web application directory C:\Program Files\Pentaho_BI\pentaho-server-ce-7.1.0.0-12\pentaho-server\tomcat\webapps\pentaho-style has finished in 940 ms
10-Jul-2018 20:39:30.910 INFO [localhost-startStop-1] org.apache.catalina.startup.HostConfig.deployDirectory Deploying web application directory C:\Program Files\Pentaho_BI\pentaho-server-ce-7.1.0.0-12\pentaho-server\tomcat\webapps\ROOT
10-Jul-2018 20:39:31.640 INFO [localhost-startStop-1] org.apache.jasper.servlet.TldScanner.scanJars At least one JAR was scanned for TLDs yet contained no TLDs. Enable debug logging for this logger for a complete list of JARs that were scanned but no TLDs were found in them. Skipping unneeded JARs during scanning can improve startup time and JSP compilation time.
10-Jul-2018 20:39:31.656 INFO [localhost-startStop-1] org.apache.catalina.startup.HostConfig.deployDirectory Deployment of web application directory C:\Program Files\Pentaho_BI\pentaho-server-ce-7.1.0.0-12\pentaho-server\tomcat\webapps\ROOT has finished in 730 ms
10-Jul-2018 20:39:31.656 INFO [localhost-startStop-1] org.apache.catalina.startup.HostConfig.deployDirectory Deploying web application directory C:\Program Files\Pentaho_BI\pentaho-server-ce-7.1.0.0-12\pentaho-server\tomcat\webapps\sw-style
10-Jul-2018 20:39:32.398 INFO [localhost-startStop-1] org.apache.jasper.servlet.TldScanner.scanJars At least one JAR was scanned for TLDs yet contained no TLDs. Enable debug logging for this logger for a complete list of JARs that were scanned but no TLDs were found in them. Skipping unneeded JARs during scanning can improve startup time and JSP compilation time.
10-Jul-2018 20:39:32.414 INFO [localhost-startStop-1] org.apache.catalina.startup.HostConfig.deployDirectory Deployment of web application directory C:\Program Files\Pentaho_BI\pentaho-server-ce-7.1.0.0-12\pentaho-server\tomcat\webapps\sw-style has finished in 758 ms
10-Jul-2018 20:39:32.445 INFO [main] org.apache.coyote.AbstractProtocol.start Starting ProtocolHandler ["http-apr-8080"]
10-Jul-2018 20:39:32.632 INFO [main] org.apache.coyote.AbstractProtocol.start Starting ProtocolHandler ["ajp-apr-8009"]
10-Jul-2018 20:39:32.632 INFO [main] org.apache.catalina.startup.Catalina.start Server startup in 298043 ms
I would appreciate anyone help in resolving my issue.
i just update the jhipster project from 4.14.3 to 4.14.4,i remained mostly code that generated on 4.14.3 the code is here,before updated it run successfully,but after updated when i use ./mvnw to start the background it return:
2018-06-04 22:07:11.189 WARN 71364 --- [ restartedMain] ationConfigEmbeddedWebApplicationContext : Exception encountered during context initialization - cancelling refresh attempt: org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'entityManagerFactory' defined in class path resource [org/springframework/boot/autoconfigure/orm/jpa/HibernateJpaAutoConfiguration.class]: Invocation of init method failed; nested exception is javax.persistence.PersistenceException: [PersistenceUnit: default] Unable to build Hibernate SessionFactory; nested exception is java.lang.IllegalStateException: All Hibernate caches should be created upfront. Please update CacheConfiguration.java to add com.sgcc.syn.domain.Authority
2018-06-04 22:07:11.209 WARN 71364 --- [ pmt-Executor-1] com.zaxxer.hikari.pool.ProxyConnection : HikariPool-1 - Connection com.mysql.jdbc.JDBC4Connection#3e5a91e1 marked as broken because of SQLSTATE(08003), ErrorCode(0)
com.mysql.jdbc.exceptions.jdbc4.MySQLNonTransientConnectionException: No operations allowed after connection closed.
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at com.mysql.jdbc.Util.handleNewInstance(Util.java:425)
at com.mysql.jdbc.Util.getInstance(Util.java:408)
at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:919)
at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:898)
at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:887)
at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:861)
at com.mysql.jdbc.ConnectionImpl.throwConnectionClosedException(ConnectionImpl.java:1184)
at com.mysql.jdbc.ConnectionImpl.checkClosed(ConnectionImpl.java:1179)
at com.mysql.jdbc.ConnectionImpl.rollback(ConnectionImpl.java:4523)
at com.zaxxer.hikari.pool.ProxyConnection.rollback(ProxyConnection.java:377)
at com.zaxxer.hikari.pool.HikariProxyConnection.rollback(HikariProxyConnection.java)
at liquibase.database.jvm.JdbcConnection.rollback(JdbcConnection.java:337)
at liquibase.database.AbstractJdbcDatabase.rollback(AbstractJdbcDatabase.java:1166)
at liquibase.lockservice.StandardLockService.acquireLock(StandardLockService.java:205)
at liquibase.lockservice.StandardLockService.waitForLock(StandardLockService.java:170)
at liquibase.Liquibase.update(Liquibase.java:196)
at liquibase.Liquibase.update(Liquibase.java:192)
at liquibase.integration.spring.SpringLiquibase.performUpdate(SpringLiquibase.java:431)
at liquibase.integration.spring.SpringLiquibase.afterPropertiesSet(SpringLiquibase.java:388)
at io.github.jhipster.config.liquibase.AsyncSpringLiquibase.initDb(AsyncSpringLiquibase.java:94)
at io.github.jhipster.config.liquibase.AsyncSpringLiquibase.lambda$afterPropertiesSet$0(AsyncSpringLiquibase.java:77)
at io.github.jhipster.async.ExceptionHandlingAsyncTaskExecutor.lambda$createWrappedRunnable$1(ExceptionHandlingAsyncTaskExecutor.java:68)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
2018-06-04 22:07:11.210 WARN 71364 --- [ pmt-Executor-1] liquibase : Failed to restore the auto commit to true
2018-06-04 22:07:11.210 ERROR 71364 --- [ pmt-Executor-1] i.g.j.c.liquibase.AsyncSpringLiquibase : Liquibase could not start correctly, your database is NOT ready: java.sql.SQLException: Connection is closed
liquibase.exception.DatabaseException: java.sql.SQLException: Connection is closed
at liquibase.database.jvm.JdbcConnection.setAutoCommit(JdbcConnection.java:363)
at liquibase.database.AbstractJdbcDatabase.close(AbstractJdbcDatabase.java:1202)
at liquibase.integration.spring.SpringLiquibase.afterPropertiesSet(SpringLiquibase.java:397)
at io.github.jhipster.config.liquibase.AsyncSpringLiquibase.initDb(AsyncSpringLiquibase.java:94)
at io.github.jhipster.config.liquibase.AsyncSpringLiquibase.lambda$afterPropertiesSet$0(AsyncSpringLiquibase.java:77)
at io.github.jhipster.async.ExceptionHandlingAsyncTaskExecutor.lambda$createWrappedRunnable$1(ExceptionHandlingAsyncTaskExecutor.java:68)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.sql.SQLException: Connection is closed
at com.zaxxer.hikari.pool.ProxyConnection$ClosedConnection.lambda$getClosedConnection$0(ProxyConnection.java:490)
at com.sun.proxy.$Proxy116.setAutoCommit(Unknown Source)
at com.zaxxer.hikari.pool.ProxyConnection.setAutoCommit(ProxyConnection.java:395)
at com.zaxxer.hikari.pool.HikariProxyConnection.setAutoCommit(HikariProxyConnection.java)
at liquibase.database.jvm.JdbcConnection.setAutoCommit(JdbcConnection.java:361)
... 8 common frames omitted
2018-06-04 22:07:11.218 ERROR 71364 --- [ restartedMain] o.s.boot.SpringApplication : Application startup failed
org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'entityManagerFactory' defined in class path resource [org/springframework/boot/autoconfigure/orm/jpa/HibernateJpaAutoConfiguration.class]: Invocation of init method failed; nested exception is javax.persistence.PersistenceException: [PersistenceUnit: default] Unable to build Hibernate SessionFactory; nested exception is java.lang.IllegalStateException: All Hibernate caches should be created upfront. Please update CacheConfiguration.java to add com.sgcc.syn.domain.Authority
at org.springframework.beans.factory.support.AbstractAutowireCapableBean
how can i resolve it.
I have a Spark (1.4.1) application, running on Yarn, that fails with the following executor log entry:
16/07/21 23:09:08 ERROR executor.CoarseGrainedExecutorBackend: Driver 9.4.136.20:55995 disassociated! Shutting down.
16/07/21 23:09:08 ERROR storage.DiskBlockObjectWriter: Uncaught exception while reverting partial writes to file /dfs1/hadoop/yarn/local/usercache/mitchus/appcache/application_1465987751317_1172/blockmgr-f367f43b-f4c8-4faf-a829-530da30fb040/1c/temp_shuffle_581adb36-1561-4db8-a556-c4ac0e6400ed
java.io.FileNotFoundException: /dfs1/hadoop/yarn/local/usercache/mitchus/appcache/application_1465987751317_1172/blockmgr-f367f43b-f4c8-4faf-a829-530da30fb040/1c/temp_shuffle_581adb36-1561-4db8-a556-c4ac0e6400ed (No such file or directory)
at java.io.FileOutputStream.open0(Native Method)
at java.io.FileOutputStream.open(FileOutputStream.java:270)
at java.io.FileOutputStream.<init>(FileOutputStream.java:213)
at org.apache.spark.storage.DiskBlockObjectWriter.revertPartialWritesAndClose(BlockObjectWriter.scala:189)
at org.apache.spark.util.collection.ExternalSorter.spillToMergeableFile(ExternalSorter.scala:328)
at org.apache.spark.util.collection.ExternalSorter.spill(ExternalSorter.scala:257)
at org.apache.spark.util.collection.ExternalSorter.spill(ExternalSorter.scala:95)
at org.apache.spark.util.collection.Spillable$class.maybeSpill(Spillable.scala:83)
at org.apache.spark.util.collection.ExternalSorter.maybeSpill(ExternalSorter.scala:95)
at org.apache.spark.util.collection.ExternalSorter.maybeSpillCollection(ExternalSorter.scala:240)
at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:220)
at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:62)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:70)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
at org.apache.spark.scheduler.Task.run(Task.scala:70)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Any clues as to what might have gone wrong?
The reason caused by temp shuffle file is deleted. There are many reasons, for one which I met is because the other executor was killed by Yarn. After the executor killed, a SHUT_DOWN signal will be sent to other executors, then the ShutdownHookManager will delete all the temp files which have registered to ShutdownHookManager. That's why you see the error. So you maybe need to check whether there are any ShutdownHookManager called log.
You can try to improve spark.yarn.executor.memoryOverhead.
We have been running Cassandra 1.0.2 in production for many months, with minimal trouble. Lately, we have started to see consistent failures in all nodes, with this error:
ERROR [CompactionExecutor:199] 2013-03-02 00:21:26,179 AbstractCassandraDaemon.java (line 133) Fatal exception in thread Thread[CompactionExecutor:199,1,RMI Runtime]
java.io.IOError: java.io.IOException: Map failed
at org.apache.cassandra.io.util.MmappedSegmentedFile$Builder.createSegments(MmappedSegmentedFile.java:225)
at org.apache.cassandra.io.util.MmappedSegmentedFile$Builder.complete(MmappedSegmentedFile.java:202)
at org.apache.cassandra.io.sstable.SSTableWriter.closeAndOpenReader(SSTableWriter.java:308)
at org.apache.cassandra.db.compaction.CompactionTask.execute(CompactionTask.java:174)
at org.apache.cassandra.db.compaction.CompactionManager$4.call(CompactionManager.java:275)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:619)
Caused by: java.io.IOException: Map failed
at sun.nio.ch.FileChannelImpl.map(FileChannelImpl.java:758)
at org.apache.cassandra.io.util.MmappedSegmentedFile$Builder.createSegments(MmappedSegmentedFile.java:217)
... 9 more
Caused by: java.lang.OutOfMemoryError: Map failed
at sun.nio.ch.FileChannelImpl.map0(Native Method)
at sun.nio.ch.FileChannelImpl.map(FileChannelImpl.java:755)
... 10 more
INFO [Thread-2] 2013-03-02 00:21:26,222 MessagingService.java (line 488) Shutting down MessageService...
The node dies after this. We remove the oldest data files and start the node again. The node runs for a while, then dies again. This happens for all eight nodes in our ring.
We are using the default compaction strategy (size tiered) and we think it is the best choice for our problem domain.
Q: Why is this happening and what should we do to fix it?