Why aren't my test properties getting picked up by spring-boot when running mvn test - spring-test

We have a Web application based on Spring-boot 2.0.x - recently enabled Spring-security for OAuth2 - security works, but I'm dealing with the aftermath of that in my project. My existing test cases are failing due to unsatisfied exception errors - can't figure out why. Plz help. I don't know why it can't resolve any of these property placeholders: w3.clientId, w3.clientSecret, etc - I have them in the application.properties file in both the main path and the test path as well. See below.
Here is the main application class definition:
package com.ibm.cio.cloud.cost.spreadsheet;
#SpringBootApplication
#ComponentScan({"com.ibm.cio.cloud.cost.spreadsheet","com.ibm.cio.cloud.cost.spreadsheet.rest.controller","com.ibm.cio.cloud.cost.spreadsheet.dao", "com.ibm.cio.cloud.cost.spreadsheet.service"})
public class BlueCostSpreadsheetUploadWebApplication {
public static void main(String[] args) {
System.setProperty("spring.devtools.restart.enabled", "false");
SpringApplication.run(BlueCostSpreadsheetUploadWebApplication.class, args);
}
}
Here is my test cases definition:
package com.ibm.cio.cloud.cost.spreadsheet.dao;
#RunWith(SpringJUnit4ClassRunner.class)
#ContextConfiguration(classes=BlueCostSpreadsheetUploadWebApplication.class)
public class UTCostSpreadsheetDAOTest {
#Autowired
private JdbcTemplate jdbcTemplate;
Here is the W3OpenIdConnectConfig class that is causing problems:
package com.ibm.cio.cloud.cost.spreadsheet.security;
#Configuration
#EnableOAuth2Client
public class W3OpenIdConnectConfig {
#Value("${w3.clientId}")
private String clientId;
#Value("${w3.clientSecret}")
private String clientSecret;
#Value("${w3.accessTokenUri}")
private String accessTokenUri;
#Value("${w3.userAuthorizationUri}")
private String userAuthorizationUri;
#Value("${w3.redirectUri}")
private String redirectUri;
#Bean
public OAuth2ProtectedResourceDetails w3OpenId() {
Finally, here is my src/test/resources/application.properties file:
spring.servlet.multipart.max-file-size=5MB
spring.servlet.multipart.max-request-size=5MB
spring.http.multipart.enabled=false
w3.clientId=ZjgzMTM3NT222NC00
w3.clientSecret=NTdmM2IyZD333iMi00
w3.accessTokenUri=https://w3id-test.sso.ibm.com/isam/oidc/endpoint/amapp-runtime-oidcidp/token
w3.userAuthorizationUri=https://w3id-test.sso.ibm.com/isam/oidc/endpoint/amapp-runtime-oidcidp/authorize
w3.redirectUri=https://localhost:8443/auth/sso/callback
w3.issuer=https://w3id-test.sso.ibm.com/isam
rsa.key=-----BEGIN PUBLIC KEY-----MIICIjANBgkqhkiG9w0BAQEFAAOCAg8AMIICCgKCAgEAodx954SU1L12SOoRH7pLhB+4S/9mLkVOMt/tyhf3TZ1TUlNRroPxLtMtVjh5LL1LlCW8XxnqbGnGc0EJamRZ3eMFpn/keJbSW6T9m+pTzY/VxUWuf4uJBDzfggOXIv+VLJ/SsPPeNKZAcriIGKG5chw6Fy9FYkk91RN2/VYoysiAqof0aXYzsbxiFbUdTFFk0CAwEAAQ==-----END PUBLIC KEY-----
# Users not allowed unless part of this bluegroup
authorized.bluegroup=BLUECOST_SPREADSHEET_UPLOAD
Here are the errors when I run a single Test case from cmd line using 'mvn test -Dtest=UTCostSpreadsheetDAOTest':
[INFO] -------------------------------------------------------
[INFO] T E S T S
[INFO] -------------------------------------------------------
[INFO] Running com.ibm.cio.cloud.cost.spreadsheet.dao.UTCostSpreadsheetDAOTest
11:19:15,034 |-INFO in ch.qos.logback.classic.LoggerContext[default] - Could NOT find resource [logback-test.xml]
11:19:15,035 |-INFO in ch.qos.logback.classic.LoggerContext[default] - Could NOT find resource [logback.groovy]
11:19:15,035 |-INFO in ch.qos.logback.classic.LoggerContext[default] - Found resource [logback.xml] at [file:/C:/BlueCostSpreadsheetUploadWeb/backend/target/classes/logback.xml]
11:19:15,038 |-WARN in ch.qos.logback.classic.LoggerContext[default] - Resource [logback.xml] occurs multiple times on the classpath.
11:19:15,038 |-WARN in ch.qos.logback.classic.LoggerContext[default] - Resource [logback.xml] occurs at [file:/C:/BlueCostSpreadsheetUploadWeb/backend/target/classes/logback.xml]
11:19:15,038 |-WARN in ch.qos.logback.classic.LoggerContext[default] - Resource [logback.xml] occurs at [jar:file:/C:/Users/JamesDePaul/.m2/repository/net/sourceforge/cobertura/cobertura/2.1.1/cobertura-2.1.1.jar!/logback.xml]
11:19:15,318 |-INFO in ch.qos.logback.classic.joran.action.ConfigurationAction - debug attribute not set
11:19:15,321 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.ConsoleAppender]
11:19:15,355 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [STDOUT]
11:19:15,592 |-WARN in ch.qos.logback.core.ConsoleAppender[STDOUT] - This appender no longer admits a layout as a sub-component, set an encoder instead.
11:19:15,592 |-WARN in ch.qos.logback.core.ConsoleAppender[STDOUT] - To ensure compatibility, wrapping your layout in LayoutWrappingEncoder.
11:19:15,592 |-WARN in ch.qos.logback.core.ConsoleAppender[STDOUT] - See also http://logback.qos.ch/codes.html#layoutInsteadOfEncoder for details
11:19:15,602 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.springframework.security.oauth2] to INFO
11:19:15,602 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting additivity of logger [org.springframework.security.oauth2] to true
11:19:15,602 |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [STDOUT] to Logger[org.springframework.security.oauth2]
11:19:15,604 |-INFO in ch.qos.logback.classic.joran.action.RootLoggerAction - Setting level of ROOT logger to ERROR
11:19:15,604 |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [STDOUT] to Logger[ROOT]
11:19:15,605 |-INFO in ch.qos.logback.classic.joran.action.RootLoggerAction - Setting level of ROOT logger to DEBUG
11:19:15,605 |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [STDOUT] to Logger[ROOT]
11:19:15,605 |-INFO in ch.qos.logback.classic.joran.action.RootLoggerAction - Setting level of ROOT logger to INFO
11:19:15,605 |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [STDOUT] to Logger[ROOT]
11:19:15,605 |-INFO in ch.qos.logback.classic.joran.action.ConfigurationAction - End of configuration.
11:19:15,607 |-INFO in ch.qos.logback.classic.joran.JoranConfigurator#588cd519 - Registering current configuration as safe fallback point
2018-09-14 11:19:16 INFO o.s.t.c.s.DefaultTestContextBootstrapper - Loaded default TestExecutionListener class names from location [META-INF/spring.factories]: [org.springframework.boot.test.mock.mockito.MockitoTestExecutionListener, org.springframework.boot.test.mock.mockito.ResetMocksTestExecutionListener, org.springframework.boot.test.autoconfigure.restdocs.RestDocsTestExecutionListener, org.springframework.boot.test.autoconfigure.web.client.MockRestServiceServerResetTestExecutionListener, org.springframework.boot.test.autoconfigure.web.servlet.MockMvcPrintOnlyOnFailureTestExecutionListener, org.springframework.boot.test.autoconfigure.web.servlet.WebDriverTestExecutionListener, org.springframework.test.context.web.ServletTestExecutionListener, org.springframework.test.context.support.DirtiesContextBeforeModesTestExecutionListener, org.springframework.test.context.support.DependencyInjectionTestExecutionListener, org.springframework.test.context.support.DirtiesContextTestExecutionListener, org.springframework.test.context.transaction.TransactionalTestExecutionListener, org.springframework.test.context.jdbc.SqlScriptsTestExecutionListener, org.springframework.security.test.context.support.WithSecurityContextTestExecutionListener, org.springframework.security.test.context.support.ReactorContextTestExecutionListener]
2018-09-14 11:19:16 INFO o.s.t.c.s.DefaultTestContextBootstrapper - Using TestExecutionListeners: [org.springframework.test.context.web.ServletTestExecutionListener#16134476, org.springframework.test.context.support.DirtiesContextBeforeModesTestExecutionListener#62b09715, org.springframework.boot.test.mock.mockito.MockitoTestExecutionListener#3e214105, org.springframework.test.context.support.DependencyInjectionTestExecutionListener#da4cf09, org.springframework.test.context.support.DirtiesContextTestExecutionListener#1980a3f, org.springframework.test.context.transaction.TransactionalTestExecutionListener#67f63d26, org.springframework.test.context.jdbc.SqlScriptsTestExecutionListener#536b71b4, org.springframework.security.test.context.support.WithSecurityContextTestExecutionListener#789c3057, org.springframework.security.test.context.support.ReactorContextTestExecutionListener#39941489, org.springframework.boot.test.mock.mockito.ResetMocksTestExecutionListener#6f5d0190, org.springframework.boot.test.autoconfigure.restdocs.RestDocsTestExecutionListener#67332b1e, org.springframework.boot.test.autoconfigure.web.client.MockRestServiceServerResetTestExecutionListener#7e34b127, org.springframework.boot.test.autoconfigure.web.servlet.MockMvcPrintOnlyOnFailureTestExecutionListener#679dd234, org.springframework.boot.test.autoconfigure.web.servlet.WebDriverTestExecutionListener#60cb1ed6]
2018-09-14 11:19:17 INFO o.s.c.s.GenericApplicationContext - Refreshing org.springframework.context.support.GenericApplicationContext#1cb9ef52: startup date [Fri Sep 14 11:19:17 MDT 2018]; root of context hierarchy
2018-09-14 11:19:25 INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration$$EnhancerBySpringCGLIB$$1e373773] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying)
2018-09-14 11:19:28 INFO o.f.c.internal.util.VersionPrinter - Flyway Community Edition 5.0.7 by Boxfuse
2018-09-14 11:19:28 INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting...
2018-09-14 11:19:28 WARN c.z.hikari.util.DriverDataSource - Registered driver with driverClassName=org.hsqldb.jdbcDriver was not found, trying direct instantiation.
2018-09-14 11:19:30 INFO com.zaxxer.hikari.pool.PoolBase - HikariPool-1 - Driver does not support get/set network timeout for connections. (feature not supported)
2018-09-14 11:19:30 INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed.
2018-09-14 11:19:30 INFO o.f.c.i.database.DatabaseFactory - Database: jdbc:hsqldb:mem:testdb (HSQL Database Engine 2.4)
2018-09-14 11:19:31 INFO o.f.core.internal.command.DbValidate - Successfully validated 3 migrations (execution time 00:00.304s)
2018-09-14 11:19:31 INFO o.f.c.i.s.JdbcTableSchemaHistory - Creating Schema History table: "PUBLIC"."flyway_schema_history"
2018-09-14 11:19:31 INFO o.f.core.internal.command.DbMigrate - Current version of schema "PUBLIC": << Empty Schema >>
2018-09-14 11:19:31 INFO o.f.core.internal.command.DbMigrate - Migrating schema "PUBLIC" to version 1 - Create BLUE COST SSCDATA Table
2018-09-14 11:19:31 INFO o.f.core.internal.command.DbMigrate - Migrating schema "PUBLIC" to version 1.1 - Create COST SPREADSHEET Table
2018-09-14 11:19:31 INFO o.f.core.internal.command.DbMigrate - Migrating schema "PUBLIC" to version 1.2 - ALTER COST SPREADSHEET Table ADD FILENAME
2018-09-14 11:19:31 INFO o.f.core.internal.command.DbMigrate - Successfully applied 3 migrations to schema "PUBLIC" (execution time 00:00.512s)
2018-09-14 11:19:32 WARN o.s.c.s.GenericApplicationContext - Exception encountered during context initialization - cancelling refresh attempt: org.springframework.beans.factory.UnsatisfiedDependencyException: Error creating bean with name 'securityConfig': Unsatisfied dependency expressed through field 'restTemplate'; nested exception is org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'w3OpenIdConnectConfig': Injection of autowired dependencies failed; nested exception is java.lang.IllegalArgumentException: Could not resolve placeholder 'w3.clientId' in value "${w3.clientId}"
2018-09-14 11:19:32 INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Shutdown initiated...
2018-09-14 11:19:32 INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Shutdown completed.
2018-09-14 11:19:32 ERROR o.s.test.context.TestContextManager - Caught exception while allowing TestExecutionListener [org.springframework.test.context.support.DependencyInjectionTestExecutionListener#da4cf09] to prepare test instance [com.ibm.cio.cloud.cost.spreadsheet.dao.UTCostSpreadsheetDAOTest#2298e5d2]
java.lang.IllegalStateException: Failed to load ApplicationContext
at org.springframework.test.context.cache.DefaultCacheAwareContextLoaderDelegate.loadContext(DefaultCacheAwareContextLoaderDelegate.java:125)
at org.springframework.test.context.support.DefaultTestContext.getApplicationContext(DefaultTestContext.java:108)
at org.springframework.test.context.support.DependencyInjectionTestExecutionListener.injectDependencies(DependencyInjectionTestExecutionListener.java:117)
at org.springframework.test.context.support.DependencyInjectionTestExecutionListener.prepareTestInstance(DependencyInjectionTestExecutionListener.java:83)
at org.springframework.test.context.TestContextManager.prepareTestInstance(TestContextManager.java:246)
at org.springframework.test.context.junit4.SpringJUnit4ClassRunner.createTest(SpringJUnit4ClassRunner.java:227)
at org.springframework.test.context.junit4.SpringJUnit4ClassRunner$1.runReflectiveCall(SpringJUnit4ClassRunner.java:289)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.springframework.test.context.junit4.SpringJUnit4ClassRunner.methodBlock(SpringJUnit4ClassRunner.java:291)
at org.springframework.test.context.junit4.SpringJUnit4ClassRunner.runChild(SpringJUnit4ClassRunner.java:246)
at org.springframework.test.context.junit4.SpringJUnit4ClassRunner.runChild(SpringJUnit4ClassRunner.java:97)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at org.springframework.test.context.junit4.statements.RunBeforeTestClassCallbacks.evaluate(RunBeforeTestClassCallbacks.java:61)
at org.springframework.test.context.junit4.statements.RunAfterTestClassCallbacks.evaluate(RunAfterTestClassCallbacks.java:70)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.springframework.test.context.junit4.SpringJUnit4ClassRunner.run(SpringJUnit4ClassRunner.java:190)
at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:369)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:275)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:239)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:160)
at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:373)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:334)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:119)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:407)
Caused by: org.springframework.beans.factory.UnsatisfiedDependencyException: Error creating bean with name 'securityConfig': Unsatisfied dependency expressed through field 'restTemplate'; nested exception is org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'w3OpenIdConnectConfig': Injection of autowired dependencies failed; nested exception is java.lang.IllegalArgumentException: Could not resolve placeholder 'w3.clientId' in value "${w3.clientId}"
at org.springframework.beans.factory.annotation.AutowiredAnnotationBeanPostProcessor$AutowiredFieldElement.inject(AutowiredAnnotationBeanPostProcessor.java:586)
at org.springframework.beans.factory.annotation.InjectionMetadata.inject(InjectionMetadata.java:91)
at org.springframework.beans.factory.annotation.AutowiredAnnotationBeanPostProcessor.postProcessPropertyValues(AutowiredAnnotationBeanPostProcessor.java:372)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.populateBean(AbstractAutowireCapableBeanFactory.java:1341)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:572)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:495)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:317)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:222)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:315)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:199)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.preInstantiateSingletons(DefaultListableBeanFactory.java:759)
at org.springframework.context.support.AbstractApplicationContext.finishBeanFactoryInitialization(AbstractApplicationContext.java:869)
at org.springframework.context.support.AbstractApplicationContext.refresh(AbstractApplicationContext.java:550)
at org.springframework.test.context.support.AbstractGenericContextLoader.loadContext(AbstractGenericContextLoader.java:128)
at org.springframework.test.context.support.AbstractGenericContextLoader.loadContext(AbstractGenericContextLoader.java:60)
at org.springframework.test.context.support.AbstractDelegatingSmartContextLoader.delegateLoading(AbstractDelegatingSmartContextLoader.java:107)
at org.springframework.test.context.support.AbstractDelegatingSmartContextLoader.loadContext(AbstractDelegatingSmartContextLoader.java:243)
at org.springframework.test.context.cache.DefaultCacheAwareContextLoaderDelegate.loadContextInternal(DefaultCacheAwareContextLoaderDelegate.java:99)
at org.springframework.test.context.cache.DefaultCacheAwareContextLoaderDelegate.loadContext(DefaultCacheAwareContextLoaderDelegate.java:117)

I solved this by adding the following annotation on top of my Test case class:
#RunWith(SpringJUnit4ClassRunner.class)
#TestPropertySource(locations = "classpath:application.properties")
#WebAppConfiguration
#ContextConfiguration(classes=BlueCostSpreadsheetUploadWebApplication.class)
public class UTCostSpreadsheetDAOTest {

Related

Spark logs configuration - log4j

I have an application and the code under com.myapplication and I am using Spark.
My log4j configuration is:
# Root logger option
log4j.rootLogger=WARN, stdout
log4j.category.com.myapplication=INFO
# Direct log messages to stdout
log4j.appender.stdout=org.apache.log4j.ConsoleAppender
log4j.appender.stdout.layout=org.apache.log4j.PatternLayout
log4j.appender.stdout.layout.ConversionPattern=%d{yyyy-MM-dd HH:mm:ss} %-5p %c{1}:%L - %m%n
When I run a script with this configuration I get log messages two times in the console, e.g.:
2022-07-26 09:47:00 INFO Spark$:49 - Configuring Spark...
1969 [main] INFO com.commerzbank.cda.spark.Spark$ - Configuring Spark...
2022-07-26 09:47:00 WARN SparkConf:66 - The configuration key 'spark.yarn.executor.memoryOverhead' has been deprecated as of Spark 2.3 and may be removed in the future. Please use the new key 'spark.executor.memoryOverhead' instead.
1971 [main] WARN org.apache.spark.SparkConf - The configuration key 'spark.yarn.executor.memoryOverhead' has been deprecated as of Spark 2.3 and may be removed in the future. Please use the new key 'spark.executor.memoryOverhead' instead.
2022-07-26 09:47:00 WARN SparkConf:66 - The configuration key 'spark.yarn.driver.memoryOverhead' has been deprecated as of Spark 2.3 and may be removed in the future. Please use the new key 'spark.driver.memoryOverhead' instead.
1971 [main] WARN org.apache.spark.SparkConf - The configuration key 'spark.yarn.driver.memoryOverhead' has been deprecated as of Spark 2.3 and may be removed in the future. Please use the new key 'spark.driver.memoryOverhead' instead.
2022-07-26 09:47:00 INFO Spark$:82 - Pick up Spark property master...
1972 [main] INFO com.commerzbank.cda.spark.Spark$ - Pick up Spark property master...
The messages are printed double one with date and one with [main]. Is this because of a wrong log4j configuration or where should I look into my code?

Why isn't my Eureka server being found by Hazelcast-Eureka-One plugin?

I have a working Hazelcast cluster configured with tcp-ip. I need it to work with Eureka discovery. I am trying to implement the hazelcast-eureka-one plugin.
The (Spring-Boot) app currently already registers itself with Eureka sucessfully, using the #EnableEurekaClient annotation. I am not concerned with whether the hazelcast eureka client is the same or a different client. I am fine with hazelcast registering itself separately from the app. As long as it works.
When I remove eureka-client.properties, the app will not start up, showing an error that eureka-client.properties can not be found. When I have the file in place, the app starts, but apparently none of the properties from eureka-client.properties are being loaded, which leaves hazelcast not knowing where the eureka server is. The logs indicate that the properties file is being found, but none of the properties seem to be imported.
Upgrading hazelcast-eureka-one to 1.1 makes no change.
Setting use-metadata-for-host-and-port to true makes no change.
Gradle:
compile group: 'com.hazelcast', name: 'hazelcast-spring', version: '3.9.4'
compile group: 'com.hazelcast', name: 'hazelcast-hibernate52', version: '1.2.3'
compile group: 'com.hazelcast', name: 'hazelcast-eureka-one', version: '1.0.1'
hazelcast.xml:
<?xml version="1.0" encoding="UTF-8"?>
<hazelcast xsi:schemaLocation="http://www.hazelcast.com/schema/config http://www.hazelcast.com/schema/config/hazelcast-config-3.9.xsd"
xmlns="http://www.hazelcast.com/schema/config"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<instance-name>app.name.hazelcast.sessions-instance</instance-name>
<group>
<name>app.name.hazelcast.sessions.local-group</name>
</group>
<network>
<join>
<multicast enabled="false"/>
<tcp-ip enabled="false"/>
<aws enabled="false"/>
<discovery-strategies>
<discovery-strategy class="com.hazelcast.eureka.one.EurekaOneDiscoveryStrategy" enabled="true">
<properties>
<property name="self-registration">true</property>
<property name="namespace">hazelcast-app-name</property>
<property name="use-metadata-for-host-and-port">false</property>
</properties>
</discovery-strategy>
</discovery-strategies>
</join>
</network>
<map name="spring:session:sessions">
<attributes>
<attribute extractor="org.springframework.session.hazelcast.PrincipalNameExtractor">principalName</attribute>
</attributes>
<indexes>
<index>principalName</index>
</indexes>
</map>
eureka-client.properties:
hazelcast.shouldUseDns=false
hazelcast.datacenter=primary
hazelcast.name=hazelcast-app-name-sessions
hazelcast.serviceUrl.default=http://username:password#svcregistry1-dev.company.com:8580/eureka/,http://username:password#svcregistry2-dev.company.com:8590/eureka/
Log file:
Loading 'hazelcast.xml' from classpath.
2019-02-15 11:19:13,935 - INFO - [localhost-startStop-1] - [,,] - com.hazelcast.instance.AddressPicker : [LOCAL] [app.name.hazelcast.sessions.local-group] [3.9.4] Prefer IPv4 stack is true.
2019-02-15 11:19:14,166 - INFO - [localhost-startStop-1] - [,,] - com.hazelcast.instance.AddressPicker : [LOCAL] [app.name.hazelcast.sessions.local-group] [3.9.4] Picked [172.28.208.1]:5701, using socket ServerSocket[addr=/0:0:0:0:0:0:0:0,localport=5701], bind any local is true
2019-02-15 11:19:14,179 - INFO - [localhost-startStop-1] - [,,] - com.hazelcast.system : [172.28.208.1]:5701 [app.name.hazelcast.sessions.local-group] [3.9.4] Hazelcast 3.9.4 (20180420 - b8001d5) starting at [172.28.208.1]:5701
2019-02-15 11:19:14,179 - INFO - [localhost-startStop-1] - [,,] - com.hazelcast.system : [172.28.208.1]:5701 [app.name.hazelcast.sessions.local-group] [3.9.4] Copyright (c) 2008-2018, Hazelcast, Inc. All Rights Reserved.
2019-02-15 11:19:14,179 - INFO - [localhost-startStop-1] - [,,] - com.hazelcast.system : [172.28.208.1]:5701 [app.name.hazelcast.sessions.local-group] [3.9.4] Configured Hazelcast Serialization version: 1
2019-02-15 11:19:14,616 - INFO - [localhost-startStop-1] - [,,] - c.h.s.i.o.impl.BackpressureRegulator : [172.28.208.1]:5701 [app.name.hazelcast.sessions.local-group] [3.9.4] Backpressure is disabled
2019-02-15 11:19:15,309 - DEBUG - [localhost-startStop-1] - [,,] - .n.c.u.OverridingPropertiesConfiguration : Base path set to file:///C:/Users/my.name/IdeaProjects/AppName/build/classes/main/
2019-02-15 11:19:15,310 - DEBUG - [localhost-startStop-1] - [,,] - .n.c.u.OverridingPropertiesConfiguration : FileName set to eureka-client.properties
2019-02-15 11:19:15,310 - DEBUG - [localhost-startStop-1] - [,,] - .n.c.u.OverridingPropertiesConfiguration : URL set to file:/C:/Users/my.name/IdeaProjects/AppName/build/classes/main/eureka-client.properties
2019-02-15 11:19:15,316 - INFO - [localhost-startStop-1] - [,,] - c.n.config.util.ConfigurationUtils : Loaded properties file file:/C:/Users/my.name/IdeaProjects/AppName/build/classes/main/eureka-client.properties
2019-02-15 11:19:15,326 - INFO - [localhost-startStop-1] - [,,] - .p.EurekaConfigBasedInstanceInfoProvider : Setting initial instance status as: STARTING
2019-02-15 11:19:15,334 - WARN - [localhost-startStop-1] - [,,] - c.n.config.util.ConfigurationUtils : file:/C:/Users/my.name/IdeaProjects/AppName/build/classes/main/eureka-client.properties is already loaded
2019-02-15 11:19:15,385 - INFO - [localhost-startStop-1] - [,,] - com.netflix.discovery.DiscoveryClient : Initializing Eureka in region us-east-1
2019-02-15 11:19:15,951 - INFO - [localhost-startStop-1] - [,,] - c.n.d.provider.DiscoveryJerseyProvider : Using JSON encoding codec LegacyJacksonJson
2019-02-15 11:19:15,951 - INFO - [localhost-startStop-1] - [,,] - c.n.d.provider.DiscoveryJerseyProvider : Using JSON decoding codec LegacyJacksonJson
2019-02-15 11:19:16,143 - INFO - [localhost-startStop-1] - [,,] - c.n.d.provider.DiscoveryJerseyProvider : Using XML encoding codec XStreamXml
2019-02-15 11:19:16,143 - INFO - [localhost-startStop-1] - [,,] - c.n.d.provider.DiscoveryJerseyProvider : Using XML decoding codec XStreamXml
2019-02-15 11:19:16,392 - INFO - [localhost-startStop-1] - [,,] - c.n.d.s.r.aws.ConfigClusterResolver : Resolving eureka endpoints via configuration
2019-02-15 11:19:16,394 - DEBUG - [localhost-startStop-1] - [,,] - c.n.discovery.endpoint.EndpointUtils : The availability zone for the given region us-east-1 are [defaultZone]
2019-02-15 11:19:16,394 - DEBUG - [localhost-startStop-1] - [,,] - c.n.d.s.r.aws.ConfigClusterResolver : Config resolved to []
2019-02-15 11:19:16,394 - ERROR - [localhost-startStop-1] - [,,] - c.n.d.s.r.aws.ConfigClusterResolver : Cannot resolve to any endpoints from provided configuration: {defaultZone=[]}
2019-02-15 11:19:16,612 - DEBUG - [localhost-startStop-1] - [,,] - c.n.d.s.r.a.ZoneAffinityClusterResolver : Local zone=defaultZone; resolved to: []
2019-02-15 11:19:16,612 - ERROR - [localhost-startStop-1] - [,,] - c.n.d.s.transport.EurekaHttpClients : Initial resolution of Eureka server endpoints failed. Check ConfigClusterResolver logs for more info
2019-02-15 11:19:16,647 - INFO - [localhost-startStop-1] - [,,] - com.netflix.discovery.DiscoveryClient : Disable delta property : false
2019-02-15 11:19:16,647 - INFO - [localhost-startStop-1] - [,,] - com.netflix.discovery.DiscoveryClient : Single vip registry refresh property : null
2019-02-15 11:19:16,647 - INFO - [localhost-startStop-1] - [,,] - com.netflix.discovery.DiscoveryClient : Force full registry fetch : false
2019-02-15 11:19:16,647 - INFO - [localhost-startStop-1] - [,,] - com.netflix.discovery.DiscoveryClient : Application is null : false
2019-02-15 11:19:16,647 - INFO - [localhost-startStop-1] - [,,] - com.netflix.discovery.DiscoveryClient : Registered Applications size is zero : true
2019-02-15 11:19:16,647 - INFO - [localhost-startStop-1] - [,,] - com.netflix.discovery.DiscoveryClient : Application version is -1: true
2019-02-15 11:19:16,647 - INFO - [localhost-startStop-1] - [,,] - com.netflix.discovery.DiscoveryClient : Getting all instance registry info from the eureka server
2019-02-15 11:19:16,648 - DEBUG - [localhost-startStop-1] - [,,] - c.n.d.s.t.d.SessionedEurekaHttpClient : Ending a session and starting anew
2019-02-15 11:19:16,655 - ERROR - [localhost-startStop-1] - [,,] - com.netflix.discovery.DiscoveryClient : DiscoveryClient_UNKNOWN/0c99d08b-8072-4fe4-a20f-c8653e10e374 - was unable to refresh its cache! status = There is no known eureka server; cluster server list is empty
com.netflix.discovery.shared.transport.TransportException: There is no known eureka server; cluster server list is empty
at com.netflix.discovery.shared.transport.decorator.RetryableEurekaHttpClient.execute(RetryableEurekaHttpClient.java:108)
at com.netflix.discovery.shared.transport.decorator.EurekaHttpClientDecorator.getApplications(EurekaHttpClientDecorator.java:134)
at com.netflix.discovery.shared.transport.decorator.EurekaHttpClientDecorator$6.execute(EurekaHttpClientDecorator.java:137)
at com.netflix.discovery.shared.transport.decorator.SessionedEurekaHttpClient.execute(SessionedEurekaHttpClient.java:77)
at com.netflix.discovery.shared.transport.decorator.EurekaHttpClientDecorator.getApplications(EurekaHttpClientDecorator.java:134)
at com.netflix.discovery.DiscoveryClient.getAndStoreFullRegistry(DiscoveryClient.java:1051)
at com.netflix.discovery.DiscoveryClient.fetchRegistry(DiscoveryClient.java:965)
at com.netflix.discovery.DiscoveryClient.<init>(DiscoveryClient.java:414)
at com.netflix.discovery.DiscoveryClient.<init>(DiscoveryClient.java:269)
at com.netflix.discovery.DiscoveryClient.<init>(DiscoveryClient.java:265)
at com.netflix.discovery.DiscoveryClient.<init>(DiscoveryClient.java:257)
at com.hazelcast.eureka.one.EurekaOneDiscoveryStrategy.<init>(EurekaOneDiscoveryStrategy.java:147)
at com.hazelcast.eureka.one.EurekaOneDiscoveryStrategy.<init>(EurekaOneDiscoveryStrategy.java:55)
at com.hazelcast.eureka.one.EurekaOneDiscoveryStrategy$EurekaOneDiscoveryStrategyBuilder.build(EurekaOneDiscoveryStrategy.java:111)
at com.hazelcast.eureka.one.EurekaOneDiscoveryStrategyFactory.newDiscoveryStrategy(EurekaOneDiscoveryStrategyFactory.java:53)
at com.hazelcast.spi.discovery.impl.DefaultDiscoveryService.buildDiscoveryStrategy(DefaultDiscoveryService.java:185)
at com.hazelcast.spi.discovery.impl.DefaultDiscoveryService.loadDiscoveryStrategies(DefaultDiscoveryService.java:145)
at com.hazelcast.spi.discovery.impl.DefaultDiscoveryService.<init>(DefaultDiscoveryService.java:60)
at com.hazelcast.spi.discovery.impl.DefaultDiscoveryServiceProvider.newDiscoveryService(DefaultDiscoveryServiceProvider.java:29)
at com.hazelcast.instance.Node.createDiscoveryService(Node.java:265)
at com.hazelcast.instance.Node.<init>(Node.java:216)
at com.hazelcast.instance.HazelcastInstanceImpl.createNode(HazelcastInstanceImpl.java:160)
at com.hazelcast.instance.HazelcastInstanceImpl.<init>(HazelcastInstanceImpl.java:128)
at com.hazelcast.instance.HazelcastInstanceFactory.constructHazelcastInstance(HazelcastInstanceFactory.java:195)
at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:174)
at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:124)
at com.hazelcast.core.Hazelcast.newHazelcastInstance(Hazelcast.java:92)
at org.springframework.boot.autoconfigure.hazelcast.HazelcastServerConfiguration$HazelcastServerConfigFileConfiguration.hazelcastInstance(HazelcastServerConfiguration.java:56)
at org.springframework.boot.autoconfigure.hazelcast.HazelcastServerConfiguration$HazelcastServerConfigFileConfiguration$$EnhancerBySpringCGLIB$$d6cfebe6.CGLIB$hazelcastInstance$0(<generated>)
at org.springframework.boot.autoconfigure.hazelcast.HazelcastServerConfiguration$HazelcastServerConfigFileConfiguration$$EnhancerBySpringCGLIB$$d6cfebe6$$FastClassBySpringCGLIB$$3a3e2869.invoke(<generated>)
at org.springframework.cglib.proxy.MethodProxy.invokeSuper(MethodProxy.java:228)
at org.springframework.context.annotation.ConfigurationClassEnhancer$BeanMethodInterceptor.intercept(ConfigurationClassEnhancer.java:365)
at org.springframework.boot.autoconfigure.hazelcast.HazelcastServerConfiguration$HazelcastServerConfigFileConfiguration$$EnhancerBySpringCGLIB$$d6cfebe6.hazelcastInstance(<generated>)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:154)
at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:583)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1246)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1096)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:535)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:495)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:317)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:222)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:315)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:199)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:251)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1135)
at org.springframework.beans.factory.support.DefaultListableBeanFactory$DependencyObjectProvider.getObject(DefaultListableBeanFactory.java:1665)
at org.springframework.session.hazelcast.config.annotation.web.http.HazelcastHttpSessionConfiguration.setHazelcastInstance(HazelcastHttpSessionConfiguration.java:96)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.springframework.beans.factory.annotation.AutowiredAnnotationBeanPostProcessor$AutowiredMethodElement.inject(AutowiredAnnotationBeanPostProcessor.java:696)
at org.springframework.beans.factory.annotation.InjectionMetadata.inject(InjectionMetadata.java:90)
at org.springframework.beans.factory.annotation.AutowiredAnnotationBeanPostProcessor.postProcessPropertyValues(AutowiredAnnotationBeanPostProcessor.java:370)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.populateBean(AbstractAutowireCapableBeanFactory.java:1336)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:572)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:495)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:317)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:222)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:315)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:199)
at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:373)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1246)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1096)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:535)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:495)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:317)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:222)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:315)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:199)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:251)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1135)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveDependency(DefaultListableBeanFactory.java:1062)
at org.springframework.beans.factory.support.ConstructorResolver.resolveAutowiredArgument(ConstructorResolver.java:819)
at org.springframework.beans.factory.support.ConstructorResolver.createArgumentArray(ConstructorResolver.java:725)
at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:475)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1246)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1096)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:535)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:495)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:317)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:222)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:315)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:204)
at org.springframework.boot.web.servlet.ServletContextInitializerBeans.getOrderedBeansOfType(ServletContextInitializerBeans.java:226)
at org.springframework.boot.web.servlet.ServletContextInitializerBeans.getOrderedBeansOfType(ServletContextInitializerBeans.java:214)
at org.springframework.boot.web.servlet.ServletContextInitializerBeans.addServletContextInitializerBeans(ServletContextInitializerBeans.java:91)
at org.springframework.boot.web.servlet.ServletContextInitializerBeans.<init>(ServletContextInitializerBeans.java:80)
at org.springframework.boot.web.servlet.context.ServletWebServerApplicationContext.getServletContextInitializerBeans(ServletWebServerApplicationContext.java:250)
at org.springframework.boot.web.servlet.context.ServletWebServerApplicationContext.selfInitialize(ServletWebServerApplicationContext.java:237)
at org.springframework.boot.web.embedded.tomcat.TomcatStarter.onStartup(TomcatStarter.java:54)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5245)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:150)
at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1420)
at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1410)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Your namespace in hazelcast.xml must be the same as the prefix of the properties in eureka-client.properties.
In other words, you need to either change the namespace to:
<property name="namespace">hazelcast</property>
Or change your eureka-client.properties to:
hazelcast-app-name.shouldUseDns=false
hazelcast-app-name.datacenter=primary
hazelcast-app-name.name=hazelcast-app-name-sessions
hazelcast-app-name.serviceUrl.default=http://username:password#svcregistry1-dev.company.com:8580/eureka/,http://username:password#svcregistry2-dev.company.com:8590/eureka/
Please read more at:
Hazelcast Eureka Plugin GH repository
Hazelcast Eureka Plugin Code Sample

Karate-Cucumber Report - No report file was added

I had to run maven clean command by mistake and the target report folder go cleansed. Now, If I build and run my feature file i get below error. How can I resolve this?
BTW, I am using karate version 0.9.0 (Even though it might not depend)
And also, How can I clean the report folder if its causing below issues?
19:07:59.921 [main] INFO com.intuit.karate.Runner - Karate version: 0.9.0
Karate version: 0.9.0
======================================================
elapsed: 0.01 | threads: 5 | thread time: 0.00
features: 0 | ignored: 0 | efficiency: 0.00
scenarios: 0 | passed: 0 | failed: 0
======================================================
Dec 13, 2018 7:08:00 PM net.masterthought.cucumber.ReportBuilder generateErrorPage
INFO: Unexpected error
net.masterthought.cucumber.ValidationException: No report file was added!
at net.masterthought.cucumber.ReportParser.parseJsonFiles(ReportParser.java:58)
at net.masterthought.cucumber.ReportBuilder.generateReports(ReportBuilder.java:88)
at PP1.RunnerClasses.Validationrunner.generateReport(Validationrunner.java:45)
at PP1.RunnerClasses.Validationrunner.testParallel(Validationrunner.java:35)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:50)
at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192)
19:08:00.696 [main] DEBUG org.apache.velocity - CommonsLogLogChute name is 'org.apache.velocity'
19:08:00.696 [main] DEBUG org.apache.velocity - Initializing Velocity, Calling init()...
19:08:00.697 [main] DEBUG org.apache.velocity - Starting Apache Velocity v1.7 (compiled: 2010-11-19 12:14:37)
19:08:00.697 [main] DEBUG org.apache.velocity - Default Properties File: org\apache\velocity\runtime\defaults\velocity.properties
19:08:00.697 [main] DEBUG org.apache.velocity - Trying to use logger class org.apache.velocity.runtime.log.AvalonLogChute
19:08:00.697 [main] DEBUG org.apache.velocity - Target log system for org.apache.velocity.runtime.log.AvalonLogChute is not available (java.lang.NoClassDefFoundError: org/apache/log/format/Formatter). Falling back to next log system...
19:08:00.697 [main] DEBUG org.apache.velocity - Trying to use logger class org.apache.velocity.runtime.log.Log4JLogChute
19:08:00.699 [main] DEBUG org.apache.velocity - Target log system for org.apache.velocity.runtime.log.Log4JLogChute is not available (java.lang.NoClassDefFoundError: org/apache/log4j/Layout). Falling back to next log system...
19:08:00.700 [main] DEBUG org.apache.velocity - Trying to use logger class org.apache.velocity.runtime.log.CommonsLogLogChute
19:08:00.700 [main] DEBUG org.apache.velocity - Using logger class org.apache.velocity.runtime.log.CommonsLogLogChute
19:08:00.709 [main] DEBUG org.apache.velocity - ResourceLoader instantiated: org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:00.735 [main] DEBUG org.apache.velocity - ResourceCache: initialized (class org.apache.velocity.runtime.resource.ResourceCacheImpl) with class java.util.Collections$SynchronizedMap cache map.
19:08:00.738 [main] DEBUG org.apache.velocity - Loaded System Directive: org.apache.velocity.runtime.directive.Stop
19:08:00.740 [main] DEBUG org.apache.velocity - Loaded System Directive: org.apache.velocity.runtime.directive.Define
19:08:00.742 [main] DEBUG org.apache.velocity - Loaded System Directive: org.apache.velocity.runtime.directive.Break
19:08:00.744 [main] DEBUG org.apache.velocity - Loaded System Directive: org.apache.velocity.runtime.directive.Evaluate
19:08:00.746 [main] DEBUG org.apache.velocity - Loaded System Directive: org.apache.velocity.runtime.directive.Literal
19:08:00.749 [main] DEBUG org.apache.velocity - Loaded System Directive: org.apache.velocity.runtime.directive.Macro
19:08:00.753 [main] DEBUG org.apache.velocity - Loaded System Directive: org.apache.velocity.runtime.directive.Parse
19:08:00.757 [main] DEBUG org.apache.velocity - Loaded System Directive: org.apache.velocity.runtime.directive.Include
19:08:00.761 [main] DEBUG org.apache.velocity - Loaded System Directive: org.apache.velocity.runtime.directive.Foreach
19:08:00.874 [main] DEBUG org.apache.velocity - Created '20' parsers.
19:08:00.889 [main] DEBUG org.apache.velocity - Velocimacro : "velocimacro.library" is not set. Trying default library: VM_global_library.vm
19:08:00.890 [main] DEBUG org.apache.velocity - Could not load resource 'VM_global_library.vm' from ResourceLoader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader: ClasspathResourceLoader Error: cannot find resource VM_global_library.vm
19:08:00.890 [main] DEBUG org.apache.velocity - Velocimacro : Default library not found.
19:08:00.890 [main] DEBUG org.apache.velocity - Velocimacro : allowInline = true : VMs can be defined inline in templates
19:08:00.890 [main] DEBUG org.apache.velocity - Velocimacro : allowInlineToOverride = false : VMs defined inline may NOT replace previous VM definitions
19:08:00.890 [main] DEBUG org.apache.velocity - Velocimacro : allowInlineLocal = false : VMs defined inline will be global in scope if allowed.
19:08:00.890 [main] DEBUG org.apache.velocity - Velocimacro : autoload off : VM system will not automatically reload global library macros
19:08:01.243 [main] DEBUG org.apache.velocity - ResourceManager : found templates/generators/errorpage.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.258 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/headers.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.259 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/head.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.265 [main] DEBUG org.apache.velocity - Velocimacro : added VM stringArray: source=/templates/macros/array.js.vm
19:08:01.265 [main] DEBUG org.apache.velocity - Velocimacro : added VM numberArray: source=/templates/macros/array.js.vm
19:08:01.265 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/array.js.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.269 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeBuildInfo: source=/templates/macros/page/buildinfo.vm
19:08:01.270 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/page/buildinfo.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.278 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeClassifications: source=/templates/macros/page/classifications.vm
19:08:01.278 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/page/classifications.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.280 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeLead: source=/templates/macros/page/lead.vm
19:08:01.280 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/page/lead.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.286 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeNavigation: source=/templates/macros/page/navigation.vm
19:08:01.286 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/page/navigation.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.291 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeReportInfo: source=/templates/macros/page/reportInfo.vm
19:08:01.292 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/page/reportInfo.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.295 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeTitle: source=/templates/macros/page/title.vm
19:08:01.295 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/page/title.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.304 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeExpandingButtons: source=/templates/macros/report/expandAllButtons.vm
19:08:01.304 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/report/expandAllButtons.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.312 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeReportTable: source=/templates/macros/report/reportTable.vm
19:08:01.313 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/report/reportTable.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.318 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeReportHeader: source=/templates/macros/report/reportHeader.vm
19:08:01.318 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/report/reportHeader.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.327 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeStatsTable: source=/templates/macros/report/statsTable.vm
19:08:01.327 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/report/statsTable.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.330 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeBrief: source=/templates/macros/json/brief.vm
19:08:01.330 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/json/brief.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.335 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeDocString: source=/templates/macros/json/docstring.vm
19:08:01.337 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/json/docstring.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.342 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeDuration: source=/templates/macros/json/duration.vm
19:08:01.342 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/json/duration.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.347 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeElement: source=/templates/macros/json/element.vm
19:08:01.347 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/json/element.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.385 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeEmbeddings: source=/templates/macros/json/embeddings.vm
19:08:01.386 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeHtmlEmbedding: source=/templates/macros/json/embeddings.vm
19:08:01.387 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeImageEmbedding: source=/templates/macros/json/embeddings.vm
19:08:01.388 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeImageReferenceEmbedding: source=/templates/macros/json/embeddings.vm
19:08:01.388 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeTextEmbedding: source=/templates/macros/json/embeddings.vm
19:08:01.389 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeUnknownEmbedding: source=/templates/macros/json/embeddings.vm
19:08:01.389 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/json/embeddings.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.392 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeHooks: source=/templates/macros/json/hooks.vm
19:08:01.392 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/json/hooks.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.397 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeOutput: source=/templates/macros/json/output.vm
19:08:01.397 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/json/output.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.401 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeMessage: source=/templates/macros/json/message.vm
19:08:01.401 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/json/message.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.406 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeSteps: source=/templates/macros/json/steps.vm
19:08:01.406 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/json/steps.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.409 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeStepName: source=/templates/macros/json/stepName.vm
19:08:01.409 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/json/stepName.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.412 [main] DEBUG org.apache.velocity - Velocimacro : added VM includeTags: source=/templates/macros/json/tags.vm
19:08:01.413 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/macros/json/tags.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
19:08:01.440 [main] DEBUG org.apache.velocity - VM #includeNavigation: too few arguments to macro. Wanted 1 got 0
19:08:01.442 [main] DEBUG org.apache.velocity - Left side ($active_tab) of '==' operation has null value. If it is a reference, it may not be in the context or its toString() returned null. /templates/macros/page/navigation.vm[line 14, column 51]
19:08:01.442 [main] DEBUG org.apache.velocity - Left side ($active_tab) of '==' operation has null value. If it is a reference, it may not be in the context or its toString() returned null. /templates/macros/page/navigation.vm[line 15, column 51]
19:08:01.442 [main] DEBUG org.apache.velocity - Left side ($active_tab) of '==' operation has null value. If it is a reference, it may not be in the context or its toString() returned null. /templates/macros/page/navigation.vm[line 16, column 51]
19:08:01.443 [main] DEBUG org.apache.velocity - Left side ($active_tab) of '==' operation has null value. If it is a reference, it may not be in the context or its toString() returned null. /templates/macros/page/navigation.vm[line 20, column 51]
19:08:01.465 [main] DEBUG org.apache.velocity - Null reference [template 'templates/generators/errorpage.vm', line 23, column 85] : $build_number cannot be resolved.
19:08:01.465 [main] DEBUG org.apache.velocity - Null reference [template 'templates/generators/errorpage.vm', line 23, column 85] : $build_number cannot be resolved.
19:08:01.480 [main] DEBUG org.apache.velocity - ResourceManager : found /templates/footer.vm with loader org.apache.velocity.runtime.resource.loader.ClasspathResourceLoader
Try this, it will work for the karate 1.0.1 version
/**
* Use this method if you want to execute entire feature files available in any folder.
*/
#Test
public void runMultipleFeatureFiles() {
System.setProperty("karate.env", "qa");
Results results = Runner.builder().outputCucumberJson(true)
.path("classpath:features/apitests/testFeatures").tags("~#Ignore").parallel(2);
Assertions.assertEquals(0, results.getFailCount());
generateReport(results.getReportDir());
}
/**
* Generate test case execution report post-execution.
*
* #param karateOutputPath - directory where you want to keep a report of test case executions.
*/
public static void generateReport(String karateOutputPath) {
Collection<File> jsonFiles =
FileUtils.listFiles(new File(karateOutputPath), new String[] {"json"}, true);
List<String> jsonPaths = new ArrayList<String>(jsonFiles.size());
jsonFiles.forEach(file -> jsonPaths.add(file.getAbsolutePath()));
Configuration config = new Configuration(new File("target"), "ReqRes.in SampleAPI Testing");
ReportBuilder reportBuilder = new ReportBuilder(jsonPaths, config);
reportBuilder.generateReports();
}
In Pom.xml add this maven dependencies:
<dependency>
<groupId>net.masterthought</groupId>
<artifactId>cucumber-reporting</artifactId>
<version>5.3.1</version>
<scope>test</scope>
</dependency>
Solved myself; Used karate-junit4 instead of karate-junit5 which is was using earlier in my .pom file.
But, I still don't understand why this error is caused if i used junit5!
//My runner file
#Test
void testParallel() {
Results results = Runner.path("classpath:conduitApp")
.outputCucumberJson(true)
.tags("~#ignore").parallel(5);
generateReport(results.getReportDir());
assertEquals(0, results.getFailCount(), results.getErrorMessages());
}
Note that outputCucumberJson(true) is called on the Runner “builder” to make it work.

Cassandra Startup Error: DseModule.java:103 - org/apache/spark/util/ByteBufferInputStream. Exiting

I installed Cassandra, DSE after working on DSE Graphs for a while I installed Spark, R. Now I am unable to start DSE. Cassandra process is not starting
I tried disabling Spark and so on. Not sure whats the issue!
$ /Users/212500500/dse_510/bin/dse cassandra -g -s
Server will start with the original workload configuration. Edit the dse script to change that (the script that printed this message).
Search: on
Graph: on
Tomcat: Logging to /Users/212500500/dse_510/logs/tomcat
Tomcat: Logging to /Users/212500500/dse_510/logs/tomcat
SFO1212500500M:dse_510 212500500$ objc[99199]: Class JavaLaunchHelper is implemented in both /Library/Java/JavaVirtualMachines/jdk1.8.0_66.jdk/Contents/Home/bin/java and /Library/Java/JavaVirtualMachines/jdk1.8.0_66.jdk/Contents/Home/jre/lib/libinstrument.dylib. One of the two will be used. Which one is undefined.
CompilerOracle: dontinline org/apache/cassandra/db/Columns$Serializer.deserializeLargeSubset (Lorg/apache/cassandra/io/util/DataInputPlus;Lorg/apache/cassandra/db/Columns;I)Lorg/apache/cassandra/db/Columns;
CompilerOracle: dontinline org/apache/cassandra/db/Columns$Serializer.serializeLargeSubset (Ljava/util/Collection;ILorg/apache/cassandra/db/Columns;ILorg/apache/cassandra/io/util/DataOutputPlus;)V
CompilerOracle: dontinline org/apache/cassandra/db/Columns$Serializer.serializeLargeSubsetSize (Ljava/util/Collection;ILorg/apache/cassandra/db/Columns;I)I
CompilerOracle: dontinline org/apache/cassandra/db/commitlog/AbstractCommitLogSegmentManager.advanceAllocatingFrom (Lorg/apache/cassandra/db/commitlog/CommitLogSegment;)V
CompilerOracle: dontinline org/apache/cassandra/db/transform/BaseIterator.tryGetMoreContents ()Z
CompilerOracle: dontinline org/apache/cassandra/db/transform/StoppingTransformation.stop ()V
CompilerOracle: dontinline org/apache/cassandra/db/transform/StoppingTransformation.stopInPartition ()V
CompilerOracle: dontinline org/apache/cassandra/io/util/BufferedDataOutputStreamPlus.doFlush (I)V
CompilerOracle: dontinline org/apache/cassandra/io/util/BufferedDataOutputStreamPlus.writeExcessSlow ()V
CompilerOracle: dontinline org/apache/cassandra/io/util/BufferedDataOutputStreamPlus.writeSlow (JI)V
CompilerOracle: dontinline org/apache/cassandra/io/util/RebufferingInputStream.readPrimitiveSlowly (I)J
CompilerOracle: inline org/apache/cassandra/db/rows/UnfilteredSerializer.serializeRowBody (Lorg/apache/cassandra/db/rows/Row;ILorg/apache/cassandra/db/SerializationHeader;Lorg/apache/cassandra/io/util/DataOutputPlus;)V
CompilerOracle: inline org/apache/cassandra/io/util/Memory.checkBounds (JJ)V
CompilerOracle: inline org/apache/cassandra/io/util/SafeMemory.checkBounds (JJ)V
CompilerOracle: inline org/apache/cassandra/utils/AsymmetricOrdering.selectBoundary (Lorg/apache/cassandra/utils/AsymmetricOrdering/Op;II)I
CompilerOracle: inline org/apache/cassandra/utils/AsymmetricOrdering.strictnessOfLessThan (Lorg/apache/cassandra/utils/AsymmetricOrdering/Op;)I
CompilerOracle: inline org/apache/cassandra/utils/BloomFilter.indexes (Lorg/apache/cassandra/utils/IFilter/FilterKey;)[J
CompilerOracle: inline org/apache/cassandra/utils/BloomFilter.setIndexes (JJIJ[J)V
CompilerOracle: inline org/apache/cassandra/utils/ByteBufferUtil.compare (Ljava/nio/ByteBuffer;[B)I
CompilerOracle: inline org/apache/cassandra/utils/ByteBufferUtil.compare ([BLjava/nio/ByteBuffer;)I
CompilerOracle: inline org/apache/cassandra/utils/ByteBufferUtil.compareUnsigned (Ljava/nio/ByteBuffer;Ljava/nio/ByteBuffer;)I
CompilerOracle: inline org/apache/cassandra/utils/FastByteOperations$UnsafeOperations.compareTo (Ljava/lang/Object;JILjava/lang/Object;JI)I
CompilerOracle: inline org/apache/cassandra/utils/FastByteOperations$UnsafeOperations.compareTo (Ljava/lang/Object;JILjava/nio/ByteBuffer;)I
CompilerOracle: inline org/apache/cassandra/utils/FastByteOperations$UnsafeOperations.compareTo (Ljava/nio/ByteBuffer;Ljava/nio/ByteBuffer;)I
CompilerOracle: inline org/apache/cassandra/utils/vint/VIntCoding.encodeVInt (JI)[B
17:31:09,605 |-INFO in ch.qos.logback.classic.LoggerContext[default] - Found resource [logback.xml] at [file:/Users/212500500/dse_510/resources/cassandra/conf/logback.xml]
17:31:09,749 |-INFO in ch.qos.logback.classic.joran.action.ConfigurationAction - debug attribute not set
17:31:09,753 |-INFO in ReconfigureOnChangeFilter{invocationCounter=0} - Will scan for changes in [[/Users/212500500/dse_510/resources/cassandra/conf/logback.xml]] every 60 seconds.
17:31:09,753 |-INFO in ch.qos.logback.classic.joran.action.ConfigurationAction - Adding ReconfigureOnChangeFilter as a turbo filter
17:31:09,764 |-INFO in ch.qos.logback.classic.joran.action.JMXConfiguratorAction - begin
17:31:09,850 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.rolling.RollingFileAppender]
17:31:09,853 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [SYSTEMLOG]
17:31:09,896 |-INFO in ch.qos.logback.core.rolling.FixedWindowRollingPolicy#795cd85e - Will use zip compression
17:31:09,906 |-INFO in ch.qos.logback.core.joran.action.NestedComplexPropertyIA - Assuming default type [ch.qos.logback.classic.encoder.PatternLayoutEncoder] for [encoder] property
17:31:09,928 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[SYSTEMLOG] - Active log file name: /Users/212500500/dse_510/logs/cassandra/system.log
17:31:09,928 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[SYSTEMLOG] - File property is set to [/Users/212500500/dse_510/logs/cassandra/system.log]
17:31:09,930 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.rolling.RollingFileAppender]
17:31:09,930 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [DEBUGLOG]
17:31:09,930 |-INFO in ch.qos.logback.core.rolling.FixedWindowRollingPolicy#59fd97a8 - Will use zip compression
17:31:09,931 |-INFO in ch.qos.logback.core.joran.action.NestedComplexPropertyIA - Assuming default type [ch.qos.logback.classic.encoder.PatternLayoutEncoder] for [encoder] property
17:31:09,931 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[DEBUGLOG] - Active log file name: /Users/212500500/dse_510/logs/cassandra/debug.log
17:31:09,931 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[DEBUGLOG] - File property is set to [/Users/212500500/dse_510/logs/cassandra/debug.log]
17:31:09,932 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.classic.AsyncAppender]
17:31:09,933 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [ASYNCDEBUGLOG]
17:31:09,938 |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [DEBUGLOG] to ch.qos.logback.classic.AsyncAppender[ASYNCDEBUGLOG]
17:31:09,938 |-INFO in ch.qos.logback.classic.AsyncAppender[ASYNCDEBUGLOG] - Attaching appender named [DEBUGLOG] to AsyncAppender.
17:31:09,938 |-INFO in ch.qos.logback.classic.AsyncAppender[ASYNCDEBUGLOG] - Setting discardingThreshold to 0
17:31:10,094 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.ConsoleAppender]
17:31:10,095 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [STDOUT]
17:31:10,097 |-INFO in ch.qos.logback.core.joran.action.NestedComplexPropertyIA - Assuming default type [ch.qos.logback.classic.encoder.PatternLayoutEncoder] for [encoder] property
17:31:10,098 |-INFO in ch.qos.logback.core.joran.util.ConfigurationWatchListUtil#f5ac9e4 - Adding [file:/Users/212500500/dse_510/SPARK_SERVER_LOGBACK_CONF_FILE_IS_UNDEFINED] to configuration watch list.
17:31:10,098 |-WARN in ch.qos.logback.core.joran.action.IncludeAction - Failed to open [file:/Users/212500500/dse_510/SPARK_SERVER_LOGBACK_CONF_FILE_IS_UNDEFINED]
17:31:10,098 |-INFO in ch.qos.logback.core.joran.util.ConfigurationWatchListUtil#f5ac9e4 - Adding [file:/Users/212500500/dse_510/resources/graph/conf/logback-gremlin-server.xml] to configuration watch list.
17:31:10,101 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.rolling.RollingFileAppender]
17:31:10,101 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [GremlinServerFileAppender]
17:31:10,104 |-INFO in ch.qos.logback.core.joran.action.NestedComplexPropertyIA - Assuming default type [ch.qos.logback.classic.boolex.JaninoEventEvaluator] for [evaluator] property
17:31:10,122 |-INFO in ch.qos.logback.core.rolling.FixedWindowRollingPolicy#123ef382 - Will use zip compression
17:31:10,123 |-INFO in ch.qos.logback.core.joran.action.NestedComplexPropertyIA - Assuming default type [ch.qos.logback.classic.encoder.PatternLayoutEncoder] for [encoder] property
17:31:10,123 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[GremlinServerFileAppender] - Active log file name: /Users/212500500/dse_510/logs/cassandra/gremlin.log
17:31:10,123 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[GremlinServerFileAppender] - File property is set to [/Users/212500500/dse_510/logs/cassandra/gremlin.log]
17:31:10,124 |-INFO in ch.qos.logback.classic.joran.action.RootLoggerAction - Setting level of ROOT logger to INFO
17:31:10,124 |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [SYSTEMLOG] to Logger[ROOT]
17:31:10,124 |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [STDOUT] to Logger[ROOT]
17:31:10,124 |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [ASYNCDEBUGLOG] to Logger[ROOT]
17:31:10,124 |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [GremlinServerFileAppender] to Logger[ROOT]
17:31:10,124 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.rolling.RollingFileAppender]
17:31:10,124 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [SLF4JAuditWriterAppender]
17:31:10,124 |-INFO in ch.qos.logback.core.joran.action.NestedComplexPropertyIA - Assuming default type [ch.qos.logback.classic.encoder.PatternLayoutEncoder] for [encoder] property
17:31:10,125 |-INFO in ch.qos.logback.core.rolling.FixedWindowRollingPolicy#dbf57b3 - Will use zip compression
17:31:10,126 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[SLF4JAuditWriterAppender] - Active log file name: /Users/212500500/dse_510/logs/cassandra/audit/audit.log
17:31:10,126 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[SLF4JAuditWriterAppender] - File property is set to [/Users/212500500/dse_510/logs/cassandra/audit/audit.log]
17:31:10,127 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [SLF4JAuditWriter] to INFO
17:31:10,127 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting additivity of logger [SLF4JAuditWriter] to false
17:31:10,127 |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [SLF4JAuditWriterAppender] to Logger[SLF4JAuditWriter]
17:31:10,127 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.rolling.RollingFileAppender]
17:31:10,127 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [DroppedAuditEventAppender]
17:31:10,127 |-INFO in ch.qos.logback.core.joran.action.NestedComplexPropertyIA - Assuming default type [ch.qos.logback.classic.encoder.PatternLayoutEncoder] for [encoder] property
17:31:10,128 |-INFO in ch.qos.logback.core.rolling.FixedWindowRollingPolicy#384ad17b - Will use zip compression
17:31:10,128 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[DroppedAuditEventAppender] - Active log file name: /Users/212500500/dse_510/logs/cassandra/audit/dropped-events.log
17:31:10,128 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[DroppedAuditEventAppender] - File property is set to [/Users/212500500/dse_510/logs/cassandra/audit/dropped-events.log]
17:31:10,128 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [DroppedAuditEventLogger] to INFO
17:31:10,128 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting additivity of logger [DroppedAuditEventLogger] to false
17:31:10,128 |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [DroppedAuditEventAppender] to Logger[DroppedAuditEventLogger]
17:31:10,129 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.cassandra] to DEBUG
17:31:10,129 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [com.thinkaurelius.thrift] to ERROR
17:31:10,129 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.rolling.RollingFileAppender]
17:31:10,129 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [SolrValidationErrorAppender]
17:31:10,129 |-INFO in ch.qos.logback.core.rolling.FixedWindowRollingPolicy#61862a7f - Will use zip compression
17:31:10,130 |-INFO in ch.qos.logback.core.joran.action.NestedComplexPropertyIA - Assuming default type [ch.qos.logback.classic.encoder.PatternLayoutEncoder] for [encoder] property
17:31:10,130 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[SolrValidationErrorAppender] - Active log file name: /Users/212500500/dse_510/logs/cassandra/solrvalidation.log
17:31:10,130 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[SolrValidationErrorAppender] - File property is set to [/Users/212500500/dse_510/logs/cassandra/solrvalidation.log]
17:31:10,131 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [SolrValidationErrorLogger] to ERROR
17:31:10,131 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting additivity of logger [SolrValidationErrorLogger] to false
17:31:10,131 |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [SolrValidationErrorAppender] to Logger[SolrValidationErrorLogger]
17:31:10,131 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [com.datastax.bdp.search.solr.metrics.MetricsWriteEventListener] to DEBUG
17:31:10,131 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.solr.core.CassandraSolrConfig] to WARN
17:31:10,131 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.solr.core.SolrCore] to WARN
17:31:10,131 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.solr.core.RequestHandlers] to WARN
17:31:10,131 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.solr.handler.component] to WARN
17:31:10,131 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.solr.search.SolrIndexSearcher] to WARN
17:31:10,131 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.solr.update] to WARN
17:31:10,131 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.lucene.index] to INFO
17:31:10,131 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [com.cryptsoft] to OFF
17:31:10,131 |-INFO in ch.qos.logback.classic.joran.action.ConfigurationAction - End of configuration.
17:31:10,131 |-INFO in ch.qos.logback.classic.joran.JoranConfigurator#441772e - Registering current configuration as safe fallback point
INFO [main] 2017-04-27 17:31:10,249 DseModule.java:90 - Loading DSE module
INFO [main] 2017-04-27 17:31:10,493 YamlConfigurationLoader.java:89 - Configuration location: file:/Users/212500500/dse_510/resources/cassandra/conf/cassandra.yaml
INFO [main] 2017-04-27 17:31:10,782 YamlConfigurationLoader.java:89 - Configuration location: file:/Users/212500500/dse_510/resources/cassandra/conf/cassandra.yaml
INFO [main] 2017-04-27 17:31:10,811 Config.java:481 - Node configuration:[allocate_tokens_for_keyspace=null; allocate_tokens_for_local_replication_factor=null; authenticator=com.datastax.bdp.cassandra.auth.DseAuthenticator; authorizer=com.datastax.bdp.cassandra.auth.DseAuthorizer; auto_bootstrap=true; auto_snapshot=true; back_pressure_enabled=false; back_pressure_strategy=org.apache.cassandra.net.RateBasedBackPressure{high_ratio=0.9, factor=5, flow=FAST}; batch_size_fail_threshold_in_kb=640; batch_size_warn_threshold_in_kb=64; batchlog_replay_throttle_in_kb=1024; broadcast_address=null; broadcast_rpc_address=null; buffer_pool_use_heap_if_exhausted=true; cas_contention_timeout_in_ms=1000; cdc_enabled=false; cdc_free_space_check_interval_ms=250; cdc_raw_directory=/var/lib/cassandra/cdc_raw; cdc_total_space_in_mb=0; client_encryption_options=<REDACTED>; cluster_name=Test Cluster; column_index_cache_size_in_kb=2; column_index_size_in_kb=64; commit_failure_policy=stop; commitlog_compression=null; commitlog_directory=/Users/212500500/dse_510/commitlog; commitlog_max_compression_buffers_in_pool=3; commitlog_periodic_queue_size=-1; commitlog_segment_size_in_mb=32; commitlog_sync=periodic; commitlog_sync_batch_window_in_ms=NaN; commitlog_sync_period_in_ms=10000; commitlog_total_space_in_mb=null; compaction_large_partition_warning_threshold_mb=100; compaction_throughput_mb_per_sec=16; concurrent_compactors=null; concurrent_counter_writes=32; concurrent_materialized_view_writes=32; concurrent_reads=32; concurrent_replicates=null; concurrent_writes=32; continuous_paging=org.apache.cassandra.config.ContinuousPagingConfig#9caca531; counter_cache_keys_to_save=2147483647; counter_cache_save_period=7200; counter_cache_size_in_mb=null; counter_write_request_timeout_in_ms=5000; credentials_cache_max_entries=1000; credentials_update_interval_in_ms=-1; credentials_validity_in_ms=2000; cross_node_timeout=false; data_file_directories=[Ljava.lang.String;#ff684e1; disk_access_mode=auto; disk_failure_policy=stop; disk_optimization_estimate_percentile=0.95; disk_optimization_page_cross_chance=0.1; disk_optimization_strategy=ssd; dynamic_snitch=true; dynamic_snitch_badness_threshold=0.1; dynamic_snitch_reset_interval_in_ms=600000; dynamic_snitch_update_interval_in_ms=100; enable_scripted_user_defined_functions=false; enable_user_defined_functions=false; enable_user_defined_functions_threads=true; encryption_options=null; endpoint_snitch=com.datastax.bdp.snitch.DseDelegateSnitch; file_cache_size_in_mb=null; gc_log_threshold_in_ms=200; gc_warn_threshold_in_ms=1000; hinted_handoff_disabled_datacenters=[]; hinted_handoff_enabled=true; hinted_handoff_throttle_in_kb=1024; hints_compression=null; hints_directory=/Users/212500500/dse_510/hints; hints_flush_period_in_ms=10000; incremental_backups=false; index_interval=null; index_summary_capacity_in_mb=null; index_summary_resize_interval_in_minutes=60; initial_token=null; inter_dc_stream_throughput_outbound_megabits_per_sec=200; inter_dc_tcp_nodelay=false; internode_authenticator=null; internode_compression=dc; internode_recv_buff_size_in_bytes=0; internode_send_buff_size_in_bytes=0; key_cache_keys_to_save=2147483647; key_cache_save_period=14400; key_cache_size_in_mb=null; listen_address=127.0.0.1; listen_interface=null; listen_interface_prefer_ipv6=false; listen_on_broadcast_address=false; max_hint_window_in_ms=10800000; max_hints_delivery_threads=2; max_hints_file_size_in_mb=128; max_mutation_size_in_kb=null; max_streaming_retries=3; max_value_size_in_mb=256; memtable_allocation_type=heap_buffers; memtable_cleanup_threshold=null; memtable_flush_writers=0; memtable_heap_space_in_mb=null; memtable_offheap_space_in_mb=null; min_free_space_per_drive_in_mb=50; native_transport_max_concurrent_connections=-1; native_transport_max_concurrent_connections_per_ip=-1; native_transport_max_frame_size_in_mb=256; native_transport_max_threads=128; native_transport_port=9042; native_transport_port_ssl=null; num_tokens=1; otc_coalescing_enough_coalesced_messages=8; otc_coalescing_strategy=DISABLED; otc_coalescing_window_us=200; partitioner=org.apache.cassandra.dht.Murmur3Partitioner; permissions_cache_max_entries=1000; permissions_update_interval_in_ms=-1; permissions_validity_in_ms=2000; phi_convict_threshold=8.0; prepared_statements_cache_size_mb=null; range_request_timeout_in_ms=10000; read_request_timeout_in_ms=5000; request_scheduler=org.apache.cassandra.scheduler.NoScheduler; request_scheduler_id=null; request_scheduler_options=null; request_timeout_in_ms=10000; role_manager=com.datastax.bdp.cassandra.auth.DseRoleManager; roles_cache_max_entries=1000; roles_update_interval_in_ms=-1; roles_validity_in_ms=2000; row_cache_class_name=org.apache.cassandra.cache.OHCProvider; row_cache_keys_to_save=2147483647; row_cache_save_period=0; row_cache_size_in_mb=0; rpc_address=127.0.0.1; rpc_interface=null; rpc_interface_prefer_ipv6=false; rpc_keepalive=true; rpc_listen_backlog=50; rpc_max_threads=2147483647; rpc_min_threads=16; rpc_port=9160; rpc_recv_buff_size_in_bytes=null; rpc_send_buff_size_in_bytes=null; rpc_server_type=sync; saved_caches_directory=/Users/212500500/dse_510/saved_caches; seed_provider=org.apache.cassandra.locator.SimpleSeedProvider{seeds=127.0.0.1}; server_encryption_options=<REDACTED>; slow_query_log_timeout_in_ms=500; snapshot_before_compaction=false; ssl_storage_port=7001; sstable_preemptive_open_interval_in_mb=50; start_native_transport=true; start_rpc=true; storage_port=7000; stream_throughput_outbound_megabits_per_sec=200; streaming_keep_alive_period_in_secs=300; streaming_socket_timeout_in_ms=86400000; thrift_framed_transport_size_in_mb=15; thrift_max_message_length_in_mb=16; thrift_prepared_statements_cache_size_mb=null; tombstone_failure_threshold=100000; tombstone_warn_threshold=1000; tracetype_query_ttl=86400; tracetype_repair_ttl=604800; transparent_data_encryption_options=org.apache.cassandra.config.TransparentDataEncryptionOptions#23aa363a; trickle_fsync=true; trickle_fsync_interval_in_kb=10240; truncate_request_timeout_in_ms=60000; unlogged_batch_across_partitions_warn_threshold=10; user_defined_function_fail_timeout=1500; user_defined_function_warn_timeout=500; user_function_timeout_policy=die; windows_timer_interval=1; write_request_timeout_in_ms=2000]
INFO [main] 2017-04-27 17:31:10,812 DatabaseDescriptor.java:366 - DiskAccessMode 'auto' determined to be mmap, indexAccessMode is mmap
INFO [main] 2017-04-27 17:31:10,813 DatabaseDescriptor.java:420 - Global memtable on-heap threshold is enabled at 1024MB
INFO [main] 2017-04-27 17:31:10,813 DatabaseDescriptor.java:424 - Global memtable off-heap threshold is enabled at 1024MB
INFO [main] 2017-04-27 17:31:10,838 RateBasedBackPressure.java:123 - Initialized back-pressure with high ratio: 0.9, factor: 5, flow: FAST, window size: 2000.
INFO [main] 2017-04-27 17:31:10,839 DatabaseDescriptor.java:717 - Back-pressure is disabled with strategy org.apache.cassandra.net.RateBasedBackPressure{high_ratio=0.9, factor=5, flow=FAST}.
INFO [main] 2017-04-27 17:31:10,861 DseDelegateSnitch.java:40 - Setting my workloads to [Cassandra, Search, Graph]
INFO [main] 2017-04-27 17:31:10,867 DseConfigYamlLoader.java:38 - Loading settings from file:/Users/212500500/dse_510/resources/dse/conf/dse.yaml
INFO [main] 2017-04-27 17:31:10,957 DseConfig.java:382 - Load of settings is done.
INFO [main] 2017-04-27 17:31:10,971 YamlConfigurationLoader.java:89 - Configuration location: file:/Users/212500500/dse_510/resources/cassandra/conf/cassandra.yaml
INFO [main] 2017-04-27 17:31:10,973 DseDelegateSnitch.java:42 - Initialized DseDelegateSnitch with workloads [Cassandra, Search, Graph], delegating to com.datastax.bdp.snitch.DseSimpleSnitch
INFO [main] 2017-04-27 17:31:10,995 YamlConfigurationLoader.java:89 - Configuration location: file:/Users/212500500/dse_510/resources/cassandra/conf/cassandra.yaml
INFO [main] 2017-04-27 17:31:11,042 DseConfig.java:402 - CQL slow log is enabled
INFO [main] 2017-04-27 17:31:11,043 DseConfig.java:403 - CQL system info tables are not enabled
INFO [main] 2017-04-27 17:31:11,043 DseConfig.java:404 - Resource level latency tracking is not enabled
INFO [main] 2017-04-27 17:31:11,043 DseConfig.java:405 - Database summary stats are not enabled
INFO [main] 2017-04-27 17:31:11,043 DseConfig.java:406 - Cluster summary stats are not enabled
INFO [main] 2017-04-27 17:31:11,043 DseConfig.java:407 - Histogram data tables are not enabled
INFO [main] 2017-04-27 17:31:11,044 DseConfig.java:408 - User level latency tracking is not enabled
INFO [main] 2017-04-27 17:31:11,044 DseConfig.java:410 - Spark cluster info tables are not enabled
INFO [main] 2017-04-27 17:31:11,044 DseConfig.java:444 - Cql solr query paging is: off
INFO [main] 2017-04-27 17:31:11,046 DseConfig.java:459 - Server ID:A4-5E-60-F4-1B-DB
INFO [main] 2017-04-27 17:31:11,334 GraphYamlConfigurationReader.java:185 - Loading graph settings from file:/Users/212500500/dse_510/resources/dse/conf/dse.yaml
INFO [main] 2017-04-27 17:31:11,335 GraphYamlConfigurationReader.java:189 - Opened stream on file:/Users/212500500/dse_510/resources/dse/conf/dse.yaml
ERROR [main] 2017-04-27 17:31:11,385 DseModule.java:103 - org/apache/spark/util/ByteBufferInputStream. Exiting...
$ ps -ef |grep dse
0 724 1 0 Mon09PM ?? 0:05.04 /usr/sbin/systemsoundserverd
348149448 99945 1448 0 5:40PM ttys001 0:00.00 grep dse
I had saperately installed Apache Spark. Problem went away when I removed Spark (unset SPARK_HOME variable) and then restarted DSE Cassandra.

having issuses using sqoop import tool from oracle to cassandra

I am trying to load data from oracle table to cassandra table. I tried to follow the same steps mentioned in document from Datastax website on Running the Sqoop demo - https://docs.datastax.com/en/datastax_enterprise/4.5/datastax_enterprise/ana/anaSqpDemo.html
Here, I am using oracle database instaed of mysql. Using Datastax enterprise 5.0.2.
dse sqoop cql-import --connect 'jdbc:oracle:thin:username/password#//ip_address_of_the_host:PORT/SERVICE_NAME' --table ORACLE_TABLE_NAME --cassandra-keyspace npa_nxx --cassandra-table npa_nxx_data --cassandra-host IP_ADDRESS_CASSANDRA --cassandra-port 9042 --cassandra-column-mapping npa:npa,nxx:nxx,latitude:lat,longitude:lon,state:state,city:city
Hadoop functionality is deprecated and may be removed in a future release.
Note: /tmp/sqoop-xxxx/compile/4657cfc531e9676b9013e057157bf522/SSFS_STAGE02_NPA_NXX.java uses or overrides a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
ERROR 13:45:08,987 Encountered IOException running import job: java.io.IOException: Failed to read the table metadata
at com.datastax.bdp.sqoop.SqoopUtil.setCqlImportOptions(SqoopUtil.java:186)
at com.datastax.bdp.sqoop.CqlImportJob.configureOutputFormat(CqlImportJob.java:120)
at org.apache.sqoop.mapreduce.ImportJobBase.runImport(ImportJobBase.java:240)
at com.datastax.bdp.sqoop.SqoopUtil.importTable(SqoopUtil.java:587)
at com.datastax.bdp.sqoop.SqlManagerAdapter.importTable(SqlManagerAdapter.java:222)
at org.apache.sqoop.tool.ImportTool.importTable(ImportTool.java:497)
at org.apache.sqoop.tool.ImportTool.run(ImportTool.java:601)
at org.apache.sqoop.Sqoop.run(Sqoop.java:143)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
at org.apache.sqoop.Sqoop.runSqoop(Sqoop.java:179)
at org.apache.sqoop.Sqoop.runTool(Sqoop.java:218)
at org.apache.sqoop.Sqoop.runTool(Sqoop.java:227)
at org.apache.sqoop.Sqoop.main(Sqoop.java:236)
at com.cloudera.sqoop.Sqoop.main(Sqoop.java:57)
Caused by: java.io.IOException: --cassandra-column-mapping contains an SQL column city that does not exist in the SQL table or query
at com.datastax.bdp.sqoop.SqoopUtil.validateColumnConsistency(SqoopUtil.java:312)
at com.datastax.bdp.sqoop.SqoopUtil.setCqlImportOptions(SqoopUtil.java:168)
... 13 more
I have tested the jdbc connection for oracle and looks good.
Please help me to understand the issue and any suggestions are welcome.
Thanks
Raghav
The error
--cassandra-column-mapping contains an SQL column city that does not exist in the SQL table or query
Suggests that your Oracle Table was not created correctly, you should double check that it has the correct schema.
dse sqoop cql-import --connect 'jdbc:oracle:thin:username/password#//ip_address_of_the_host:PORT/SERVICE_NAME' --table ORACLE_TABLE_NAME --cassandra-keyspace npa_nxx --cassandra-table npa_nxx_data --cassandra-host IP_ADDRESS_CASSANDRA --cassandra-port 9042 --cassandra-column-mapping npa:NPA,nxx:NXX,latitude:LAT,longitude:LON,state:STATE,city:CITY
Hadoop functionality is deprecated and may be removed in a future release.
Note: /tmp/sqoop-tmhmadm/compile/06a746beb5d4af8ac13b60568fedbcbd/SSFS_STAGE02_NPA_NXX.java uses or overrides a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
WARN 11:23:58,503 Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
java.lang.Throwable: Child Error
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
Caused by: java.io.IOException: Task process exit with nonzero status of -1.
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
java.lang.Throwable: Child Error
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
Caused by: java.io.IOException: Task process exit with nonzero status of -1.
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
java.lang.Throwable: Child Error
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
Caused by: java.io.IOException: Task process exit with nonzero status of -1.
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
java.lang.Throwable: Child Error
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
Caused by: java.io.IOException: Task process exit with nonzero status of -1.
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
attempt_201610061121_0001_m_000001_0: 11:24:15,977 |-INFO in ch.qos.logback.classic.LoggerContext[default] - Could NOT find resource [logback.groovy]
attempt_201610061121_0001_m_000001_0: 11:24:15,978 |-INFO in ch.qos.logback.classic.LoggerContext[default] - Could NOT find resource [logback-test.xml]
attempt_201610061121_0001_m_000001_0: 11:24:15,979 |-INFO in ch.qos.logback.classic.LoggerContext[default] - Found resource [logback.xml] at [file:/etc/dse/cassandra/logback.xml]
attempt_201610061121_0001_m_000001_0: 11:24:16,412 |-INFO in ch.qos.logback.classic.joran.action.ConfigurationAction - debug attribute not set
attempt_201610061121_0001_m_000001_0: 11:24:16,418 |-INFO in ReconfigureOnChangeFilter{invocationCounter=0} - Will scan for changes in [[/etc/dse/cassandra/logback.xml]] every 60 seconds.
attempt_201610061121_0001_m_000001_0: 11:24:16,418 |-INFO in ch.qos.logback.classic.joran.action.ConfigurationAction - Adding ReconfigureOnChangeFilter as a turbo filter
attempt_201610061121_0001_m_000001_0: 11:24:16,432 |-INFO in ch.qos.logback.classic.joran.action.JMXConfiguratorAction - begin
attempt_201610061121_0001_m_000001_0: 11:24:16,617 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.rolling.RollingFileAppender]
attempt_201610061121_0001_m_000001_0: 11:24:16,624 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [SYSTEMLOG]
attempt_201610061121_0001_m_000001_0: 11:24:17,683 |-INFO in ch.qos.logback.core.rolling.FixedWindowRollingPolicy#74294adb - Will use zip compression
attempt_201610061121_0001_m_000001_0: 11:24:17,835 |-INFO in ch.qos.logback.core.joran.action.NestedComplexPropertyIA - Assuming default type [ch.qos.logback.classic.encoder.PatternLayoutEncoder] for [encoder] property
attempt_201610061121_0001_m_000001_0: 11:24:17,919 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[SYSTEMLOG] - Active log file name: cassandra.logdir_IS_UNDEFINED/system.log
attempt_201610061121_0001_m_000001_0: 11:24:17,919 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[SYSTEMLOG] - File property is set to [cassandra.logdir_IS_UNDEFINED/system.log]
attempt_201610061121_0001_m_000001_0: 11:24:17,921 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.rolling.RollingFileAppender]
attempt_201610061121_0001_m_000001_0: 11:24:17,921 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [DEBUGLOG]
.
.
.
.
attempt_201610061121_0001_m_000002_2: 11:24:50,116 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [com.thinkaurelius.thrift] to ERROR
attempt_201610061121_0001_m_000002_2: 11:24:50,116 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.rolling.RollingFileAppender]
attempt_201610061121_0001_m_000002_2: 11:24:50,116 |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [SolrValidationErrorAppender]
attempt_201610061121_0001_m_000002_2: 11:24:50,117 |-INFO in ch.qos.logback.core.rolling.FixedWindowRollingPolicy#276438c9 - Will use zip compression
attempt_201610061121_0001_m_000002_2: 11:24:50,117 |-WARN in ch.qos.logback.core.rolling.FixedWindowRollingPolicy#276438c9 - Large window sizes are not allowed.
attempt_201610061121_0001_m_000002_2: 11:24:50,118 |-WARN in ch.qos.logback.core.rolling.FixedWindowRollingPolicy#276438c9 - MaxIndex reduced to 21
attempt_201610061121_0001_m_000002_2: 11:24:50,118 |-INFO in ch.qos.logback.core.joran.action.NestedComplexPropertyIA - Assuming default type [ch.qos.logback.classic.encoder.PatternLayoutEncoder] for [encoder] property
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[SolrValidationErrorAppender] - Active log file name: cassandra.logdir_IS_UNDEFINED/solrvalidation.log
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[SolrValidationErrorAppender] - File property is set to [cassandra.logdir_IS_UNDEFINED/solrvalidation.log]
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [SolrValidationErrorLogger] to ERROR
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting additivity of logger [SolrValidationErrorLogger] to false
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [SolrValidationErrorAppender] to Logger[SolrValidationErrorLogger]
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [com.datastax.bdp.search.solr.metrics.MetricsWriteEventListener] to DEBUG
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.solr.core.CassandraSolrConfig] to WARN
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.solr.core.SolrCore] to WARN
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.solr.core.RequestHandlers] to WARN
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.solr.handler.component] to WARN
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.solr.search.SolrIndexSearcher] to WARN
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.solr.update] to WARN
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.lucene.index] to INFO
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [com.cryptsoft] to OFF
attempt_201610061121_0001_m_000002_2: 11:24:50,120 |-INFO in ch.qos.logback.classic.joran.action.ConfigurationAction - End of configuration.
attempt_201610061121_0001_m_000002_2: 11:24:50,121 |-INFO in ch.qos.logback.classic.joran.JoranConfigurator#588df31b - Registering current configuration as safe fallback point
attempt_201610061121_0001_m_000002_2: INFO 11:24:51,231 NativeCodeLoader.java:43 - Loaded the native-hadoop library
ERROR 11:25:28,355 Error during import: Import job failed!
Here is the DDL for Oracle table:
CREATE TABLE "SCHEMA_NAME"."NPA_NXX"
( "NPA_NXX_KEY" NUMBER(*,0) NOT NULL ENABLE,
"NPA" NUMBER(*,0),
"NXX" NUMBER(*,0),
"LAT" FLOAT(126),
"LON" FLOAT(126),
"LINETYPE" CHAR(1 BYTE),
"STATE" VARCHAR2(2 BYTE),
"CITY" VARCHAR2(36 BYTE),
CONSTRAINT "NPA_NXX_PK" PRIMARY KEY ("NPA_NXX_KEY")
USING INDEX PCTFREE 10 INITRANS 2 MAXTRANS 255
STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1
BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT)
TABLESPACE "XXXXXX_DATA" ENABLE
) SEGMENT CREATION IMMEDIATE
PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255
NOCOMPRESS LOGGING
STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1
BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT)
TABLESPACE "XXXXX_DATA" ;
Cassandra keyspace and table:
cqlsh> CREATE KEYSPACE npa_nxx WITH REPLICATION =
{'class':'NetworkTopologyStrategy', 'Analytics':1};
cqlsh> CREATE TABLE npa_nxx.npa_nxx_data (npa int, nxx int,
latitude float, longitude float, state text, city text,
PRIMARY KEY(npa, nxx));

Resources