Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- java.security.AccessControlException: access denied org.apache.derby.security.SystemPermission( "engine", "usederbyinternals" )
- at java.security.AccessControlContext.checkPermission(AccessControlContext.java:472)
- at java.security.AccessController.checkPermission(AccessController.java:884)
- at org.apache.derby.iapi.security.SecurityUtil.checkDerbyInternalsPrivilege(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.startMonitor(Unknown Source)
- at org.apache.derby.iapi.jdbc.JDBCBoot$1.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.iapi.jdbc.JDBCBoot.boot(Unknown Source)
- at org.apache.derby.iapi.jdbc.JDBCBoot.boot(Unknown Source)
- at org.apache.derby.jdbc.EmbeddedDriver.boot(Unknown Source)
- at org.apache.derby.jdbc.EmbeddedDriver.<clinit>(Unknown Source)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at java.lang.Class.newInstance(Class.java:442)
- at org.datanucleus.store.rdbms.connectionpool.AbstractConnectionPoolFactory.loadDriver(AbstractConnectionPoolFactory.java:47)
- at org.datanucleus.store.rdbms.connectionpool.BoneCPConnectionPoolFactory.createConnectionPool(BoneCPConnectionPoolFactory.java:54)
- at org.datanucleus.store.rdbms.ConnectionFactoryImpl.generateDataSources(ConnectionFactoryImpl.java:238)
- at org.datanucleus.store.rdbms.ConnectionFactoryImpl.initialiseDataSources(ConnectionFactoryImpl.java:131)
- at org.datanucleus.store.rdbms.ConnectionFactoryImpl.<init>(ConnectionFactoryImpl.java:85)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
- at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:325)
- at org.datanucleus.store.AbstractStoreManager.registerConnectionFactory(AbstractStoreManager.java:282)
- at org.datanucleus.store.AbstractStoreManager.<init>(AbstractStoreManager.java:240)
- at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:286)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
- at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
- at org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
- at org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
- at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
- at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
- at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
- at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:365)
- at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:394)
- at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:291)
- at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:258)
- at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76)
- at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136)
- at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:57)
- at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:66)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:593)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:571)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:620)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:461)
- at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:66)
- at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:72)
- at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:5762)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:199)
- at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1521)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:132)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104)
- at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3005)
- at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024)
- at org.apache.hadoop.hive.ql.metadata.Hive.getAllDatabases(Hive.java:1234)
- at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:174)
- at org.apache.hadoop.hive.ql.metadata.Hive.<clinit>(Hive.java:166)
- at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:503)
- at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:191)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:264)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:362)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:266)
- at org.apache.spark.sql.hive.HiveExternalCatalog.client$lzycompute(HiveExternalCatalog.scala:66)
- at org.apache.spark.sql.hive.HiveExternalCatalog.client(HiveExternalCatalog.scala:65)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:195)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:195)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:195)
- at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:97)
- at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:194)
- at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:105)
- at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:93)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:39)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog$lzycompute(HiveSessionStateBuilder.scala:54)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog(HiveSessionStateBuilder.scala:52)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog(HiveSessionStateBuilder.scala:35)
- at org.apache.spark.sql.internal.BaseSessionStateBuilder.build(BaseSessionStateBuilder.scala:289)
- at org.apache.spark.sql.SparkSession$.org$apache$spark$sql$SparkSession$$instantiateSessionState(SparkSession.scala:1059)
- at org.apache.spark.sql.SparkSession$$anonfun$sessionState$2.apply(SparkSession.scala:137)
- at org.apache.spark.sql.SparkSession$$anonfun$sessionState$2.apply(SparkSession.scala:136)
- at scala.Option.getOrElse(Option.scala:121)
- at org.apache.spark.sql.SparkSession.sessionState$lzycompute(SparkSession.scala:136)
- at org.apache.spark.sql.SparkSession.sessionState(SparkSession.scala:133)
- at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:66)
- at org.apache.spark.sql.SparkSession.createDataFrame(SparkSession.scala:587)
- at org.apache.spark.sql.SparkSession.createDataFrame(SparkSession.scala:344)
- at scala.it.agilelab.bigdata.wasp.contacthistory.consumers.spark.strategies.InnerNormalizationStrategySpec$$anonfun$1$$anonfun$apply$mcV$sp$1$$anonfun$apply$mcV$sp$9.apply(InnerNormalizationStrategySpec.scala:82)
- at scala.it.agilelab.bigdata.wasp.contacthistory.consumers.spark.strategies.InnerNormalizationStrategySpec$$anonfun$1$$anonfun$apply$mcV$sp$1$$anonfun$apply$mcV$sp$9.apply(InnerNormalizationStrategySpec.scala:78)
- at it.agilelab.bigdata.wasp.contacthistory.models.SparkSetup$class.withSparkSession(SparkSetup.scala:22)
- at scala.it.agilelab.bigdata.wasp.contacthistory.consumers.spark.strategies.InnerNormalizationStrategySpec.withSparkSession(InnerNormalizationStrategySpec.scala:17)
- at scala.it.agilelab.bigdata.wasp.contacthistory.consumers.spark.strategies.InnerNormalizationStrategySpec$$anonfun$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(InnerNormalizationStrategySpec.scala:78)
- at scala.it.agilelab.bigdata.wasp.contacthistory.consumers.spark.strategies.InnerNormalizationStrategySpec$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(InnerNormalizationStrategySpec.scala:78)
- at scala.it.agilelab.bigdata.wasp.contacthistory.consumers.spark.strategies.InnerNormalizationStrategySpec$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(InnerNormalizationStrategySpec.scala:78)
- at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85)
- at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
- at org.scalatest.Transformer.apply(Transformer.scala:22)
- at org.scalatest.Transformer.apply(Transformer.scala:20)
- at org.scalatest.WordSpecLike$$anon$1.apply(WordSpecLike.scala:1078)
- at org.scalatest.TestSuite$class.withFixture(TestSuite.scala:196)
- at org.scalatest.WordSpec.withFixture(WordSpec.scala:1881)
- at org.scalatest.WordSpecLike$class.invokeWithFixture$1(WordSpecLike.scala:1075)
- at org.scalatest.WordSpecLike$$anonfun$runTest$1.apply(WordSpecLike.scala:1088)
- at org.scalatest.WordSpecLike$$anonfun$runTest$1.apply(WordSpecLike.scala:1088)
- at org.scalatest.SuperEngine.runTestImpl(Engine.scala:289)
- at org.scalatest.WordSpecLike$class.runTest(WordSpecLike.scala:1088)
- at org.scalatest.WordSpec.runTest(WordSpec.scala:1881)
- at org.scalatest.WordSpecLike$$anonfun$runTests$1.apply(WordSpecLike.scala:1147)
- at org.scalatest.WordSpecLike$$anonfun$runTests$1.apply(WordSpecLike.scala:1147)
- at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:396)
- at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:384)
- at scala.collection.immutable.List.foreach(List.scala:392)
- at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:384)
- at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:373)
- at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:410)
- at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:384)
- at scala.collection.immutable.List.foreach(List.scala:392)
- at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:384)
- at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:379)
- at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:461)
- at org.scalatest.WordSpecLike$class.runTests(WordSpecLike.scala:1147)
- at org.scalatest.WordSpec.runTests(WordSpec.scala:1881)
- at org.scalatest.Suite$class.run(Suite.scala:1147)
- at org.scalatest.WordSpec.org$scalatest$WordSpecLike$$super$run(WordSpec.scala:1881)
- at org.scalatest.WordSpecLike$$anonfun$run$1.apply(WordSpecLike.scala:1192)
- at org.scalatest.WordSpecLike$$anonfun$run$1.apply(WordSpecLike.scala:1192)
- at org.scalatest.SuperEngine.runImpl(Engine.scala:521)
- at org.scalatest.WordSpecLike$class.run(WordSpecLike.scala:1192)
- at scala.it.agilelab.bigdata.wasp.contacthistory.consumers.spark.strategies.InnerNormalizationStrategySpec.org$scalatest$BeforeAndAfterAll$$super$run(InnerNormalizationStrategySpec.scala:17)
- at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:213)
- at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:210)
- at scala.it.agilelab.bigdata.wasp.contacthistory.consumers.spark.strategies.InnerNormalizationStrategySpec.run(InnerNormalizationStrategySpec.scala:17)
- at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:314)
- at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:480)
- at sbt.TestRunner.runTest$1(TestFramework.scala:106)
- at sbt.TestRunner.run(TestFramework.scala:117)
- at sbt.TestFramework$$anon$2$$anonfun$$lessinit$greater$1.$anonfun$apply$1(TestFramework.scala:262)
- at sbt.TestFramework$.sbt$TestFramework$$withContextLoader(TestFramework.scala:233)
- at sbt.TestFramework$$anon$2$$anonfun$$lessinit$greater$1.apply(TestFramework.scala:262)
- at sbt.TestFramework$$anon$2$$anonfun$$lessinit$greater$1.apply(TestFramework.scala:262)
- at sbt.TestFunction.apply(TestFramework.scala:271)
- at sbt.Tests$.$anonfun$toTask$1(Tests.scala:281)
- at sbt.std.Transform$$anon$3.$anonfun$apply$2(System.scala:46)
- at sbt.std.Transform$$anon$4.work(System.scala:66)
- at sbt.Execute.$anonfun$submit$2(Execute.scala:262)
- at sbt.internal.util.ErrorHandling$.wideConvert(ErrorHandling.scala:16)
- at sbt.Execute.work(Execute.scala:271)
- at sbt.Execute.$anonfun$submit$1(Execute.scala:262)
- at sbt.ConcurrentRestrictions$$anon$4.$anonfun$submitValid$1(ConcurrentRestrictions.scala:174)
- at sbt.CompletionService$$anon$2.call(CompletionService.scala:36)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
- at java.lang.Thread.run(Thread.java:748)
- - should be normalized without errors *** FAILED ***
- [info] java.lang.IllegalArgumentException: Error while instantiating 'org.apache.spark.sql.hive.HiveSessionStateBuilder':
- [info] at org.apache.spark.sql.SparkSession$.org$apache$spark$sql$SparkSession$$instantiateSessionState(SparkSession.scala:1062)
- [info] at org.apache.spark.sql.SparkSession$$anonfun$sessionState$2.apply(SparkSession.scala:137)
- [info] at org.apache.spark.sql.SparkSession$$anonfun$sessionState$2.apply(SparkSession.scala:136)
- [info] at scala.Option.getOrElse(Option.scala:121)
- [info] at org.apache.spark.sql.SparkSession.sessionState$lzycompute(SparkSession.scala:136)
- [info] at org.apache.spark.sql.SparkSession.sessionState(SparkSession.scala:133)
- [info] at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:66)
- [info] at org.apache.spark.sql.SparkSession.createDataFrame(SparkSession.scala:587)
- [info] at org.apache.spark.sql.SparkSession.createDataFrame(SparkSession.scala:344)
- [info] at scala.it.agilelab.bigdata.wasp.contacthistory.consumers.spark.strategies.InnerNormalizationStrategySpec$$anonfun$1$$anonfun$apply$mcV$sp$1$$anonfun$apply$mcV$sp$9.apply(InnerNormalizationStrategySpec.scala:82)
- [info] ...
- [info] Cause: org.apache.spark.sql.AnalysisException: java.lang.RuntimeException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient;
- [info] at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:106)
- [info] at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:194)
- [info] at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:105)
- [info] at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:93)
- [info] at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:39)
- [info] at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog$lzycompute(HiveSessionStateBuilder.scala:54)
- [info] at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog(HiveSessionStateBuilder.scala:52)
- [info] at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog(HiveSessionStateBuilder.scala:35)
- [info] at org.apache.spark.sql.internal.BaseSessionStateBuilder.build(BaseSessionStateBuilder.scala:289)
- [info] at org.apache.spark.sql.SparkSession$.org$apache$spark$sql$SparkSession$$instantiateSessionState(SparkSession.scala:1059)
- [info] ...
- [info] Cause: java.lang.RuntimeException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient
- [info] at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:522)
- [info] at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:191)
- [info] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- [info] at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- [info] at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- [info] at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- [info] at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:264)
- [info] at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:362)
- [info] at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:266)
- [info] at org.apache.spark.sql.hive.HiveExternalCatalog.client$lzycompute(HiveExternalCatalog.scala:66)
- [info] ...
- [info] Cause: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient
- [info] at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1523)
- [info] at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86)
- [info] at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:132)
- [info] at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104)
- [info] at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3005)
- [info] at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024)
- [info] at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:503)
- [info] at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:191)
- [info] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- [info] at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- [info] ...
- [info] Cause: java.lang.reflect.InvocationTargetException:
- [info] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- [info] at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- [info] at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- [info] at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- [info] at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1521)
- [info] at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86)
- [info] at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:132)
- [info] at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104)
- [info] at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3005)
- [info] at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024)
- [info] ...
- [info] Cause: javax.jdo.JDOFatalInternalException: Error creating transactional connection factory
- [info] at org.datanucleus.api.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:587)
- [info] at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:788)
- [info] at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
- [info] at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
- [info] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- [info] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- [info] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- [info] at java.lang.reflect.Method.invoke(Method.java:498)
- [info] at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
- [info] at java.security.AccessController.doPrivileged(Native Method)
- [info] ...
- [info] Cause: java.lang.reflect.InvocationTargetException:
- [info] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- [info] at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- [info] at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- [info] at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- [info] at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
- [info] at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:325)
- [info] at org.datanucleus.store.AbstractStoreManager.registerConnectionFactory(AbstractStoreManager.java:282)
- [info] at org.datanucleus.store.AbstractStoreManager.<init>(AbstractStoreManager.java:240)
- [info] at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:286)
- [info] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- [info] ...
- [info] Cause: java.lang.NoClassDefFoundError: Could not initialize class org.apache.derby.jdbc.EmbeddedDriver
- [info] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- [info] at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- [info] at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- [info] at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- [info] at java.lang.Class.newInstance(Class.java:442)
- [info] at org.datanucleus.store.rdbms.connectionpool.AbstractConnectionPoolFactory.loadDriver(AbstractConnectionPoolFactory.java:47)
- [info] at org.datanucleus.store.rdbms.connectionpool.BoneCPConnectionPoolFactory.createConnectionPool(BoneCPConnectionPoolFactory.java:54)
- [info] at org.datanucleus.store.rdbms.ConnectionFactoryImpl.generateDataSources(ConnectionFactoryImpl.java:238)
- [info] at org.datanucleus.store.rdbms.ConnectionFactoryImpl.initialiseDataSources(ConnectionFactoryImpl.java:131)
- [info] at org.datanucleus.store.rdbms.ConnectionFactoryImpl.<init>(ConnectionFactoryImpl.java:85)
- [info] ...
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement