Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- [2017-09-27 08:28:44,352] INFO k.jobserver.JobResultActor [] [akka://JobServer/user/context-supervisor/$c] - Starting actor spark.jobserver.JobResultActor
- [2017-09-27 08:28:49,606] INFO AkkaClusterSupervisorActor [] [] - Failed to initialize context Actor[akka.tcp://JobServer@127.0.0.1:64373/user/jobManager-82-be8b-64748bc3e87b#1874780438]
- java.lang.IllegalArgumentException: Error while instantiating 'org.apache.spark.sql.hive.HiveSessionStateBuilder':
- at org.apache.spark.sql.SparkSession$.org$apache$spark$sql$SparkSession$$instantiateSessionState(SparkSession.scala:1053)
- at org.apache.spark.sql.SparkSession$$anonfun$sessionState$2.apply(SparkSession.scala:130)
- at org.apache.spark.sql.SparkSession$$anonfun$sessionState$2.apply(SparkSession.scala:130)
- at scala.Option.getOrElse(Option.scala:121)
- at org.apache.spark.sql.SparkSession.sessionState$lzycompute(SparkSession.scala:129)
- at org.apache.spark.sql.SparkSession.sessionState(SparkSession.scala:126)
- at org.apache.spark.sql.SparkSession$Builder$$anonfun$getOrCreate$5.apply(SparkSession.scala:938)
- at org.apache.spark.sql.SparkSession$Builder$$anonfun$getOrCreate$5.apply(SparkSession.scala:938)
- at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:99)
- at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:99)
- at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230)
- at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
- at scala.collection.mutable.HashMap.foreach(HashMap.scala:99)
- at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:938)
- at spark.jobserver.python.PythonSessionContextFactory.makeContext(PythonContextFactory.scala:187)
- at spark.jobserver.python.PythonSessionContextFactory.makeContext(PythonContextFactory.scala:163)
- at spark.jobserver.context.SparkContextFactory$class.makeContext(SparkContextFactory.scala:64)
- at spark.jobserver.python.PythonSessionContextFactory.makeContext(PythonContextFactory.scala:163)
- at spark.jobserver.JobManagerActor$$anonfun$wrappedReceive$1.applyOrElse(JobManagerActor.scala:163)
- at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
- at spark.jobserver.common.akka.ActorStack$$anonfun$receive$1.applyOrElse(ActorStack.scala:33)
- at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
- at spark.jobserver.common.akka.Slf4jLogging$$anonfun$receive$1$$anonfun$applyOrElse$1.apply$mcV$sp(Slf4jLogging.scala:25)
- at spark.jobserver.common.akka.Slf4jLogging$class.spark$jobserver$common$akka$Slf4jLogging$$withAkkaSourceLogging(Slf4jLogging.scala:34)
- at spark.jobserver.common.akka.Slf4jLogging$$anonfun$receive$1.applyOrElse(Slf4jLogging.scala:24)
- at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
- at spark.jobserver.common.akka.ActorMetrics$$anonfun$receive$1.applyOrElse(ActorMetrics.scala:23)
- at akka.actor.Actor$class.aroundReceive(Actor.scala:484)
- at spark.jobserver.common.akka.InstrumentedActor.aroundReceive(InstrumentedActor.scala:8)
- at akka.actor.ActorCell.receiveMessage(ActorCell.scala:526)
- at akka.actor.ActorCell.invoke(ActorCell.scala:495)
- at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:257)
- at akka.dispatch.Mailbox.run(Mailbox.scala:224)
- at akka.dispatch.Mailbox.exec(Mailbox.scala:234)
- at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
- at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
- at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
- at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
- Caused by: org.apache.spark.sql.AnalysisException: java.lang.RuntimeException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient;
- at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:106)
- at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:193)
- at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:105)
- at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:93)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:39)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog$lzycompute(HiveSessionStateBuilder.scala:54)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog(HiveSessionStateBuilder.scala:52)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog(HiveSessionStateBuilder.scala:35)
- at org.apache.spark.sql.internal.BaseSessionStateBuilder.build(BaseSessionStateBuilder.scala:289)
- at org.apache.spark.sql.SparkSession$.org$apache$spark$sql$SparkSession$$instantiateSessionState(SparkSession.scala:1050)
- ... 37 more
- Caused by: java.lang.RuntimeException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient
- at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:522)
- at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:191)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:264)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:362)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:266)
- at org.apache.spark.sql.hive.HiveExternalCatalog.client$lzycompute(HiveExternalCatalog.scala:66)
- at org.apache.spark.sql.hive.HiveExternalCatalog.client(HiveExternalCatalog.scala:65)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:194)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:194)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:194)
- at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:97)
- ... 46 more
- Caused by: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient
- at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1523)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:132)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104)
- at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3005)
- at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024)
- at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:503)
- ... 60 more
- Caused by: java.lang.reflect.InvocationTargetException
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1521)
- ... 66 more
- Caused by: javax.jdo.JDOFatalDataStoreException: Unable to open a test connection to the given database. JDBC url = jdbc:derby:;databaseName=metastore_db;create=true, username = APP. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------
- java.sql.SQLException: Failed to start database 'metastore_db' with class loader org.apache.spark.sql.hive.client.IsolatedClientLoader$$anon$1@759f2543, see the next exception for details.
- at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
- at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
- at org.apache.derby.impl.jdbc.Util.seeNextException(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection.bootDatabase(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver$1.run(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver$1.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
- at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
- at java.sql.DriverManager.getConnection(DriverManager.java:664)
- at java.sql.DriverManager.getConnection(DriverManager.java:208)
- at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361)
- at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
- at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
- at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
- at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
- at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
- at org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
- at org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
- at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
- at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
- at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
- at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:365)
- at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:394)
- at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:291)
- at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:258)
- at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76)
- at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136)
- at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:57)
- at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:66)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:593)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:571)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:624)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:461)
- at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:66)
- at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:72)
- at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:5762)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:199)
- at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1521)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:132)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104)
- at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3005)
- at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024)
- at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:503)
- at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:191)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:264)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:362)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:266)
- at org.apache.spark.sql.hive.HiveExternalCatalog.client$lzycompute(HiveExternalCatalog.scala:66)
- at org.apache.spark.sql.hive.HiveExternalCatalog.client(HiveExternalCatalog.scala:65)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:194)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:194)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:194)
- at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:97)
- at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:193)
- at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:105)
- at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:93)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:39)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog$lzycompute(HiveSessionStateBuilder.scala:54)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog(HiveSessionStateBuilder.scala:52)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog(HiveSessionStateBuilder.scala:35)
- at org.apache.spark.sql.internal.BaseSessionStateBuilder.build(BaseSessionStateBuilder.scala:289)
- at org.apache.spark.sql.SparkSession$.org$apache$spark$sql$SparkSession$$instantiateSessionState(SparkSession.scala:1050)
- at org.apache.spark.sql.SparkSession$$anonfun$sessionState$2.apply(SparkSession.scala:130)
- at org.apache.spark.sql.SparkSession$$anonfun$sessionState$2.apply(SparkSession.scala:130)
- at scala.Option.getOrElse(Option.scala:121)
- at org.apache.spark.sql.SparkSession.sessionState$lzycompute(SparkSession.scala:129)
- at org.apache.spark.sql.SparkSession.sessionState(SparkSession.scala:126)
- at org.apache.spark.sql.SparkSession$Builder$$anonfun$getOrCreate$5.apply(SparkSession.scala:938)
- at org.apache.spark.sql.SparkSession$Builder$$anonfun$getOrCreate$5.apply(SparkSession.scala:938)
- at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:99)
- at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:99)
- at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230)
- at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
- at scala.collection.mutable.HashMap.foreach(HashMap.scala:99)
- at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:938)
- at spark.jobserver.python.PythonSessionContextFactory.makeContext(PythonContextFactory.scala:187)
- at spark.jobserver.python.PythonSessionContextFactory.makeContext(PythonContextFactory.scala:163)
- at spark.jobserver.context.SparkContextFactory$class.makeContext(SparkContextFactory.scala:64)
- at spark.jobserver.python.PythonSessionContextFactory.makeContext(PythonContextFactory.scala:163)
- at spark.jobserver.JobManagerActor$$anonfun$wrappedReceive$1.applyOrElse(JobManagerActor.scala:163)
- at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
- at spark.jobserver.common.akka.ActorStack$$anonfun$receive$1.applyOrElse(ActorStack.scala:33)
- at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
- at spark.jobserver.common.akka.Slf4jLogging$$anonfun$receive$1$$anonfun$applyOrElse$1.apply$mcV$sp(Slf4jLogging.scala:25)
- at spark.jobserver.common.akka.Slf4jLogging$class.spark$jobserver$common$akka$Slf4jLogging$$withAkkaSourceLogging(Slf4jLogging.scala:34)
- at spark.jobserver.common.akka.Slf4jLogging$$anonfun$receive$1.applyOrElse(Slf4jLogging.scala:24)
- at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
- at spark.jobserver.common.akka.ActorMetrics$$anonfun$receive$1.applyOrElse(ActorMetrics.scala:23)
- at akka.actor.Actor$class.aroundReceive(Actor.scala:484)
- at spark.jobserver.common.akka.InstrumentedActor.aroundReceive(InstrumentedActor.scala:8)
- at akka.actor.ActorCell.receiveMessage(ActorCell.scala:526)
- at akka.actor.ActorCell.invoke(ActorCell.scala:495)
- at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:257)
- at akka.dispatch.Mailbox.run(Mailbox.scala:224)
- at akka.dispatch.Mailbox.exec(Mailbox.scala:234)
- at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
- at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
- at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
- at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
- Caused by: ERROR XJ040: Failed to start database 'metastore_db' with class loader org.apache.spark.sql.hive.client.IsolatedClientLoader$$anon$1@759f2543, see the next exception for details.
- at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
- at org.apache.derby.impl.jdbc.SQLExceptionFactory.wrapArgsForTransportAcrossDRDA(Unknown Source)
- ... 128 more
- Caused by: ERROR XSDB6: Another instance of Derby may have already booted the database /Users/nislam/projects/spark-jobserver-latest/job-server-extras/metastore_db.
- at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
- at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.privGetJBMSLockOnDB(Unknown Source)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.getJBMSLockOnDB(Unknown Source)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.raw.RawStore$6.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.store.raw.RawStore.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.raw.RawStore.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.access.RAMAccessManager$5.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.store.access.RAMAccessManager.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.access.RAMAccessManager.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.db.BasicDatabase$5.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.db.BasicDatabase.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.db.BasicDatabase.bootStore(Unknown Source)
- at org.apache.derby.impl.db.BasicDatabase.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startProviderService(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.findProviderAndStartService(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startPersistentService(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.startPersistentService(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection$4.run(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection$4.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.jdbc.EmbedConnection.startPersistentService(Unknown Source)
- ... 125 more
- ------
- NestedThrowables:
- java.sql.SQLException: Unable to open a test connection to the given database. JDBC url = jdbc:derby:;databaseName=metastore_db;create=true, username = APP. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------
- java.sql.SQLException: Failed to start database 'metastore_db' with class loader org.apache.spark.sql.hive.client.IsolatedClientLoader$$anon$1@759f2543, see the next exception for details.
- at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
- at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
- at org.apache.derby.impl.jdbc.Util.seeNextException(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection.bootDatabase(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver$1.run(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver$1.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
- at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
- at java.sql.DriverManager.getConnection(DriverManager.java:664)
- at java.sql.DriverManager.getConnection(DriverManager.java:208)
- at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361)
- at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
- at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
- at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
- at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
- at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
- at org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
- at org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
- at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
- at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
- at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
- at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:365)
- at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:394)
- at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:291)
- at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:258)
- at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76)
- at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136)
- at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:57)
- at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:66)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:593)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:571)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:624)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:461)
- at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:66)
- at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:72)
- at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:5762)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:199)
- at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1521)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:132)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104)
- at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3005)
- at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024)
- at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:503)
- at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:191)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:264)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:362)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:266)
- at org.apache.spark.sql.hive.HiveExternalCatalog.client$lzycompute(HiveExternalCatalog.scala:66)
- at org.apache.spark.sql.hive.HiveExternalCatalog.client(HiveExternalCatalog.scala:65)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:194)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:194)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:194)
- at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:97)
- at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:193)
- at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:105)
- at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:93)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:39)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog$lzycompute(HiveSessionStateBuilder.scala:54)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog(HiveSessionStateBuilder.scala:52)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog(HiveSessionStateBuilder.scala:35)
- at org.apache.spark.sql.internal.BaseSessionStateBuilder.build(BaseSessionStateBuilder.scala:289)
- at org.apache.spark.sql.SparkSession$.org$apache$spark$sql$SparkSession$$instantiateSessionState(SparkSession.scala:1050)
- at org.apache.spark.sql.SparkSession$$anonfun$sessionState$2.apply(SparkSession.scala:130)
- at org.apache.spark.sql.SparkSession$$anonfun$sessionState$2.apply(SparkSession.scala:130)
- at scala.Option.getOrElse(Option.scala:121)
- at org.apache.spark.sql.SparkSession.sessionState$lzycompute(SparkSession.scala:129)
- at org.apache.spark.sql.SparkSession.sessionState(SparkSession.scala:126)
- at org.apache.spark.sql.SparkSession$Builder$$anonfun$getOrCreate$5.apply(SparkSession.scala:938)
- at org.apache.spark.sql.SparkSession$Builder$$anonfun$getOrCreate$5.apply(SparkSession.scala:938)
- at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:99)
- at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:99)
- at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230)
- at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
- at scala.collection.mutable.HashMap.foreach(HashMap.scala:99)
- at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:938)
- at spark.jobserver.python.PythonSessionContextFactory.makeContext(PythonContextFactory.scala:187)
- at spark.jobserver.python.PythonSessionContextFactory.makeContext(PythonContextFactory.scala:163)
- at spark.jobserver.context.SparkContextFactory$class.makeContext(SparkContextFactory.scala:64)
- at spark.jobserver.python.PythonSessionContextFactory.makeContext(PythonContextFactory.scala:163)
- at spark.jobserver.JobManagerActor$$anonfun$wrappedReceive$1.applyOrElse(JobManagerActor.scala:163)
- at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
- at spark.jobserver.common.akka.ActorStack$$anonfun$receive$1.applyOrElse(ActorStack.scala:33)
- at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
- at spark.jobserver.common.akka.Slf4jLogging$$anonfun$receive$1$$anonfun$applyOrElse$1.apply$mcV$sp(Slf4jLogging.scala:25)
- at spark.jobserver.common.akka.Slf4jLogging$class.spark$jobserver$common$akka$Slf4jLogging$$withAkkaSourceLogging(Slf4jLogging.scala:34)
- at spark.jobserver.common.akka.Slf4jLogging$$anonfun$receive$1.applyOrElse(Slf4jLogging.scala:24)
- at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
- at spark.jobserver.common.akka.ActorMetrics$$anonfun$receive$1.applyOrElse(ActorMetrics.scala:23)
- at akka.actor.Actor$class.aroundReceive(Actor.scala:484)
- at spark.jobserver.common.akka.InstrumentedActor.aroundReceive(InstrumentedActor.scala:8)
- at akka.actor.ActorCell.receiveMessage(ActorCell.scala:526)
- at akka.actor.ActorCell.invoke(ActorCell.scala:495)
- at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:257)
- at akka.dispatch.Mailbox.run(Mailbox.scala:224)
- at akka.dispatch.Mailbox.exec(Mailbox.scala:234)
- at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
- at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
- at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
- at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
- Caused by: ERROR XJ040: Failed to start database 'metastore_db' with class loader org.apache.spark.sql.hive.client.IsolatedClientLoader$$anon$1@759f2543, see the next exception for details.
- at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
- at org.apache.derby.impl.jdbc.SQLExceptionFactory.wrapArgsForTransportAcrossDRDA(Unknown Source)
- ... 128 more
- Caused by: ERROR XSDB6: Another instance of Derby may have already booted the database /Users/nislam/projects/spark-jobserver-latest/job-server-extras/metastore_db.
- at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
- at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.privGetJBMSLockOnDB(Unknown Source)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.getJBMSLockOnDB(Unknown Source)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.raw.RawStore$6.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.store.raw.RawStore.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.raw.RawStore.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.access.RAMAccessManager$5.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.store.access.RAMAccessManager.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.access.RAMAccessManager.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.db.BasicDatabase$5.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.db.BasicDatabase.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.db.BasicDatabase.bootStore(Unknown Source)
- at org.apache.derby.impl.db.BasicDatabase.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startProviderService(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.findProviderAndStartService(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startPersistentService(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.startPersistentService(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection$4.run(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection$4.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.jdbc.EmbedConnection.startPersistentService(Unknown Source)
- ... 125 more
- ------
- at org.datanucleus.api.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:436)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:788)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
- at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
- at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
- at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
- at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:365)
- at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:394)
- at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:291)
- at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:258)
- at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76)
- at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136)
- at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:57)
- at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:66)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:593)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:571)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:624)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:461)
- at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:66)
- at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:72)
- at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:5762)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:199)
- at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
- ... 71 more
- Caused by: java.sql.SQLException: Unable to open a test connection to the given database. JDBC url = jdbc:derby:;databaseName=metastore_db;create=true, username = APP. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------
- java.sql.SQLException: Failed to start database 'metastore_db' with class loader org.apache.spark.sql.hive.client.IsolatedClientLoader$$anon$1@759f2543, see the next exception for details.
- at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
- at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
- at org.apache.derby.impl.jdbc.Util.seeNextException(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection.bootDatabase(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver$1.run(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver$1.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
- at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
- at java.sql.DriverManager.getConnection(DriverManager.java:664)
- at java.sql.DriverManager.getConnection(DriverManager.java:208)
- at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361)
- at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
- at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
- at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
- at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
- at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
- at org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
- at org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
- at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
- at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
- at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
- at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:365)
- at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:394)
- at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:291)
- at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:258)
- at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76)
- at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136)
- at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:57)
- at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:66)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:593)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:571)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:624)
- at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:461)
- at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:66)
- at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:72)
- at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:5762)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:199)
- at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1521)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:132)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104)
- at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3005)
- at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024)
- at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:503)
- at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:191)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:264)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:362)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:266)
- at org.apache.spark.sql.hive.HiveExternalCatalog.client$lzycompute(HiveExternalCatalog.scala:66)
- at org.apache.spark.sql.hive.HiveExternalCatalog.client(HiveExternalCatalog.scala:65)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:194)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:194)
- at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:194)
- at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:97)
- at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:193)
- at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:105)
- at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:93)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:39)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog$lzycompute(HiveSessionStateBuilder.scala:54)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog(HiveSessionStateBuilder.scala:52)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog(HiveSessionStateBuilder.scala:35)
- at org.apache.spark.sql.internal.BaseSessionStateBuilder.build(BaseSessionStateBuilder.scala:289)
- at org.apache.spark.sql.SparkSession$.org$apache$spark$sql$SparkSession$$instantiateSessionState(SparkSession.scala:1050)
- at org.apache.spark.sql.SparkSession$$anonfun$sessionState$2.apply(SparkSession.scala:130)
- at org.apache.spark.sql.SparkSession$$anonfun$sessionState$2.apply(SparkSession.scala:130)
- at scala.Option.getOrElse(Option.scala:121)
- at org.apache.spark.sql.SparkSession.sessionState$lzycompute(SparkSession.scala:129)
- at org.apache.spark.sql.SparkSession.sessionState(SparkSession.scala:126)
- at org.apache.spark.sql.SparkSession$Builder$$anonfun$getOrCreate$5.apply(SparkSession.scala:938)
- at org.apache.spark.sql.SparkSession$Builder$$anonfun$getOrCreate$5.apply(SparkSession.scala:938)
- at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:99)
- at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:99)
- at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230)
- at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
- at scala.collection.mutable.HashMap.foreach(HashMap.scala:99)
- at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:938)
- at spark.jobserver.python.PythonSessionContextFactory.makeContext(PythonContextFactory.scala:187)
- at spark.jobserver.python.PythonSessionContextFactory.makeContext(PythonContextFactory.scala:163)
- at spark.jobserver.context.SparkContextFactory$class.makeContext(SparkContextFactory.scala:64)
- at spark.jobserver.python.PythonSessionContextFactory.makeContext(PythonContextFactory.scala:163)
- at spark.jobserver.JobManagerActor$$anonfun$wrappedReceive$1.applyOrElse(JobManagerActor.scala:163)
- at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
- at spark.jobserver.common.akka.ActorStack$$anonfun$receive$1.applyOrElse(ActorStack.scala:33)
- at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
- at spark.jobserver.common.akka.Slf4jLogging$$anonfun$receive$1$$anonfun$applyOrElse$1.apply$mcV$sp(Slf4jLogging.scala:25)
- at spark.jobserver.common.akka.Slf4jLogging$class.spark$jobserver$common$akka$Slf4jLogging$$withAkkaSourceLogging(Slf4jLogging.scala:34)
- at spark.jobserver.common.akka.Slf4jLogging$$anonfun$receive$1.applyOrElse(Slf4jLogging.scala:24)
- at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
- at spark.jobserver.common.akka.ActorMetrics$$anonfun$receive$1.applyOrElse(ActorMetrics.scala:23)
- at akka.actor.Actor$class.aroundReceive(Actor.scala:484)
- at spark.jobserver.common.akka.InstrumentedActor.aroundReceive(InstrumentedActor.scala:8)
- at akka.actor.ActorCell.receiveMessage(ActorCell.scala:526)
- at akka.actor.ActorCell.invoke(ActorCell.scala:495)
- at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:257)
- at akka.dispatch.Mailbox.run(Mailbox.scala:224)
- at akka.dispatch.Mailbox.exec(Mailbox.scala:234)
- at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
- at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
- at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
- at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
- Caused by: ERROR XJ040: Failed to start database 'metastore_db' with class loader org.apache.spark.sql.hive.client.IsolatedClientLoader$$anon$1@759f2543, see the next exception for details.
- at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
- at org.apache.derby.impl.jdbc.SQLExceptionFactory.wrapArgsForTransportAcrossDRDA(Unknown Source)
- ... 128 more
- Caused by: ERROR XSDB6: Another instance of Derby may have already booted the database /Users/nislam/projects/spark-jobserver-latest/job-server-extras/metastore_db.
- at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
- at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.privGetJBMSLockOnDB(Unknown Source)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.getJBMSLockOnDB(Unknown Source)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.raw.RawStore$6.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.store.raw.RawStore.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.raw.RawStore.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.access.RAMAccessManager$5.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.store.access.RAMAccessManager.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.access.RAMAccessManager.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.db.BasicDatabase$5.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.db.BasicDatabase.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.db.BasicDatabase.bootStore(Unknown Source)
- at org.apache.derby.impl.db.BasicDatabase.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startProviderService(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.findProviderAndStartService(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startPersistentService(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.startPersistentService(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection$4.run(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection$4.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.jdbc.EmbedConnection.startPersistentService(Unknown Source)
- ... 125 more
- ------
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at com.jolbox.bonecp.PoolUtil.generateSQLException(PoolUtil.java:192)
- at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:422)
- at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
- at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
- at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
- at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
- at org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
- at org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
- at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
- ... 100 more
- Caused by: java.sql.SQLException: Failed to start database 'metastore_db' with class loader org.apache.spark.sql.hive.client.IsolatedClientLoader$$anon$1@759f2543, see the next exception for details.
- at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
- at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
- at org.apache.derby.impl.jdbc.Util.seeNextException(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection.bootDatabase(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver$1.run(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver$1.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
- at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
- at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
- at java.sql.DriverManager.getConnection(DriverManager.java:664)
- at java.sql.DriverManager.getConnection(DriverManager.java:208)
- at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361)
- at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
- ... 112 more
- Caused by: ERROR XJ040: Failed to start database 'metastore_db' with class loader org.apache.spark.sql.hive.client.IsolatedClientLoader$$anon$1@759f2543, see the next exception for details.
- at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
- at org.apache.derby.impl.jdbc.SQLExceptionFactory.wrapArgsForTransportAcrossDRDA(Unknown Source)
- ... 128 more
- Caused by: ERROR XSDB6: Another instance of Derby may have already booted the database /Users/nislam/projects/spark-jobserver-latest/job-server-extras/metastore_db.
- at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
- at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.privGetJBMSLockOnDB(Unknown Source)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.getJBMSLockOnDB(Unknown Source)
- at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.raw.RawStore$6.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.store.raw.RawStore.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.raw.RawStore.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.access.RAMAccessManager$5.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.store.access.RAMAccessManager.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.store.access.RAMAccessManager.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.db.BasicDatabase$5.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.db.BasicDatabase.bootServiceModule(Unknown Source)
- at org.apache.derby.impl.db.BasicDatabase.bootStore(Unknown Source)
- at org.apache.derby.impl.db.BasicDatabase.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
- at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startProviderService(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.findProviderAndStartService(Unknown Source)
- at org.apache.derby.impl.services.monitor.BaseMonitor.startPersistentService(Unknown Source)
- at org.apache.derby.iapi.services.monitor.Monitor.startPersistentService(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection$4.run(Unknown Source)
- at org.apache.derby.impl.jdbc.EmbedConnection$4.run(Unknown Source)
- at java.security.AccessController.doPrivileged(Native Method)
- at org.apache.derby.impl.jdbc.EmbedConnection.startPersistentService(Unknown Source)
- ... 125 more
- [2017-09-27 08:28:49,616] INFO .Cluster(akka://JobServer) [] [akka.cluster.Cluster(akka://JobServer)] - Cluster Node [akka.tcp://JobServer@127.0.0.1:63698] - Marking node [akka.tcp://JobServer@127.0.0.1:64373] as [Down]
- [2017-09-27 08:28:54,298] WARN .cluster.ClusterCoreDaemon [] [akka.tcp://JobServer@127.0.0.1:63698/system/cluster/core/daemon] - Cluster Node [akka.tcp://JobServer@127.0.0.1:63698] - Marking node(s) as UNREACHABLE [Member(address = akka.tcp://JobServer@127.0.0.1:64373, status = Down)]. Node roles [supervisor]
- [2017-09-27 08:28:54,300] INFO .Cluster(akka://JobServer) [] [akka.cluster.Cluster(akka://JobServer)] - Cluster Node [akka.tcp://JobServer@127.0.0.1:63698] - Leader is removing unreachable node [akka.tcp://JobServer@127.0.0.1:64373]
Add Comment
Please, Sign In to add comment