Guest User

Untitled

a guest
May 19th, 2017
295
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 35.84 KB | None | 0 0
  1. Spark Command: /opt/jdk1.8.0_112/bin/java -cp /etc/dse/spark/:/usr/share/dse/spark/jars/*:/etc/dse/hadoop2-client/ -Djava.library.path=/usr/share/dse/hadoop2-client/lib/native:/usr/share/dse/cassandra/lib/sigar-bin: -Dcassandra.logdir=/var/log/cassandra -XX:MaxHeapFreeRatio=50 -XX:MinHeapFreeRatio=20 -Dguice_include_stack_traces=OFF -Ddse.system_memory_in_mb=32174 -Dcassandra.config.loader=com.datastax.bdp.config.DseConfigurationLoader -Dlogback.configurationFile=/etc/dse/spark/logback-spark.xml -Dcassandra.logdir=/var/log/cassandra -Ddse.client.configuration.impl=com.datastax.bdp.transport.client.HadoopBasedClientConfiguration -Dderby.stream.error.method=com.datastax.bdp.derby.LogbackBridge.getLogger -Xmx1024M org.apache.spark.deploy.SparkSubmit --class org.apache.spark.sql.hive.thriftserver.HiveThriftServer2 spark-internal
  2. ========================================
  3. WARN 2017-05-19 14:11:31 org.apache.spark.SparkContext: Use an existing SparkContext, some configuration may not take effect.
  4. WARN 2017-05-19 14:11:36 org.apache.hadoop.hive.metastore.HiveMetaStore: Retrying creating default database after error: Unexpected exception caught.
  5. javax.jdo.JDOFatalInternalException: Unexpected exception caught.
  6. at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1193) ~[jdo-api-3.0.1.jar:3.0.1]
  7. at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) ~[jdo-api-3.0.1.jar:3.0.1]
  8. at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) ~[jdo-api-3.0.1.jar:3.0.1]
  9. at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:365) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  10. at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:394) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  11. at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:291) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  12. at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:258) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  13. at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) ~[hadoop-common-2.7.1.3.jar:na]
  14. at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) ~[hadoop-common-2.7.1.3.jar:na]
  15. at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:57) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  16. at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:66) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  17. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:593) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  18. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:571) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  19. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:620) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  20. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:461) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  21. at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:66) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  22. at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:72) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  23. at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:5762) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  24. at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:199) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  25. at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74) [hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  26. at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) [na:1.8.0_112]
  27. at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) [na:1.8.0_112]
  28. at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) [na:1.8.0_112]
  29. at java.lang.reflect.Constructor.newInstance(Constructor.java:423) [na:1.8.0_112]
  30. at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1521) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  31. at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  32. at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:132) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  33. at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  34. at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3005) [hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  35. at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024) [hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  36. at org.apache.hadoop.hive.ql.metadata.Hive.getAllDatabases(Hive.java:1234) [hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  37. at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:174) [hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  38. at org.apache.hadoop.hive.ql.metadata.Hive.<clinit>(Hive.java:166) [hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  39. at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:503) [hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  40. at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:189) [spark-hive_2.11-2.0.2.6.jar:2.0.2.6]
  41. at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:247) [spark-hive_2.11-2.0.2.6.jar:2.0.2.6]
  42. at org.apache.spark.sql.hive.HiveUtils$.newClientForExecution(HiveUtils.scala:250) [spark-hive_2.11-2.0.2.6.jar:2.0.2.6]
  43. at org.apache.spark.sql.hive.thriftserver.HiveThriftServer2$.main(HiveThriftServer2.scala:88) [spark-hive-thriftserver_2.11-2.0.2.6.jar:2.0.2.6]
  44. at org.apache.spark.sql.hive.thriftserver.HiveThriftServer2.main(HiveThriftServer2.scala) [spark-hive-thriftserver_2.11-2.0.2.6.jar:2.0.2.6]
  45. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:1.8.0_112]
  46. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:1.8.0_112]
  47. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:1.8.0_112]
  48. at java.lang.reflect.Method.invoke(Method.java:498) ~[na:1.8.0_112]
  49. at org.apache.spark.deploy.DseSparkSubmit$.org$apache$spark$deploy$DseSparkSubmit$$runMain(DseSparkSubmit.scala:730) [dse-spark-5.1.0.jar:5.1.0]
  50. at org.apache.spark.deploy.DseSparkSubmit$.doRunMain$1(DseSparkSubmit.scala:175) [dse-spark-5.1.0.jar:5.1.0]
  51. at org.apache.spark.deploy.DseSparkSubmit$.submit(DseSparkSubmit.scala:200) [dse-spark-5.1.0.jar:5.1.0]
  52. at org.apache.spark.deploy.DseSparkSubmit$.main(DseSparkSubmit.scala:109) [dse-spark-5.1.0.jar:5.1.0]
  53. at org.apache.spark.deploy.DseSparkSubmitBootstrapper$.main(DseSparkSubmitBootstrapper.scala:74) [dse-spark-5.1.0.jar:5.1.0]
  54. at org.apache.spark.deploy.DseSparkSubmitBootstrapper.main(DseSparkSubmitBootstrapper.scala) [dse-spark-5.1.0.jar:5.1.0]
  55. Caused by: java.lang.reflect.InvocationTargetException: null
  56. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:1.8.0_112]
  57. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:1.8.0_112]
  58. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:1.8.0_112]
  59. at java.lang.reflect.Method.invoke(Method.java:498) ~[na:1.8.0_112]
  60. at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) ~[jdo-api-3.0.1.jar:3.0.1]
  61. at java.security.AccessController.doPrivileged(Native Method) ~[na:1.8.0_112]
  62. at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) ~[jdo-api-3.0.1.jar:3.0.1]
  63. at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) ~[jdo-api-3.0.1.jar:3.0.1]
  64. ... 48 common frames omitted
  65. Caused by: java.lang.NoClassDefFoundError: org/apache/log4j/or/RendererMap
  66. at org.apache.log4j.Hierarchy.<init>(Hierarchy.java:97) ~[apache-log4j-extras-1.2.17.jar:na]
  67. at org.apache.log4j.LogManager.<clinit>(LogManager.java:82) ~[apache-log4j-extras-1.2.17.jar:na]
  68. at org.apache.log4j.Logger.getLogger(Logger.java:104) ~[apache-log4j-extras-1.2.17.jar:na]
  69. at org.datanucleus.util.Log4JLogger.<init>(Log4JLogger.java:49) ~[datanucleus-core-3.2.10.jar:na]
  70. at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) [na:1.8.0_112]
  71. at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) [na:1.8.0_112]
  72. at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) [na:1.8.0_112]
  73. at java.lang.reflect.Constructor.newInstance(Constructor.java:423) [na:1.8.0_112]
  74. at org.datanucleus.util.NucleusLogger.getLoggerInstance(NucleusLogger.java:237) ~[datanucleus-core-3.2.10.jar:na]
  75. at org.datanucleus.util.NucleusLogger.<clinit>(NucleusLogger.java:205) ~[datanucleus-core-3.2.10.jar:na]
  76. at org.datanucleus.plugin.PluginRegistryFactory.newPluginRegistry(PluginRegistryFactory.java:74) ~[datanucleus-core-3.2.10.jar:na]
  77. at org.datanucleus.plugin.PluginManager.<init>(PluginManager.java:61) ~[datanucleus-core-3.2.10.jar:na]
  78. at org.datanucleus.plugin.PluginManager.createPluginManager(PluginManager.java:427) ~[datanucleus-core-3.2.10.jar:na]
  79. at org.datanucleus.NucleusContext.<init>(NucleusContext.java:266) ~[datanucleus-core-3.2.10.jar:na]
  80. at org.datanucleus.NucleusContext.<init>(NucleusContext.java:247) ~[datanucleus-core-3.2.10.jar:na]
  81. at org.datanucleus.NucleusContext.<init>(NucleusContext.java:225) ~[datanucleus-core-3.2.10.jar:na]
  82. at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.<init>(JDOPersistenceManagerFactory.java:416) ~[datanucleus-api-jdo-3.2.6.jar:na]
  83. at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:301) ~[datanucleus-api-jdo-3.2.6.jar:na]
  84. at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202) ~[datanucleus-api-jdo-3.2.6.jar:na]
  85. ... 56 common frames omitted
  86. Caused by: java.lang.ClassNotFoundException: org.apache.log4j.or.RendererMap
  87. at java.net.URLClassLoader.findClass(URLClassLoader.java:381) ~[na:1.8.0_112]
  88. at java.lang.ClassLoader.loadClass(ClassLoader.java:424) ~[na:1.8.0_112]
  89. at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331) ~[na:1.8.0_112]
  90. at java.lang.ClassLoader.loadClass(ClassLoader.java:357) ~[na:1.8.0_112]
  91. ... 75 common frames omitted
  92. WARN 2017-05-19 14:11:36 hive.ql.metadata.Hive: Failed to access metastore. This class should not accessed in runtime.
  93. org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient
  94. at org.apache.hadoop.hive.ql.metadata.Hive.getAllDatabases(Hive.java:1236) ~[hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  95. at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:174) ~[hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  96. at org.apache.hadoop.hive.ql.metadata.Hive.<clinit>(Hive.java:166) ~[hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  97. at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:503) [hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  98. at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:189) [spark-hive_2.11-2.0.2.6.jar:2.0.2.6]
  99. at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:247) [spark-hive_2.11-2.0.2.6.jar:2.0.2.6]
  100. at org.apache.spark.sql.hive.HiveUtils$.newClientForExecution(HiveUtils.scala:250) [spark-hive_2.11-2.0.2.6.jar:2.0.2.6]
  101. at org.apache.spark.sql.hive.thriftserver.HiveThriftServer2$.main(HiveThriftServer2.scala:88) [spark-hive-thriftserver_2.11-2.0.2.6.jar:2.0.2.6]
  102. at org.apache.spark.sql.hive.thriftserver.HiveThriftServer2.main(HiveThriftServer2.scala) [spark-hive-thriftserver_2.11-2.0.2.6.jar:2.0.2.6]
  103. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:1.8.0_112]
  104. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:1.8.0_112]
  105. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:1.8.0_112]
  106. at java.lang.reflect.Method.invoke(Method.java:498) ~[na:1.8.0_112]
  107. at org.apache.spark.deploy.DseSparkSubmit$.org$apache$spark$deploy$DseSparkSubmit$$runMain(DseSparkSubmit.scala:730) [dse-spark-5.1.0.jar:5.1.0]
  108. at org.apache.spark.deploy.DseSparkSubmit$.doRunMain$1(DseSparkSubmit.scala:175) [dse-spark-5.1.0.jar:5.1.0]
  109. at org.apache.spark.deploy.DseSparkSubmit$.submit(DseSparkSubmit.scala:200) [dse-spark-5.1.0.jar:5.1.0]
  110. at org.apache.spark.deploy.DseSparkSubmit$.main(DseSparkSubmit.scala:109) [dse-spark-5.1.0.jar:5.1.0]
  111. at org.apache.spark.deploy.DseSparkSubmitBootstrapper$.main(DseSparkSubmitBootstrapper.scala:74) [dse-spark-5.1.0.jar:5.1.0]
  112. at org.apache.spark.deploy.DseSparkSubmitBootstrapper.main(DseSparkSubmitBootstrapper.scala) [dse-spark-5.1.0.jar:5.1.0]
  113. Caused by: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient
  114. at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1523) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  115. at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  116. at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:132) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  117. at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  118. at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3005) ~[hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  119. at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024) ~[hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  120. at org.apache.hadoop.hive.ql.metadata.Hive.getAllDatabases(Hive.java:1234) ~[hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  121. ... 18 common frames omitted
  122. Caused by: java.lang.reflect.InvocationTargetException: null
  123. at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[na:1.8.0_112]
  124. at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) ~[na:1.8.0_112]
  125. at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[na:1.8.0_112]
  126. at java.lang.reflect.Constructor.newInstance(Constructor.java:423) ~[na:1.8.0_112]
  127. at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1521) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  128. ... 24 common frames omitted
  129. Caused by: javax.jdo.JDOFatalInternalException: Unexpected exception caught.
  130. at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1193) ~[jdo-api-3.0.1.jar:3.0.1]
  131. at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) ~[jdo-api-3.0.1.jar:3.0.1]
  132. at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) ~[jdo-api-3.0.1.jar:3.0.1]
  133. at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:365) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  134. at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:394) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  135. at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:291) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  136. at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:258) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  137. at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) ~[hadoop-common-2.7.1.3.jar:na]
  138. at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) ~[hadoop-common-2.7.1.3.jar:na]
  139. at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:57) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  140. at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:66) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  141. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:593) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  142. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:571) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  143. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:624) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  144. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:461) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  145. at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:66) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  146. at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:72) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  147. at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:5762) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  148. at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:199) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  149. at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74) ~[hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  150. ... 29 common frames omitted
  151. Caused by: java.lang.reflect.InvocationTargetException: null
  152. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:1.8.0_112]
  153. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:1.8.0_112]
  154. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:1.8.0_112]
  155. at java.lang.reflect.Method.invoke(Method.java:498) ~[na:1.8.0_112]
  156. at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) ~[jdo-api-3.0.1.jar:3.0.1]
  157. at java.security.AccessController.doPrivileged(Native Method) ~[na:1.8.0_112]
  158. at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) ~[jdo-api-3.0.1.jar:3.0.1]
  159. at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) ~[jdo-api-3.0.1.jar:3.0.1]
  160. ... 48 common frames omitted
  161. Caused by: java.lang.NoClassDefFoundError: Could not initialize class org.datanucleus.util.NucleusLogger
  162. at org.datanucleus.plugin.PluginRegistryFactory.newPluginRegistry(PluginRegistryFactory.java:74) ~[datanucleus-core-3.2.10.jar:na]
  163. at org.datanucleus.plugin.PluginManager.<init>(PluginManager.java:61) ~[datanucleus-core-3.2.10.jar:na]
  164. at org.datanucleus.plugin.PluginManager.createPluginManager(PluginManager.java:427) ~[datanucleus-core-3.2.10.jar:na]
  165. at org.datanucleus.NucleusContext.<init>(NucleusContext.java:266) ~[datanucleus-core-3.2.10.jar:na]
  166. at org.datanucleus.NucleusContext.<init>(NucleusContext.java:247) ~[datanucleus-core-3.2.10.jar:na]
  167. at org.datanucleus.NucleusContext.<init>(NucleusContext.java:225) ~[datanucleus-core-3.2.10.jar:na]
  168. at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.<init>(JDOPersistenceManagerFactory.java:416) ~[datanucleus-api-jdo-3.2.6.jar:na]
  169. at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:301) ~[datanucleus-api-jdo-3.2.6.jar:na]
  170. at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202) ~[datanucleus-api-jdo-3.2.6.jar:na]
  171. ... 56 common frames omitted
  172. WARN 2017-05-19 14:11:36 org.apache.hadoop.hive.metastore.HiveMetaStore: Retrying creating default database after error: Unexpected exception caught.
  173. javax.jdo.JDOFatalInternalException: Unexpected exception caught.
  174. at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1193) ~[jdo-api-3.0.1.jar:3.0.1]
  175. at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) ~[jdo-api-3.0.1.jar:3.0.1]
  176. at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) ~[jdo-api-3.0.1.jar:3.0.1]
  177. at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:365) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  178. at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:394) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  179. at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:291) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  180. at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:258) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  181. at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) ~[hadoop-common-2.7.1.3.jar:na]
  182. at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) ~[hadoop-common-2.7.1.3.jar:na]
  183. at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:57) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  184. at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:66) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  185. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:593) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  186. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:571) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  187. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:620) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  188. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:461) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  189. at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:66) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  190. at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:72) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  191. at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:5762) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  192. at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:199) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  193. at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74) [hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  194. at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) [na:1.8.0_112]
  195. at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) [na:1.8.0_112]
  196. at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) [na:1.8.0_112]
  197. at java.lang.reflect.Constructor.newInstance(Constructor.java:423) [na:1.8.0_112]
  198. at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1521) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  199. at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  200. at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:132) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  201. at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) [hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  202. at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3005) [hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  203. at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024) [hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  204. at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:503) [hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  205. at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:189) [spark-hive_2.11-2.0.2.6.jar:2.0.2.6]
  206. at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:247) [spark-hive_2.11-2.0.2.6.jar:2.0.2.6]
  207. at org.apache.spark.sql.hive.HiveUtils$.newClientForExecution(HiveUtils.scala:250) [spark-hive_2.11-2.0.2.6.jar:2.0.2.6]
  208. at org.apache.spark.sql.hive.thriftserver.HiveThriftServer2$.main(HiveThriftServer2.scala:88) [spark-hive-thriftserver_2.11-2.0.2.6.jar:2.0.2.6]
  209. at org.apache.spark.sql.hive.thriftserver.HiveThriftServer2.main(HiveThriftServer2.scala) [spark-hive-thriftserver_2.11-2.0.2.6.jar:2.0.2.6]
  210. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:1.8.0_112]
  211. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:1.8.0_112]
  212. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:1.8.0_112]
  213. at java.lang.reflect.Method.invoke(Method.java:498) ~[na:1.8.0_112]
  214. at org.apache.spark.deploy.DseSparkSubmit$.org$apache$spark$deploy$DseSparkSubmit$$runMain(DseSparkSubmit.scala:730) [dse-spark-5.1.0.jar:5.1.0]
  215. at org.apache.spark.deploy.DseSparkSubmit$.doRunMain$1(DseSparkSubmit.scala:175) [dse-spark-5.1.0.jar:5.1.0]
  216. at org.apache.spark.deploy.DseSparkSubmit$.submit(DseSparkSubmit.scala:200) [dse-spark-5.1.0.jar:5.1.0]
  217. at org.apache.spark.deploy.DseSparkSubmit$.main(DseSparkSubmit.scala:109) [dse-spark-5.1.0.jar:5.1.0]
  218. at org.apache.spark.deploy.DseSparkSubmitBootstrapper$.main(DseSparkSubmitBootstrapper.scala:74) [dse-spark-5.1.0.jar:5.1.0]
  219. at org.apache.spark.deploy.DseSparkSubmitBootstrapper.main(DseSparkSubmitBootstrapper.scala) [dse-spark-5.1.0.jar:5.1.0]
  220. Caused by: java.lang.reflect.InvocationTargetException: null
  221. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:1.8.0_112]
  222. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:1.8.0_112]
  223. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:1.8.0_112]
  224. at java.lang.reflect.Method.invoke(Method.java:498) ~[na:1.8.0_112]
  225. at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) ~[jdo-api-3.0.1.jar:3.0.1]
  226. at java.security.AccessController.doPrivileged(Native Method) ~[na:1.8.0_112]
  227. at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) ~[jdo-api-3.0.1.jar:3.0.1]
  228. at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) ~[jdo-api-3.0.1.jar:3.0.1]
  229. ... 45 common frames omitted
  230. Caused by: java.lang.NoClassDefFoundError: Could not initialize class org.datanucleus.util.NucleusLogger
  231. at org.datanucleus.plugin.PluginRegistryFactory.newPluginRegistry(PluginRegistryFactory.java:74) ~[datanucleus-core-3.2.10.jar:na]
  232. at org.datanucleus.plugin.PluginManager.<init>(PluginManager.java:61) ~[datanucleus-core-3.2.10.jar:na]
  233. at org.datanucleus.plugin.PluginManager.createPluginManager(PluginManager.java:427) ~[datanucleus-core-3.2.10.jar:na]
  234. at org.datanucleus.NucleusContext.<init>(NucleusContext.java:266) ~[datanucleus-core-3.2.10.jar:na]
  235. at org.datanucleus.NucleusContext.<init>(NucleusContext.java:247) ~[datanucleus-core-3.2.10.jar:na]
  236. at org.datanucleus.NucleusContext.<init>(NucleusContext.java:225) ~[datanucleus-core-3.2.10.jar:na]
  237. at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.<init>(JDOPersistenceManagerFactory.java:416) ~[datanucleus-api-jdo-3.2.6.jar:na]
  238. at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:301) ~[datanucleus-api-jdo-3.2.6.jar:na]
  239. at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202) ~[datanucleus-api-jdo-3.2.6.jar:na]
  240. ... 53 common frames omitted
  241. ERROR 2017-05-19 14:11:36 org.apache.spark.deploy.DseSparkSubmitBootstrapper: Failed to start or submit Spark application
  242. java.lang.RuntimeException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient
  243. at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:522) ~[hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  244. at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:189) ~[spark-hive_2.11-2.0.2.6.jar:2.0.2.6]
  245. at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:247) ~[spark-hive_2.11-2.0.2.6.jar:2.0.2.6]
  246. at org.apache.spark.sql.hive.HiveUtils$.newClientForExecution(HiveUtils.scala:250) ~[spark-hive_2.11-2.0.2.6.jar:2.0.2.6]
  247. at org.apache.spark.sql.hive.thriftserver.HiveThriftServer2$.main(HiveThriftServer2.scala:88) ~[spark-hive-thriftserver_2.11-2.0.2.6.jar:2.0.2.6]
  248. at org.apache.spark.sql.hive.thriftserver.HiveThriftServer2.main(HiveThriftServer2.scala) ~[spark-hive-thriftserver_2.11-2.0.2.6.jar:2.0.2.6]
  249. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:1.8.0_112]
  250. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:1.8.0_112]
  251. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:1.8.0_112]
  252. at java.lang.reflect.Method.invoke(Method.java:498) ~[na:1.8.0_112]
  253. at org.apache.spark.deploy.DseSparkSubmit$.org$apache$spark$deploy$DseSparkSubmit$$runMain(DseSparkSubmit.scala:730) ~[dse-spark-5.1.0.jar:5.1.0]
  254. at org.apache.spark.deploy.DseSparkSubmit$.doRunMain$1(DseSparkSubmit.scala:175) ~[dse-spark-5.1.0.jar:5.1.0]
  255. at org.apache.spark.deploy.DseSparkSubmit$.submit(DseSparkSubmit.scala:200) ~[dse-spark-5.1.0.jar:5.1.0]
  256. at org.apache.spark.deploy.DseSparkSubmit$.main(DseSparkSubmit.scala:109) ~[dse-spark-5.1.0.jar:5.1.0]
  257. at org.apache.spark.deploy.DseSparkSubmitBootstrapper$.main(DseSparkSubmitBootstrapper.scala:74) ~[dse-spark-5.1.0.jar:5.1.0]
  258. at org.apache.spark.deploy.DseSparkSubmitBootstrapper.main(DseSparkSubmitBootstrapper.scala) [dse-spark-5.1.0.jar:5.1.0]
  259. Caused by: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient
  260. at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1523) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  261. at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  262. at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:132) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  263. at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  264. at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3005) ~[hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  265. at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024) ~[hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  266. at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:503) ~[hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  267. ... 15 common frames omitted
  268. Caused by: java.lang.reflect.InvocationTargetException: null
  269. at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[na:1.8.0_112]
  270. at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) ~[na:1.8.0_112]
  271. at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[na:1.8.0_112]
  272. at java.lang.reflect.Constructor.newInstance(Constructor.java:423) ~[na:1.8.0_112]
  273. at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1521) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  274. ... 21 common frames omitted
  275. Caused by: javax.jdo.JDOFatalInternalException: Unexpected exception caught.
  276. at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1193) ~[jdo-api-3.0.1.jar:3.0.1]
  277. at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) ~[jdo-api-3.0.1.jar:3.0.1]
  278. at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) ~[jdo-api-3.0.1.jar:3.0.1]
  279. at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:365) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  280. at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:394) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  281. at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:291) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  282. at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:258) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  283. at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) ~[hadoop-common-2.7.1.3.jar:na]
  284. at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) ~[hadoop-common-2.7.1.3.jar:na]
  285. at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:57) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  286. at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:66) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  287. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:593) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  288. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:571) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  289. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:624) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  290. at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:461) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  291. at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:66) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  292. at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:72) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  293. at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:5762) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  294. at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:199) ~[hive-metastore-1.2.1.spark2.jar:1.2.1.spark2]
  295. at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74) ~[hive-exec-1.2.1.spark2.jar:1.2.1.spark2]
  296. ... 26 common frames omitted
  297. Caused by: java.lang.reflect.InvocationTargetException: null
  298. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:1.8.0_112]
  299. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:1.8.0_112]
  300. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:1.8.0_112]
  301. at java.lang.reflect.Method.invoke(Method.java:498) ~[na:1.8.0_112]
  302. at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) ~[jdo-api-3.0.1.jar:3.0.1]
  303. at java.security.AccessController.doPrivileged(Native Method) ~[na:1.8.0_112]
  304. at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) ~[jdo-api-3.0.1.jar:3.0.1]
  305. at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) ~[jdo-api-3.0.1.jar:3.0.1]
  306. ... 45 common frames omitted
  307. Caused by: java.lang.NoClassDefFoundError: Could not initialize class org.datanucleus.util.NucleusLogger
  308. at org.datanucleus.plugin.PluginRegistryFactory.newPluginRegistry(PluginRegistryFactory.java:74) ~[datanucleus-core-3.2.10.jar:na]
  309. at org.datanucleus.plugin.PluginManager.<init>(PluginManager.java:61) ~[datanucleus-core-3.2.10.jar:na]
  310. at org.datanucleus.plugin.PluginManager.createPluginManager(PluginManager.java:427) ~[datanucleus-core-3.2.10.jar:na]
  311. at org.datanucleus.NucleusContext.<init>(NucleusContext.java:266) ~[datanucleus-core-3.2.10.jar:na]
  312. at org.datanucleus.NucleusContext.<init>(NucleusContext.java:247) ~[datanucleus-core-3.2.10.jar:na]
  313. at org.datanucleus.NucleusContext.<init>(NucleusContext.java:225) ~[datanucleus-core-3.2.10.jar:na]
  314. at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.<init>(JDOPersistenceManagerFactory.java:416) ~[datanucleus-api-jdo-3.2.6.jar:na]
  315. at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:301) ~[datanucleus-api-jdo-3.2.6.jar:na]
  316. at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202) ~[datanucleus-api-jdo-3.2.6.jar:na]
  317. ... 53 common frames omitted
Add Comment
Please, Sign In to add comment