Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- 20/08/11 12:12:03 WARN metastore: set_ugi() not successful, Likely cause: new client talking to old server. Continuing without it.
- org.apache.thrift.transport.TTransportException
- at org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
- at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86)
- at org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:425)
- at org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:321)
- at org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:225)
- at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77)
- at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_set_ugi(ThriftHiveMetastore.java:4247)
- at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.set_ugi(ThriftHiveMetastore.java:4233)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:498)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:247)
- at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:70)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:83)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104)
- at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600)
- at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652)
- at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632)
- at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894)
- at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248)
- at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231)
- at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:388)
- at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332)
- at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312)
- at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288)
- at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:288)
- at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:314)
- at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:255)
- at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:254)
- at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:304)
- at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:413)
- at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221)
- at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23)
- at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
- at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221)
- at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:154)
- at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:144)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:40)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$resourceLoader$1(HiveSessionStateBuilder.scala:47)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.client$lzycompute(HiveSessionStateBuilder.scala:115)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.client(HiveSessionStateBuilder.scala:115)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.addJar(HiveSessionStateBuilder.scala:117)
- at org.apache.spark.sql.execution.command.AddJarCommand.run(resources.scala:40)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79)
- at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:226)
- at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3472)
- at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$4(SQLExecution.scala:100)
- at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160)
- at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:87)
- at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3468)
- at org.apache.spark.sql.Dataset.<init>(Dataset.scala:226)
- at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:96)
- at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:607)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
- at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
- at py4j.Gateway.invoke(Gateway.java:282)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:238)
- at java.lang.Thread.run(Thread.java:748)
- 20/08/11 12:12:03 INFO metastore: Connected to metastore.
- 20/08/11 12:12:03 WARN RetryingMetaStoreClient: MetaStoreClient lost connection. Attempting to reconnect (1 of 1) after 1s. getAllFunctions
- org.apache.thrift.transport.TTransportException
- at org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
- at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86)
- at org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:425)
- at org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:321)
- at org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:225)
- at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77)
- at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_all_functions(ThriftHiveMetastore.java:3845)
- at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_all_functions(ThriftHiveMetastore.java:3833)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getAllFunctions(HiveMetaStoreClient.java:2399)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:173)
- at com.sun.proxy.$Proxy32.getAllFunctions(Unknown Source)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:2336)
- at com.sun.proxy.$Proxy32.getAllFunctions(Unknown Source)
- at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894)
- at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248)
- at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231)
- at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:388)
- at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332)
- at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312)
- at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288)
- at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:288)
- at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:314)
- at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:255)
- at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:254)
- at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:304)
- at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:413)
- at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221)
- at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23)
- at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
- at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221)
- at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:154)
- at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:144)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:40)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$resourceLoader$1(HiveSessionStateBuilder.scala:47)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.client$lzycompute(HiveSessionStateBuilder.scala:115)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.client(HiveSessionStateBuilder.scala:115)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.addJar(HiveSessionStateBuilder.scala:117)
- at org.apache.spark.sql.execution.command.AddJarCommand.run(resources.scala:40)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79)
- at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:226)
- at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3472)
- at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$4(SQLExecution.scala:100)
- at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160)
- at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:87)
- at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3468)
- at org.apache.spark.sql.Dataset.<init>(Dataset.scala:226)
- at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:96)
- at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:607)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
- at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
- at py4j.Gateway.invoke(Gateway.java:282)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:238)
- at java.lang.Thread.run(Thread.java:748)
- 20/08/11 12:12:04 WARN TIOStreamTransport: Error closing output stream.
- java.net.SocketException: Socket closed
- at java.net.SocketOutputStream.socketWrite(SocketOutputStream.java:118)
- at java.net.SocketOutputStream.write(SocketOutputStream.java:155)
- at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
- at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
- at java.io.FilterOutputStream.close(FilterOutputStream.java:158)
- at org.apache.thrift.transport.TIOStreamTransport.close(TIOStreamTransport.java:110)
- at org.apache.thrift.transport.TSocket.close(TSocket.java:235)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.close(HiveMetaStoreClient.java:563)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.reconnect(HiveMetaStoreClient.java:335)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:163)
- at com.sun.proxy.$Proxy32.getAllFunctions(Unknown Source)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:2336)
- at com.sun.proxy.$Proxy32.getAllFunctions(Unknown Source)
- at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894)
- at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248)
- at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231)
- at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:388)
- at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332)
- at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312)
- at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288)
- at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:288)
- at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:314)
- at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:255)
- at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:254)
- at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:304)
- at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:413)
- at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221)
- at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23)
- at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
- at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221)
- at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:154)
- at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:144)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:40)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$resourceLoader$1(HiveSessionStateBuilder.scala:47)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.client$lzycompute(HiveSessionStateBuilder.scala:115)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.client(HiveSessionStateBuilder.scala:115)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.addJar(HiveSessionStateBuilder.scala:117)
- at org.apache.spark.sql.execution.command.AddJarCommand.run(resources.scala:40)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79)
- at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:226)
- at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3472)
- at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$4(SQLExecution.scala:100)
- at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160)
- at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:87)
- at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3468)
- at org.apache.spark.sql.Dataset.<init>(Dataset.scala:226)
- at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:96)
- at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:607)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
- at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
- at py4j.Gateway.invoke(Gateway.java:282)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:238)
- at java.lang.Thread.run(Thread.java:748)
- 20/08/11 12:12:04 INFO metastore: Closed a connection to metastore, current connections: 0
- 20/08/11 12:12:04 INFO metastore: Trying to connect to metastore with URI thrift://10.32.74.254:9083
- 20/08/11 12:12:04 INFO metastore: Opened a connection to metastore, current connections: 1
- 20/08/11 12:12:04 WARN metastore: set_ugi() not successful, Likely cause: new client talking to old server. Continuing without it.
- org.apache.thrift.transport.TTransportException
- at org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
- at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86)
- at org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:425)
- at org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:321)
- at org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:225)
- at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77)
- at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_set_ugi(ThriftHiveMetastore.java:4247)
- at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.set_ugi(ThriftHiveMetastore.java:4233)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:498)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.reconnect(HiveMetaStoreClient.java:340)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:163)
- at com.sun.proxy.$Proxy32.getAllFunctions(Unknown Source)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:2336)
- at com.sun.proxy.$Proxy32.getAllFunctions(Unknown Source)
- at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894)
- at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248)
- at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231)
- at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:388)
- at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332)
- at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312)
- at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288)
- at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:288)
- at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:314)
- at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:255)
- at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:254)
- at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:304)
- at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:413)
- at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221)
- at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23)
- at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
- at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221)
- at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:154)
- at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:144)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:40)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$resourceLoader$1(HiveSessionStateBuilder.scala:47)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.client$lzycompute(HiveSessionStateBuilder.scala:115)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.client(HiveSessionStateBuilder.scala:115)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.addJar(HiveSessionStateBuilder.scala:117)
- at org.apache.spark.sql.execution.command.AddJarCommand.run(resources.scala:40)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79)
- at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:226)
- at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3472)
- at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$4(SQLExecution.scala:100)
- at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160)
- at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:87)
- at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3468)
- at org.apache.spark.sql.Dataset.<init>(Dataset.scala:226)
- at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:96)
- at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:607)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
- at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
- at py4j.Gateway.invoke(Gateway.java:282)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:238)
- at java.lang.Thread.run(Thread.java:748)
- 20/08/11 12:12:04 INFO metastore: Connected to metastore.
- 20/08/11 12:12:04 WARN Hive: Failed to register all functions.
- org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.thrift.transport.TTransportException
- at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3897)
- at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248)
- at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231)
- at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:388)
- at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332)
- at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312)
- at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288)
- at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:288)
- at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:314)
- at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:255)
- at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:254)
- at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:304)
- at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:413)
- at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221)
- at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23)
- at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
- at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221)
- at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:154)
- at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:144)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:40)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$resourceLoader$1(HiveSessionStateBuilder.scala:47)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.client$lzycompute(HiveSessionStateBuilder.scala:115)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.client(HiveSessionStateBuilder.scala:115)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.addJar(HiveSessionStateBuilder.scala:117)
- at org.apache.spark.sql.execution.command.AddJarCommand.run(resources.scala:40)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79)
- at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:226)
- at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3472)
- at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$4(SQLExecution.scala:100)
- at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160)
- at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:87)
- at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3468)
- at org.apache.spark.sql.Dataset.<init>(Dataset.scala:226)
- at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:96)
- at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:607)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
- at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
- at py4j.Gateway.invoke(Gateway.java:282)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:238)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: org.apache.thrift.transport.TTransportException
- at org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
- at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86)
- at org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:425)
- at org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:321)
- at org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:225)
- at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77)
- at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_all_functions(ThriftHiveMetastore.java:3845)
- at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_all_functions(ThriftHiveMetastore.java:3833)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getAllFunctions(HiveMetaStoreClient.java:2399)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:173)
- at com.sun.proxy.$Proxy32.getAllFunctions(Unknown Source)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:2336)
- at com.sun.proxy.$Proxy32.getAllFunctions(Unknown Source)
- at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894)
- ... 47 more
- 20/08/11 12:12:04 WARN HiveClientImpl: HiveClient got thrift exception, destroying client and retrying (0 tries remaining)
- org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.thrift.transport.TTransportException
- at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:236)
- at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:388)
- at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332)
- at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312)
- at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288)
- at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:288)
- at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:314)
- at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:255)
- at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:254)
- at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:304)
- at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:413)
- at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221)
- at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23)
- at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
- at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221)
- at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:154)
- at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:144)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:40)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$resourceLoader$1(HiveSessionStateBuilder.scala:47)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.client$lzycompute(HiveSessionStateBuilder.scala:115)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.client(HiveSessionStateBuilder.scala:115)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.addJar(HiveSessionStateBuilder.scala:117)
- at org.apache.spark.sql.execution.command.AddJarCommand.run(resources.scala:40)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79)
- at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:226)
- at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3472)
- at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$4(SQLExecution.scala:100)
- at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160)
- at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:87)
- at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3468)
- at org.apache.spark.sql.Dataset.<init>(Dataset.scala:226)
- at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:96)
- at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:607)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
- at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
- at py4j.Gateway.invoke(Gateway.java:282)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:238)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.thrift.transport.TTransportException
- at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3897)
- at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248)
- at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231)
- ... 45 more
- Caused by: org.apache.thrift.transport.TTransportException
- at org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
- at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86)
- at org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:425)
- at org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:321)
- at org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:225)
- at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77)
- at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_all_functions(ThriftHiveMetastore.java:3845)
- at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_all_functions(ThriftHiveMetastore.java:3833)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getAllFunctions(HiveMetaStoreClient.java:2399)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:173)
- at com.sun.proxy.$Proxy32.getAllFunctions(Unknown Source)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:2336)
- at com.sun.proxy.$Proxy32.getAllFunctions(Unknown Source)
- at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894)
- ... 47 more
- 20/08/11 12:12:05 WARN HiveClientImpl: Deadline exceeded
- Traceback (most recent call last):
- File "/var/lib/spark/python/lib/pyspark.zip/pyspark/sql/utils.py", line 98, in deco
- File "/var/lib/spark/python/lib/py4j-0.10.8.1-src.zip/py4j/protocol.py", line 328, in get_return_value
- py4j.protocol.Py4JJavaError: An error occurred while calling o53.sql.
- : org.apache.spark.sql.AnalysisException: org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.thrift.transport.TTransportException;
- at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:109)
- at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221)
- at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:154)
- at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:144)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:40)
- at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$resourceLoader$1(HiveSessionStateBuilder.scala:47)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.client$lzycompute(HiveSessionStateBuilder.scala:115)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.client(HiveSessionStateBuilder.scala:115)
- at org.apache.spark.sql.hive.HiveSessionResourceLoader.addJar(HiveSessionStateBuilder.scala:117)
- at org.apache.spark.sql.execution.command.AddJarCommand.run(resources.scala:40)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79)
- at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:226)
- at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3472)
- at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$4(SQLExecution.scala:100)
- at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160)
- at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:87)
- at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3468)
- at org.apache.spark.sql.Dataset.<init>(Dataset.scala:226)
- at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:96)
- at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:607)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
- at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
- at py4j.Gateway.invoke(Gateway.java:282)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:238)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.thrift.transport.TTransportException
- at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:236)
- at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:388)
- at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332)
- at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312)
- at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288)
- at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:288)
- at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:314)
- at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:255)
- at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:254)
- at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:304)
- at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:413)
- at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221)
- at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23)
- at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
- ... 32 more
- Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.thrift.transport.TTransportException
- at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3897)
- at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248)
- at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231)
- ... 45 more
- Caused by: org.apache.thrift.transport.TTransportException
- at org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
- at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86)
- at org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:425)
- at org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:321)
- at org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:225)
- at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77)
- at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_all_functions(ThriftHiveMetastore.java:3845)
- at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_all_functions(ThriftHiveMetastore.java:3833)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getAllFunctions(HiveMetaStoreClient.java:2399)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:173)
- at com.sun.proxy.$Proxy32.getAllFunctions(Unknown Source)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:2336)
- at com.sun.proxy.$Proxy32.getAllFunctions(Unknown Source)
- at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894)
- ... 47 more
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement