Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- hive>
- >
- >
- >
- >
- >
- >
- >
- > select count(*) from logparsersample;
- 17/06/28 00:12:25 [main]: INFO log.PerfLogger: <PERFLOG method=Driver.run from=org.apache.hadoop.hive.ql.Driver>
- 17/06/28 00:12:25 [main]: INFO log.PerfLogger: <PERFLOG method=TimeToSubmit from=org.apache.hadoop.hive.ql.Driver>
- 17/06/28 00:12:25 [main]: INFO log.PerfLogger: <PERFLOG method=compile from=org.apache.hadoop.hive.ql.Driver>
- 17/06/28 00:12:25 [main]: INFO log.PerfLogger: <PERFLOG method=parse from=org.apache.hadoop.hive.ql.Driver>
- 17/06/28 00:12:25 [main]: INFO parse.ParseDriver: Parsing command: select count(*) from logparsersample
- 17/06/28 00:12:25 [main]: INFO parse.ParseDriver: Parse Completed
- 17/06/28 00:12:25 [main]: INFO log.PerfLogger: </PERFLOG method=parse start=1498588945656 end=1498588945712 duration=56 from=org.apache.hadoop.hive.ql.Driver>
- 17/06/28 00:12:25 [main]: INFO log.PerfLogger: <PERFLOG method=semanticAnalyze from=org.apache.hadoop.hive.ql.Driver>
- 17/06/28 00:12:25 [main]: INFO parse.CalcitePlanner: Starting Semantic Analysis
- 17/06/28 00:12:25 [main]: INFO parse.CalcitePlanner: Completed phase 1 of Semantic Analysis
- 17/06/28 00:12:25 [main]: INFO parse.CalcitePlanner: Get metadata for source tables
- 17/06/28 00:12:25 [main]: INFO metastore.HiveMetaStore: 0: get_table : db=test tbl=logparsersample
- 17/06/28 00:12:25 [main]: INFO HiveMetaStore.audit: ugi=chamith ip=unknown-ip-addr cmd=get_table : db=test tbl=logparsersample
- 17/06/28 00:12:25 [main]: INFO parse.CalcitePlanner: Get metadata for subqueries
- 17/06/28 00:12:25 [main]: INFO parse.CalcitePlanner: Get metadata for destination tables
- 17/06/28 00:12:25 [main]: ERROR hdfs.KeyProviderCache: Could not find uri with key [hadoop.security.key.provider.path] to create a keyProvider !!
- 17/06/28 00:12:25 [main]: WARN ipc.Client: Failed to connect to server: localhost/127.0.0.1:9000: try once and fail.
- java.net.ConnectException: Connection refused
- at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
- at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
- at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
- at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
- at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
- at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:681)
- at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:777)
- at org.apache.hadoop.ipc.Client$Connection.access$3500(Client.java:409)
- at org.apache.hadoop.ipc.Client.getConnection(Client.java:1542)
- at org.apache.hadoop.ipc.Client.call(Client.java:1373)
- at org.apache.hadoop.ipc.Client.call(Client.java:1337)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:227)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
- at com.sun.proxy.$Proxy16.getEZForPath(Unknown Source)
- at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getEZForPath(ClientNamenodeProtocolTranslatorPB.java:1403)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:398)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:163)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:155)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
- at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:335)
- at com.sun.proxy.$Proxy17.getEZForPath(Unknown Source)
- at org.apache.hadoop.hdfs.DFSClient.getEZForPath(DFSClient.java:2808)
- at org.apache.hadoop.hdfs.DistributedFileSystem$51.doCall(DistributedFileSystem.java:2252)
- at org.apache.hadoop.hdfs.DistributedFileSystem$51.doCall(DistributedFileSystem.java:2249)
- at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
- at org.apache.hadoop.hdfs.DistributedFileSystem.getEZForPath(DistributedFileSystem.java:2249)
- at org.apache.hadoop.hdfs.client.HdfsAdmin.getEncryptionZoneForPath(HdfsAdmin.java:338)
- at org.apache.hadoop.hive.shims.Hadoop23Shims$HdfsEncryptionShim.isPathEncrypted(Hadoop23Shims.java:1221)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.isPathEncrypted(SemanticAnalyzer.java:1884)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getStrongestEncryptedTablePath(SemanticAnalyzer.java:1965)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getStagingDirectoryPathname(SemanticAnalyzer.java:1997)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getMetaData(SemanticAnalyzer.java:1810)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getMetaData(SemanticAnalyzer.java:1545)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genResolvedParseTree(SemanticAnalyzer.java:10077)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:10128)
- at org.apache.hadoop.hive.ql.parse.CalcitePlanner.analyzeInternal(CalcitePlanner.java:210)
- at org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:227)
- at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:425)
- at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:309)
- at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1145)
- at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1193)
- at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1082)
- at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1072)
- at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:213)
- at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:165)
- at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
- at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:736)
- at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:681)
- at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:621)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.util.RunJar.run(RunJar.java:234)
- at org.apache.hadoop.util.RunJar.main(RunJar.java:148)
- 17/06/28 00:12:25 [main]: ERROR parse.CalcitePlanner: org.apache.hadoop.hive.ql.metadata.HiveException: Unable to determine if hdfs://localhost:9000/user/chamith/es-hadoop is encrypted: java.net.ConnectException: Call From ubuntu/127.0.1.1 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.isPathEncrypted(SemanticAnalyzer.java:1888)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getStrongestEncryptedTablePath(SemanticAnalyzer.java:1965)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getStagingDirectoryPathname(SemanticAnalyzer.java:1997)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getMetaData(SemanticAnalyzer.java:1810)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getMetaData(SemanticAnalyzer.java:1545)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genResolvedParseTree(SemanticAnalyzer.java:10077)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:10128)
- at org.apache.hadoop.hive.ql.parse.CalcitePlanner.analyzeInternal(CalcitePlanner.java:210)
- at org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:227)
- at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:425)
- at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:309)
- at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1145)
- at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1193)
- at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1082)
- at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1072)
- at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:213)
- at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:165)
- at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
- at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:736)
- at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:681)
- at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:621)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.util.RunJar.run(RunJar.java:234)
- at org.apache.hadoop.util.RunJar.main(RunJar.java:148)
- Caused by: java.net.ConnectException: Call From ubuntu/127.0.1.1 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:801)
- at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:732)
- at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1485)
- at org.apache.hadoop.ipc.Client.call(Client.java:1427)
- at org.apache.hadoop.ipc.Client.call(Client.java:1337)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:227)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
- at com.sun.proxy.$Proxy16.getEZForPath(Unknown Source)
- at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getEZForPath(ClientNamenodeProtocolTranslatorPB.java:1403)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:398)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:163)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:155)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
- at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:335)
- at com.sun.proxy.$Proxy17.getEZForPath(Unknown Source)
- at org.apache.hadoop.hdfs.DFSClient.getEZForPath(DFSClient.java:2808)
- at org.apache.hadoop.hdfs.DistributedFileSystem$51.doCall(DistributedFileSystem.java:2252)
- at org.apache.hadoop.hdfs.DistributedFileSystem$51.doCall(DistributedFileSystem.java:2249)
- at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
- at org.apache.hadoop.hdfs.DistributedFileSystem.getEZForPath(DistributedFileSystem.java:2249)
- at org.apache.hadoop.hdfs.client.HdfsAdmin.getEncryptionZoneForPath(HdfsAdmin.java:338)
- at org.apache.hadoop.hive.shims.Hadoop23Shims$HdfsEncryptionShim.isPathEncrypted(Hadoop23Shims.java:1221)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.isPathEncrypted(SemanticAnalyzer.java:1884)
- ... 26 more
- Caused by: java.net.ConnectException: Connection refused
- at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
- at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
- at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
- at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
- at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
- at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:681)
- at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:777)
- at org.apache.hadoop.ipc.Client$Connection.access$3500(Client.java:409)
- at org.apache.hadoop.ipc.Client.getConnection(Client.java:1542)
- at org.apache.hadoop.ipc.Client.call(Client.java:1373)
- ... 49 more
- FAILED: SemanticException Unable to determine if hdfs://localhost:9000/user/chamith/es-hadoop is encrypted: java.net.ConnectException: Call From ubuntu/127.0.1.1 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
- 17/06/28 00:12:25 [main]: ERROR ql.Driver: FAILED: SemanticException Unable to determine if hdfs://localhost:9000/user/chamith/es-hadoop is encrypted: java.net.ConnectException: Call From ubuntu/127.0.1.1 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
- org.apache.hadoop.hive.ql.parse.SemanticException: Unable to determine if hdfs://localhost:9000/user/chamith/es-hadoop is encrypted: java.net.ConnectException: Call From ubuntu/127.0.1.1 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getMetaData(SemanticAnalyzer.java:1868)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getMetaData(SemanticAnalyzer.java:1545)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genResolvedParseTree(SemanticAnalyzer.java:10077)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:10128)
- at org.apache.hadoop.hive.ql.parse.CalcitePlanner.analyzeInternal(CalcitePlanner.java:210)
- at org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:227)
- at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:425)
- at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:309)
- at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1145)
- at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1193)
- at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1082)
- at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1072)
- at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:213)
- at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:165)
- at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
- at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:736)
- at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:681)
- at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:621)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.util.RunJar.run(RunJar.java:234)
- at org.apache.hadoop.util.RunJar.main(RunJar.java:148)
- Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Unable to determine if hdfs://localhost:9000/user/chamith/es-hadoop is encrypted: java.net.ConnectException: Call From ubuntu/127.0.1.1 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.isPathEncrypted(SemanticAnalyzer.java:1888)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getStrongestEncryptedTablePath(SemanticAnalyzer.java:1965)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getStagingDirectoryPathname(SemanticAnalyzer.java:1997)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getMetaData(SemanticAnalyzer.java:1810)
- ... 23 more
- Caused by: java.net.ConnectException: Call From ubuntu/127.0.1.1 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:801)
- at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:732)
- at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1485)
- at org.apache.hadoop.ipc.Client.call(Client.java:1427)
- at org.apache.hadoop.ipc.Client.call(Client.java:1337)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:227)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
- at com.sun.proxy.$Proxy16.getEZForPath(Unknown Source)
- at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getEZForPath(ClientNamenodeProtocolTranslatorPB.java:1403)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:398)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:163)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:155)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
- at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:335)
- at com.sun.proxy.$Proxy17.getEZForPath(Unknown Source)
- at org.apache.hadoop.hdfs.DFSClient.getEZForPath(DFSClient.java:2808)
- at org.apache.hadoop.hdfs.DistributedFileSystem$51.doCall(DistributedFileSystem.java:2252)
- at org.apache.hadoop.hdfs.DistributedFileSystem$51.doCall(DistributedFileSystem.java:2249)
- at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
- at org.apache.hadoop.hdfs.DistributedFileSystem.getEZForPath(DistributedFileSystem.java:2249)
- at org.apache.hadoop.hdfs.client.HdfsAdmin.getEncryptionZoneForPath(HdfsAdmin.java:338)
- at org.apache.hadoop.hive.shims.Hadoop23Shims$HdfsEncryptionShim.isPathEncrypted(Hadoop23Shims.java:1221)
- at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.isPathEncrypted(SemanticAnalyzer.java:1884)
- ... 26 more
- Caused by: java.net.ConnectException: Connection refused
- at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
- at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
- at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
- at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
- at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
- at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:681)
- at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:777)
- at org.apache.hadoop.ipc.Client$Connection.access$3500(Client.java:409)
- at org.apache.hadoop.ipc.Client.getConnection(Client.java:1542)
- at org.apache.hadoop.ipc.Client.call(Client.java:1373)
- ... 49 more
- 17/06/28 00:12:25 [main]: INFO log.PerfLogger: </PERFLOG method=compile start=1498588945655 end=1498588945924 duration=269 from=org.apache.hadoop.hive.ql.Driver>
- 17/06/28 00:12:25 [main]: INFO log.PerfLogger: <PERFLOG method=releaseLocks from=org.apache.hadoop.hive.ql.Driver>
- 17/06/28 00:12:25 [main]: INFO log.PerfLogger: </PERFLOG method=releaseLocks start=1498588945924 end=1498588945924 duration=0 from=org.apache.hadoop.hive.ql.Driver>
- 17/06/28 00:12:25 [main]: INFO log.PerfLogger: <PERFLOG method=releaseLocks from=org.apache.hadoop.hive.ql.Driver>
- 17/06/28 00:12:25 [main]: INFO log.PerfLogger: </PERFLOG method=releaseLocks start=1498588945924 end=1498588945925 duration=1 from=org.apache.hadoop.hive.ql.Driver>
- hive>
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement