Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- C:\Users\wrlxs\park\python\lib\py4j-0.10.3-src.zip\py4j\java_gateway.py in __call__(self, *args)
- 1131 answer = self.gateway_client.send_command(command)
- 1132 return_value = get_return_value(
- -> 1133 answer, self.gateway_client, self.target_id, self.name)
- 1134
- 1135 for temp_arg in temp_args:
- C:\Users\wrlxs\park\python\pyspark\sql\utils.py in deco(*a, **kw)
- 61 def deco(*a, **kw):
- 62 try:
- ---> 63 return f(*a, **kw)
- 64 except py4j.protocol.Py4JJavaError as e:
- 65 s = e.java_exception.toString()
- C:\Users\wrlxs\park\python\lib\py4j-0.10.3-src.zip\py4j\protocol.py in get_return_value(answer, gateway_client, target_id, name)
- 317 raise Py4JJavaError(
- 318 "An error occurred while calling {0}{1}{2}.\n".
- --> 319 format(target_id, ".", name), value)
- 320 else:
- 321 raise Py4JError(
- Py4JJavaError: An error occurred while calling o25.applySchemaToPythonRDD.
- : java.lang.RuntimeException: java.lang.RuntimeException: Error while running command to get file permissions : java.io.IOException: (null) entry in command string: null ls -F C:\tmp\hive
- at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:770)
- at org.apache.hadoop.util.Shell.execCommand(Shell.java:866)
- at org.apache.hadoop.util.Shell.execCommand(Shell.java:849)
- at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097)
- at org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:659)
- at org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.getPermission(RawLocalFileSystem.java:634)
- at org.apache.hadoop.hive.ql.session.SessionState.createRootHDFSDir(SessionState.java:599)
- at org.apache.hadoop.hive.ql.session.SessionState.createSessionDirs(SessionState.java:554)
- at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:508)
- at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:189)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown Source)
- at java.lang.reflect.Constructor.newInstance(Unknown Source)
- at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:258)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:359)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:263)
- at org.apache.spark.sql.hive.HiveSharedState.metadataHive$lzycompute(HiveSharedState.scala:39)
- at org.apache.spark.sql.hive.HiveSharedState.metadataHive(HiveSharedState.scala:38)
- at org.apache.spark.sql.hive.HiveSharedState.externalCatalog$lzycompute(HiveSharedState.scala:46)
- at org.apache.spark.sql.hive.HiveSharedState.externalCatalog(HiveSharedState.scala:45)
- at org.apache.spark.sql.hive.HiveSessionState.catalog$lzycompute(HiveSessionState.scala:50)
- at org.apache.spark.sql.hive.HiveSessionState.catalog(HiveSessionState.scala:48)
- at org.apache.spark.sql.hive.HiveSessionState$$anon$1.<init>(HiveSessionState.scala:63)
- at org.apache.spark.sql.hive.HiveSessionState.analyzer$lzycompute(HiveSessionState.scala:63)
- at org.apache.spark.sql.hive.HiveSessionState.analyzer(HiveSessionState.scala:62)
- at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:49)
- at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
- at org.apache.spark.sql.SparkSession.applySchemaToPythonRDD(SparkSession.scala:666)
- at org.apache.spark.sql.SparkSession.applySchemaToPythonRDD(SparkSession.scala:656)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
- at java.lang.reflect.Method.invoke(Unknown Source)
- at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:237)
- at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
- at py4j.Gateway.invoke(Gateway.java:280)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:214)
- at java.lang.Thread.run(Unknown Source)
- at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:522)
- at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:189)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown Source)
- at java.lang.reflect.Constructor.newInstance(Unknown Source)
- at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:258)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:359)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:263)
- at org.apache.spark.sql.hive.HiveSharedState.metadataHive$lzycompute(HiveSharedState.scala:39)
- at org.apache.spark.sql.hive.HiveSharedState.metadataHive(HiveSharedState.scala:38)
- at org.apache.spark.sql.hive.HiveSharedState.externalCatalog$lzycompute(HiveSharedState.scala:46)
- at org.apache.spark.sql.hive.HiveSharedState.externalCatalog(HiveSharedState.scala:45)
- at org.apache.spark.sql.hive.HiveSessionState.catalog$lzycompute(HiveSessionState.scala:50)
- at org.apache.spark.sql.hive.HiveSessionState.catalog(HiveSessionState.scala:48)
- at org.apache.spark.sql.hive.HiveSessionState$$anon$1.<init>(HiveSessionState.scala:63)
- at org.apache.spark.sql.hive.HiveSessionState.analyzer$lzycompute(HiveSessionState.scala:63)
- at org.apache.spark.sql.hive.HiveSessionState.analyzer(HiveSessionState.scala:62)
- at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:49)
- at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
- at org.apache.spark.sql.SparkSession.applySchemaToPythonRDD(SparkSession.scala:666)
- at org.apache.spark.sql.SparkSession.applySchemaToPythonRDD(SparkSession.scala:656)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
- at java.lang.reflect.Method.invoke(Unknown Source)
- at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:237)
- at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
- at py4j.Gateway.invoke(Gateway.java:280)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:214)
- at java.lang.Thread.run(Unknown Source)
- Caused by: java.lang.RuntimeException: Error while running command to get file permissions : java.io.IOException: (null) entry in command string: null ls -F C:\tmp\hive
- at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:770)
- at org.apache.hadoop.util.Shell.execCommand(Shell.java:866)
- at org.apache.hadoop.util.Shell.execCommand(Shell.java:849)
- at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097)
- at org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:659)
- at org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.getPermission(RawLocalFileSystem.java:634)
- at org.apache.hadoop.hive.ql.session.SessionState.createRootHDFSDir(SessionState.java:599)
- at org.apache.hadoop.hive.ql.session.SessionState.createSessionDirs(SessionState.java:554)
- at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:508)
- at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:189)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown Source)
- at java.lang.reflect.Constructor.newInstance(Unknown Source)
- at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:258)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:359)
- at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:263)
- at org.apache.spark.sql.hive.HiveSharedState.metadataHive$lzycompute(HiveSharedState.scala:39)
- at org.apache.spark.sql.hive.HiveSharedState.metadataHive(HiveSharedState.scala:38)
- at org.apache.spark.sql.hive.HiveSharedState.externalCatalog$lzycompute(HiveSharedState.scala:46)
- at org.apache.spark.sql.hive.HiveSharedState.externalCatalog(HiveSharedState.scala:45)
- at org.apache.spark.sql.hive.HiveSessionState.catalog$lzycompute(HiveSessionState.scala:50)
- at org.apache.spark.sql.hive.HiveSessionState.catalog(HiveSessionState.scala:48)
- at org.apache.spark.sql.hive.HiveSessionState$$anon$1.<init>(HiveSessionState.scala:63)
- at org.apache.spark.sql.hive.HiveSessionState.analyzer$lzycompute(HiveSessionState.scala:63)
- at org.apache.spark.sql.hive.HiveSessionState.analyzer(HiveSessionState.scala:62)
- at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:49)
- at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
- at org.apache.spark.sql.SparkSession.applySchemaToPythonRDD(SparkSession.scala:666)
- at org.apache.spark.sql.SparkSession.applySchemaToPythonRDD(SparkSession.scala:656)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
- at java.lang.reflect.Method.invoke(Unknown Source)
- at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:237)
- at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
- at py4j.Gateway.invoke(Gateway.java:280)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:214)
- at java.lang.Thread.run(Unknown Source)
- at org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:699)
- at org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.getPermission(RawLocalFileSystem.java:634)
- at org.apache.hadoop.hive.ql.session.SessionState.createRootHDFSDir(SessionState.java:599)
- at org.apache.hadoop.hive.ql.session.SessionState.createSessionDirs(SessionState.java:554)
- at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:508)
- ... 32 more
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement