Advertisement
Guest User

ranger

a guest
Feb 7th, 2016
432
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 27.39 KB | None | 0 0
  1. phdns01.cloud.hortonworks.com
  2. Tasks
  3. Copy Open Restart NameNode
  4. stderr: /var/lib/ambari-agent/data/errors-1419.txt
  5.  
  6. 2016-02-07 07:36:10,083 - Error creating ambari admin user. Http status code - 401.
  7. {"statusCode":401,"msgDesc":"Authentication Failed"}
  8. 2016-02-07 07:36:10,083 - Ambari admin user creation failed
  9. stdout: /var/lib/ambari-agent/data/output-1419.txt
  10.  
  11. 2016-02-07 07:36:03,541 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.3.4.0-3276
  12. 2016-02-07 07:36:03,542 - Checking if need to create versioned conf dir /etc/hadoop/2.3.4.0-3276/0
  13. 2016-02-07 07:36:03,542 - call['conf-select create-conf-dir --package hadoop --stack-version 2.3.4.0-3276 --conf-version 0'] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1}
  14. 2016-02-07 07:36:03,580 - call returned (1, '/etc/hadoop/2.3.4.0-3276/0 exist already', '')
  15. 2016-02-07 07:36:03,581 - checked_call['conf-select set-conf-dir --package hadoop --stack-version 2.3.4.0-3276 --conf-version 0'] {'logoutput': False, 'sudo': True, 'quiet': False}
  16. 2016-02-07 07:36:03,616 - checked_call returned (0, '/usr/hdp/2.3.4.0-3276/hadoop/conf -> /etc/hadoop/2.3.4.0-3276/0')
  17. 2016-02-07 07:36:03,616 - Ensuring that hadoop has the correct symlink structure
  18. 2016-02-07 07:36:03,617 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf
  19. 2016-02-07 07:36:03,817 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.3.4.0-3276
  20. 2016-02-07 07:36:03,817 - Checking if need to create versioned conf dir /etc/hadoop/2.3.4.0-3276/0
  21. 2016-02-07 07:36:03,817 - call['conf-select create-conf-dir --package hadoop --stack-version 2.3.4.0-3276 --conf-version 0'] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1}
  22. 2016-02-07 07:36:03,855 - call returned (1, '/etc/hadoop/2.3.4.0-3276/0 exist already', '')
  23. 2016-02-07 07:36:03,855 - checked_call['conf-select set-conf-dir --package hadoop --stack-version 2.3.4.0-3276 --conf-version 0'] {'logoutput': False, 'sudo': True, 'quiet': False}
  24. 2016-02-07 07:36:03,891 - checked_call returned (0, '/usr/hdp/2.3.4.0-3276/hadoop/conf -> /etc/hadoop/2.3.4.0-3276/0')
  25. 2016-02-07 07:36:03,891 - Ensuring that hadoop has the correct symlink structure
  26. 2016-02-07 07:36:03,891 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf
  27. 2016-02-07 07:36:03,894 - Group['flink'] {}
  28. 2016-02-07 07:36:03,896 - Group['spark'] {}
  29. 2016-02-07 07:36:03,896 - Group['ranger'] {}
  30. 2016-02-07 07:36:03,896 - Group['zeppelin'] {}
  31. 2016-02-07 07:36:03,897 - Group['hadoop'] {}
  32. 2016-02-07 07:36:03,897 - Group['users'] {}
  33. 2016-02-07 07:36:03,897 - Group['knox'] {}
  34. 2016-02-07 07:36:03,898 - User['hive'] {'gid': 'hadoop', 'groups': ['hadoop']}
  35. 2016-02-07 07:36:03,899 - User['storm'] {'gid': 'hadoop', 'groups': ['hadoop']}
  36. 2016-02-07 07:36:03,900 - User['zookeeper'] {'gid': 'hadoop', 'groups': ['hadoop']}
  37. 2016-02-07 07:36:03,901 - User['oozie'] {'gid': 'hadoop', 'groups': ['users']}
  38. 2016-02-07 07:36:03,902 - User['atlas'] {'gid': 'hadoop', 'groups': ['hadoop']}
  39. 2016-02-07 07:36:03,903 - User['ams'] {'gid': 'hadoop', 'groups': ['hadoop']}
  40. 2016-02-07 07:36:03,904 - User['falcon'] {'gid': 'hadoop', 'groups': ['users']}
  41. 2016-02-07 07:36:03,905 - User['ranger'] {'gid': 'hadoop', 'groups': ['ranger']}
  42. 2016-02-07 07:36:03,906 - User['tez'] {'gid': 'hadoop', 'groups': ['users']}
  43. 2016-02-07 07:36:03,907 - User['zeppelin'] {'gid': 'hadoop', 'groups': ['hadoop']}
  44. 2016-02-07 07:36:03,908 - User['flink'] {'gid': 'hadoop', 'groups': ['hadoop']}
  45. 2016-02-07 07:36:03,909 - User['spark'] {'gid': 'hadoop', 'groups': ['hadoop']}
  46. 2016-02-07 07:36:03,910 - User['ambari-qa'] {'gid': 'hadoop', 'groups': ['users']}
  47. 2016-02-07 07:36:03,910 - User['flume'] {'gid': 'hadoop', 'groups': ['hadoop']}
  48. 2016-02-07 07:36:03,911 - User['kafka'] {'gid': 'hadoop', 'groups': ['hadoop']}
  49. 2016-02-07 07:36:03,912 - User['hdfs'] {'gid': 'hadoop', 'groups': ['hadoop']}
  50. 2016-02-07 07:36:03,914 - User['sqoop'] {'gid': 'hadoop', 'groups': ['hadoop']}
  51. 2016-02-07 07:36:03,914 - User['yarn'] {'gid': 'hadoop', 'groups': ['hadoop']}
  52. 2016-02-07 07:36:03,915 - User['mapred'] {'gid': 'hadoop', 'groups': ['hadoop']}
  53. 2016-02-07 07:36:03,916 - User['hbase'] {'gid': 'hadoop', 'groups': ['hadoop']}
  54. 2016-02-07 07:36:03,917 - User['knox'] {'gid': 'hadoop', 'groups': ['hadoop']}
  55. 2016-02-07 07:36:03,918 - User['hcat'] {'gid': 'hadoop', 'groups': ['hadoop']}
  56. 2016-02-07 07:36:03,919 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
  57. 2016-02-07 07:36:03,922 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] {'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'}
  58. 2016-02-07 07:36:03,930 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] due to not_if
  59. 2016-02-07 07:36:03,930 - Directory['/tmp/hbase-hbase'] {'owner': 'hbase', 'recursive': True, 'mode': 0775, 'cd_access': 'a'}
  60. 2016-02-07 07:36:03,935 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
  61. 2016-02-07 07:36:03,936 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] {'not_if': '(test $(id -u hbase) -gt 1000) || (false)'}
  62. 2016-02-07 07:36:03,943 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] due to not_if
  63. 2016-02-07 07:36:03,944 - Group['hdfs'] {'ignore_failures': False}
  64. 2016-02-07 07:36:03,944 - User['hdfs'] {'ignore_failures': False, 'groups': ['hadoop', 'hdfs']}
  65. 2016-02-07 07:36:03,945 - Directory['/etc/hadoop'] {'mode': 0755}
  66. 2016-02-07 07:36:03,966 - File['/usr/hdp/current/hadoop-client/conf/hadoop-env.sh'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'}
  67. 2016-02-07 07:36:03,967 - Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] {'owner': 'hdfs', 'group': 'hadoop', 'mode': 0777}
  68. 2016-02-07 07:36:03,986 - Execute[('setenforce', '0')] {'not_if': '(! which getenforce ) || (which getenforce && getenforce | grep -q Disabled)', 'sudo': True, 'only_if': 'test -f /selinux/enforce'}
  69. 2016-02-07 07:36:04,012 - Directory['/var/log/hadoop'] {'owner': 'root', 'mode': 0775, 'group': 'hadoop', 'recursive': True, 'cd_access': 'a'}
  70. 2016-02-07 07:36:04,017 - Directory['/var/run/hadoop'] {'owner': 'root', 'group': 'root', 'recursive': True, 'cd_access': 'a'}
  71. 2016-02-07 07:36:04,017 - Directory['/tmp/hadoop-hdfs'] {'owner': 'hdfs', 'recursive': True, 'cd_access': 'a'}
  72. 2016-02-07 07:36:04,025 - File['/usr/hdp/current/hadoop-client/conf/commons-logging.properties'] {'content': Template('commons-logging.properties.j2'), 'owner': 'hdfs'}
  73. 2016-02-07 07:36:04,027 - File['/usr/hdp/current/hadoop-client/conf/health_check'] {'content': Template('health_check.j2'), 'owner': 'hdfs'}
  74. 2016-02-07 07:36:04,028 - File['/usr/hdp/current/hadoop-client/conf/log4j.properties'] {'content': ..., 'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644}
  75. 2016-02-07 07:36:04,042 - File['/usr/hdp/current/hadoop-client/conf/hadoop-metrics2.properties'] {'content': Template('hadoop-metrics2.properties.j2'), 'owner': 'hdfs'}
  76. 2016-02-07 07:36:04,043 - File['/usr/hdp/current/hadoop-client/conf/task-log4j.properties'] {'content': StaticFile('task-log4j.properties'), 'mode': 0755}
  77. 2016-02-07 07:36:04,044 - File['/usr/hdp/current/hadoop-client/conf/configuration.xsl'] {'owner': 'hdfs', 'group': 'hadoop'}
  78. 2016-02-07 07:36:04,052 - File['/etc/hadoop/conf/topology_mappings.data'] {'owner': 'hdfs', 'content': Template('topology_mappings.data.j2'), 'only_if': 'test -d /etc/hadoop/conf', 'group': 'hadoop'}
  79. 2016-02-07 07:36:04,058 - File['/etc/hadoop/conf/topology_script.py'] {'content': StaticFile('topology_script.py'), 'only_if': 'test -d /etc/hadoop/conf', 'mode': 0755}
  80. 2016-02-07 07:36:04,330 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.3.4.0-3276
  81. 2016-02-07 07:36:04,330 - Checking if need to create versioned conf dir /etc/hadoop/2.3.4.0-3276/0
  82. 2016-02-07 07:36:04,330 - call['conf-select create-conf-dir --package hadoop --stack-version 2.3.4.0-3276 --conf-version 0'] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1}
  83. 2016-02-07 07:36:04,364 - call returned (1, '/etc/hadoop/2.3.4.0-3276/0 exist already', '')
  84. 2016-02-07 07:36:04,364 - checked_call['conf-select set-conf-dir --package hadoop --stack-version 2.3.4.0-3276 --conf-version 0'] {'logoutput': False, 'sudo': True, 'quiet': False}
  85. 2016-02-07 07:36:04,406 - checked_call returned (0, '/usr/hdp/2.3.4.0-3276/hadoop/conf -> /etc/hadoop/2.3.4.0-3276/0')
  86. 2016-02-07 07:36:04,407 - Ensuring that hadoop has the correct symlink structure
  87. 2016-02-07 07:36:04,407 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf
  88. 2016-02-07 07:36:04,409 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.3.4.0-3276
  89. 2016-02-07 07:36:04,409 - Checking if need to create versioned conf dir /etc/hadoop/2.3.4.0-3276/0
  90. 2016-02-07 07:36:04,410 - call['conf-select create-conf-dir --package hadoop --stack-version 2.3.4.0-3276 --conf-version 0'] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1}
  91. 2016-02-07 07:36:04,448 - call returned (1, '/etc/hadoop/2.3.4.0-3276/0 exist already', '')
  92. 2016-02-07 07:36:04,449 - checked_call['conf-select set-conf-dir --package hadoop --stack-version 2.3.4.0-3276 --conf-version 0'] {'logoutput': False, 'sudo': True, 'quiet': False}
  93. 2016-02-07 07:36:04,484 - checked_call returned (0, '/usr/hdp/2.3.4.0-3276/hadoop/conf -> /etc/hadoop/2.3.4.0-3276/0')
  94. 2016-02-07 07:36:04,485 - Ensuring that hadoop has the correct symlink structure
  95. 2016-02-07 07:36:04,485 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf
  96. 2016-02-07 07:36:04,498 - Execute['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'ulimit -c unlimited ; /usr/hdp/current/hadoop-client/sbin/hadoop-daemon.sh --config /usr/hdp/current/hadoop-client/conf stop namenode''] {'environment': {'HADOOP_LIBEXEC_DIR': '/usr/hdp/current/hadoop-client/libexec'}, 'only_if': 'ambari-sudo.sh -H -E test -f /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid && ambari-sudo.sh -H -E pgrep -F /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid'}
  97. 2016-02-07 07:36:09,606 - File['/var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid'] {'action': ['delete']}
  98. 2016-02-07 07:36:09,612 - Directory['/etc/security/limits.d'] {'owner': 'root', 'group': 'root', 'recursive': True}
  99. 2016-02-07 07:36:09,620 - File['/etc/security/limits.d/hdfs.conf'] {'content': Template('hdfs.conf.j2'), 'owner': 'root', 'group': 'root', 'mode': 0644}
  100. 2016-02-07 07:36:09,621 - XmlConfig['hadoop-policy.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'configuration_attributes': {}, 'configurations': ...}
  101. 2016-02-07 07:36:09,637 - Generating config: /usr/hdp/current/hadoop-client/conf/hadoop-policy.xml
  102. 2016-02-07 07:36:09,637 - File['/usr/hdp/current/hadoop-client/conf/hadoop-policy.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'}
  103. 2016-02-07 07:36:09,652 - XmlConfig['ssl-client.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'configuration_attributes': {}, 'configurations': ...}
  104. 2016-02-07 07:36:09,663 - Generating config: /usr/hdp/current/hadoop-client/conf/ssl-client.xml
  105. 2016-02-07 07:36:09,664 - File['/usr/hdp/current/hadoop-client/conf/ssl-client.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'}
  106. 2016-02-07 07:36:09,673 - Directory['/usr/hdp/current/hadoop-client/conf/secure'] {'owner': 'root', 'group': 'hadoop', 'recursive': True, 'cd_access': 'a'}
  107. 2016-02-07 07:36:09,674 - XmlConfig['ssl-client.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf/secure', 'configuration_attributes': {}, 'configurations': ...}
  108. 2016-02-07 07:36:09,686 - Generating config: /usr/hdp/current/hadoop-client/conf/secure/ssl-client.xml
  109. 2016-02-07 07:36:09,686 - File['/usr/hdp/current/hadoop-client/conf/secure/ssl-client.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'}
  110. 2016-02-07 07:36:09,695 - XmlConfig['ssl-server.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'configuration_attributes': {}, 'configurations': ...}
  111. 2016-02-07 07:36:09,711 - Generating config: /usr/hdp/current/hadoop-client/conf/ssl-server.xml
  112. 2016-02-07 07:36:09,712 - File['/usr/hdp/current/hadoop-client/conf/ssl-server.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'}
  113. 2016-02-07 07:36:09,722 - XmlConfig['hdfs-site.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'configuration_attributes': {}, 'configurations': ...}
  114. 2016-02-07 07:36:09,734 - Generating config: /usr/hdp/current/hadoop-client/conf/hdfs-site.xml
  115. 2016-02-07 07:36:09,735 - File['/usr/hdp/current/hadoop-client/conf/hdfs-site.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'}
  116. 2016-02-07 07:36:09,799 - XmlConfig['core-site.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'mode': 0644, 'configuration_attributes': {}, 'owner': 'hdfs', 'configurations': ...}
  117. 2016-02-07 07:36:09,812 - Generating config: /usr/hdp/current/hadoop-client/conf/core-site.xml
  118. 2016-02-07 07:36:09,812 - File['/usr/hdp/current/hadoop-client/conf/core-site.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'}
  119. 2016-02-07 07:36:09,846 - File['/usr/hdp/current/hadoop-client/conf/slaves'] {'content': Template('slaves.j2'), 'owner': 'hdfs'}
  120. 2016-02-07 07:36:09,848 - Directory['/hadoop/hdfs/namenode'] {'owner': 'hdfs', 'recursive': True, 'group': 'hadoop', 'mode': 0755, 'cd_access': 'a'}
  121. 2016-02-07 07:36:09,849 - Called service start with upgrade_type: None
  122. 2016-02-07 07:36:09,849 - HDFS: Setup ranger: command retry not enabled thus skipping if ranger admin is down !
  123. 2016-02-07 07:36:09,850 - File['/var/lib/ambari-agent/tmp/mysql-connector-java.jar'] {'content': DownloadSource('http://phdns01.cloud.hortonworks.com:8080/resources//mysql-jdbc-driver.jar'), 'mode': 0644}
  124. 2016-02-07 07:36:09,850 - Not downloading the file from http://phdns01.cloud.hortonworks.com:8080/resources//mysql-jdbc-driver.jar, because /var/lib/ambari-agent/tmp/mysql-jdbc-driver.jar already exists
  125. 2016-02-07 07:36:09,853 - Execute[('cp', '--remove-destination', '/var/lib/ambari-agent/tmp/mysql-connector-java.jar', '/usr/hdp/current/hadoop-client/lib/mysql-connector-java.jar')] {'path': ['/bin', '/usr/bin/'], 'sudo': True}
  126. 2016-02-07 07:36:09,865 - File['/usr/hdp/current/hadoop-client/lib/mysql-connector-java.jar'] {'mode': 0644}
  127. 2016-02-07 07:36:09,902 - Rangeradmin: Skip ranger admin if it's down !
  128. 2016-02-07 07:36:10,083 - Error creating ambari admin user. Http status code - 401.
  129. {"statusCode":401,"msgDesc":"Authentication Failed"}
  130. 2016-02-07 07:36:10,083 - Ambari admin user creation failed
  131. 2016-02-07 07:36:10,086 - File['/usr/hdp/current/hadoop-client/conf/ranger-security.xml'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644}
  132. 2016-02-07 07:36:10,087 - Writing File['/usr/hdp/current/hadoop-client/conf/ranger-security.xml'] because contents don't match
  133. 2016-02-07 07:36:10,087 - Directory['/etc/ranger/HDPTEST_hadoop'] {'owner': 'hdfs', 'cd_access': 'a', 'group': 'hadoop', 'recursive': True, 'mode': 0775}
  134. 2016-02-07 07:36:10,088 - Directory['/etc/ranger/HDPTEST_hadoop/policycache'] {'owner': 'hdfs', 'recursive': True, 'group': 'hadoop', 'mode': 0775, 'cd_access': 'a'}
  135. 2016-02-07 07:36:10,089 - File['/etc/ranger/HDPTEST_hadoop/policycache/hdfs_HDPTEST_hadoop.json'] {'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644}
  136. 2016-02-07 07:36:10,090 - XmlConfig['ranger-hdfs-audit.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'mode': 0744, 'configuration_attributes': {}, 'owner': 'hdfs', 'configurations': ...}
  137. 2016-02-07 07:36:10,103 - Generating config: /usr/hdp/current/hadoop-client/conf/ranger-hdfs-audit.xml
  138. 2016-02-07 07:36:10,104 - File['/usr/hdp/current/hadoop-client/conf/ranger-hdfs-audit.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0744, 'encoding': 'UTF-8'}
  139. 2016-02-07 07:36:10,123 - XmlConfig['ranger-hdfs-security.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'mode': 0744, 'configuration_attributes': {}, 'owner': 'hdfs', 'configurations': ...}
  140. 2016-02-07 07:36:10,135 - Generating config: /usr/hdp/current/hadoop-client/conf/ranger-hdfs-security.xml
  141. 2016-02-07 07:36:10,135 - File['/usr/hdp/current/hadoop-client/conf/ranger-hdfs-security.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0744, 'encoding': 'UTF-8'}
  142. 2016-02-07 07:36:10,145 - XmlConfig['ranger-policymgr-ssl.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'mode': 0744, 'configuration_attributes': {}, 'owner': 'hdfs', 'configurations': ...}
  143. 2016-02-07 07:36:10,159 - Generating config: /usr/hdp/current/hadoop-client/conf/ranger-policymgr-ssl.xml
  144. 2016-02-07 07:36:10,160 - File['/usr/hdp/current/hadoop-client/conf/ranger-policymgr-ssl.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0744, 'encoding': 'UTF-8'}
  145. 2016-02-07 07:36:10,168 - Execute[('/usr/hdp/2.3.4.0-3276/ranger-hdfs-plugin/ranger_credential_helper.py', '-l', '/usr/hdp/2.3.4.0-3276/ranger-hdfs-plugin/install/lib/*', '-f', '/etc/ranger/HDPTEST_hadoop/cred.jceks', '-k', 'auditDBCred', '-v', [PROTECTED], '-c', '1')] {'logoutput': True, 'environment': {'JAVA_HOME': '/usr/jdk64/jdk1.8.0_60'}, 'sudo': True}
  146. Using Java:/usr/jdk64/jdk1.8.0_60/bin/java
  147. Alias auditDBCred created successfully!
  148. 2016-02-07 07:36:11,826 - Execute[('/usr/hdp/2.3.4.0-3276/ranger-hdfs-plugin/ranger_credential_helper.py', '-l', '/usr/hdp/2.3.4.0-3276/ranger-hdfs-plugin/install/lib/*', '-f', '/etc/ranger/HDPTEST_hadoop/cred.jceks', '-k', 'sslKeyStore', '-v', [PROTECTED], '-c', '1')] {'logoutput': True, 'environment': {'JAVA_HOME': '/usr/jdk64/jdk1.8.0_60'}, 'sudo': True}
  149. Using Java:/usr/jdk64/jdk1.8.0_60/bin/java
  150. Alias sslKeyStore created successfully!
  151. 2016-02-07 07:36:13,923 - Execute[('/usr/hdp/2.3.4.0-3276/ranger-hdfs-plugin/ranger_credential_helper.py', '-l', '/usr/hdp/2.3.4.0-3276/ranger-hdfs-plugin/install/lib/*', '-f', '/etc/ranger/HDPTEST_hadoop/cred.jceks', '-k', 'sslTrustStore', '-v', [PROTECTED], '-c', '1')] {'logoutput': True, 'environment': {'JAVA_HOME': '/usr/jdk64/jdk1.8.0_60'}, 'sudo': True}
  152. Using Java:/usr/jdk64/jdk1.8.0_60/bin/java
  153. Alias sslTrustStore created successfully!
  154. 2016-02-07 07:36:15,202 - File['/etc/ranger/HDPTEST_hadoop/cred.jceks'] {'owner': 'hdfs', 'group': 'hadoop', 'mode': 0640}
  155. /hadoop/hdfs/namenode/namenode-formatted/ exists. Namenode DFS already formatted
  156. 2016-02-07 07:36:15,204 - Directory['/hadoop/hdfs/namenode/namenode-formatted/'] {'recursive': True}
  157. 2016-02-07 07:36:15,209 - File['/etc/hadoop/conf/dfs.exclude'] {'owner': 'hdfs', 'content': Template('exclude_hosts_list.j2'), 'group': 'hadoop'}
  158. 2016-02-07 07:36:15,210 - Option for start command:
  159. 2016-02-07 07:36:15,211 - Directory['/var/run/hadoop'] {'owner': 'hdfs', 'group': 'hadoop', 'mode': 0755}
  160. 2016-02-07 07:36:15,211 - Changing owner for /var/run/hadoop from 0 to hdfs
  161. 2016-02-07 07:36:15,211 - Changing group for /var/run/hadoop from 0 to hadoop
  162. 2016-02-07 07:36:15,212 - Directory['/var/run/hadoop/hdfs'] {'owner': 'hdfs', 'recursive': True}
  163. 2016-02-07 07:36:15,212 - Directory['/var/log/hadoop/hdfs'] {'owner': 'hdfs', 'recursive': True}
  164. 2016-02-07 07:36:15,213 - File['/var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid'] {'action': ['delete'], 'not_if': 'ambari-sudo.sh -H -E test -f /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid && ambari-sudo.sh -H -E pgrep -F /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid'}
  165. 2016-02-07 07:36:15,222 - Execute['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'ulimit -c unlimited ; /usr/hdp/current/hadoop-client/sbin/hadoop-daemon.sh --config /usr/hdp/current/hadoop-client/conf start namenode''] {'environment': {'HADOOP_LIBEXEC_DIR': '/usr/hdp/current/hadoop-client/libexec'}, 'not_if': 'ambari-sudo.sh -H -E test -f /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid && ambari-sudo.sh -H -E pgrep -F /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid'}
  166. 2016-02-07 07:36:19,354 - Must wait to leave safemode since High Availability is not enabled.
  167. 2016-02-07 07:36:19,354 - Stay in safe mode: False
  168. 2016-02-07 07:36:19,354 - Wait to leafe safemode since must transition from ON to OFF.
  169. 2016-02-07 07:36:19,355 - Execute['/usr/hdp/current/hadoop-hdfs-namenode/bin/hdfs dfsadmin -fs hdfs://phdns01.cloud.hortonworks.com:8020 -safemode get | grep 'Safe mode is OFF''] {'logoutput': True, 'tries': 180, 'user': 'hdfs', 'try_sleep': 10}
  170. safemode: NameNode still not started
  171. 2016-02-07 07:36:23,943 - Retrying after 10 seconds. Reason: Execution of '/usr/hdp/current/hadoop-hdfs-namenode/bin/hdfs dfsadmin -fs hdfs://phdns01.cloud.hortonworks.com:8020 -safemode get | grep 'Safe mode is OFF'' returned 1. safemode: NameNode still not started
  172. 2016-02-07 07:36:36,959 - Retrying after 10 seconds. Reason: Execution of '/usr/hdp/current/hadoop-hdfs-namenode/bin/hdfs dfsadmin -fs hdfs://phdns01.cloud.hortonworks.com:8020 -safemode get | grep 'Safe mode is OFF'' returned 1.
  173. 2016-02-07 07:36:49,939 - Retrying after 10 seconds. Reason: Execution of '/usr/hdp/current/hadoop-hdfs-namenode/bin/hdfs dfsadmin -fs hdfs://phdns01.cloud.hortonworks.com:8020 -safemode get | grep 'Safe mode is OFF'' returned 1.
  174. 2016-02-07 07:37:03,217 - Retrying after 10 seconds. Reason: Execution of '/usr/hdp/current/hadoop-hdfs-namenode/bin/hdfs dfsadmin -fs hdfs://phdns01.cloud.hortonworks.com:8020 -safemode get | grep 'Safe mode is OFF'' returned 1.
  175. 2016-02-07 07:37:16,129 - Retrying after 10 seconds. Reason: Execution of '/usr/hdp/current/hadoop-hdfs-namenode/bin/hdfs dfsadmin -fs hdfs://phdns01.cloud.hortonworks.com:8020 -safemode get | grep 'Safe mode is OFF'' returned 1.
  176. 2016-02-07 07:37:29,084 - Retrying after 10 seconds. Reason: Execution of '/usr/hdp/current/hadoop-hdfs-namenode/bin/hdfs dfsadmin -fs hdfs://phdns01.cloud.hortonworks.com:8020 -safemode get | grep 'Safe mode is OFF'' returned 1.
  177. 2016-02-07 07:37:42,205 - Retrying after 10 seconds. Reason: Execution of '/usr/hdp/current/hadoop-hdfs-namenode/bin/hdfs dfsadmin -fs hdfs://phdns01.cloud.hortonworks.com:8020 -safemode get | grep 'Safe mode is OFF'' returned 1.
  178. 2016-02-07 07:37:55,234 - Retrying after 10 seconds. Reason: Execution of '/usr/hdp/current/hadoop-hdfs-namenode/bin/hdfs dfsadmin -fs hdfs://phdns01.cloud.hortonworks.com:8020 -safemode get | grep 'Safe mode is OFF'' returned 1.
  179. 2016-02-07 07:38:08,674 - Retrying after 10 seconds. Reason: Execution of '/usr/hdp/current/hadoop-hdfs-namenode/bin/hdfs dfsadmin -fs hdfs://phdns01.cloud.hortonworks.com:8020 -safemode get | grep 'Safe mode is OFF'' returned 1.
  180. 2016-02-07 07:38:21,634 - Retrying after 10 seconds. Reason: Execution of '/usr/hdp/current/hadoop-hdfs-namenode/bin/hdfs dfsadmin -fs hdfs://phdns01.cloud.hortonworks.com:8020 -safemode get | grep 'Safe mode is OFF'' returned 1.
  181. Safe mode is OFF
  182. 2016-02-07 07:38:34,713 - HdfsResource['/tmp'] {'security_enabled': False, 'only_if': True, 'keytab': [EMPTY], 'hadoop_bin_dir': '/usr/hdp/current/hadoop-client/bin', 'default_fs': 'hdfs://phdns01.cloud.hortonworks.com:8020', 'hdfs_site': ..., 'kinit_path_local': '/usr/bin/kinit', 'principal_name': None, 'user': 'hdfs', 'owner': 'hdfs', 'hadoop_conf_dir': '/usr/hdp/current/hadoop-client/conf', 'type': 'directory', 'action': ['create_on_execute'], 'mode': 0777}
  183. 2016-02-07 07:38:34,718 - call['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET '"'"'http://phdns01.cloud.hortonworks.com:50070/webhdfs/v1/tmp?op=GETFILESTATUS&user.name=hdfs'"'"' 1>/tmp/tmpDlGsyc 2>/tmp/tmpTFZ2fO''] {'logoutput': None, 'quiet': False}
  184. 2016-02-07 07:38:36,604 - call returned (0, '')
  185. 2016-02-07 07:38:36,606 - HdfsResource['/user/ambari-qa'] {'security_enabled': False, 'only_if': True, 'keytab': [EMPTY], 'hadoop_bin_dir': '/usr/hdp/current/hadoop-client/bin', 'default_fs': 'hdfs://phdns01.cloud.hortonworks.com:8020', 'hdfs_site': ..., 'kinit_path_local': '/usr/bin/kinit', 'principal_name': None, 'user': 'hdfs', 'owner': 'ambari-qa', 'hadoop_conf_dir': '/usr/hdp/current/hadoop-client/conf', 'type': 'directory', 'action': ['create_on_execute'], 'mode': 0770}
  186. 2016-02-07 07:38:36,608 - call['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET '"'"'http://phdns01.cloud.hortonworks.com:50070/webhdfs/v1/user/ambari-qa?op=GETFILESTATUS&user.name=hdfs'"'"' 1>/tmp/tmp3ACj8X 2>/tmp/tmpoYVXsr''] {'logoutput': None, 'quiet': False}
  187. 2016-02-07 07:38:36,669 - call returned (0, '')
  188. 2016-02-07 07:38:36,733 - HdfsResource[None] {'security_enabled': False, 'only_if': True, 'keytab': [EMPTY], 'hadoop_bin_dir': '/usr/hdp/current/hadoop-client/bin', 'default_fs': 'hdfs://phdns01.cloud.hortonworks.com:8020', 'hdfs_site': ..., 'kinit_path_local': '/usr/bin/kinit', 'principal_name': None, 'user': 'hdfs', 'action': ['execute'], 'hadoop_conf_dir': '/usr/hdp/current/hadoop-client/conf'}
  189. 2016-02-07 07:38:36,734 - HdfsResource['/ranger/audit'] {'security_enabled': False, 'only_if': True, 'keytab': [EMPTY], 'hadoop_bin_dir': '/usr/hdp/current/hadoop-client/bin', 'default_fs': 'hdfs://phdns01.cloud.hortonworks.com:8020', 'user': 'hdfs', 'hdfs_site': ..., 'kinit_path_local': '/usr/bin/kinit', 'principal_name': None, 'recursive_chmod': True, 'owner': 'hdfs', 'group': 'hdfs', 'hadoop_conf_dir': '/usr/hdp/current/hadoop-client/conf', 'type': 'directory', 'action': ['create_on_execute'], 'mode': 0755}
  190. 2016-02-07 07:38:36,786 - call['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET '"'"'http://phdns01.cloud.hortonworks.com:50070/webhdfs/v1/ranger/audit?op=GETFILESTATUS&user.name=hdfs'"'"' 1>/tmp/tmpJSPrSM 2>/tmp/tmpQTwIwx''] {'logoutput': None, 'quiet': False}
  191. 2016-02-07 07:38:36,851 - call returned (0, '')
  192. 2016-02-07 07:38:36,853 - HdfsResource['/ranger/audit/hdfs'] {'security_enabled': False, 'only_if': True, 'keytab': [EMPTY], 'hadoop_bin_dir': '/usr/hdp/current/hadoop-client/bin', 'default_fs': 'hdfs://phdns01.cloud.hortonworks.com:8020', 'user': 'hdfs', 'hdfs_site': ..., 'kinit_path_local': '/usr/bin/kinit', 'principal_name': None, 'recursive_chmod': True, 'owner': 'hdfs', 'group': 'hdfs', 'hadoop_conf_dir': '/usr/hdp/current/hadoop-client/conf', 'type': 'directory', 'action': ['create_on_execute'], 'mode': 0700}
  193. 2016-02-07 07:38:36,855 - call['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET '"'"'http://phdns01.cloud.hortonworks.com:50070/webhdfs/v1/ranger/audit/hdfs?op=GETFILESTATUS&user.name=hdfs'"'"' 1>/tmp/tmpfyMTg5 2>/tmp/tmpaBXx99''] {'logoutput': None, 'quiet': False}
  194. 2016-02-07 07:38:36,922 - call returned (0, '')
  195. 2016-02-07 07:38:36,924 - HdfsResource[None] {'security_enabled': False, 'only_if': True, 'keytab': [EMPTY], 'hadoop_bin_dir': '/usr/hdp/current/hadoop-client/bin', 'default_fs': 'hdfs://phdns01.cloud.hortonworks.com:8020', 'hdfs_site': ..., 'kinit_path_local': '/usr/bin/kinit', 'principal_name': None, 'user': 'hdfs', 'action': ['execute'], 'hadoop_conf_dir': '/usr/hdp/current/hadoop-client/conf'}
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement