Advertisement
zamotivator

Untitled

Sep 24th, 2014
309
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 3.98 KB | None | 0 0
  1. o.tsarev 1327 0.0 0.0 125796 1560 ? Ss 14:49 0:00 SCREEN
  2. o.tsarev 1328 0.0 0.0 106504 1968 pts/18 Ss 14:49 0:00 \_ /bin/bash
  3. o.tsarev 4210 0.0 0.0 106504 828 pts/18 S+ 14:50 0:00 \_ /bin/bash
  4. o.tsarev 4212 0.0 0.0 190912 11264 pts/18 S+ 14:50 0:00 | \_ python2.7 ./start.py --worker_count=8
  5. o.tsarev 4213 0.0 0.0 190968 9024 pts/18 S+ 14:50 0:00 | \_ python /usr/local/sbin/hd.py hdtask.py
  6. o.tsarev 4370 0.0 0.0 106272 1348 pts/18 S+ 14:50 0:00 | \_ /bin/sh -c yarn jar /usr/lib/hadoop-mapreduce/hadoop-streaming.jar -D mapreduce.task.timeout="-1000" -D mapreduce.map.memory.mb="128" -D mapred.max.split.size="268435456" -D mapreduce.job.reduces="1024" -D mapreduce.reduce.memory.mb="1024" -D mapred.reduce.slowstart.completed.maps="0.90" -D mapreduce.map.maxattempts="16" -D mapreduce.reduce.java.opts="-Xmx256m" -D mapred.job.queue.name="o.tsarev" -D typedinput.compressed="true" -D stream.map.input="typedbytes" -D mapreduce.map.java.opts="-Xmx256m" -D mapreduce.job.name="XXX_current_2014-09-24-14" -libjars "reader.jar" -input "/data/YYY/ZZZ_XXX/*.pb*" -output "/user/o.tsarev/XXX_current" -file "/home/o.tsarev/WWW/hdtask.py" -file "/usr/local/sbin/hd.py" -file "/usr/local/YYY/bin/READER" -file "WWW.cfg" -file "WWW.dump" -file "common.py" -file "config.py" -file "convert.py" -file "delta.py" -file "YYY.py" -file "XXX.py" -file "hdfs.py" -file "start.py" -file "test_common.py" -file "test_delta.py" -file "test_YYY.py" -file "test_XXX.py" -file "WWW_pb2.py" -file "reader.jar" -inputformat "reader.OVCFInputFormat" -reducer "python hd.py --reducer hdtask.py" -mapper "python hd.py --mapper hdtask.py" 2>&1 | /bin/sed "s/hp3.hadoop.domain.name:8088/hadoop.hadoop.domain.name/g"
  7. o.tsarev 4371 25.8 1.0 2889512 713960 pts/18 Sl+ 14:50 0:31 | \_ /usr/lib/jvm/jre/bin/java -Dproc_jar -Xmx2048m -Dhadoop.log.dir=/usr/lib/hadoop-yarn/logs -Dyarn.log.dir=/usr/lib/hadoop-yarn/logs -Dhadoop.log.file=yarn.log -Dyarn.log.file=yarn.log -Dyarn.home.dir= -Dyarn.id.str= -Dhadoop.root.logger=INFO,console -Dyarn.root.logger=INFO,console -Djava.library.path=/usr/lib/hadoop/lib/native -Dyarn.policy.file=hadoop-policy.xml -Dhadoop.log.dir=/usr/lib/hadoop-yarn/logs -Dyarn.log.dir=/usr/lib/hadoop-yarn/logs -Dhadoop.log.file=yarn.log -Dyarn.log.file=yarn.log -Dyarn.home.dir=/usr/lib/hadoop-yarn -Dhadoop.home.dir=/usr/lib/hadoop-yarn -Dhadoop.root.logger=INFO,console -Dyarn.root.logger=INFO,console -Djava.library.path=/usr/lib/hadoop/lib/native -classpath /etc/hadoop/conf:/etc/hadoop/conf:/etc/hadoop/conf:/usr/lib/hadoop/lib/*:/usr/lib/hadoop/.//*:/usr/lib/hadoop-hdfs/./:/usr/lib/hadoop-hdfs/lib/*:/usr/lib/hadoop-hdfs/.//*:/usr/lib/hadoop-yarn/lib/*:/usr/lib/hadoop-yarn/.//*:/usr/lib/hadoop-mapreduce/lib/*:/usr/lib/hadoop-mapreduce/.//*:/usr/lib/hadoop-yarn/.//*:/usr/lib/hadoop-yarn/lib/* org.apache.hadoop.util.RunJar /usr/lib/hadoop-mapreduce/hadoop-streaming.jar -D mapreduce.task.timeout=-1000 -D mapreduce.map.memory.mb=128 -D mapred.max.split.size=268435456 -D mapreduce.job.reduces=1024 -D mapreduce.reduce.memory.mb=1024 -D mapred.reduce.slowstart.completed.maps=0.90 -D mapreduce.map.maxattempts=16 -D mapreduce.reduce.java.opts=-Xmx256m -D mapred.job.queue.name=o.tsarev -D typedinput.compressed=true -D stream.map.input=typedbytes -D mapreduce.map.java.opts=-Xmx256m -D mapreduce.job.name=XXX_current_2014-09-24-14 -libjars reader.jar -input /data/YYY/ZZZ_XXX/*.pb* -output /user/o.tsarev/XXX_current -file /home/o.tsarev/WWW/hdtask.py -file /usr/local/sbin/hd.py -file /usr/local/YYY/bin/READER -file WWW.cfg -file WWW.dump -file common.py -file config.py -file convert.py -file delta.py -file YYY.py -file XXX.py -file hdfs.py -file start.py -file test_common.py -file test_delta.py -file test_YYY.py -file test_XXX.py -file WWW_pb2.py -file reader.jar -inputformat reader.OVCFInputFormat -reducer python hd.py --reducer hdtask.py -mapper python hd.py --mapper hdtask.py
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement