Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

mrjob cannot be run in company with pymongo on hadoop #913

Closed
whzhcahzxh opened this issue Jun 9, 2014 · 5 comments
Closed

mrjob cannot be run in company with pymongo on hadoop #913

whzhcahzxh opened this issue Jun 9, 2014 · 5 comments
Labels

Comments

@whzhcahzxh
Copy link

This is hard work. I write a map-reduce script runs well on my local pc. When i try to run it on hadoop (by -r hadoop), it gets error like:
no configs found; falling back on auto-configuration
no configs found; falling back on auto-configuration
creating tmp directory /tmp/step1.root.20140606.091711.815391
writing wrapper script to /tmp/step1.root.20140606.091711.815391/setup-wrapper.sh
reading from STDIN
Copying local files into hdfs:///user/root/tmp/mrjob/step1.root.20140606.091711.815391/files/
Using Hadoop version 2.0.0
HADOOP: packageJobJar: [] [/opt/cloudera/parcels/CDH-4.6.0-1.cdh4.6.0.p0.26/lib/hadoop/hadoop-streaming.jar] /tmp/streamjob8615643898520402804.jar tmpDir=null
HADOOP: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.
HADOOP: Total input paths to process : 1
HADOOP: getLocalDirs(): [/tmp/hadoop-root/mapred/local]
HADOOP: Running job: job_201405161502_0059
HADOOP: To kill this job, run:
HADOOP: /opt/cloudera/parcels/CDH-4.6.0-1.cdh4.6.0.p0.26/lib/hadoop/bin/hadoop job -Dmapred.job.tracker=v-lab-110:8021 -kill job_201405161502_0059
HADOOP: Tracking URL: http://v-lab-110:50030/jobdetails.jsp?jobid=job_201405161502_0059
HADOOP: map 0% reduce 0%
HADOOP: map 100% reduce 100%
HADOOP: To kill this job, run:
HADOOP: /opt/cloudera/parcels/CDH-4.6.0-1.cdh4.6.0.p0.26/lib/hadoop/bin/hadoop job -Dmapred.job.tracker=v-lab-110:8021 -kill job_201405161502_0059
HADOOP: Tracking URL: http://v-lab-110:50030/jobdetails.jsp?jobid=job_201405161502_0059
HADOOP: Job not successful. Error: NA
HADOOP: killJob...
HADOOP: Streaming Command Failed!
Job failed with return code 256: ['/opt/cloudera/parcels/CDH-4.6.0-1.cdh4.6.0.p0.26/lib/hadoop/bin/hadoop', 'jar', '/opt/cloudera/parcels/CDH-4.6.0-1.cdh4.6.0.p0.26/lib/hadoop/hadoop-streaming.jar', '-files', 'hdfs:///user/root/tmp/mrjob/step1.root.20140606.091711.815391/files/setup-wrapper.sh#setup-wrapper.sh,hdfs:///user/root/tmp/mrjob/step1.root.20140606.091711.815391/files/step1.py#step1.py', '-archives', 'hdfs:///user/root/tmp/mrjob/step1.root.20140606.091711.815391/files/mrjob.tar.gz#mrjob.tar.gz', '-input', 'hdfs:///user/root/tmp/mrjob/step1.root.20140606.091711.815391/files/STDIN', '-output', 'hdfs:///user/root/tmp/mrjob/step1.root.20140606.091711.815391/step-output/1', '-mapper', 'sh -e setup-wrapper.sh python step1.py --step-num=0 --mapper', '-combiner', 'sh -e setup-wrapper.sh python step1.py --step-num=0 --combiner', '-reducer', 'sh -e setup-wrapper.sh python step1.py --step-num=0 --reducer']
Scanning logs for probable cause of failure
Traceback (most recent call last):
File "step1.py", line 176, in
step.run()
File "/usr/local/lib/python2.7/site-packages/mrjob/job.py", line 494, in run
mr_job.execute()
File "/usr/local/lib/python2.7/site-packages/mrjob/job.py", line 512, in execute
super(MRJob, self).execute()
File "/usr/local/lib/python2.7/site-packages/mrjob/launch.py", line 147, in execute
self.run_job()
File "/usr/local/lib/python2.7/site-packages/mrjob/launch.py", line 208, in run_job
runner.run()
File "/usr/local/lib/python2.7/site-packages/mrjob/runner.py", line 458, in run
self._run()
File "/usr/local/lib/python2.7/site-packages/mrjob/hadoop.py", line 239, in _run
self._run_job_in_hadoop()
File "/usr/local/lib/python2.7/site-packages/mrjob/hadoop.py", line 358, in _run_job_in_hadoop
raise CalledProcessError(returncode, step_args)

subprocess.CalledProcessError: Command '['/opt/cloudera/parcels/CDH-4.6.0-1.cdh4.6.0.p0.26/lib/hadoop/bin/hadoop', 'jar', '/opt/cloudera/parcels/CDH-4.6.0-1.cdh4.6.0.p0.26/lib/hadoop/hadoop-streaming.jar', '-files', 'hdfs:///user/root/tmp/mrjob/step1.root.20140606.091711.815391/files/setup-wrapper.sh#setup-wrapper.sh,hdfs:///user/root/tmp/mrjob/step1.root.20140606.091711.815391/files/step1.py#step1.py', '-archives', 'hdfs:///user/root/tmp/mrjob/step1.root.20140606.091711.815391/files/mrjob.tar.gz#mrjob.tar.gz', '-input', 'hdfs:///user/root/tmp/mrjob/step1.root.20140606.091711.815391/files/STDIN', '-output', 'hdfs:///user/root/tmp/mrjob/step1.root.20140606.091711.815391/step-output/1', '-mapper', 'sh -e setup-wrapper.sh python step1.py --step-num=0 --mapper', '-combiner', 'sh -e setup-wrapper.sh python step1.py --step-num=0 --combiner', '-reducer', 'sh -e setup-wrapper.sh python step1.py --step-num=0 --reducer']' returned non-zero exit status 256

I tried every situation and found out that if i import pymongo in step1.py, it gets this error. I think it is a bug.

@hophacker
Copy link

How did you solve this issue, since I encountered the same problem.

@tarnfeld
Copy link
Contributor

tarnfeld commented Aug 8, 2014

You'll see in the HADOOP: output from your logs, a link to the jobtracker -> http://v-lab-110:50030/jobdetails.jsp?jobid=job_201405161502_0059

That's the web page for your job, if you go to that link you'll be able to see the status of your job, as well as which map/reduce tasks are failing. If you click the little number (under the Failed/Killed section) you'll be able to see the STDOUT/STDERR logs for each individual task, and also the python exception that's causing your job to fail.

Hope this helps.

@autumnjolitz
Copy link
Contributor

Without the stderr/stdout/syslog, it is not possible to determine cause of failure.

@whzhcahzxh -- did you install pip and pip install pymongo (and any yum dependencies) in your bootstrap?

@coyotemarin
Copy link
Collaborator

Thanks for letting me know, but I can't proceed without more information, so closing this ticket for now.

@iarroyof
Copy link

Hi all... I have the same problem:

raise CalledProcessError(proc.returncode, args) subprocess.CalledProcessError: Command '['/usr/local/hadoop/bin/hadoop', 'fs', '-mkdir', '-p', 'hdfs:///user/hduser/tmp/mrjob/wordcount.hduser.20151210.202210.421677/files/']' returned non-zero exit status 1

my log is:

python wordcount.py -r hadoop wikiEn.txt > wcEn.txt
no configs found; falling back on auto-configuration
no configs found; falling back on auto-configuration
creating tmp directory /tmp/wordcount.hduser.20151210.204707.587853
writing wrapper script to /tmp/wordcount.hduser.20151210.204707.587853/setup-wrapper.sh
Using Hadoop version 2.7.1
Copying local files into hdfs:///user/hduser/tmp/mrjob/wordcount.hduser.20151210.204707.587853/files/

PLEASE NOTE: Starting in mrjob v0.5.0, protocols will be strict by default. It's recommended you run your job with --strict-protocols or set up mrjob.conf as described at https://pythonhosted.org/mrjob/whats-new.html#ready-for-strict-protocols

HADOOP: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
HADOOP: packageJobJar: [/tmp/hadoop-unjar6688635533667031039/] [] /tmp/streamjob2800164258139328597.jar tmpDir=null
HADOOP: Connecting to ResourceManager at /0.0.0.0:8032
HADOOP: Connecting to ResourceManager at /0.0.0.0:8032
HADOOP: Total input paths to process : 1
HADOOP: number of splits:18
HADOOP: Submitting tokens for job: job_1449780411314_0001
HADOOP: Submitted application application_1449780411314_0001
HADOOP: The url to track the job: http://elc13x5000a040w:8088/proxy/application_1449780411314_0001/
HADOOP: Running job: job_1449780411314_0001
HADOOP: Job job_1449780411314_0001 running in uber mode : false
HADOOP:  map 0% reduce 0%
HADOOP:  map 1% reduce 0%
HADOOP:  map 6% reduce 0%
HADOOP: Task Id : attempt_1449780411314_0001_m_000005_0, Status : FAILED
HADOOP: Error: java.lang.RuntimeException: PipeMapRed.waitOutputThreads(): subprocess failed with code 1
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.waitOutputThreads(PipeMapRed.java:322)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.mapRedFinished(PipeMapRed.java:535)
HADOOP:     at org.apache.hadoop.streaming.PipeMapper.close(PipeMapper.java:130)
HADOOP:     at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:61)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRunner.run(PipeMapRunner.java:34)
HADOOP:     at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453)
HADOOP:     at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
HADOOP:     at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
HADOOP:     at java.security.AccessController.doPrivileged(Native Method)
HADOOP:     at javax.security.auth.Subject.doAs(Subject.java:422)
HADOOP:     at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
HADOOP:     at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
HADOOP: 
HADOOP:  map 1% reduce 0%
HADOOP:  map 2% reduce 0%
HADOOP:  map 3% reduce 0%
HADOOP: Task Id : attempt_1449780411314_0001_m_000004_0, Status : FAILED
HADOOP: Error: java.lang.RuntimeException: PipeMapRed.waitOutputThreads(): subprocess failed with code 1
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.waitOutputThreads(PipeMapRed.java:322)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.mapRedFinished(PipeMapRed.java:535)
HADOOP:     at org.apache.hadoop.streaming.PipeMapper.close(PipeMapper.java:130)
HADOOP:     at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:61)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRunner.run(PipeMapRunner.java:34)
HADOOP:     at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453)
HADOOP:     at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
HADOOP:     at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
HADOOP:     at java.security.AccessController.doPrivileged(Native Method)
HADOOP:     at javax.security.auth.Subject.doAs(Subject.java:422)
HADOOP:     at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
HADOOP:     at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
HADOOP: 
HADOOP:  map 4% reduce 0%
HADOOP: Task Id : attempt_1449780411314_0001_m_000005_1, Status : FAILED
HADOOP: Error: java.lang.RuntimeException: PipeMapRed.waitOutputThreads(): subprocess failed with code 1
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.waitOutputThreads(PipeMapRed.java:322)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.mapRedFinished(PipeMapRed.java:535)
HADOOP:     at org.apache.hadoop.streaming.PipeMapper.close(PipeMapper.java:130)
HADOOP:     at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:61)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRunner.run(PipeMapRunner.java:34)
HADOOP:     at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453)
HADOOP:     at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
HADOOP:     at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
HADOOP:     at java.security.AccessController.doPrivileged(Native Method)
HADOOP:     at javax.security.auth.Subject.doAs(Subject.java:422)
HADOOP:     at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
HADOOP:     at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
HADOOP: 
HADOOP:  map 5% reduce 0%
HADOOP:  map 6% reduce 0%
HADOOP:  map 7% reduce 0%
HADOOP: Task Id : attempt_1449780411314_0001_m_000000_0, Status : FAILED
HADOOP: Error: java.lang.RuntimeException: PipeMapRed.waitOutputThreads(): subprocess failed with code 1
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.waitOutputThreads(PipeMapRed.java:322)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.mapRedFinished(PipeMapRed.java:535)
HADOOP:     at org.apache.hadoop.streaming.PipeMapper.close(PipeMapper.java:130)
HADOOP:     at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:61)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRunner.run(PipeMapRunner.java:34)
HADOOP:     at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453)
HADOOP:     at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
HADOOP:     at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
HADOOP:     at java.security.AccessController.doPrivileged(Native Method)
HADOOP:     at javax.security.auth.Subject.doAs(Subject.java:422)
HADOOP:     at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
HADOOP:     at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
HADOOP: 
HADOOP:  map 5% reduce 0%
HADOOP: Task Id : attempt_1449780411314_0001_m_000004_1, Status : FAILED
HADOOP: Error: java.lang.RuntimeException: PipeMapRed.waitOutputThreads(): subprocess failed with code 1
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.waitOutputThreads(PipeMapRed.java:322)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.mapRedFinished(PipeMapRed.java:535)
HADOOP:     at org.apache.hadoop.streaming.PipeMapper.close(PipeMapper.java:130)
HADOOP:     at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:61)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRunner.run(PipeMapRunner.java:34)
HADOOP:     at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453)
HADOOP:     at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
HADOOP:     at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
HADOOP:     at java.security.AccessController.doPrivileged(Native Method)
HADOOP:     at javax.security.auth.Subject.doAs(Subject.java:422)
HADOOP:     at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
HADOOP:     at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
HADOOP: 
HADOOP:  map 6% reduce 0%
HADOOP: Task Id : attempt_1449780411314_0001_m_000005_2, Status : FAILED
HADOOP: Error: java.lang.RuntimeException: PipeMapRed.waitOutputThreads(): subprocess failed with code 1
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.waitOutputThreads(PipeMapRed.java:322)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.mapRedFinished(PipeMapRed.java:535)
HADOOP:     at org.apache.hadoop.streaming.PipeMapper.close(PipeMapper.java:130)
HADOOP:     at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:61)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRunner.run(PipeMapRunner.java:34)
HADOOP:     at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453)
HADOOP:     at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
HADOOP:     at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
HADOOP:     at java.security.AccessController.doPrivileged(Native Method)
HADOOP:     at javax.security.auth.Subject.doAs(Subject.java:422)
HADOOP:     at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
HADOOP:     at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
HADOOP: 
HADOOP:  map 7% reduce 0%
HADOOP: Task Id : attempt_1449780411314_0001_m_000003_0, Status : FAILED
HADOOP: Error: java.lang.RuntimeException: PipeMapRed.waitOutputThreads(): subprocess failed with code 1
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.waitOutputThreads(PipeMapRed.java:322)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.mapRedFinished(PipeMapRed.java:535)
HADOOP:     at org.apache.hadoop.streaming.PipeMapper.close(PipeMapper.java:130)
HADOOP:     at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:61)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRunner.run(PipeMapRunner.java:34)
HADOOP:     at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453)
HADOOP:     at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
HADOOP:     at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
HADOOP:     at java.security.AccessController.doPrivileged(Native Method)
HADOOP:     at javax.security.auth.Subject.doAs(Subject.java:422)
HADOOP:     at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
HADOOP:     at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
HADOOP: 
HADOOP:  map 5% reduce 0%
HADOOP: Task Id : attempt_1449780411314_0001_m_000002_0, Status : FAILED
HADOOP: Error: java.lang.RuntimeException: PipeMapRed.waitOutputThreads(): subprocess failed with code 1
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.waitOutputThreads(PipeMapRed.java:322)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRed.mapRedFinished(PipeMapRed.java:535)
HADOOP:     at org.apache.hadoop.streaming.PipeMapper.close(PipeMapper.java:130)
HADOOP:     at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:61)
HADOOP:     at org.apache.hadoop.streaming.PipeMapRunner.run(PipeMapRunner.java:34)
HADOOP:     at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453)
HADOOP:     at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
HADOOP:     at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
HADOOP:     at java.security.AccessController.doPrivileged(Native Method)
HADOOP:     at javax.security.auth.Subject.doAs(Subject.java:422)
HADOOP:     at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
HADOOP:     at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
HADOOP: 
HADOOP:  map 3% reduce 0%
HADOOP:  map 4% reduce 0%
HADOOP:  map 100% reduce 100%
HADOOP: Job job_1449780411314_0001 failed with state FAILED due to: Task failed task_1449780411314_0001_m_000005
HADOOP: Job failed as tasks failed. failedMaps:1 failedReduces:0
HADOOP: 
HADOOP: Counters: 10
HADOOP:     Job Counters 
HADOOP:         Failed map tasks=9
HADOOP:         Killed map tasks=5
HADOOP:         Launched map tasks=14
HADOOP:         Other local map tasks=7
HADOOP:         Data-local map tasks=7
HADOOP:         Total time spent by all maps in occupied slots (ms)=1602014
HADOOP:         Total time spent by all reduces in occupied slots (ms)=0
HADOOP:         Total time spent by all map tasks (ms)=1602014
HADOOP:         Total vcore-seconds taken by all map tasks=1602014
HADOOP:         Total megabyte-seconds taken by all map tasks=1640462336
HADOOP: Job not successful!
HADOOP: Streaming Command Failed!
Job failed with return code 256: ['/usr/local/hadoop/bin/hadoop', 'jar', '/usr/local/hadoop/share/hadoop/tools/lib/hadoop-streaming-2.7.1.jar', '-files', 'hdfs:///user/hduser/tmp/mrjob/wordcount.hduser.20151210.204707.587853/files/wordcount.py#wordcount.py,hdfs:///user/hduser/tmp/mrjob/wordcount.hduser.20151210.204707.587853/files/setup-wrapper.sh#setup-wrapper.sh', '-archives', 'hdfs:///user/hduser/tmp/mrjob/wordcount.hduser.20151210.204707.587853/files/mrjob.tar.gz#mrjob.tar.gz', '-input', 'hdfs:///user/hduser/tmp/mrjob/wordcount.hduser.20151210.204707.587853/files/wikiEn_cadenas.txt', '-output', 'hdfs:///user/hduser/tmp/mrjob/wordcount.hduser.20151210.204707.587853/output', '-mapper', 'sh -ex setup-wrapper.sh python wordcount.py --step-num=0 --mapper', '-reducer', 'sh -ex setup-wrapper.sh python wordcount.py --step-num=0 --reducer']
Scanning logs for probable cause of failure
Traceback (most recent call last):
  File "wordcount.py", line 59, in <module>
    wc.run()
  File "/usr/local/lib/python2.7/dist-packages/mrjob/job.py", line 461, in run
    mr_job.execute()
  File "/usr/local/lib/python2.7/dist-packages/mrjob/job.py", line 479, in execute
    super(MRJob, self).execute()
  File "/usr/local/lib/python2.7/dist-packages/mrjob/launch.py", line 153, in execute
    self.run_job()
  File "/usr/local/lib/python2.7/dist-packages/mrjob/launch.py", line 216, in run_job
    runner.run()
  File "/usr/local/lib/python2.7/dist-packages/mrjob/runner.py", line 470, in run
    self._run()
  File "/usr/local/lib/python2.7/dist-packages/mrjob/hadoop.py", line 237, in _run
    self._run_job_in_hadoop()
  File "/usr/local/lib/python2.7/dist-packages/mrjob/hadoop.py", line 372, in _run_job_in_hadoop
    raise CalledProcessError(returncode, step_args)
subprocess.CalledProcessError: Command '['/usr/local/hadoop/bin/hadoop', 'jar', '/usr/local/hadoop/share/hadoop/tools/lib/hadoop-streaming-2.7.1.jar', '-files', 'hdfs:///user/hduser/tmp/mrjob/wordcount.hduser.20151210.204707.587853/files/wordcount.py#wordcount.py,hdfs:///user/hduser/tmp/mrjob/wordcount.hduser.20151210.204707.587853/files/setup-wrapper.sh#setup-wrapper.sh', '-archives', 'hdfs:///user/hduser/tmp/mrjob/wordcount.hduser.20151210.204707.587853/files/mrjob.tar.gz#mrjob.tar.gz', '-input', 'hdfs:///user/hduser/tmp/mrjob/wordcount.hduser.20151210.204707.587853/files/wikiEn_cadenas.txt', '-output', 'hdfs:///user/hduser/tmp/mrjob/wordcount.hduser.20151210.204707.587853/output', '-mapper', 'sh -ex setup-wrapper.sh python wordcount.py --step-num=0 --mapper', '-reducer', 'sh -ex setup-wrapper.sh python wordcount.py --step-num=0 --reducer']' returned non-zero exit status 256

So I don't know what is wrong.

Thank you very much for helping.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

6 participants