Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

pydoop mapreduce job hangs because pydoop is not supported by Cloudera Hadoop 2.6 CDH5 #216

Closed
igorvishnevskiy opened this issue Jun 17, 2016 · 21 comments
Labels

Comments

@igorvishnevskiy
Copy link

igorvishnevskiy commented Jun 17, 2016

Hello guys. I have already tried all examples possible. I tried to set job to pick up files from local name node and from HDFS. The same problem. Can not find any online reference to the same behavior. What am I doing wrong? I am trying to run a simple word count example and that won't run. It hangs as soon as mapper looks like trying to read from the file. When I would specify wrong path, job does fail. Therefore I think the problem is on pydoop reading from file. I tried similar mapreduce on mrjob and it worked fine, but I want to use pydoop. What am I missing possibly? I am on the Hadoop 2.6 CDH5. I tried all examples I could find online. The same issue with all of them. Thank you.

Code of the Runner:

import sys, os, argparse, logging
logging.basicConfig(level=logging.INFO)

import pydoop
import pydoop.hadut as hadut
import pydoop.test_support as pts


CONF = {
  "mapred.map.tasks": "2",
  "mapred.reduce.tasks": "2",
  "mapred.job.name": "igor_test_pydoop",
  }
HADOOP_CONF_DIR = pydoop.hadoop_conf()
PREFIX = os.getenv("PREFIX", pts.get_wd_prefix())


def update_conf(args):
  if args.D:
    for kv_pair in args.D:
      k, v = [_.strip() for _ in kv_pair.split("=")]
      CONF[k] = v


def make_parser():
  parser = argparse.ArgumentParser()
  parser.add_argument("pipes_exe", metavar="PIPES_EXE",
                      help="python script to be run by pipes")
  parser.add_argument("local_input", metavar="INPUT_DIR",
                      help="local input directory")
  parser.add_argument("-D", metavar="NAME=VALUE", action="append",
                      help="additional Hadoop configuration parameters")
  return parser


def main(argv):
  parser = make_parser()
  args = parser.parse_args(argv)
  update_conf(args)
  logger = logging.getLogger("main")
  logger.setLevel(logging.DEBUG)
  runner = hadut.PipesRunner(prefix=PREFIX, logger=logger)
  with open(args.pipes_exe) as f:
    pipes_code = pts.add_sys_path(f.read())
  runner.set_input(args.local_input, put=False)
  runner.set_exe(pipes_code)
  runner.run(properties=CONF, hadoop_conf_dir=HADOOP_CONF_DIR, logger=logger)
  res = runner.collect_output()
  runner.clean()
  local_wc = pts.LocalWordCount(args.local_input)
  logging.info(local_wc.check(res))


if __name__ == "__main__":
  main(sys.argv[1:])

Code of the mapreduce example:

import pydoop.pipes as pp


class Mapper(pp.Mapper):

  def map(self, context):
    words = context.getInputValue().split()
    for w in words:
      context.emit(w, "1")


class Reducer(pp.Reducer):

  def reduce(self, context):
    s = 0
    while context.nextValue():
      s += int(context.getInputValue())
    context.emit(context.getInputKey(), str(s))


if __name__ == "__main__":
  pp.runTask(pp.Factory(Mapper, Reducer))

These are logs from console (more syslog logs are at the very bottom):

[hdfs@server01 bin]$ python pydoop_runner.py wc.py hdfs://domain/user/data/test.txt
INFO:main:assigning input to hdfs://domain/user/data/test.txt
INFO:main:running MapReduce application
DEBUG:main:final args: ['/usr/bin/hadoop', '--config', '/etc/hadoop/conf', 'pipes', '-D', 'hadoop.pipes.java.recordwriter=true', '-D', 'mapred.map.tasks=2', '-D', 'mapred.job.name=igor_test_pydoop', '-D', 'hadoop.pipes.java.recordreader=true', '-D', 'mapred.reduce.tasks=2', '-program', 'hdfs://domain/user/hdfs/pydoop_f055a031a2b944069ae0af74b1c449d2/pydoop_4770ab1ab519444fb57c0dc0bbbc4cff', '-input', 'hdfs://domain/user/data/test.txt', '-output', 'pydoop_f055a031a2b944069ae0af74b1c449d2/output']
DEPRECATED: Use of this script to execute mapred command is deprecated.
Instead use the mapred command for it.

16/06/16 19:30:57 INFO client.RMProxy: Connecting to ResourceManager at dnjbaname-ib01.barnesandnoble.com/192.168.5.10:8032
16/06/16 19:30:57 INFO client.RMProxy: Connecting to ResourceManager at dnjbaname-ib01.barnesandnoble.com/192.168.5.10:8032
16/06/16 19:30:58 WARN mapreduce.JobSubmitter: No job jar file set.  User classes may not be found. See Job or Job#setJar(String).
16/06/16 19:30:58 INFO mapred.FileInputFormat: Total input paths to process : 1
16/06/16 19:30:58 INFO mapreduce.JobSubmitter: number of splits:2
16/06/16 19:30:58 INFO Configuration.deprecation: mapred.job.name is deprecated. Instead, use mapreduce.job.name
16/06/16 19:30:58 INFO Configuration.deprecation: hadoop.pipes.java.recordreader is deprecated. Instead, use mapreduce.pipes.isjavarecordreader
16/06/16 19:30:58 INFO Configuration.deprecation: mapred.map.tasks is deprecated. Instead, use mapreduce.job.maps
16/06/16 19:30:58 INFO Configuration.deprecation: hadoop.pipes.java.recordwriter is deprecated. Instead, use mapreduce.pipes.isjavarecordwriter
16/06/16 19:30:58 INFO Configuration.deprecation: mapred.reduce.tasks is deprecated. Instead, use mapreduce.job.reduces
16/06/16 19:30:58 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1465331659466_0626
16/06/16 19:30:58 INFO mapred.YARNRunner: Job jar is not present. Not adding any jar to the list of resources.
16/06/16 19:30:58 INFO impl.YarnClientImpl: Submitted application application_1465331659466_0626
16/06/16 19:30:58 INFO mapreduce.Job: The url to track the job: http://url
16/06/16 19:30:58 INFO mapreduce.Job: Running job: job_1465331659466_0626
16/06/16 19:31:04 INFO mapreduce.Job: Job job_1465331659466_0626 running in uber mode : false
16/06/16 19:31:04 INFO mapreduce.Job:  map 0% reduce 0%
16/06/16 19:41:33 INFO mapreduce.Job: Task Id : attempt_1465331659466_0626_m_000001_0, Status : FAILED
AttemptID:attempt_1465331659466_0626_m_000001_0 Timed out after 600 secs
16/06/16 19:41:33 INFO mapreduce.Job: Task Id : attempt_1465331659466_0626_m_000000_0, Status : FAILED
AttemptID:attempt_1465331659466_0626_m_000000_0 Timed out after 600 secs

stderr logs:

Jun 16, 2016 7:31:10 PM com.google.inject.servlet.InternalServletModule$BackwardsCompatibleServletContextProvider get
WARNING: You are attempting to use a deprecated API (specifically, attempting to @Inject ServletContext inside an eagerly created singleton. While we allow this for backwards compatibility, be warned that this MAY have unexpected behavior if you have more than one injector (with ServletModule) running in the same JVM. Please consult the Guice documentation at http://code.google.com/p/google-guice/wiki/Servlets for more information.
Jun 16, 2016 7:31:10 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory register
INFO: Registering org.apache.hadoop.mapreduce.v2.app.webapp.JAXBContextResolver as a provider class
Jun 16, 2016 7:31:10 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory register
INFO: Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class
Jun 16, 2016 7:31:10 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory register
INFO: Registering org.apache.hadoop.mapreduce.v2.app.webapp.AMWebServices as a root resource class
Jun 16, 2016 7:31:10 PM com.sun.jersey.server.impl.application.WebApplicationImpl _initiate
INFO: Initiating Jersey application, version 'Jersey: 1.9 09/02/2011 11:17 AM'
Jun 16, 2016 7:31:10 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory getComponentProvider
INFO: Binding org.apache.hadoop.mapreduce.v2.app.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton"
Jun 16, 2016 7:31:10 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory getComponentProvider
INFO: Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton"
Jun 16, 2016 7:31:11 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory getComponentProvider
INFO: Binding org.apache.hadoop.mapreduce.v2.app.webapp.AMWebServices to GuiceManagedComponentProvider with the scope "PerRequest"

syslog logs:

16-06-16 19:31:00,322 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Created MRAppMaster for application appattempt_1465331659466_0626_000001
2016-06-16 19:31:00,597 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Executing with tokens:
2016-06-16 19:31:00,597 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Kind: YARN_AM_RM_TOKEN, Service: , Ident: (org.apache.hadoop.yarn.security.AMRMTokenIdentifier@3cd3e762)
2016-06-16 19:31:01,163 WARN [main] org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
2016-06-16 19:31:01,288 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter set in config null
2016-06-16 19:31:01,289 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter is org.apache.hadoop.mapred.FileOutputCommitter
2016-06-16 19:31:01,346 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.jobhistory.EventType for class org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler
2016-06-16 19:31:01,347 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.job.event.JobEventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobEventDispatcher
2016-06-16 19:31:01,348 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.job.event.TaskEventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskEventDispatcher
2016-06-16 19:31:01,349 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.job.event.TaskAttemptEventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskAttemptEventDispatcher
2016-06-16 19:31:01,349 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventType for class org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler
2016-06-16 19:31:01,350 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.speculate.Speculator$EventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$SpeculatorEventDispatcher
2016-06-16 19:31:01,350 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.rm.ContainerAllocator$EventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerAllocatorRouter
2016-06-16 19:31:01,351 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncher$EventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerLauncherRouter
2016-06-16 19:31:01,389 INFO [main] org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system [hdfs://masked-url:8020]
2016-06-16 19:31:01,408 INFO [main] org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system [hdfs://masked-url:8020]
2016-06-16 19:31:01,428 INFO [main] org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system [hdfs://masked-url:8020]
2016-06-16 19:31:01,437 INFO [main] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Emitting job history data to the timeline server is not enabled
2016-06-16 19:31:01,523 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.job.event.JobFinishEvent$Type for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobFinishEventHandler
2016-06-16 19:31:01,697 INFO [main] org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from hadoop-metrics2.properties
2016-06-16 19:31:01,757 INFO [main] org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s).
2016-06-16 19:31:01,757 INFO [main] org.apache.hadoop.metrics2.impl.MetricsSystemImpl: MRAppMaster metrics system started
2016-06-16 19:31:01,766 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Adding job token for job_1465331659466_0626 to jobTokenSecretManager
2016-06-16 19:31:01,890 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Not uberizing job_1465331659466_0626 because: not enabled; too many reduces;
2016-06-16 19:31:01,917 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Input size for job job_1465331659466_0626 = 41764. Number of splits = 2
2016-06-16 19:31:01,918 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Number of reduces for job job_1465331659466_0626 = 2
2016-06-16 19:31:01,918 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: job_1465331659466_0626Job Transitioned from NEW to INITED
2016-06-16 19:31:01,919 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: MRAppMaster launching normal, non-uberized, multi-container job job_1465331659466_0626.
2016-06-16 19:31:01,947 INFO [main] org.apache.hadoop.ipc.CallQueueManager: Using callQueue class java.util.concurrent.LinkedBlockingQueue
2016-06-16 19:31:01,956 INFO [Socket Reader #1 for port 57330] org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 57330
2016-06-16 19:31:01,974 INFO [main] org.apache.hadoop.yarn.factories.impl.pb.RpcServerFactoryPBImpl: Adding protocol org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB to the server
2016-06-16 19:31:01,974 INFO [IPC Server Responder] org.apache.hadoop.ipc.Server: IPC Server Responder: starting
2016-06-16 19:31:01,975 INFO [IPC Server listener on 57330] org.apache.hadoop.ipc.Server: IPC Server listener on 57330: starting
2016-06-16 19:31:01,976 INFO [main] org.apache.hadoop.mapreduce.v2.app.client.MRClientService: Instantiated MRClientService at masked-url/192.168.5.9:57330
2016-06-16 19:31:02,030 INFO [main] org.mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog
2016-06-16 19:31:02,033 INFO [main] org.apache.hadoop.http.HttpRequestLog: Http request log for http.requests.mapreduce is not defined
2016-06-16 19:31:02,042 INFO [main] org.apache.hadoop.http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)
2016-06-16 19:31:02,045 INFO [main] org.apache.hadoop.http.HttpServer2: Added filter AM_PROXY_FILTER (class=org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter) to context mapreduce
2016-06-16 19:31:02,046 INFO [main] org.apache.hadoop.http.HttpServer2: Added filter AM_PROXY_FILTER (class=org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter) to context static
2016-06-16 19:31:02,048 INFO [main] org.apache.hadoop.http.HttpServer2: adding path spec: /mapreduce/*
2016-06-16 19:31:02,048 INFO [main] org.apache.hadoop.http.HttpServer2: adding path spec: /ws/*
2016-06-16 19:31:02,055 INFO [main] org.apache.hadoop.http.HttpServer2: Jetty bound to port 46957
2016-06-16 19:31:02,055 INFO [main] org.mortbay.log: jetty-6.1.26.cloudera.4
2016-06-16 19:31:02,080 INFO [main] org.mortbay.log: Extract jar:file:/usr/lib/hadoop-yarn/hadoop-yarn-common-2.6.0-cdh5.4.4.jar!/webapps/mapreduce to /tmp/Jetty_0_0_0_0_46957_mapreduce____ttoakg/webapp
2016-06-16 19:31:02,320 INFO [main] org.mortbay.log: Started HttpServer2$SelectChannelConnectorWithSafeStartup@0.0.0.0:46957
2016-06-16 19:31:02,320 INFO [main] org.apache.hadoop.yarn.webapp.WebApps: Web app /mapreduce started at 46957
2016-06-16 19:31:02,622 INFO [main] org.apache.hadoop.yarn.webapp.WebApps: Registered webapp guice modules
2016-06-16 19:31:02,627 INFO [main] org.apache.hadoop.ipc.CallQueueManager: Using callQueue class java.util.concurrent.LinkedBlockingQueue
2016-06-16 19:31:02,627 INFO [Socket Reader #1 for port 39832] org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 39832
2016-06-16 19:31:02,631 INFO [IPC Server Responder] org.apache.hadoop.ipc.Server: IPC Server Responder: starting
2016-06-16 19:31:02,631 INFO [IPC Server listener on 39832] org.apache.hadoop.ipc.Server: IPC Server listener on 39832: starting
2016-06-16 19:31:02,650 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: nodeBlacklistingEnabled:true
2016-06-16 19:31:02,651 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: maxTaskFailuresPerNode is 3
2016-06-16 19:31:02,651 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: blacklistDisablePercent is 33
2016-06-16 19:31:02,708 INFO [main] org.apache.hadoop.yarn.client.RMProxy: Connecting to ResourceManager at masked-url/192.168.5.10:8030
2016-06-16 19:31:02,785 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: maxContainerCapability: <memory:8192, vCores:16>
2016-06-16 19:31:02,785 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: queue: root.hdfs
2016-06-16 19:31:02,812 INFO [main] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Upper limit on the thread pool size is 500
2016-06-16 19:31:02,814 INFO [main] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: yarn.client.max-cached-nodemanagers-proxies : 0
2016-06-16 19:31:02,821 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: job_1465331659466_0626Job Transitioned from INITED to SETUP
2016-06-16 19:31:02,823 INFO [CommitterEvent Processor #0] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: JOB_SETUP
2016-06-16 19:31:02,826 INFO [CommitterEvent Processor #0] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: File Output Committer Algorithm version is 1
2016-06-16 19:31:02,832 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: job_1465331659466_0626Job Transitioned from SETUP to RUNNING
2016-06-16 19:31:02,877 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:31:02,891 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Event Writer setup for JobId: job_1465331659466_0626, File: hdfs://masked-url:8020/user/hdfs/.staging/job_1465331659466_0626/job_1465331659466_0626_1.jhist
2016-06-16 19:31:02,898 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:31:02,920 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:31:02,925 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1465331659466_0626_m_000000 Task Transitioned from NEW to SCHEDULED
2016-06-16 19:31:02,925 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:31:02,925 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:31:02,925 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:31:02,925 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1465331659466_0626_m_000001 Task Transitioned from NEW to SCHEDULED
2016-06-16 19:31:02,926 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1465331659466_0626_r_000000 Task Transitioned from NEW to SCHEDULED
2016-06-16 19:31:02,926 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1465331659466_0626_r_000001 Task Transitioned from NEW to SCHEDULED
2016-06-16 19:31:02,928 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_0 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-16 19:31:02,928 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_0 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-16 19:31:02,928 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_r_000000_0 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-16 19:31:02,928 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_r_000001_0 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-16 19:31:02,929 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: mapResourceRequest:<memory:1024, vCores:1>
2016-06-16 19:31:02,937 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: reduceResourceRequest:<memory:1024, vCores:1>
2016-06-16 19:31:03,122 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system [hdfs://masked-url:8020]
2016-06-16 19:31:03,807 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before Scheduling: PendingReds:2 ScheduledMaps:2 ScheduledReds:0 AssignedMaps:0 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:0 ContRel:0 HostLocal:0 RackLocal:0
2016-06-16 19:31:03,895 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_0626: ask=5 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:23866, vCores:63> knownNMs=4
2016-06-16 19:31:03,896 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:23866, vCores:63>
2016-06-16 19:31:03,897 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 2
2016-06-16 19:31:04,909 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated containers 2
2016-06-16 19:31:04,911 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1465331659466_0626_01_000002 to attempt_1465331659466_0626_m_000000_0
2016-06-16 19:31:04,914 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1465331659466_0626_01_000003 to attempt_1465331659466_0626_m_000001_0
2016-06-16 19:31:04,914 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:21818, vCores:61>
2016-06-16 19:31:04,914 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 2
2016-06-16 19:31:04,914 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:2 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:2 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:2 ContRel:0 HostLocal:2 RackLocal:0
2016-06-16 19:31:04,958 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:31:04,959 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Job jar is not present. Not adding any jar to the list of resources.
2016-06-16 19:31:04,977 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: The job-conf file on the remote FS is /user/hdfs/.staging/job_1465331659466_0626/job.xml
2016-06-16 19:31:04,980 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Adding #0 tokens and #1 secret keys for NM use for launching container
2016-06-16 19:31:04,980 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Size of containertokens_dob is 1
2016-06-16 19:31:04,980 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Putting shuffle token in serviceData
2016-06-16 19:31:05,013 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_0 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED
2016-06-16 19:31:05,016 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:31:05,017 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_0 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED
2016-06-16 19:31:05,019 INFO [ContainerLauncher #0] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1465331659466_0626_01_000002 taskAttempt attempt_1465331659466_0626_m_000000_0
2016-06-16 19:31:05,019 INFO [ContainerLauncher #1] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1465331659466_0626_01_000003 taskAttempt attempt_1465331659466_0626_m_000001_0
2016-06-16 19:31:05,023 INFO [ContainerLauncher #1] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1465331659466_0626_m_000001_0
2016-06-16 19:31:05,023 INFO [ContainerLauncher #0] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1465331659466_0626_m_000000_0
2016-06-16 19:31:05,024 INFO [ContainerLauncher #1] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : masked-url:8041
2016-06-16 19:31:05,046 INFO [ContainerLauncher #0] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : masked-url:8041
2016-06-16 19:31:05,089 INFO [ContainerLauncher #1] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1465331659466_0626_m_000001_0 : 13562
2016-06-16 19:31:05,089 INFO [ContainerLauncher #0] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1465331659466_0626_m_000000_0 : 13562
2016-06-16 19:31:05,091 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1465331659466_0626_m_000001_0] using containerId: [container_1465331659466_0626_01_000003 on NM: [masked-url:8041]
2016-06-16 19:31:05,094 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_0 TaskAttempt Transitioned from ASSIGNED to RUNNING
2016-06-16 19:31:05,095 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1465331659466_0626_m_000000_0] using containerId: [container_1465331659466_0626_01_000002 on NM: [masked-url:8041]
2016-06-16 19:31:05,095 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_0 TaskAttempt Transitioned from ASSIGNED to RUNNING
2016-06-16 19:31:05,096 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1465331659466_0626_m_000001 Task Transitioned from SCHEDULED to RUNNING
2016-06-16 19:31:05,096 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1465331659466_0626_m_000000 Task Transitioned from SCHEDULED to RUNNING
2016-06-16 19:31:05,916 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_0626: ask=5 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:21818, vCores:61> knownNMs=4
2016-06-16 19:31:06,156 INFO [Socket Reader #1 for port 39832] SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for job_1465331659466_0626 (auth:SIMPLE)
2016-06-16 19:31:06,177 INFO [IPC Server handler 0 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : jvm_1465331659466_0626_m_000003 asked for a task
2016-06-16 19:31:06,177 INFO [IPC Server handler 0 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: jvm_1465331659466_0626_m_000003 given task: attempt_1465331659466_0626_m_000001_0
2016-06-16 19:31:06,464 INFO [Socket Reader #1 for port 39832] SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for job_1465331659466_0626 (auth:SIMPLE)
2016-06-16 19:31:06,481 INFO [IPC Server handler 1 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : jvm_1465331659466_0626_m_000002 asked for a task
2016-06-16 19:31:06,481 INFO [IPC Server handler 1 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: jvm_1465331659466_0626_m_000002 given task: attempt_1465331659466_0626_m_000000_0
2016-06-16 19:31:12,913 INFO [IPC Server handler 1 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of TaskAttempt attempt_1465331659466_0626_m_000001_0 is : 0.0
2016-06-16 19:31:13,278 INFO [IPC Server handler 6 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of TaskAttempt attempt_1465331659466_0626_m_000000_0 is : 0.0
2016-06-16 19:41:32,641 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_0626_m_000001_0: AttemptID:attempt_1465331659466_0626_m_000001_0 Timed out after 600 secs
2016-06-16 19:41:32,642 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_0 TaskAttempt Transitioned from RUNNING to FAIL_CONTAINER_CLEANUP
2016-06-16 19:41:32,643 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_0626_m_000000_0: AttemptID:attempt_1465331659466_0626_m_000000_0 Timed out after 600 secs
2016-06-16 19:41:32,643 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_0 TaskAttempt Transitioned from RUNNING to FAIL_CONTAINER_CLEANUP
2016-06-16 19:41:32,644 INFO [ContainerLauncher #2] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1465331659466_0626_01_000003 taskAttempt attempt_1465331659466_0626_m_000001_0
2016-06-16 19:41:32,645 INFO [ContainerLauncher #2] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1465331659466_0626_m_000001_0
2016-06-16 19:41:32,645 INFO [ContainerLauncher #3] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1465331659466_0626_01_000002 taskAttempt attempt_1465331659466_0626_m_000000_0
2016-06-16 19:41:32,645 INFO [ContainerLauncher #3] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1465331659466_0626_m_000000_0
2016-06-16 19:41:32,645 INFO [ContainerLauncher #2] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : masked-url:8041
2016-06-16 19:41:32,647 INFO [ContainerLauncher #3] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : masked-url:8041
2016-06-16 19:41:32,667 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_0 TaskAttempt Transitioned from FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
2016-06-16 19:41:32,667 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_0 TaskAttempt Transitioned from FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
2016-06-16 19:41:32,667 INFO [CommitterEvent Processor #1] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: TASK_ABORT
2016-06-16 19:41:32,668 INFO [CommitterEvent Processor #2] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: TASK_ABORT
2016-06-16 19:41:32,675 WARN [CommitterEvent Processor #2] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not delete hdfs://masked-url:8020/user/hdfs/pydoop_f055a031a2b944069ae0af74b1c449d2/output/_temporary/1/_temporary/attempt_1465331659466_0626_m_000001_0
2016-06-16 19:41:32,675 WARN [CommitterEvent Processor #1] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not delete hdfs://masked-url:8020/user/hdfs/pydoop_f055a031a2b944069ae0af74b1c449d2/output/_temporary/1/_temporary/attempt_1465331659466_0626_m_000000_0
2016-06-16 19:41:32,678 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_0 TaskAttempt Transitioned from FAIL_TASK_CLEANUP to FAILED
2016-06-16 19:41:32,678 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_0 TaskAttempt Transitioned from FAIL_TASK_CLEANUP to FAILED
2016-06-16 19:41:32,689 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:41:32,690 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:41:32,690 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:41:32,691 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:41:32,691 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:41:32,691 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:41:32,691 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 1 failures on node masked-url
2016-06-16 19:41:32,694 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_1 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-16 19:41:32,694 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 1 failures on node masked-url
2016-06-16 19:41:32,694 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_1 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-16 19:41:32,695 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added attempt_1465331659466_0626_m_000001_1 to list of failed maps
2016-06-16 19:41:32,695 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added attempt_1465331659466_0626_m_000000_1 to list of failed maps
2016-06-16 19:41:32,756 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before Scheduling: PendingReds:2 ScheduledMaps:2 ScheduledReds:0 AssignedMaps:2 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:2 ContRel:0 HostLocal:2 RackLocal:0
2016-06-16 19:41:32,759 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_0626: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:21818, vCores:61> knownNMs=4
2016-06-16 19:41:32,759 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:21818, vCores:61>
2016-06-16 19:41:32,759 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 2
2016-06-16 19:41:33,767 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received completed container container_1465331659466_0626_01_000002
2016-06-16 19:41:33,767 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received completed container container_1465331659466_0626_01_000003
2016-06-16 19:41:33,768 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_0626_m_000000_0: Container killed by the ApplicationMaster.
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143

2016-06-16 19:41:33,768 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated containers 2
2016-06-16 19:41:33,768 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_0626_m_000001_0: Container killed by the ApplicationMaster.
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143

2016-06-16 19:41:33,768 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning container Container: [ContainerId: container_1465331659466_0626_01_000004, NodeId: masked-url:8041, NodeHttpAddress: masked-url:8042, Resource: <memory:1024, vCores:1>, Priority: 5, Token: Token { kind: ContainerToken, service: 192.168.5.9:8041 }, ] to fast fail map
2016-06-16 19:41:33,768 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned from earlierFailedMaps
2016-06-16 19:41:33,768 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1465331659466_0626_01_000004 to attempt_1465331659466_0626_m_000001_1
2016-06-16 19:41:33,768 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning container Container: [ContainerId: container_1465331659466_0626_01_000005, NodeId: masked-url:8041, NodeHttpAddress: masked-url:8042, Resource: <memory:1024, vCores:1>, Priority: 5, Token: Token { kind: ContainerToken, service: 192.168.5.12:8041 }, ] to fast fail map
2016-06-16 19:41:33,768 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned from earlierFailedMaps
2016-06-16 19:41:33,768 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1465331659466_0626_01_000005 to attempt_1465331659466_0626_m_000000_1
2016-06-16 19:41:33,769 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:21818, vCores:61>
2016-06-16 19:41:33,769 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 2
2016-06-16 19:41:33,769 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:2 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:2 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:4 ContRel:0 HostLocal:2 RackLocal:0
2016-06-16 19:41:33,793 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:41:33,794 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_1 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED
2016-06-16 19:41:33,794 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:41:33,795 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_1 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED
2016-06-16 19:41:33,796 INFO [ContainerLauncher #4] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1465331659466_0626_01_000004 taskAttempt attempt_1465331659466_0626_m_000001_1
2016-06-16 19:41:33,796 INFO [ContainerLauncher #4] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1465331659466_0626_m_000001_1
2016-06-16 19:41:33,796 INFO [ContainerLauncher #5] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1465331659466_0626_01_000005 taskAttempt attempt_1465331659466_0626_m_000000_1
2016-06-16 19:41:33,796 INFO [ContainerLauncher #5] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1465331659466_0626_m_000000_1
2016-06-16 19:41:33,796 INFO [ContainerLauncher #4] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : masked-url:8041
2016-06-16 19:41:33,797 INFO [ContainerLauncher #5] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : masked-url:8041
2016-06-16 19:41:33,805 INFO [ContainerLauncher #4] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1465331659466_0626_m_000001_1 : 13562
2016-06-16 19:41:33,805 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1465331659466_0626_m_000001_1] using containerId: [container_1465331659466_0626_01_000004 on NM: [masked-url:8041]
2016-06-16 19:41:33,805 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_1 TaskAttempt Transitioned from ASSIGNED to RUNNING
2016-06-16 19:41:33,817 INFO [ContainerLauncher #5] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1465331659466_0626_m_000000_1 : 13562
2016-06-16 19:41:33,818 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1465331659466_0626_m_000000_1] using containerId: [container_1465331659466_0626_01_000005 on NM: [masked-url:8041]
2016-06-16 19:41:33,818 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_1 TaskAttempt Transitioned from ASSIGNED to RUNNING
2016-06-16 19:41:34,773 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_0626: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:21818, vCores:61> knownNMs=4
2016-06-16 19:41:34,842 INFO [Socket Reader #1 for port 39832] SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for job_1465331659466_0626 (auth:SIMPLE)
2016-06-16 19:41:34,852 INFO [IPC Server handler 28 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : jvm_1465331659466_0626_m_000005 asked for a task
2016-06-16 19:41:34,852 INFO [IPC Server handler 28 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: jvm_1465331659466_0626_m_000005 given task: attempt_1465331659466_0626_m_000000_1
2016-06-16 19:41:35,016 INFO [Socket Reader #1 for port 39832] SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for job_1465331659466_0626 (auth:SIMPLE)
2016-06-16 19:41:35,028 INFO [IPC Server handler 9 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : jvm_1465331659466_0626_m_000004 asked for a task
2016-06-16 19:41:35,028 INFO [IPC Server handler 9 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: jvm_1465331659466_0626_m_000004 given task: attempt_1465331659466_0626_m_000001_1
2016-06-16 19:41:41,529 INFO [IPC Server handler 28 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of TaskAttempt attempt_1465331659466_0626_m_000000_1 is : 0.0
2016-06-16 19:41:41,863 INFO [IPC Server handler 9 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of TaskAttempt attempt_1465331659466_0626_m_000001_1 is : 0.0
2016-06-16 19:52:02,645 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_0626_m_000000_1: AttemptID:attempt_1465331659466_0626_m_000000_1 Timed out after 600 secs
2016-06-16 19:52:02,645 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_1 TaskAttempt Transitioned from RUNNING to FAIL_CONTAINER_CLEANUP
2016-06-16 19:52:02,645 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_0626_m_000001_1: AttemptID:attempt_1465331659466_0626_m_000001_1 Timed out after 600 secs
2016-06-16 19:52:02,645 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_1 TaskAttempt Transitioned from RUNNING to FAIL_CONTAINER_CLEANUP
2016-06-16 19:52:02,646 INFO [ContainerLauncher #6] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1465331659466_0626_01_000005 taskAttempt attempt_1465331659466_0626_m_000000_1
2016-06-16 19:52:02,647 INFO [ContainerLauncher #6] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1465331659466_0626_m_000000_1
2016-06-16 19:52:02,647 INFO [ContainerLauncher #7] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1465331659466_0626_01_000004 taskAttempt attempt_1465331659466_0626_m_000001_1
2016-06-16 19:52:02,647 INFO [ContainerLauncher #6] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : masked-url:8041
2016-06-16 19:52:02,648 INFO [ContainerLauncher #7] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1465331659466_0626_m_000001_1
2016-06-16 19:52:02,649 INFO [ContainerLauncher #7] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : masked-url:8041
2016-06-16 19:52:02,658 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_1 TaskAttempt Transitioned from FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
2016-06-16 19:52:02,658 INFO [CommitterEvent Processor #3] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: TASK_ABORT
2016-06-16 19:52:02,659 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_1 TaskAttempt Transitioned from FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
2016-06-16 19:52:02,659 INFO [CommitterEvent Processor #4] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: TASK_ABORT
2016-06-16 19:52:02,662 WARN [CommitterEvent Processor #3] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not delete hdfs://masked-url:8020/user/hdfs/pydoop_f055a031a2b944069ae0af74b1c449d2/output/_temporary/1/_temporary/attempt_1465331659466_0626_m_000001_1
2016-06-16 19:52:02,662 WARN [CommitterEvent Processor #4] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not delete hdfs://masked-url:8020/user/hdfs/pydoop_f055a031a2b944069ae0af74b1c449d2/output/_temporary/1/_temporary/attempt_1465331659466_0626_m_000000_1
2016-06-16 19:52:02,663 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_1 TaskAttempt Transitioned from FAIL_TASK_CLEANUP to FAILED
2016-06-16 19:52:02,663 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_1 TaskAttempt Transitioned from FAIL_TASK_CLEANUP to FAILED
2016-06-16 19:52:02,664 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:52:02,664 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:52:02,664 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:52:02,665 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:52:02,665 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:52:02,665 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:52:02,665 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 1 failures on node masked-url
2016-06-16 19:52:02,666 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_2 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-16 19:52:02,666 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 1 failures on node masked-url
2016-06-16 19:52:02,666 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_2 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-16 19:52:02,667 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added attempt_1465331659466_0626_m_000001_2 to list of failed maps
2016-06-16 19:52:02,667 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added attempt_1465331659466_0626_m_000000_2 to list of failed maps
2016-06-16 19:52:03,469 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before Scheduling: PendingReds:2 ScheduledMaps:2 ScheduledReds:0 AssignedMaps:2 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:4 ContRel:0 HostLocal:2 RackLocal:0
2016-06-16 19:52:03,470 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_0626: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:21818, vCores:61> knownNMs=4
2016-06-16 19:52:03,471 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:21818, vCores:61>
2016-06-16 19:52:03,471 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 2
2016-06-16 19:52:04,472 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received completed container container_1465331659466_0626_01_000004
2016-06-16 19:52:04,472 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received completed container container_1465331659466_0626_01_000005
2016-06-16 19:52:04,473 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated containers 2
2016-06-16 19:52:04,473 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_0626_m_000001_1: Container killed by the ApplicationMaster.
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143

2016-06-16 19:52:04,473 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning container Container: [ContainerId: container_1465331659466_0626_01_000006, NodeId: masked-url:8041, NodeHttpAddress: masked-url:8042, Resource: <memory:1024, vCores:1>, Priority: 5, Token: Token { kind: ContainerToken, service: 192.168.5.9:8041 }, ] to fast fail map
2016-06-16 19:52:04,473 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned from earlierFailedMaps
2016-06-16 19:52:04,473 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_0626_m_000000_1: Container killed by the ApplicationMaster.
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143

2016-06-16 19:52:04,473 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1465331659466_0626_01_000006 to attempt_1465331659466_0626_m_000001_2
2016-06-16 19:52:04,473 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning container Container: [ContainerId: container_1465331659466_0626_01_000007, NodeId: masked-url:8041, NodeHttpAddress: masked-url:8042, Resource: <memory:1024, vCores:1>, Priority: 5, Token: Token { kind: ContainerToken, service: 192.168.5.13:8041 }, ] to fast fail map
2016-06-16 19:52:04,473 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned from earlierFailedMaps
2016-06-16 19:52:04,473 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1465331659466_0626_01_000007 to attempt_1465331659466_0626_m_000000_2
2016-06-16 19:52:04,473 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:21818, vCores:61>
2016-06-16 19:52:04,473 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 2
2016-06-16 19:52:04,473 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:2 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:2 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:6 ContRel:0 HostLocal:2 RackLocal:0
2016-06-16 19:52:04,473 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:52:04,474 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_2 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED
2016-06-16 19:52:04,474 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 19:52:04,475 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_2 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED
2016-06-16 19:52:04,476 INFO [ContainerLauncher #8] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1465331659466_0626_01_000006 taskAttempt attempt_1465331659466_0626_m_000001_2
2016-06-16 19:52:04,476 INFO [ContainerLauncher #8] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1465331659466_0626_m_000001_2
2016-06-16 19:52:04,476 INFO [ContainerLauncher #8] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : masked-url:8041
2016-06-16 19:52:04,476 INFO [ContainerLauncher #9] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1465331659466_0626_01_000007 taskAttempt attempt_1465331659466_0626_m_000000_2
2016-06-16 19:52:04,476 INFO [ContainerLauncher #9] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1465331659466_0626_m_000000_2
2016-06-16 19:52:04,477 INFO [ContainerLauncher #9] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : masked-url:8041
2016-06-16 19:52:04,484 INFO [ContainerLauncher #8] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1465331659466_0626_m_000001_2 : 13562
2016-06-16 19:52:04,484 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1465331659466_0626_m_000001_2] using containerId: [container_1465331659466_0626_01_000006 on NM: [masked-url:8041]
2016-06-16 19:52:04,484 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_2 TaskAttempt Transitioned from ASSIGNED to RUNNING
2016-06-16 19:52:04,485 INFO [ContainerLauncher #9] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1465331659466_0626_m_000000_2 : 13562
2016-06-16 19:52:04,485 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1465331659466_0626_m_000000_2] using containerId: [container_1465331659466_0626_01_000007 on NM: [masked-url:8041]
2016-06-16 19:52:04,486 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_2 TaskAttempt Transitioned from ASSIGNED to RUNNING
2016-06-16 19:52:05,477 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_0626: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:21818, vCores:61> knownNMs=4
2016-06-16 19:52:05,720 INFO [Socket Reader #1 for port 39832] SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for job_1465331659466_0626 (auth:SIMPLE)
2016-06-16 19:52:05,732 INFO [IPC Server handler 25 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : jvm_1465331659466_0626_m_000006 asked for a task
2016-06-16 19:52:05,732 INFO [IPC Server handler 25 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: jvm_1465331659466_0626_m_000006 given task: attempt_1465331659466_0626_m_000001_2
2016-06-16 19:52:05,936 INFO [Socket Reader #1 for port 39832] SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for job_1465331659466_0626 (auth:SIMPLE)
2016-06-16 19:52:05,953 INFO [IPC Server handler 27 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : jvm_1465331659466_0626_m_000007 asked for a task
2016-06-16 19:52:05,953 INFO [IPC Server handler 27 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: jvm_1465331659466_0626_m_000007 given task: attempt_1465331659466_0626_m_000000_2
2016-06-16 19:52:12,562 INFO [IPC Server handler 24 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of TaskAttempt attempt_1465331659466_0626_m_000001_2 is : 0.0
2016-06-16 19:52:12,779 INFO [IPC Server handler 20 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of TaskAttempt attempt_1465331659466_0626_m_000000_2 is : 0.0
2016-06-16 20:02:32,649 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_0626_m_000001_2: AttemptID:attempt_1465331659466_0626_m_000001_2 Timed out after 600 secs
2016-06-16 20:02:32,649 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_2 TaskAttempt Transitioned from RUNNING to FAIL_CONTAINER_CLEANUP
2016-06-16 20:02:32,649 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_0626_m_000000_2: AttemptID:attempt_1465331659466_0626_m_000000_2 Timed out after 600 secs
2016-06-16 20:02:32,649 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_2 TaskAttempt Transitioned from RUNNING to FAIL_CONTAINER_CLEANUP
2016-06-16 20:02:32,650 INFO [ContainerLauncher #1] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1465331659466_0626_01_000006 taskAttempt attempt_1465331659466_0626_m_000001_2
2016-06-16 20:02:32,650 INFO [ContainerLauncher #0] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1465331659466_0626_01_000007 taskAttempt attempt_1465331659466_0626_m_000000_2
2016-06-16 20:02:32,650 INFO [ContainerLauncher #1] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1465331659466_0626_m_000001_2
2016-06-16 20:02:32,651 INFO [ContainerLauncher #0] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1465331659466_0626_m_000000_2
2016-06-16 20:02:32,651 INFO [ContainerLauncher #1] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : masked-url:8041
2016-06-16 20:02:32,651 INFO [ContainerLauncher #0] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : masked-url:8041
2016-06-16 20:02:32,658 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_2 TaskAttempt Transitioned from FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
2016-06-16 20:02:32,658 INFO [CommitterEvent Processor #0] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: TASK_ABORT
2016-06-16 20:02:32,658 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_2 TaskAttempt Transitioned from FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
2016-06-16 20:02:32,659 INFO [CommitterEvent Processor #2] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: TASK_ABORT
2016-06-16 20:02:32,662 WARN [CommitterEvent Processor #0] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not delete hdfs://masked-url:8020/user/hdfs/pydoop_f055a031a2b944069ae0af74b1c449d2/output/_temporary/1/_temporary/attempt_1465331659466_0626_m_000001_2
2016-06-16 20:02:32,662 WARN [CommitterEvent Processor #2] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not delete hdfs://masked-url:8020/user/hdfs/pydoop_f055a031a2b944069ae0af74b1c449d2/output/_temporary/1/_temporary/attempt_1465331659466_0626_m_000000_2
2016-06-16 20:02:32,662 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_2 TaskAttempt Transitioned from FAIL_TASK_CLEANUP to FAILED
2016-06-16 20:02:32,663 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_2 TaskAttempt Transitioned from FAIL_TASK_CLEANUP to FAILED
2016-06-16 20:02:32,663 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 20:02:32,663 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 20:02:32,663 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 20:02:32,665 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 20:02:32,666 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 20:02:32,666 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 20:02:32,666 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 2 failures on node masked-url
2016-06-16 20:02:32,666 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_3 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-16 20:02:32,666 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 2 failures on node masked-url
2016-06-16 20:02:32,666 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_3 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-16 20:02:32,667 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added attempt_1465331659466_0626_m_000001_3 to list of failed maps
2016-06-16 20:02:32,667 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added attempt_1465331659466_0626_m_000000_3 to list of failed maps
2016-06-16 20:02:33,177 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before Scheduling: PendingReds:2 ScheduledMaps:2 ScheduledReds:0 AssignedMaps:2 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:6 ContRel:0 HostLocal:2 RackLocal:0
2016-06-16 20:02:33,179 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_0626: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:21818, vCores:61> knownNMs=4
2016-06-16 20:02:33,180 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:21818, vCores:61>
2016-06-16 20:02:33,180 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 2
2016-06-16 20:02:34,181 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received completed container container_1465331659466_0626_01_000007
2016-06-16 20:02:34,181 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received completed container container_1465331659466_0626_01_000006
2016-06-16 20:02:34,181 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated containers 2
2016-06-16 20:02:34,181 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_0626_m_000000_2: Container killed by the ApplicationMaster.
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143

2016-06-16 20:02:34,181 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning container Container: [ContainerId: container_1465331659466_0626_01_000008, NodeId: masked-url:8041, NodeHttpAddress: masked-url:8042, Resource: <memory:1024, vCores:1>, Priority: 5, Token: Token { kind: ContainerToken, service: 192.168.5.9:8041 }, ] to fast fail map
2016-06-16 20:02:34,182 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned from earlierFailedMaps
2016-06-16 20:02:34,182 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_0626_m_000001_2: Container killed by the ApplicationMaster.
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143

2016-06-16 20:02:34,182 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1465331659466_0626_01_000008 to attempt_1465331659466_0626_m_000001_3
2016-06-16 20:02:34,182 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning container Container: [ContainerId: container_1465331659466_0626_01_000009, NodeId: masked-url:8041, NodeHttpAddress: masked-url:8042, Resource: <memory:1024, vCores:1>, Priority: 5, Token: Token { kind: ContainerToken, service: 192.168.5.11:8041 }, ] to fast fail map
2016-06-16 20:02:34,182 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned from earlierFailedMaps
2016-06-16 20:02:34,182 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1465331659466_0626_01_000009 to attempt_1465331659466_0626_m_000000_3
2016-06-16 20:02:34,182 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:21818, vCores:61>
2016-06-16 20:02:34,182 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 2
2016-06-16 20:02:34,182 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:2 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:2 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:8 ContRel:0 HostLocal:2 RackLocal:0
2016-06-16 20:02:34,182 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 20:02:34,183 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_3 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED
2016-06-16 20:02:34,183 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved masked-url to /default
2016-06-16 20:02:34,183 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_3 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED
2016-06-16 20:02:34,184 INFO [ContainerLauncher #3] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1465331659466_0626_01_000008 taskAttempt attempt_1465331659466_0626_m_000001_3
2016-06-16 20:02:34,184 INFO [ContainerLauncher #2] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1465331659466_0626_01_000009 taskAttempt attempt_1465331659466_0626_m_000000_3
2016-06-16 20:02:34,184 INFO [ContainerLauncher #3] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1465331659466_0626_m_000001_3
2016-06-16 20:02:34,184 INFO [ContainerLauncher #2] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1465331659466_0626_m_000000_3
2016-06-16 20:02:34,184 INFO [ContainerLauncher #3] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : masked-url:8041
2016-06-16 20:02:34,185 INFO [ContainerLauncher #2] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : masked-url:8041
2016-06-16 20:02:34,193 INFO [ContainerLauncher #3] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1465331659466_0626_m_000001_3 : 13562
2016-06-16 20:02:34,194 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1465331659466_0626_m_000001_3] using containerId: [container_1465331659466_0626_01_000008 on NM: [masked-url:8041]
2016-06-16 20:02:34,194 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000001_3 TaskAttempt Transitioned from ASSIGNED to RUNNING
2016-06-16 20:02:34,194 INFO [ContainerLauncher #2] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1465331659466_0626_m_000000_3 : 13562
2016-06-16 20:02:34,194 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1465331659466_0626_m_000000_3] using containerId: [container_1465331659466_0626_01_000009 on NM: [masked-url:8041]
2016-06-16 20:02:34,194 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_0626_m_000000_3 TaskAttempt Transitioned from ASSIGNED to RUNNING
2016-06-16 20:02:35,183 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_0626: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:21818, vCores:61> knownNMs=4
2016-06-16 20:02:35,256 INFO [Socket Reader #1 for port 39832] SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for job_1465331659466_0626 (auth:SIMPLE)
2016-06-16 20:02:35,267 INFO [IPC Server handler 28 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : jvm_1465331659466_0626_m_000009 asked for a task
2016-06-16 20:02:35,267 INFO [IPC Server handler 28 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: jvm_1465331659466_0626_m_000009 given task: attempt_1465331659466_0626_m_000000_3
2016-06-16 20:02:35,401 INFO [Socket Reader #1 for port 39832] SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for job_1465331659466_0626 (auth:SIMPLE)
2016-06-16 20:02:35,413 INFO [IPC Server handler 20 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : jvm_1465331659466_0626_m_000008 asked for a task
2016-06-16 20:02:35,413 INFO [IPC Server handler 20 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: jvm_1465331659466_0626_m_000008 given task: attempt_1465331659466_0626_m_000001_3
2016-06-16 20:02:41,956 INFO [IPC Server handler 22 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of TaskAttempt attempt_1465331659466_0626_m_000000_3 is : 0.0
2016-06-16 20:02:42,237 INFO [IPC Server handler 20 on 39832] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of TaskAttempt attempt_1465331659466_0626_m_000001_3 is : 0.0
@elzaggo
Copy link
Contributor

elzaggo commented Jun 17, 2016

Hi!
Do you have the job history server running? It would be helpful if you could recover the actual task execution log files from the node manager.

--gianluigi

@igorvishnevskiy
Copy link
Author

igorvishnevskiy commented Jun 22, 2016

Hi Gianluigi. The logs that I included above are the ones from: "<<<host_address>>>:19888/jobhistory/". That's all that's there. Nothing else. Thank you for your help.

Job basically gets stuck on:
16/06/16 19:31:04 INFO mapreduce.Job: map 0% reduce 0%
...and no movement at all.

I tried examples from the following page as well:
http://crs4.github.io/pydoop/tutorial/hdfs_api.html#hdfs-api-tutorial

import pydoop.hdfs as hdfs
hdfs.dump("hello", "hdfs://path/test.txt")
text = hdfs.load("hdfs://path/test.txt")
print text

It works good. For some reason jobs hang only on mapreduce job. Am I missing something in a setup maybe or my runner?

@igorvishnevskiy
Copy link
Author

igorvishnevskiy commented Jun 22, 2016

I also just tried to run this simple example as a script:

def mapper(_, text, writer):
    wordlist = text.split()
    for word in wordlist:
        writer.emit(word, "1")
    writer.count("num words", len(wordlist))

def reducer(word, count, writer):
    writer.emit(word, sum(map(int, count)))

Command to start a script was:
pydoop script wc.py hdfs://path/temp/test.txt hdfs://path/temp/test_test/

The same problem. There is something wrong with mapreduce API in pydoop. hdfs api works good for me, but mapreduce api does not. Or there are some settings that are required for Hadoop 2.6 CDH5 environment for pydoop's mapreduce api to start working. What are they, please if anybody knows, let me and everyone else who might face this issue know. Thank you for your help!

On the following page:
https://crs4.github.io/pydoop/installation.html
At the very bottom under "Hadoop2 / CDH4". There is a specification to edit hdfs-site.xml and set dfs.namenode.fs-limits.min-block-size to a low value:

<property>
  <name>dfs.namenode.fs-limits.min-block-size</name>
  <value>512</value>
</property>

then restart Hadoop daemons.
Do we need to do that as well for combo of Hadoop 2.6 / CDH5?

Thank you.

@igorvishnevskiy
Copy link
Author

igorvishnevskiy commented Jun 22, 2016

Hi @elzaggo. These are additional fresh set of logs from "....:19888/jobhistory/".

2016-06-22 16:48:59,665 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Created MRAppMaster for application appattempt_1465331659466_1048_000001
2016-06-22 16:48:59,882 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Executing with tokens:
2016-06-22 16:48:59,882 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Kind: YARN_AM_RM_TOKEN, Service: , Ident: (org.apache.hadoop.yarn.security.AMRMTokenIdentifier@7eac9008)
2016-06-22 16:48:59,930 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Using mapred newApiCommitter.
2016-06-22 16:49:00,420 WARN [main] org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
2016-06-22 16:49:00,541 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter set in config null
2016-06-22 16:49:00,593 INFO [main] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: File Output Committer Algorithm version is 1
2016-06-22 16:49:00,596 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter is org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
2016-06-22 16:49:00,618 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.jobhistory.EventType for class org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler
2016-06-22 16:49:00,618 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.job.event.JobEventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobEventDispatcher
2016-06-22 16:49:00,619 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.job.event.TaskEventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskEventDispatcher
2016-06-22 16:49:00,620 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.job.event.TaskAttemptEventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskAttemptEventDispatcher
2016-06-22 16:49:00,620 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventType for class org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler
2016-06-22 16:49:00,621 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.speculate.Speculator$EventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$SpeculatorEventDispatcher
2016-06-22 16:49:00,622 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.rm.ContainerAllocator$EventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerAllocatorRouter
2016-06-22 16:49:00,623 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncher$EventType for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerLauncherRouter
2016-06-22 16:49:00,702 INFO [main] org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system [hdfs://node-name-ib01.domain-name.com:8020]
2016-06-22 16:49:00,723 INFO [main] org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system [hdfs://node-name-ib01.domain-name.com:8020]
2016-06-22 16:49:00,742 INFO [main] org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system [hdfs://node-name-ib01.domain-name.com:8020]
2016-06-22 16:49:00,749 INFO [main] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Emitting job history data to the timeline server is not enabled
2016-06-22 16:49:00,789 INFO [main] org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class org.apache.hadoop.mapreduce.v2.app.job.event.JobFinishEvent$Type for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobFinishEventHandler
2016-06-22 16:49:00,963 INFO [main] org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from hadoop-metrics2.properties
2016-06-22 16:49:01,023 INFO [main] org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s).
2016-06-22 16:49:01,023 INFO [main] org.apache.hadoop.metrics2.impl.MetricsSystemImpl: MRAppMaster metrics system started
2016-06-22 16:49:01,030 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Adding job token for job_1465331659466_1048 to jobTokenSecretManager
2016-06-22 16:49:01,140 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Not uberizing job_1465331659466_1048 because: not enabled;
2016-06-22 16:49:01,154 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Input size for job job_1465331659466_1048 = 5. Number of splits = 1
2016-06-22 16:49:01,155 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Number of reduces for job job_1465331659466_1048 = 1
2016-06-22 16:49:01,155 INFO [main] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: job_1465331659466_1048Job Transitioned from NEW to INITED
2016-06-22 16:49:01,157 INFO [main] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: MRAppMaster launching normal, non-uberized, multi-container job job_1465331659466_1048.
2016-06-22 16:49:01,185 INFO [main] org.apache.hadoop.ipc.CallQueueManager: Using callQueue class java.util.concurrent.LinkedBlockingQueue
2016-06-22 16:49:01,194 INFO [Socket Reader #1 for port 46701] org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 46701
2016-06-22 16:49:01,212 INFO [main] org.apache.hadoop.yarn.factories.impl.pb.RpcServerFactoryPBImpl: Adding protocol org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB to the server
2016-06-22 16:49:01,213 INFO [IPC Server Responder] org.apache.hadoop.ipc.Server: IPC Server Responder: starting
2016-06-22 16:49:01,213 INFO [IPC Server listener on 46701] org.apache.hadoop.ipc.Server: IPC Server listener on 46701: starting
2016-06-22 16:49:01,214 INFO [main] org.apache.hadoop.mapreduce.v2.app.client.MRClientService: Instantiated MRClientService at node-name-ib02.domain-name.com/192.168.5.9:46701
2016-06-22 16:49:01,269 INFO [main] org.mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog
2016-06-22 16:49:01,272 INFO [main] org.apache.hadoop.http.HttpRequestLog: Http request log for http.requests.mapreduce is not defined
2016-06-22 16:49:01,280 INFO [main] org.apache.hadoop.http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)
2016-06-22 16:49:01,284 INFO [main] org.apache.hadoop.http.HttpServer2: Added filter AM_PROXY_FILTER (class=org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter) to context mapreduce
2016-06-22 16:49:01,285 INFO [main] org.apache.hadoop.http.HttpServer2: Added filter AM_PROXY_FILTER (class=org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter) to context static
2016-06-22 16:49:01,287 INFO [main] org.apache.hadoop.http.HttpServer2: adding path spec: /mapreduce/*
2016-06-22 16:49:01,287 INFO [main] org.apache.hadoop.http.HttpServer2: adding path spec: /ws/*
2016-06-22 16:49:01,294 INFO [main] org.apache.hadoop.http.HttpServer2: Jetty bound to port 40320
2016-06-22 16:49:01,294 INFO [main] org.mortbay.log: jetty-6.1.26.cloudera.4
2016-06-22 16:49:01,317 INFO [main] org.mortbay.log: Extract jar:file:/usr/lib/hadoop-yarn/hadoop-yarn-common-2.6.0-cdh5.4.4.jar!/webapps/mapreduce to /tmp/Jetty_0_0_0_0_40320_mapreduce____yhnhg/webapp
2016-06-22 16:49:01,566 INFO [main] org.mortbay.log: Started HttpServer2$SelectChannelConnectorWithSafeStartup@0.0.0.0:40320
2016-06-22 16:49:01,566 INFO [main] org.apache.hadoop.yarn.webapp.WebApps: Web app /mapreduce started at 40320
2016-06-22 16:49:01,862 INFO [main] org.apache.hadoop.yarn.webapp.WebApps: Registered webapp guice modules
2016-06-22 16:49:01,867 INFO [main] org.apache.hadoop.ipc.CallQueueManager: Using callQueue class java.util.concurrent.LinkedBlockingQueue
2016-06-22 16:49:01,867 INFO [Socket Reader #1 for port 40919] org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 40919
2016-06-22 16:49:01,870 INFO [IPC Server Responder] org.apache.hadoop.ipc.Server: IPC Server Responder: starting
2016-06-22 16:49:01,871 INFO [IPC Server listener on 40919] org.apache.hadoop.ipc.Server: IPC Server listener on 40919: starting
2016-06-22 16:49:01,891 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: nodeBlacklistingEnabled:true
2016-06-22 16:49:01,891 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: maxTaskFailuresPerNode is 3
2016-06-22 16:49:01,891 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: blacklistDisablePercent is 33
2016-06-22 16:49:01,936 INFO [main] org.apache.hadoop.yarn.client.RMProxy: Connecting to ResourceManager at node-name-ib01.domain-name.com/192.168.5.10:8030
2016-06-22 16:49:02,023 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: maxContainerCapability: <memory:8192, vCores:16>
2016-06-22 16:49:02,024 INFO [main] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: queue: root.hdfs
2016-06-22 16:49:02,029 INFO [main] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Upper limit on the thread pool size is 500
2016-06-22 16:49:02,031 INFO [main] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: yarn.client.max-cached-nodemanagers-proxies : 0
2016-06-22 16:49:02,038 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: job_1465331659466_1048Job Transitioned from INITED to SETUP
2016-06-22 16:49:02,040 INFO [CommitterEvent Processor #0] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: JOB_SETUP
2016-06-22 16:49:02,047 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: job_1465331659466_1048Job Transitioned from SETUP to RUNNING
2016-06-22 16:49:02,085 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved dnjbadata-ib01.domain-name.com to /default
2016-06-22 16:49:02,108 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Event Writer setup for JobId: job_1465331659466_1048, File: hdfs://node-name-ib01.domain-name.com:8020/user/hdfs/.staging/job_1465331659466_1048/job_1465331659466_1048_1.jhist
2016-06-22 16:49:02,110 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved dnjbadata-ib02.domain-name.com to /default
2016-06-22 16:49:02,133 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved dnjbadata-ib03.domain-name.com to /default
2016-06-22 16:49:02,136 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1465331659466_1048_m_000000 Task Transitioned from NEW to SCHEDULED
2016-06-22 16:49:02,137 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1465331659466_1048_r_000000 Task Transitioned from NEW to SCHEDULED
2016-06-22 16:49:02,139 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_0 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-22 16:49:02,139 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_r_000000_0 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-22 16:49:02,140 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: mapResourceRequest:<memory:1024, vCores:1>
2016-06-22 16:49:02,146 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: reduceResourceRequest:<memory:1024, vCores:1>
2016-06-22 16:49:02,333 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system [hdfs://node-name-ib01.domain-name.com:8020]
2016-06-22 16:49:03,027 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:0 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:0 ContRel:0 HostLocal:0 RackLocal:0
2016-06-22 16:49:03,115 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_1048: ask=5 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:23866, vCores:63> knownNMs=4
2016-06-22 16:49:03,116 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:23866, vCores:63>
2016-06-22 16:49:03,117 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 1
2016-06-22 16:49:04,127 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated containers 1
2016-06-22 16:49:04,129 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1465331659466_1048_01_000002 to attempt_1465331659466_1048_m_000000_0
2016-06-22 16:49:04,130 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:22842, vCores:62>
2016-06-22 16:49:04,130 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 1
2016-06-22 16:49:04,131 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:1 ContRel:0 HostLocal:1 RackLocal:0
2016-06-22 16:49:04,180 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved dnjbadata-ib03.domain-name.com to /default
2016-06-22 16:49:04,181 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Job jar is not present. Not adding any jar to the list of resources.
2016-06-22 16:49:04,199 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: The job-conf file on the remote FS is /user/hdfs/.staging/job_1465331659466_1048/job.xml
2016-06-22 16:49:04,205 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Adding #0 tokens and #1 secret keys for NM use for launching container
2016-06-22 16:49:04,205 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Size of containertokens_dob is 1
2016-06-22 16:49:04,205 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Putting shuffle token in serviceData
2016-06-22 16:49:04,232 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_0 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED
2016-06-22 16:49:04,236 INFO [ContainerLauncher #0] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1465331659466_1048_01_000002 taskAttempt attempt_1465331659466_1048_m_000000_0
2016-06-22 16:49:04,238 INFO [ContainerLauncher #0] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1465331659466_1048_m_000000_0
2016-06-22 16:49:04,239 INFO [ContainerLauncher #0] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : dnjbadata-ib03.domain-name.com:8041
2016-06-22 16:49:04,294 INFO [ContainerLauncher #0] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1465331659466_1048_m_000000_0 : 13562
2016-06-22 16:49:04,295 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1465331659466_1048_m_000000_0] using containerId: [container_1465331659466_1048_01_000002 on NM: [dnjbadata-ib03.domain-name.com:8041]
2016-06-22 16:49:04,298 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_0 TaskAttempt Transitioned from ASSIGNED to RUNNING
2016-06-22 16:49:04,298 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1465331659466_1048_m_000000 Task Transitioned from SCHEDULED to RUNNING
2016-06-22 16:49:05,133 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_1048: ask=5 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:22842, vCores:62> knownNMs=4
2016-06-22 16:49:05,764 INFO [Socket Reader #1 for port 40919] SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for job_1465331659466_1048 (auth:SIMPLE)
2016-06-22 16:49:05,782 INFO [IPC Server handler 1 on 40919] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : jvm_1465331659466_1048_m_000002 asked for a task
2016-06-22 16:49:05,783 INFO [IPC Server handler 1 on 40919] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: jvm_1465331659466_1048_m_000002 given task: attempt_1465331659466_1048_m_000000_0
2016-06-22 16:59:31,881 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_1048_m_000000_0: AttemptID:attempt_1465331659466_1048_m_000000_0 Timed out after 600 secs
2016-06-22 16:59:31,887 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_0 TaskAttempt Transitioned from RUNNING to FAIL_CONTAINER_CLEANUP
2016-06-22 16:59:31,888 INFO [ContainerLauncher #1] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1465331659466_1048_01_000002 taskAttempt attempt_1465331659466_1048_m_000000_0
2016-06-22 16:59:31,889 INFO [ContainerLauncher #1] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1465331659466_1048_m_000000_0
2016-06-22 16:59:31,889 INFO [ContainerLauncher #1] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : dnjbadata-ib03.domain-name.com:8041
2016-06-22 16:59:31,909 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_0 TaskAttempt Transitioned from FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
2016-06-22 16:59:31,910 INFO [CommitterEvent Processor #1] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: TASK_ABORT
2016-06-22 16:59:31,920 WARN [CommitterEvent Processor #1] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not delete hdfs://node-name-ib01.domain-name.com:8020/user/temp/test_test/_temporary/1/_temporary/attempt_1465331659466_1048_m_000000_0
2016-06-22 16:59:31,922 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_0 TaskAttempt Transitioned from FAIL_TASK_CLEANUP to FAILED
2016-06-22 16:59:31,938 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved dnjbadata-ib01.domain-name.com to /default
2016-06-22 16:59:31,938 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved dnjbadata-ib02.domain-name.com to /default
2016-06-22 16:59:31,938 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved dnjbadata-ib03.domain-name.com to /default
2016-06-22 16:59:31,938 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 1 failures on node dnjbadata-ib03.domain-name.com
2016-06-22 16:59:31,941 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_1 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-22 16:59:31,941 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added attempt_1465331659466_1048_m_000000_1 to list of failed maps
2016-06-22 16:59:31,959 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:1 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:1 ContRel:0 HostLocal:1 RackLocal:0
2016-06-22 16:59:31,961 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_1048: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:22842, vCores:62> knownNMs=4
2016-06-22 16:59:31,962 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:22842, vCores:62>
2016-06-22 16:59:31,962 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 1
2016-06-22 16:59:32,970 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received completed container container_1465331659466_1048_01_000002
2016-06-22 16:59:32,970 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated containers 1
2016-06-22 16:59:32,971 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_1048_m_000000_0: Container killed by the ApplicationMaster.
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143

2016-06-22 16:59:32,971 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning container Container: [ContainerId: container_1465331659466_1048_01_000003, NodeId: node-name-ib02.domain-name.com:8041, NodeHttpAddress: node-name-ib02.domain-name.com:8042, Resource: <memory:1024, vCores:1>, Priority: 5, Token: Token { kind: ContainerToken, service: 192.168.5.9:8041 }, ] to fast fail map
2016-06-22 16:59:32,971 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned from earlierFailedMaps
2016-06-22 16:59:32,971 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1465331659466_1048_01_000003 to attempt_1465331659466_1048_m_000000_1
2016-06-22 16:59:32,971 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:22842, vCores:62>
2016-06-22 16:59:32,972 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 1
2016-06-22 16:59:32,972 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:2 ContRel:0 HostLocal:1 RackLocal:0
2016-06-22 16:59:32,999 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved node-name-ib02.domain-name.com to /default
2016-06-22 16:59:33,000 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_1 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED
2016-06-22 16:59:33,001 INFO [ContainerLauncher #2] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1465331659466_1048_01_000003 taskAttempt attempt_1465331659466_1048_m_000000_1
2016-06-22 16:59:33,001 INFO [ContainerLauncher #2] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1465331659466_1048_m_000000_1
2016-06-22 16:59:33,001 INFO [ContainerLauncher #2] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : node-name-ib02.domain-name.com:8041
2016-06-22 16:59:33,012 INFO [ContainerLauncher #2] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1465331659466_1048_m_000000_1 : 13562
2016-06-22 16:59:33,012 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1465331659466_1048_m_000000_1] using containerId: [container_1465331659466_1048_01_000003 on NM: [node-name-ib02.domain-name.com:8041]
2016-06-22 16:59:33,013 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_1 TaskAttempt Transitioned from ASSIGNED to RUNNING
2016-06-22 16:59:33,975 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_1048: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:22842, vCores:62> knownNMs=4
2016-06-22 16:59:34,226 INFO [Socket Reader #1 for port 40919] SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for job_1465331659466_1048 (auth:SIMPLE)
2016-06-22 16:59:34,239 INFO [IPC Server handler 29 on 40919] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : jvm_1465331659466_1048_m_000003 asked for a task
2016-06-22 16:59:34,239 INFO [IPC Server handler 29 on 40919] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: jvm_1465331659466_1048_m_000003 given task: attempt_1465331659466_1048_m_000000_1
2016-06-22 17:10:01,885 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_1048_m_000000_1: AttemptID:attempt_1465331659466_1048_m_000000_1 Timed out after 600 secs
2016-06-22 17:10:01,886 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_1 TaskAttempt Transitioned from RUNNING to FAIL_CONTAINER_CLEANUP
2016-06-22 17:10:01,887 INFO [ContainerLauncher #3] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1465331659466_1048_01_000003 taskAttempt attempt_1465331659466_1048_m_000000_1
2016-06-22 17:10:01,887 INFO [ContainerLauncher #3] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1465331659466_1048_m_000000_1
2016-06-22 17:10:01,888 INFO [ContainerLauncher #3] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : node-name-ib02.domain-name.com:8041
2016-06-22 17:10:01,895 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_1 TaskAttempt Transitioned from FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
2016-06-22 17:10:01,896 INFO [CommitterEvent Processor #2] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: TASK_ABORT
2016-06-22 17:10:01,900 WARN [CommitterEvent Processor #2] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not delete hdfs://node-name-ib01.domain-name.com:8020/user/temp/test_test/_temporary/1/_temporary/attempt_1465331659466_1048_m_000000_1
2016-06-22 17:10:01,900 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_1 TaskAttempt Transitioned from FAIL_TASK_CLEANUP to FAILED
2016-06-22 17:10:01,901 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved dnjbadata-ib01.domain-name.com to /default
2016-06-22 17:10:01,901 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved dnjbadata-ib02.domain-name.com to /default
2016-06-22 17:10:01,901 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved dnjbadata-ib03.domain-name.com to /default
2016-06-22 17:10:01,902 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 1 failures on node node-name-ib02.domain-name.com
2016-06-22 17:10:01,902 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_2 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-22 17:10:01,903 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added attempt_1465331659466_1048_m_000000_2 to list of failed maps
2016-06-22 17:10:02,717 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:1 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:2 ContRel:0 HostLocal:1 RackLocal:0
2016-06-22 17:10:02,719 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_1048: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:22842, vCores:62> knownNMs=4
2016-06-22 17:10:02,719 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:22842, vCores:62>
2016-06-22 17:10:02,719 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 1
2016-06-22 17:10:03,721 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received completed container container_1465331659466_1048_01_000003
2016-06-22 17:10:03,721 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated containers 1
2016-06-22 17:10:03,721 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning container Container: [ContainerId: container_1465331659466_1048_01_000004, NodeId: node-name-ib02.domain-name.com:8041, NodeHttpAddress: node-name-ib02.domain-name.com:8042, Resource: <memory:1024, vCores:1>, Priority: 5, Token: Token { kind: ContainerToken, service: 192.168.5.9:8041 }, ] to fast fail map
2016-06-22 17:10:03,721 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_1048_m_000000_1: Container killed by the ApplicationMaster.
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143

2016-06-22 17:10:03,721 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned from earlierFailedMaps
2016-06-22 17:10:03,722 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1465331659466_1048_01_000004 to attempt_1465331659466_1048_m_000000_2
2016-06-22 17:10:03,722 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:22842, vCores:62>
2016-06-22 17:10:03,722 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 1
2016-06-22 17:10:03,722 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:3 ContRel:0 HostLocal:1 RackLocal:0
2016-06-22 17:10:03,722 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved node-name-ib02.domain-name.com to /default
2016-06-22 17:10:03,723 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_2 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED
2016-06-22 17:10:03,724 INFO [ContainerLauncher #4] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1465331659466_1048_01_000004 taskAttempt attempt_1465331659466_1048_m_000000_2
2016-06-22 17:10:03,724 INFO [ContainerLauncher #4] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1465331659466_1048_m_000000_2
2016-06-22 17:10:03,724 INFO [ContainerLauncher #4] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : node-name-ib02.domain-name.com:8041
2016-06-22 17:10:03,732 INFO [ContainerLauncher #4] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1465331659466_1048_m_000000_2 : 13562
2016-06-22 17:10:03,732 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1465331659466_1048_m_000000_2] using containerId: [container_1465331659466_1048_01_000004 on NM: [node-name-ib02.domain-name.com:8041]
2016-06-22 17:10:03,732 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_2 TaskAttempt Transitioned from ASSIGNED to RUNNING
2016-06-22 17:10:04,725 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_1048: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:22842, vCores:62> knownNMs=4
2016-06-22 17:10:04,914 INFO [Socket Reader #1 for port 40919] SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for job_1465331659466_1048 (auth:SIMPLE)
2016-06-22 17:10:04,927 INFO [IPC Server handler 26 on 40919] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : jvm_1465331659466_1048_m_000004 asked for a task
2016-06-22 17:10:04,927 INFO [IPC Server handler 26 on 40919] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: jvm_1465331659466_1048_m_000004 given task: attempt_1465331659466_1048_m_000000_2
2016-06-22 17:20:31,889 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_1048_m_000000_2: AttemptID:attempt_1465331659466_1048_m_000000_2 Timed out after 600 secs
2016-06-22 17:20:31,890 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_2 TaskAttempt Transitioned from RUNNING to FAIL_CONTAINER_CLEANUP
2016-06-22 17:20:31,891 INFO [ContainerLauncher #5] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1465331659466_1048_01_000004 taskAttempt attempt_1465331659466_1048_m_000000_2
2016-06-22 17:20:31,891 INFO [ContainerLauncher #5] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1465331659466_1048_m_000000_2
2016-06-22 17:20:31,892 INFO [ContainerLauncher #5] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : node-name-ib02.domain-name.com:8041
2016-06-22 17:20:31,898 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_2 TaskAttempt Transitioned from FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
2016-06-22 17:20:31,898 INFO [CommitterEvent Processor #3] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: TASK_ABORT
2016-06-22 17:20:31,902 WARN [CommitterEvent Processor #3] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not delete hdfs://node-name-ib01.domain-name.com:8020/user/temp/test_test/_temporary/1/_temporary/attempt_1465331659466_1048_m_000000_2
2016-06-22 17:20:31,903 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_2 TaskAttempt Transitioned from FAIL_TASK_CLEANUP to FAILED
2016-06-22 17:20:31,903 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved dnjbadata-ib01.domain-name.com to /default
2016-06-22 17:20:31,903 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved dnjbadata-ib02.domain-name.com to /default
2016-06-22 17:20:31,904 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved dnjbadata-ib03.domain-name.com to /default
2016-06-22 17:20:31,904 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_3 TaskAttempt Transitioned from NEW to UNASSIGNED
2016-06-22 17:20:31,904 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 2 failures on node node-name-ib02.domain-name.com
2016-06-22 17:20:31,905 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Added attempt_1465331659466_1048_m_000000_3 to list of failed maps
2016-06-22 17:20:32,477 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:1 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:3 ContRel:0 HostLocal:1 RackLocal:0
2016-06-22 17:20:32,479 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_1048: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:22842, vCores:62> knownNMs=4
2016-06-22 17:20:32,479 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:22842, vCores:62>
2016-06-22 17:20:32,479 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 1
2016-06-22 17:20:33,481 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received completed container container_1465331659466_1048_01_000004
2016-06-22 17:20:33,481 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got allocated containers 1
2016-06-22 17:20:33,481 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning container Container: [ContainerId: container_1465331659466_1048_01_000005, NodeId: node-name-ib02.domain-name.com:8041, NodeHttpAddress: node-name-ib02.domain-name.com:8042, Resource: <memory:1024, vCores:1>, Priority: 5, Token: Token { kind: ContainerToken, service: 192.168.5.9:8041 }, ] to fast fail map
2016-06-22 17:20:33,481 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_1048_m_000000_2: Container killed by the ApplicationMaster.
Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143

2016-06-22 17:20:33,481 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned from earlierFailedMaps
2016-06-22 17:20:33,481 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned container container_1465331659466_1048_01_000005 to attempt_1465331659466_1048_m_000000_3
2016-06-22 17:20:33,481 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Recalculating schedule, headroom=<memory:22842, vCores:62>
2016-06-22 17:20:33,482 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow start threshold not met. completedMapsForReduceSlowstart 1
2016-06-22 17:20:33,482 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:4 ContRel:0 HostLocal:1 RackLocal:0
2016-06-22 17:20:33,482 INFO [AsyncDispatcher event handler] org.apache.hadoop.yarn.util.RackResolver: Resolved node-name-ib02.domain-name.com to /default
2016-06-22 17:20:33,483 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_3 TaskAttempt Transitioned from UNASSIGNED to ASSIGNED
2016-06-22 17:20:33,484 INFO [ContainerLauncher #6] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container container_1465331659466_1048_01_000005 taskAttempt attempt_1465331659466_1048_m_000000_3
2016-06-22 17:20:33,484 INFO [ContainerLauncher #6] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Launching attempt_1465331659466_1048_m_000000_3
2016-06-22 17:20:33,484 INFO [ContainerLauncher #6] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : node-name-ib02.domain-name.com:8041
2016-06-22 17:20:33,491 INFO [ContainerLauncher #6] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Shuffle port returned by ContainerManager for attempt_1465331659466_1048_m_000000_3 : 13562
2016-06-22 17:20:33,491 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: TaskAttempt: [attempt_1465331659466_1048_m_000000_3] using containerId: [container_1465331659466_1048_01_000005 on NM: [node-name-ib02.domain-name.com:8041]
2016-06-22 17:20:33,491 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_3 TaskAttempt Transitioned from ASSIGNED to RUNNING
2016-06-22 17:20:34,483 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1465331659466_1048: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:22842, vCores:62> knownNMs=4
2016-06-22 17:20:34,719 INFO [Socket Reader #1 for port 40919] SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for job_1465331659466_1048 (auth:SIMPLE)
2016-06-22 17:20:34,734 INFO [IPC Server handler 10 on 40919] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : jvm_1465331659466_1048_m_000005 asked for a task
2016-06-22 17:20:34,734 INFO [IPC Server handler 10 on 40919] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: jvm_1465331659466_1048_m_000005 given task: attempt_1465331659466_1048_m_000000_3
2016-06-22 17:31:01,893 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1465331659466_1048_m_000000_3: AttemptID:attempt_1465331659466_1048_m_000000_3 Timed out after 600 secs
2016-06-22 17:31:01,893 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_3 TaskAttempt Transitioned from RUNNING to FAIL_CONTAINER_CLEANUP
2016-06-22 17:31:01,894 INFO [ContainerLauncher #7] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container container_1465331659466_1048_01_000005 taskAttempt attempt_1465331659466_1048_m_000000_3
2016-06-22 17:31:01,895 INFO [ContainerLauncher #7] org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: KILLING attempt_1465331659466_1048_m_000000_3
2016-06-22 17:31:01,895 INFO [ContainerLauncher #7] org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: Opening proxy : node-name-ib02.domain-name.com:8041
2016-06-22 17:31:01,902 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_3 TaskAttempt Transitioned from FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP
2016-06-22 17:31:01,902 INFO [CommitterEvent Processor #4] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: TASK_ABORT
2016-06-22 17:31:01,906 WARN [CommitterEvent Processor #4] org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not delete hdfs://node-name-ib01.domain-name.com:8020/user/temp/test_test/_temporary/1/_temporary/attempt_1465331659466_1048_m_000000_3
2016-06-22 17:31:01,907 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_m_000000_3 TaskAttempt Transitioned from FAIL_TASK_CLEANUP to FAILED
2016-06-22 17:31:01,910 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1465331659466_1048_m_000000 Task Transitioned from RUNNING to FAILED
2016-06-22 17:31:01,910 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: 3 failures on node node-name-ib02.domain-name.com
2016-06-22 17:31:01,910 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: Blacklisted host node-name-ib02.domain-name.com
2016-06-22 17:31:01,910 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Num completed Tasks: 1
2016-06-22 17:31:01,910 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Job failed as tasks failed. failedMaps:1 failedReduces:0
2016-06-22 17:31:01,912 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: job_1465331659466_1048Job Transitioned from RUNNING to FAIL_WAIT
2016-06-22 17:31:01,912 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1465331659466_1048_r_000000 Task Transitioned from SCHEDULED to KILL_WAIT
2016-06-22 17:31:01,912 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: attempt_1465331659466_1048_r_000000_0 TaskAttempt Transitioned from UNASSIGNED to KILLED
2016-06-22 17:31:01,913 INFO [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Processing the event EventType: CONTAINER_DEALLOCATE
2016-06-22 17:31:01,913 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1465331659466_1048_r_000000 Task Transitioned from KILL_WAIT to KILLED
2016-06-22 17:31:01,913 ERROR [Thread-50] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Could not deallocate container for task attemptId attempt_1465331659466_1048_r_000000_0
2016-06-22 17:31:01,914 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: job_1465331659466_1048Job Transitioned from FAIL_WAIT to FAIL_ABORT
2016-06-22 17:31:01,914 INFO [CommitterEvent Processor #0] org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: Processing the event EventType: JOB_ABORT
2016-06-22 17:31:01,921 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: job_1465331659466_1048Job Transitioned from FAIL_ABORT to FAILED
2016-06-22 17:31:01,922 INFO [Thread-158] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: We are finishing cleanly so this is the last retry
2016-06-22 17:31:01,922 INFO [Thread-158] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Notify RMCommunicator isAMLastRetry: true
2016-06-22 17:31:01,922 INFO [Thread-158] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: RMCommunicator notified that shouldUnregistered is: true
2016-06-22 17:31:01,922 INFO [Thread-158] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Notify JHEH isAMLastRetry: true
2016-06-22 17:31:01,922 INFO [Thread-158] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: JobHistoryEventHandler notified that forceJobCompletion is true
2016-06-22 17:31:01,922 INFO [Thread-158] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Calling stop for all the services
2016-06-22 17:31:01,923 INFO [Thread-158] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Stopping JobHistoryEventHandler. Size of the outstanding queue size is 0
2016-06-22 17:31:01,949 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copying hdfs://node-name-ib01.domain-name.com:8020/user/hdfs/.staging/job_1465331659466_1048/job_1465331659466_1048_1.jhist to hdfs://node-name-ib01.domain-name.com:8020/user/history/done_intermediate/hdfs/job_1465331659466_1048-1466628538199-hdfs-pydoop-1466631061910-0-0-FAILED-root.hdfs-1466628542033.jhist_tmp
2016-06-22 17:31:01,971 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copied to done location: hdfs://node-name-ib01.domain-name.com:8020/user/history/done_intermediate/hdfs/job_1465331659466_1048-1466628538199-hdfs-pydoop-1466631061910-0-0-FAILED-root.hdfs-1466628542033.jhist_tmp
2016-06-22 17:31:01,973 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copying hdfs://node-name-ib01.domain-name.com:8020/user/hdfs/.staging/job_1465331659466_1048/job_1465331659466_1048_1_conf.xml to hdfs://node-name-ib01.domain-name.com:8020/user/history/done_intermediate/hdfs/job_1465331659466_1048_conf.xml_tmp
2016-06-22 17:31:01,996 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copied to done location: hdfs://node-name-ib01.domain-name.com:8020/user/history/done_intermediate/hdfs/job_1465331659466_1048_conf.xml_tmp
2016-06-22 17:31:02,001 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved tmp to done: hdfs://node-name-ib01.domain-name.com:8020/user/history/done_intermediate/hdfs/job_1465331659466_1048.summary_tmp to hdfs://node-name-ib01.domain-name.com:8020/user/history/done_intermediate/hdfs/job_1465331659466_1048.summary
2016-06-22 17:31:02,002 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved tmp to done: hdfs://node-name-ib01.domain-name.com:8020/user/history/done_intermediate/hdfs/job_1465331659466_1048_conf.xml_tmp to hdfs://node-name-ib01.domain-name.com:8020/user/history/done_intermediate/hdfs/job_1465331659466_1048_conf.xml
2016-06-22 17:31:02,003 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved tmp to done: hdfs://node-name-ib01.domain-name.com:8020/user/history/done_intermediate/hdfs/job_1465331659466_1048-1466628538199-hdfs-pydoop-1466631061910-0-0-FAILED-root.hdfs-1466628542033.jhist_tmp to hdfs://node-name-ib01.domain-name.com:8020/user/history/done_intermediate/hdfs/job_1465331659466_1048-1466628538199-hdfs-pydoop-1466631061910-0-0-FAILED-root.hdfs-1466628542033.jhist
2016-06-22 17:31:02,003 INFO [Thread-158] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Stopped JobHistoryEventHandler. super.stop()
2016-06-22 17:31:02,007 INFO [Thread-158] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Setting job diagnostics to Task failed task_1465331659466_1048_m_000000
Job failed as tasks failed. failedMaps:1 failedReduces:0

2016-06-22 17:31:02,007 INFO [Thread-158] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: History url is http://node-name-ib01.domain-name.com:19888/jobhistory/job/job_1465331659466_1048
2016-06-22 17:31:02,014 INFO [Thread-158] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Waiting for application to be successfully unregistered.
2016-06-22 17:31:03,016 INFO [Thread-158] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Final Stats: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1 AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:4 ContRel:0 HostLocal:1 RackLocal:0
2016-06-22 17:31:03,016 INFO [Thread-158] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Deleting staging directory hdfs://node-name-ib01.domain-name.com:8020 /user/hdfs/.staging/job_1465331659466_1048
2016-06-22 17:31:03,018 INFO [Thread-158] org.apache.hadoop.ipc.Server: Stopping server on 40919
2016-06-22 17:31:03,020 INFO [IPC Server listener on 40919] org.apache.hadoop.ipc.Server: Stopping IPC Server listener on 40919
2016-06-22 17:31:03,020 INFO [IPC Server Responder] org.apache.hadoop.ipc.Server: Stopping IPC Server Responder
2016-06-22 17:31:03,020 INFO [TaskHeartbeatHandler PingChecker] org.apache.hadoop.mapreduce.v2.app.TaskHeartbeatHandler: TaskHeartbeatHandler thread interrupted

@ilveroluca
Copy link
Member

Hi. Do you see anything in the stderr or stdout logs for a task? Often this sort of failure coincides with the Python interpreter spitting out a message on stderr that can give us a clue as to exactly what's going on.

Try running an example from the Pydoop installation. For instance, from the root Pydoop directory,

(cd examples/pydoop_submit && PATH=$HADOOP_PREFIX/bin:$PATH ./run_wordcount_minimal)

If it runs, then we'll know that Pydoop is properly installed on your cluster. If it doesn't, then we can try to figure out what's wrong with the installation.

Luca

@igorvishnevskiy
Copy link
Author

igorvishnevskiy commented Jun 23, 2016

Hello @ilveroluca. I did just what you have said. stderr logs are bellow. stdout logs bucket is empty, not a single line of log there.
I ran the example using command:
(cd igor/wordcount/ && PATH=$HADOOP_PREFIX/bin:$PATH ./run_wc_minimal)
I checked out example from the following repo:
git clone https://github.com/crs4/pydoop.git

There is also example with new pydoop's api.
(cd igor/wordcount/ && PATH=$HADOOP_PREFIX/bin:$PATH ./run_wc_minimal_new_api)
I tried that as well, the same problem.

I wanted to add that I installed pydoop using pip install, maybe I should reinstall it using source code from the repository? Maybe there is something wrong with a build installed by pip? Is there any way to get an older build of pydoop to try and see if that would work? Thank you.

Jun 23, 2016 2:29:25 PM com.google.inject.servlet.InternalServletModule$BackwardsCompatibleServletContextProvider get
WARNING: You are attempting to use a deprecated API (specifically, attempting to @Inject ServletContext inside an eagerly created singleton. While we allow this for backwards compatibility, be warned that this MAY have unexpected behavior if you have more than one injector (with ServletModule) running in the same JVM. Please consult the Guice documentation at http://code.google.com/p/google-guice/wiki/Servlets for more information.
Jun 23, 2016 2:29:26 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory register
INFO: Registering org.apache.hadoop.mapreduce.v2.app.webapp.JAXBContextResolver as a provider class
Jun 23, 2016 2:29:26 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory register
INFO: Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class
Jun 23, 2016 2:29:26 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory register
INFO: Registering org.apache.hadoop.mapreduce.v2.app.webapp.AMWebServices as a root resource class
Jun 23, 2016 2:29:26 PM com.sun.jersey.server.impl.application.WebApplicationImpl _initiate
INFO: Initiating Jersey application, version 'Jersey: 1.9 09/02/2011 11:17 AM'
Jun 23, 2016 2:29:26 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory getComponentProvider
INFO: Binding org.apache.hadoop.mapreduce.v2.app.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton"
Jun 23, 2016 2:29:26 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory getComponentProvider
INFO: Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton"
Jun 23, 2016 2:29:26 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory getComponentProvider
INFO: Binding org.apache.hadoop.mapreduce.v2.app.webapp.AMWebServices to GuiceManagedComponentProvider with the scope "PerRequest"

Thank you so much for your help guys!

@igorvishnevskiy
Copy link
Author

igorvishnevskiy commented Jun 23, 2016

Hello @ilveroluca.

After setting paths for:
export HADOOP_HOME=/usr/lib/hadoop/bin/
export HADOOP_CONF_DIR=/usr/lib/hadoop/etc/hadoop
export HDFS_HOST=HOST_ADDRESS
export HDFS_PORT=PORT
I started getting the following exception. Not sure if it's a good sign, but at least now we see errors.

cd igor/wordcount/ && PATH=/usr/lib/hadoop/bin:$PATH ./run_wc_minimal
loadFileSystems error:
(unable to get stack trace for java.lang.NoClassDefFoundError exception: ExceptionUtils::getStackTrace error.)
hdfsBuilderConnect(forceNewInstance=0, nn=default, port=0, kerbTicketCachePath=(NULL), userName=(NULL)) error:
(unable to get stack trace for java.lang.NoClassDefFoundError exception: ExceptionUtils::getStackTrace error.)
Traceback (most recent call last):
  File "run_wc.py", line 79, in <module>
    main(sys.argv[1:])
  File "run_wc.py", line 66, in main
    runner = hadut.PipesRunner(prefix=PREFIX, logger=logger)
  File "/usr/lib64/python2.7/site-packages/pydoop/hadut.py", line 457, in __init__
    hdfs.mkdir(self.wd)
  File "/usr/lib64/python2.7/site-packages/pydoop/hdfs/__init__.py", line 262, in mkdir
    fs = hdfs(host, port, user)
  File "/usr/lib64/python2.7/site-packages/pydoop/hdfs/fs.py", line 150, in __init__
    h, p, u, fs = _get_connection_info(host, port, user)
  File "/usr/lib64/python2.7/site-packages/pydoop/hdfs/fs.py", line 64, in _get_connection_info
    fs = core_hdfs_fs(host, port, user)
  File "/usr/lib64/python2.7/site-packages/pydoop/hdfs/core/__init__.py", line 57, in core_hdfs_fs
    return _CORE_MODULE.CoreHdfsFs(host, port, user)
RuntimeError: (255, 'Unknown error 255')

Found one more online thread where someone else faced similar, ran through the commands, environment is set correctly looks like, still I get above exception:
(https://sourceforge.net/p/pydoop/discussion/990018/thread/123f50b8/)

>>> import pydoop
>>> pydoop.hadoop_home()
'/usr/lib/hadoop'
>>> pydoop.hadoop_exec()
'/usr/lib/hadoop/bin/hadoop'
>>> pydoop.hadoop_version()
'2.6.0-cdh5.4.4'
>>> for p in pydoop.hadoop_classpath().split(":"): print p
... 
/usr/lib/hadoop/parquet-thrift.jar
/usr/lib/hadoop/parquet-generator.jar
/usr/lib/hadoop/hadoop-auth-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop/hadoop-common-2.6.0-cdh5.4.4-tests.jar
/usr/lib/hadoop/parquet-scrooge_2.10.jar
/usr/lib/hadoop/parquet-format-javadoc.jar
/usr/lib/hadoop/hadoop-annotations-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop/parquet-hadoop-bundle.jar
/usr/lib/hadoop/hadoop-common-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop/parquet-encoding.jar
/usr/lib/hadoop/hadoop-annotations.jar
/usr/lib/hadoop/parquet-format.jar
/usr/lib/hadoop/parquet-format-sources.jar
/usr/lib/hadoop/parquet-common.jar
/usr/lib/hadoop/hadoop-aws.jar
/usr/lib/hadoop/parquet-pig-bundle.jar
/usr/lib/hadoop/hadoop-auth.jar
/usr/lib/hadoop/hadoop-nfs.jar
/usr/lib/hadoop/parquet-hadoop.jar
/usr/lib/hadoop/parquet-test-hadoop2.jar
/usr/lib/hadoop/hadoop-common-tests.jar
/usr/lib/hadoop/parquet-pig.jar
/usr/lib/hadoop/parquet-protobuf.jar
/usr/lib/hadoop/hadoop-nfs-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop/parquet-jackson.jar
/usr/lib/hadoop/hadoop-aws-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop/parquet-cascading.jar
/usr/lib/hadoop/hadoop-common.jar
/usr/lib/hadoop/parquet-scala_2.10.jar
/usr/lib/hadoop/parquet-tools.jar
/usr/lib/hadoop/hadoop-streaming-1.1.1.jar
/usr/lib/hadoop/parquet-avro.jar
/usr/lib/hadoop/parquet-column.jar
/usr/lib/hadoop/lib/aws-java-sdk-1.7.4.jar
/usr/lib/hadoop/lib/jsch-0.1.42.jar
/usr/lib/hadoop/lib/mysql-connector-java-5.1.36-bin.jar
/usr/lib/hadoop/lib/apacheds-kerberos-codec-2.0.0-M15.jar
/usr/lib/hadoop/lib/zookeeper.jar
/usr/lib/hadoop/lib/gson-2.1.jar
/usr/lib/hadoop/lib/mockito-all-1.8.5.jar
/usr/lib/hadoop/lib/jersey-core-1.9.jar
/usr/lib/hadoop/lib/commons-net-3.1.jar
/usr/lib/hadoop/lib/commons-compress-1.4.1.jar
/usr/lib/hadoop/lib/slf4j-api-1.7.5.jar
/usr/lib/hadoop/lib/avro.jar
/usr/lib/hadoop/lib/api-asn1-api-1.0.0-M20.jar
/usr/lib/hadoop/lib/guava-11.0.2.jar
/usr/lib/hadoop/lib/jets3t-0.9.0.jar
/usr/lib/hadoop/lib/jersey-json-1.9.jar
/usr/lib/hadoop/lib/jaxb-impl-2.2.3-1.jar
/usr/lib/hadoop/lib/jetty-util-6.1.26.cloudera.4.jar
/usr/lib/hadoop/lib/jsr305-3.0.0.jar
/usr/lib/hadoop/lib/apacheds-i18n-2.0.0-M15.jar
/usr/lib/hadoop/lib/netty-3.6.2.Final.jar
/usr/lib/hadoop/lib/hue-plugins-3.7.0-cdh5.4.4.jar
/usr/lib/hadoop/lib/htrace-core-3.0.4.jar
/usr/lib/hadoop/lib/servlet-api-2.5.jar
/usr/lib/hadoop/lib/jackson-jaxrs-1.8.8.jar
/usr/lib/hadoop/lib/stax-api-1.0-2.jar
/usr/lib/hadoop/lib/commons-collections-3.2.1.jar
/usr/lib/hadoop/lib/commons-logging-1.1.3.jar
/usr/lib/hadoop/lib/hamcrest-core-1.3.jar
/usr/lib/hadoop/lib/slf4j-log4j12.jar
/usr/lib/hadoop/lib/java-xmlbuilder-0.4.jar
/usr/lib/hadoop/lib/httpclient-4.2.5.jar
/usr/lib/hadoop/lib/asm-3.2.jar
/usr/lib/hadoop/lib/httpcore-4.2.5.jar
/usr/lib/hadoop/lib/commons-beanutils-1.7.0.jar
/usr/lib/hadoop/lib/commons-el-1.0.jar
/usr/lib/hadoop/lib/xz-1.0.jar
/usr/lib/hadoop/lib/curator-client-2.7.1.jar
/usr/lib/hadoop/lib/commons-codec-1.4.jar
/usr/lib/hadoop/lib/api-util-1.0.0-M20.jar
/usr/lib/hadoop/lib/commons-beanutils-core-1.8.0.jar
/usr/lib/hadoop/lib/gson-2.2.4.jar
/usr/lib/hadoop/lib/jasper-runtime-5.5.23.jar
/usr/lib/hadoop/lib/snappy-java-1.0.4.1.jar
/usr/lib/hadoop/lib/protobuf-java-2.5.0.jar
/usr/lib/hadoop/lib/commons-io-2.4.jar
/usr/lib/hadoop/lib/jackson-core-asl-1.8.8.jar
/usr/lib/hadoop/lib/logredactor-1.0.3.jar
/usr/lib/hadoop/lib/jetty-6.1.26.cloudera.4.jar
/usr/lib/hadoop/lib/curator-recipes-2.7.1.jar
/usr/lib/hadoop/lib/jasper-compiler-5.5.23.jar
/usr/lib/hadoop/lib/jaxb-api-2.2.2.jar
/usr/lib/hadoop/lib/jackson-mapper-asl-1.8.8.jar
/usr/lib/hadoop/lib/junit-4.11.jar
/usr/lib/hadoop/lib/jsp-api-2.1.jar
/usr/lib/hadoop/lib/commons-math3-3.1.1.jar
/usr/lib/hadoop/lib/activation-1.1.jar
/usr/lib/hadoop/lib/curator-framework-2.7.1.jar
/usr/lib/hadoop/lib/xmlenc-0.52.jar
/usr/lib/hadoop/lib/jackson-xc-1.8.8.jar
/usr/lib/hadoop/lib/commons-configuration-1.6.jar
/usr/lib/hadoop/lib/commons-httpclient-3.1.jar
/usr/lib/hadoop/lib/jersey-server-1.9.jar
/usr/lib/hadoop/lib/commons-cli-1.2.jar
/usr/lib/hadoop/lib/commons-digester-1.8.jar
/usr/lib/hadoop/lib/paranamer-2.3.jar
/usr/lib/hadoop/lib/log4j-1.2.17.jar
/usr/lib/hadoop/lib/jettison-1.1.jar
/usr/lib/hadoop/lib/commons-lang-2.6.jar
/usr/lib/hadoop-hdfs/hadoop-hdfs-tests.jar
/usr/lib/hadoop-hdfs/hadoop-hdfs.jar
/usr/lib/hadoop-hdfs/hadoop-hdfs-2.6.0-cdh5.4.4-tests.jar
/usr/lib/hadoop-hdfs/hadoop-hdfs-nfs-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-hdfs/hadoop-hdfs-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-hdfs/hadoop-hdfs-nfs.jar
/usr/lib/hadoop-hdfs/lib/leveldbjni-all-1.8.jar
/usr/lib/hadoop-hdfs/lib/jersey-core-1.9.jar
/usr/lib/hadoop-hdfs/lib/guava-11.0.2.jar
/usr/lib/hadoop-hdfs/lib/jetty-util-6.1.26.cloudera.4.jar
/usr/lib/hadoop-hdfs/lib/jsr305-3.0.0.jar
/usr/lib/hadoop-hdfs/lib/netty-3.6.2.Final.jar
/usr/lib/hadoop-hdfs/lib/htrace-core-3.0.4.jar
/usr/lib/hadoop-hdfs/lib/servlet-api-2.5.jar
/usr/lib/hadoop-hdfs/lib/commons-logging-1.1.3.jar
/usr/lib/hadoop-hdfs/lib/asm-3.2.jar
/usr/lib/hadoop-hdfs/lib/commons-el-1.0.jar
/usr/lib/hadoop-hdfs/lib/commons-codec-1.4.jar
/usr/lib/hadoop-hdfs/lib/jasper-runtime-5.5.23.jar
/usr/lib/hadoop-hdfs/lib/protobuf-java-2.5.0.jar
/usr/lib/hadoop-hdfs/lib/commons-io-2.4.jar
/usr/lib/hadoop-hdfs/lib/jackson-core-asl-1.8.8.jar
/usr/lib/hadoop-hdfs/lib/jetty-6.1.26.cloudera.4.jar
/usr/lib/hadoop-hdfs/lib/jackson-mapper-asl-1.8.8.jar
/usr/lib/hadoop-hdfs/lib/jsp-api-2.1.jar
/usr/lib/hadoop-hdfs/lib/xmlenc-0.52.jar
/usr/lib/hadoop-hdfs/lib/commons-daemon-1.0.13.jar
/usr/lib/hadoop-hdfs/lib/jersey-server-1.9.jar
/usr/lib/hadoop-hdfs/lib/commons-cli-1.2.jar
/usr/lib/hadoop-hdfs/lib/log4j-1.2.17.jar
/usr/lib/hadoop-hdfs/lib/commons-lang-2.6.jar
/usr/lib/hadoop-yarn/hadoop-yarn-api.jar
/usr/lib/hadoop-yarn/hadoop-yarn-server-tests-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-yarn/hadoop-yarn-api-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-yarn/hadoop-yarn-applications-unmanaged-am-launcher.jar
/usr/lib/hadoop-yarn/hadoop-yarn-server-applicationhistoryservice.jar
/usr/lib/hadoop-yarn/hadoop-yarn-registry-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-yarn/hadoop-yarn-server-nodemanager.jar
/usr/lib/hadoop-yarn/hadoop-yarn-applications-distributedshell-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-yarn/hadoop-yarn-common.jar
/usr/lib/hadoop-yarn/hadoop-yarn-registry.jar
/usr/lib/hadoop-yarn/hadoop-yarn-applications-unmanaged-am-launcher-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-yarn/hadoop-yarn-client-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-yarn/hadoop-yarn-server-common.jar
/usr/lib/hadoop-yarn/hadoop-yarn-server-nodemanager-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-yarn/hadoop-yarn-server-web-proxy-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-yarn/hadoop-yarn-server-resourcemanager-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-yarn/hadoop-yarn-common-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-yarn/hadoop-yarn-server-applicationhistoryservice-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-yarn/hadoop-yarn-server-resourcemanager.jar
/usr/lib/hadoop-yarn/hadoop-yarn-client.jar
/usr/lib/hadoop-yarn/hadoop-yarn-server-common-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-yarn/hadoop-yarn-applications-distributedshell.jar
/usr/lib/hadoop-yarn/hadoop-yarn-server-tests.jar
/usr/lib/hadoop-yarn/hadoop-yarn-server-web-proxy.jar
/usr/lib/hadoop-yarn/lib/leveldbjni-all-1.8.jar
/usr/lib/hadoop-yarn/lib/zookeeper.jar
/usr/lib/hadoop-yarn/lib/jersey-core-1.9.jar
/usr/lib/hadoop-yarn/lib/commons-compress-1.4.1.jar
/usr/lib/hadoop-yarn/lib/javax.inject-1.jar
/usr/lib/hadoop-yarn/lib/guava-11.0.2.jar
/usr/lib/hadoop-yarn/lib/jersey-json-1.9.jar
/usr/lib/hadoop-yarn/lib/jaxb-impl-2.2.3-1.jar
/usr/lib/hadoop-yarn/lib/jetty-util-6.1.26.cloudera.4.jar
/usr/lib/hadoop-yarn/lib/jsr305-3.0.0.jar
/usr/lib/hadoop-yarn/lib/servlet-api-2.5.jar
/usr/lib/hadoop-yarn/lib/jackson-jaxrs-1.8.8.jar
/usr/lib/hadoop-yarn/lib/stax-api-1.0-2.jar
/usr/lib/hadoop-yarn/lib/commons-collections-3.2.1.jar
/usr/lib/hadoop-yarn/lib/commons-logging-1.1.3.jar
/usr/lib/hadoop-yarn/lib/asm-3.2.jar
/usr/lib/hadoop-yarn/lib/aopalliance-1.0.jar
/usr/lib/hadoop-yarn/lib/xz-1.0.jar
/usr/lib/hadoop-yarn/lib/commons-codec-1.4.jar
/usr/lib/hadoop-yarn/lib/guice-servlet-3.0.jar
/usr/lib/hadoop-yarn/lib/guice-3.0.jar
/usr/lib/hadoop-yarn/lib/protobuf-java-2.5.0.jar
/usr/lib/hadoop-yarn/lib/commons-io-2.4.jar
/usr/lib/hadoop-yarn/lib/jackson-core-asl-1.8.8.jar
/usr/lib/hadoop-yarn/lib/jetty-6.1.26.cloudera.4.jar
/usr/lib/hadoop-yarn/lib/jaxb-api-2.2.2.jar
/usr/lib/hadoop-yarn/lib/jackson-mapper-asl-1.8.8.jar
/usr/lib/hadoop-yarn/lib/activation-1.1.jar
/usr/lib/hadoop-yarn/lib/jackson-xc-1.8.8.jar
/usr/lib/hadoop-yarn/lib/commons-httpclient-3.1.jar
/usr/lib/hadoop-yarn/lib/jersey-server-1.9.jar
/usr/lib/hadoop-yarn/lib/commons-cli-1.2.jar
/usr/lib/hadoop-yarn/lib/jersey-guice-1.9.jar
/usr/lib/hadoop-yarn/lib/jersey-client-1.9.jar
/usr/lib/hadoop-yarn/lib/log4j-1.2.17.jar
/usr/lib/hadoop-yarn/lib/jettison-1.1.jar
/usr/lib/hadoop-yarn/lib/spark-1.3.0-cdh5.4.4-yarn-shuffle.jar
/usr/lib/hadoop-yarn/lib/commons-lang-2.6.jar
/usr/lib/hadoop-yarn/lib/jline-2.11.jar
/usr/lib/hadoop-mapreduce/jsch-0.1.42.jar
/usr/lib/hadoop-mapreduce/hadoop-azure.jar
/usr/lib/hadoop-mapreduce/apacheds-kerberos-codec-2.0.0-M15.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-nativetask.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-app-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/hadoop-streaming.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-hs.jar
/usr/lib/hadoop-mapreduce/zookeeper.jar
/usr/lib/hadoop-mapreduce/hadoop-auth-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-jobclient.jar
/usr/lib/hadoop-mapreduce/mockito-all-1.8.5.jar
/usr/lib/hadoop-mapreduce/jersey-core-1.9.jar
/usr/lib/hadoop-mapreduce/commons-net-3.1.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-app.jar
/usr/lib/hadoop-mapreduce/commons-compress-1.4.1.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-common.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-nativetask-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/hadoop-gridmix.jar
/usr/lib/hadoop-mapreduce/hadoop-sls-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/avro.jar
/usr/lib/hadoop-mapreduce/api-asn1-api-1.0.0-M20.jar
/usr/lib/hadoop-mapreduce/guava-11.0.2.jar
/usr/lib/hadoop-mapreduce/hadoop-datajoin.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-hs-plugins.jar
/usr/lib/hadoop-mapreduce/hadoop-sls.jar
/usr/lib/hadoop-mapreduce/hadoop-archives-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/jets3t-0.9.0.jar
/usr/lib/hadoop-mapreduce/jersey-json-1.9.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-hs-plugins-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/jaxb-impl-2.2.3-1.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/jetty-util-6.1.26.cloudera.4.jar
/usr/lib/hadoop-mapreduce/jsr305-3.0.0.jar
/usr/lib/hadoop-mapreduce/apacheds-i18n-2.0.0-M15.jar
/usr/lib/hadoop-mapreduce/hadoop-gridmix-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/htrace-core-3.0.4.jar
/usr/lib/hadoop-mapreduce/servlet-api-2.5.jar
/usr/lib/hadoop-mapreduce/jackson-jaxrs-1.8.8.jar
/usr/lib/hadoop-mapreduce/stax-api-1.0-2.jar
/usr/lib/hadoop-mapreduce/commons-collections-3.2.1.jar
/usr/lib/hadoop-mapreduce/commons-logging-1.1.3.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-shuffle.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-shuffle-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/hamcrest-core-1.3.jar
/usr/lib/hadoop-mapreduce/jackson-core-2.2.3.jar
/usr/lib/hadoop-mapreduce/java-xmlbuilder-0.4.jar
/usr/lib/hadoop-mapreduce/httpclient-4.2.5.jar
/usr/lib/hadoop-mapreduce/asm-3.2.jar
/usr/lib/hadoop-mapreduce/httpcore-4.2.5.jar
/usr/lib/hadoop-mapreduce/hadoop-auth.jar
/usr/lib/hadoop-mapreduce/hadoop-distcp-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/commons-beanutils-1.7.0.jar
/usr/lib/hadoop-mapreduce/metrics-core-3.0.1.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-common-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/commons-el-1.0.jar
/usr/lib/hadoop-mapreduce/xz-1.0.jar
/usr/lib/hadoop-mapreduce/hadoop-distcp.jar
/usr/lib/hadoop-mapreduce/curator-client-2.7.1.jar
/usr/lib/hadoop-mapreduce/commons-codec-1.4.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-hs-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/hadoop-ant.jar
/usr/lib/hadoop-mapreduce/api-util-1.0.0-M20.jar
/usr/lib/hadoop-mapreduce/jackson-databind-2.2.3.jar
/usr/lib/hadoop-mapreduce/commons-beanutils-core-1.8.0.jar
/usr/lib/hadoop-mapreduce/gson-2.2.4.jar
/usr/lib/hadoop-mapreduce/hadoop-streaming-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/jasper-runtime-5.5.23.jar
/usr/lib/hadoop-mapreduce/snappy-java-1.0.4.1.jar
/usr/lib/hadoop-mapreduce/protobuf-java-2.5.0.jar
/usr/lib/hadoop-mapreduce/commons-io-2.4.jar
/usr/lib/hadoop-mapreduce/jackson-core-asl-1.8.8.jar
/usr/lib/hadoop-mapreduce/jetty-6.1.26.cloudera.4.jar
/usr/lib/hadoop-mapreduce/curator-recipes-2.7.1.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-jobclient-2.6.0-cdh5.4.4-tests.jar
/usr/lib/hadoop-mapreduce/hadoop-extras-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/jasper-compiler-5.5.23.jar
/usr/lib/hadoop-mapreduce/jaxb-api-2.2.2.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-core-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/jackson-mapper-asl-1.8.8.jar
/usr/lib/hadoop-mapreduce/junit-4.11.jar
/usr/lib/hadoop-mapreduce/jsp-api-2.1.jar
/usr/lib/hadoop-mapreduce/jackson-annotations-2.2.3.jar
/usr/lib/hadoop-mapreduce/commons-math3-3.1.1.jar
/usr/lib/hadoop-mapreduce/activation-1.1.jar
/usr/lib/hadoop-mapreduce/curator-framework-2.7.1.jar
/usr/lib/hadoop-mapreduce/hadoop-azure-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/xmlenc-0.52.jar
/usr/lib/hadoop-mapreduce/microsoft-windowsazure-storage-sdk-0.6.0.jar
/usr/lib/hadoop-mapreduce/jackson-xc-1.8.8.jar
/usr/lib/hadoop-mapreduce/commons-configuration-1.6.jar
/usr/lib/hadoop-mapreduce/hadoop-ant-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/commons-httpclient-3.1.jar
/usr/lib/hadoop-mapreduce/jersey-server-1.9.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-core.jar
/usr/lib/hadoop-mapreduce/hadoop-rumen.jar
/usr/lib/hadoop-mapreduce/commons-cli-1.2.jar
/usr/lib/hadoop-mapreduce/commons-digester-1.8.jar
/usr/lib/hadoop-mapreduce/hadoop-archives.jar
/usr/lib/hadoop-mapreduce/hadoop-datajoin-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/hadoop-extras.jar
/usr/lib/hadoop-mapreduce/paranamer-2.3.jar
/usr/lib/hadoop-mapreduce/hadoop-rumen-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/log4j-1.2.17.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar
/usr/lib/hadoop-mapreduce/jettison-1.1.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-jobclient-2.6.0-cdh5.4.4.jar
/usr/lib/hadoop-mapreduce/commons-lang-2.6.jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-jobclient-tests.jar
/usr/lib/hadoop-mapreduce/lib/leveldbjni-all-1.8.jar
/usr/lib/hadoop-mapreduce/lib/jersey-core-1.9.jar
/usr/lib/hadoop-mapreduce/lib/commons-compress-1.4.1.jar
/usr/lib/hadoop-mapreduce/lib/avro.jar
/usr/lib/hadoop-mapreduce/lib/javax.inject-1.jar
/usr/lib/hadoop-mapreduce/lib/netty-3.6.2.Final.jar
/usr/lib/hadoop-mapreduce/lib/hamcrest-core-1.3.jar
/usr/lib/hadoop-mapreduce/lib/asm-3.2.jar
/usr/lib/hadoop-mapreduce/lib/aopalliance-1.0.jar
/usr/lib/hadoop-mapreduce/lib/xz-1.0.jar
/usr/lib/hadoop-mapreduce/lib/guice-servlet-3.0.jar
/usr/lib/hadoop-mapreduce/lib/snappy-java-1.0.4.1.jar
/usr/lib/hadoop-mapreduce/lib/guice-3.0.jar
/usr/lib/hadoop-mapreduce/lib/protobuf-java-2.5.0.jar
/usr/lib/hadoop-mapreduce/lib/commons-io-2.4.jar
/usr/lib/hadoop-mapreduce/lib/jackson-core-asl-1.8.8.jar
/usr/lib/hadoop-mapreduce/lib/jackson-mapper-asl-1.8.8.jar
/usr/lib/hadoop-mapreduce/lib/junit-4.11.jar
/usr/lib/hadoop-mapreduce/lib/jersey-server-1.9.jar
/usr/lib/hadoop-mapreduce/lib/jersey-guice-1.9.jar
/usr/lib/hadoop-mapreduce/lib/paranamer-2.3.jar
/usr/lib/hadoop-mapreduce/lib/log4j-1.2.17.jar
/usr/lib/hadoop/lib/native
/etc/hadoop/conf

@igorvishnevskiy
Copy link
Author

igorvishnevskiy commented Jun 23, 2016

Now I got it to pick up host address and port. Still throws the following error. I masked host address by "HOST_ADDRESS" in the below error, but there is a correct one when exception is thrown.

loadFileSystems error:
(unable to get stack trace for java.lang.NoClassDefFoundError exception: ExceptionUtils::getStackTrace error.)
hdfsBuilderConnect(forceNewInstance=0, nn=HOST_ADDRESS, port=8020, kerbTicketCachePath=(NULL), userName=(NULL)) error:
(unable to get stack trace for java.lang.NoClassDefFoundError exception: ExceptionUtils::getStackTrace error.)
ERROR - RuntimeError:  (255, 'Unknown error 255')

@igorvishnevskiy
Copy link
Author

igorvishnevskiy commented Jun 23, 2016

OK, moving on. Got last issue resolved looks like, I am hoping :). Now I get the next exception. Stuck on this one now. Looks like it is looking for "mapred" in the wrong path, even though following returns correct path. And here we run on "Cloudera Distribution of Hadoop" if that helps.

>>> import pydoop
>>> pydoop.hadoop_home()
'/usr/lib/hadoop

Exception:

[wordcount]$ sh run_wc_minimal
Activated service account credentials for: [account-1@analytics-platform-sandbox.iam.gserviceaccount.com]
INFO:main:copying input data to HDFS
INFO:main:running MapReduce application
DEPRECATED: Use of this script to execute mapred command is deprecated.
Instead use the mapred command for it.

/usr/lib/hadoop-mapreduce/bin/mapred: line 27: /usr/lib/hadoop-mapreduce/bin/mapred-config.sh: No such file or directory
/usr/lib/hadoop-mapreduce/bin/mapred: line 152: exec: : not found
Traceback (most recent call last):
  File "run_wc.py", line 79, in <module>
    main(sys.argv[1:])
  File "run_wc.py", line 71, in main
    runner.run(properties=CONF, hadoop_conf_dir=HADOOP_CONF_DIR, logger=logger)
  File "/usr/lib64/python2.7/site-packages/pydoop/hadut.py", line 505, in run
    run_pipes(self.exe, self.input, self.output, **kwargs)
  File "/usr/lib64/python2.7/site-packages/pydoop/hadut.py", line 371, in run_pipes
    keep_streams=keep_streams)
  File "/usr/lib64/python2.7/site-packages/pydoop/hadut.py", line 173, in run_cmd
    raise RunCmdError(p.returncode, ' '.join(_args), error)
pydoop.hadut.RunCmdError: command exited with 127 status

And when I checked again, there is mapred at the correct path, but then why do I get above exception?

[wordcount]$ ls -la /usr/lib/hadoop-mapreduce/bin/
total 608
drwxr-xr-x 2 hdfs hdfs   4096 Jun 23 03:09 .
drwxr-xr-x 6 hdfs hdfs   4096 Dec 10  2015 ..
-rwxr-xr-x 1 hdfs hdfs   5205 Jul  7  2015 mapred
-rwxr-xr-x 1 hdfs hdfs 145512 Jul  7  2015 pipes-sort
-rwxr-xr-x 1 hdfs hdfs 155928 Jul  7  2015 wordcount-nopipe
-rwxr-xr-x 1 hdfs hdfs 150616 Jul  7  2015 wordcount-part
-rwxr-xr-x 1 hdfs hdfs 146720 Jul  7  2015 wordcount-simple

@igorvishnevskiy
Copy link
Author

igorvishnevskiy commented Jun 30, 2016

So. I contacted Cloudera and this is what they told me:

After asking numerous engineers and support team, we do not support Pydoop in CDH5.

So as of June 30st, 2016, Clouderas Distribution of Hadoop version CDH5 does not support Python's Hadoop Library "Pydoop". Though on CHD4 it was supported.

I definitely wish CHD5 did support it. Maybe later support will be added. If Pydoop's support on Cloudera ever gets added and someone notices it before me, please send me a note to my LinkedIn. Thank you.
https://www.linkedin.com/in/igor-vishnevskiy-b6551a86

@igorvishnevskiy igorvishnevskiy changed the title pydoop job hangs on the "INFO mapreduce.Job: map 0% reduce 0%" step and times out. pydoop job hangs on Cloudera Hadoop 2.6 CDH5 Jun 30, 2016
@igorvishnevskiy igorvishnevskiy changed the title pydoop job hangs on Cloudera Hadoop 2.6 CDH5 pydoop job hangs because pydoop is not supported by Cloudera Hadoop 2.6 CDH5 Jun 30, 2016
@igorvishnevskiy igorvishnevskiy changed the title pydoop job hangs because pydoop is not supported by Cloudera Hadoop 2.6 CDH5 pydoop mapreduce job hangs because pydoop is not supported by Cloudera Hadoop 2.6 CDH5 Jun 30, 2016
@ilveroluca
Copy link
Member

Hi @wwwmaster2k . Sorry about the delay. I lost track of your issue.

I'm afraid we never had "official" support from Cloudera, and they have a tendency to do things their own way -- different from the standard Hadoop distribution. We do our best to try to cope with their changes, but it costs us a lot of time and effort and since we're not Cloudera users issues can still slip by.

Nevertheless, though they don't support our project, our tests on Travis run on various CDH5 releases. Have a look here:

https://travis-ci.org/crs4/pydoop/builds/112216921

That's the automated test results for the latest official Pydoop release (1.2.0). In theory your set-up should work. We'll take a closer look at your logs; hopefully we'll notice something.

Oh, I saw how you're setting your environment variables. HADOOP_HOME should be set to /usr/lib/hadoop (notice there's not bin).

Actually, not that I think of it, it might help if you have a look at .travis/install_hadoop.sh. In there you'll find a function install_cdh5 that might help you spot the difference between your set-up and the that's working for us on Travis.

Cheers,

Luca

@ilveroluca
Copy link
Member

How did you install CDH? From packages, Cloudera Agent, or a tarball?

@igorvishnevskiy
Copy link
Author

Hello @ilveroluca. Thank you for your help.
We installed Cloudera using:
Cloudera manager… using rpms.

@mdrio
Copy link
Member

mdrio commented Jul 4, 2016

Hi, we tested Pydoop on CDH 5 installed on Ubuntu using deb packages. Maybe there is something different in other OS/installations methods that produces your bug.

@ilveroluca
Copy link
Member

ilveroluca commented Jul 4, 2016

Ok @wwwmaster2k . You'll have to tell us where mapred-config.sh gets installed in your set-up.

Reading the error you got, what's happening is that the mapred executable is complaining because it tried to run or source /usr/lib/hadoop-mapreduce/bin/mapred-config.sh and didn't find it.

@igorvishnevskiy
Copy link
Author

igorvishnevskiy commented Jul 5, 2016

Thank you so much for your help @ilveroluca and @mdrio. I ran find command and found: mapred-config.sh at the following path.
/usr/lib/hadoop/libexec/mapred-config.sh

@ilveroluca
Copy link
Member

And what happens if you run the mapred executable from the command line? Do you get the same error?

@igorvishnevskiy
Copy link
Author

[hdfs]$ sh mapred
mapred: line 28: /usr/bin/mapred-config.sh: No such file or directory

@igorvishnevskiy
Copy link
Author

I fixed path to mapred-config.sh in mapred executable and now I get following:

[hdfs]$ sh mapred
Usage: mapred [--config confdir] [--loglevel loglevel] COMMAND
       where COMMAND is one of:
  pipes                run a Pipes job
  job                  manipulate MapReduce jobs
  queue                get information regarding JobQueues
  classpath            prints the class path needed for running
                       mapreduce subcommands
  historyserver        run job history servers as a standalone daemon
  distcp <srcurl> <desturl> copy file or directories recursively
  archive -archiveName NAME -p <parent path> <src>* <dest> create a hadoop archive
  hsadmin              job history server admin interface

Most commands print help when invoked w/o parameters.

@ilveroluca
Copy link
Member

mapred should be executable, so you shouldn't need to run it through sh. So now that you've fixed that script, do the Pydoop examples run?

@simleo simleo added the CDH label Nov 22, 2017
@simleo
Copy link
Member

simleo commented May 27, 2019

As of 2.0.0, Pydoop no longer tries to explicitly support the various layouts of customized Hadoop releases such as CDH or HDP. However, if the hadoop command is in the PATH and HADOOP_CONF_DIR points to the correct location, everything should work.

@simleo simleo closed this as completed May 27, 2019
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

5 participants