Skip to content
This repository has been archived by the owner on Feb 9, 2021. It is now read-only.

Commit

Permalink
Merge r1480440 through r1480820 from trunk.
Browse files Browse the repository at this point in the history
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/HDFS-2802@1480824 13f79535-47bb-0310-9956-ffa450edef68
  • Loading branch information
szetszwo committed May 9, 2013
2 parents 4039a9c + fa20931 commit 725c34b
Show file tree
Hide file tree
Showing 56 changed files with 1,191 additions and 292 deletions.
4 changes: 0 additions & 4 deletions hadoop-client/pom.xml
Original file line number Original file line Diff line number Diff line change
Expand Up @@ -99,10 +99,6 @@
<groupId>org.eclipse.jdt</groupId> <groupId>org.eclipse.jdt</groupId>
<artifactId>core</artifactId> <artifactId>core</artifactId>
</exclusion> </exclusion>
<exclusion>
<groupId>org.aspectj</groupId>
<artifactId>aspectjrt</artifactId>
</exclusion>
<exclusion> <exclusion>
<groupId>org.apache.avro</groupId> <groupId>org.apache.avro</groupId>
<artifactId>avro-ipc</artifactId> <artifactId>avro-ipc</artifactId>
Expand Down
4 changes: 3 additions & 1 deletion hadoop-common-project/hadoop-common/CHANGES.txt
Original file line number Original file line Diff line number Diff line change
Expand Up @@ -694,7 +694,9 @@ Release 2.0.5-beta - UNRELEASED
HADOOP-9455. HADOOP_CLIENT_OPTS appended twice causes JVM failures. HADOOP-9455. HADOOP_CLIENT_OPTS appended twice causes JVM failures.
(Chris Nauroth via suresh) (Chris Nauroth via suresh)


Release 2.0.4-alpha - UNRELEASED HADOOP-9550. Remove aspectj dependency. (kkambatl via tucu)

Release 2.0.4-alpha - 2013-04-25


INCOMPATIBLE CHANGES INCOMPATIBLE CHANGES


Expand Down
2 changes: 1 addition & 1 deletion hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
Original file line number Original file line Diff line number Diff line change
Expand Up @@ -630,7 +630,7 @@ Release 2.0.5-beta - UNRELEASED


HDFS-4784. NPE in FSDirectory.resolvePath(). (Brandon Li via suresh) HDFS-4784. NPE in FSDirectory.resolvePath(). (Brandon Li via suresh)


Release 2.0.4-alpha - UNRELEASED Release 2.0.4-alpha - 2013-04-25


INCOMPATIBLE CHANGES INCOMPATIBLE CHANGES


Expand Down
17 changes: 16 additions & 1 deletion hadoop-mapreduce-project/CHANGES.txt
Original file line number Original file line Diff line number Diff line change
Expand Up @@ -224,6 +224,9 @@ Release 2.0.5-beta - UNRELEASED
MAPREDUCE-5159. Change ValueAggregatorJob to add APIs which can support MAPREDUCE-5159. Change ValueAggregatorJob to add APIs which can support
binary compatibility with hadoop-1 examples. (Zhijie Shen via vinodkv) binary compatibility with hadoop-1 examples. (Zhijie Shen via vinodkv)


MAPREDUCE-5157. Bring back old sampler related code so that we can support
binary compatibility with hadoop-1 sorter example. (Zhijie Shen via vinodkv)

OPTIMIZATIONS OPTIMIZATIONS


MAPREDUCE-4974. Optimising the LineRecordReader initialize() method MAPREDUCE-4974. Optimising the LineRecordReader initialize() method
Expand Down Expand Up @@ -383,7 +386,19 @@ Release 2.0.5-beta - UNRELEASED
MAPREDUCE-5204. Handling YarnRemoteException separately from IOException in MAPREDUCE-5204. Handling YarnRemoteException separately from IOException in
MR app after YARN-629. (Xuan Gong via vinodkv) MR app after YARN-629. (Xuan Gong via vinodkv)


Release 2.0.4-alpha - UNRELEASED MAPREDUCE-5209. Fix units in a ShuffleScheduler log message.
(Tsuyoshi OZAWA via cdouglas)

MAPREDUCE-5212. Handling YarnRemoteException separately from IOException in
MR App's use of ClientRMProtocol after YARN-631. (Xuan Gong via vinodkv)

MAPREDUCE-5226. Handling YarnRemoteException separately from IOException in
MR App's use of AMRMProtocol after YARN-630. (Xuan Gong via vinodkv)

MAPREDUCE-4942. mapreduce.Job has a bunch of methods that throw
InterruptedException so its incompatible with MR1. (rkanter via tucu)

Release 2.0.4-alpha - 2013-04-25


INCOMPATIBLE CHANGES INCOMPATIBLE CHANGES


Expand Down
Original file line number Original file line Diff line number Diff line change
Expand Up @@ -18,6 +18,7 @@


package org.apache.hadoop.mapreduce.v2.app.rm; package org.apache.hadoop.mapreduce.v2.app.rm;


import java.io.IOException;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.Collections; import java.util.Collections;
import java.util.HashMap; import java.util.HashMap;
Expand Down Expand Up @@ -144,7 +145,8 @@ public void init(Configuration conf) {
LOG.info("blacklistDisablePercent is " + blacklistDisablePercent); LOG.info("blacklistDisablePercent is " + blacklistDisablePercent);
} }


protected AllocateResponse makeRemoteRequest() throws YarnRemoteException { protected AllocateResponse makeRemoteRequest() throws YarnRemoteException,
IOException {
AllocateRequest allocateRequest = BuilderUtils.newAllocateRequest( AllocateRequest allocateRequest = BuilderUtils.newAllocateRequest(
applicationAttemptId, lastResponseID, super.getApplicationProgress(), applicationAttemptId, lastResponseID, super.getApplicationProgress(),
new ArrayList<ResourceRequest>(ask), new ArrayList<ContainerId>( new ArrayList<ResourceRequest>(ask), new ArrayList<ContainerId>(
Expand Down
Original file line number Original file line Diff line number Diff line change
Expand Up @@ -18,6 +18,7 @@


package org.apache.hadoop.mapreduce.v2.app; package org.apache.hadoop.mapreduce.v2.app;


import java.io.IOException;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.List; import java.util.List;
import java.util.concurrent.BlockingQueue; import java.util.concurrent.BlockingQueue;
Expand Down Expand Up @@ -202,7 +203,7 @@ protected AMRMProtocol createSchedulerProxy() {
public RegisterApplicationMasterResponse public RegisterApplicationMasterResponse
registerApplicationMaster( registerApplicationMaster(
RegisterApplicationMasterRequest request) RegisterApplicationMasterRequest request)
throws YarnRemoteException { throws YarnRemoteException, IOException {
RegisterApplicationMasterResponse response = RegisterApplicationMasterResponse response =
Records.newRecord(RegisterApplicationMasterResponse.class); Records.newRecord(RegisterApplicationMasterResponse.class);
response.setMinimumResourceCapability(BuilderUtils response.setMinimumResourceCapability(BuilderUtils
Expand All @@ -215,15 +216,15 @@ protected AMRMProtocol createSchedulerProxy() {
@Override @Override
public FinishApplicationMasterResponse finishApplicationMaster( public FinishApplicationMasterResponse finishApplicationMaster(
FinishApplicationMasterRequest request) FinishApplicationMasterRequest request)
throws YarnRemoteException { throws YarnRemoteException, IOException {
FinishApplicationMasterResponse response = FinishApplicationMasterResponse response =
Records.newRecord(FinishApplicationMasterResponse.class); Records.newRecord(FinishApplicationMasterResponse.class);
return response; return response;
} }


@Override @Override
public AllocateResponse allocate(AllocateRequest request) public AllocateResponse allocate(AllocateRequest request)
throws YarnRemoteException { throws YarnRemoteException, IOException {


AllocateResponse response = AllocateResponse response =
Records.newRecord(AllocateResponse.class); Records.newRecord(AllocateResponse.class);
Expand Down
Original file line number Original file line Diff line number Diff line change
Expand Up @@ -100,6 +100,7 @@ protected AMRMProtocol createSchedulerProxy() {
when(scheduler.allocate(isA(AllocateRequest.class))) when(scheduler.allocate(isA(AllocateRequest.class)))
.thenThrow(RPCUtil.getRemoteException(new IOException("forcefail"))); .thenThrow(RPCUtil.getRemoteException(new IOException("forcefail")));
} catch (YarnRemoteException e) { } catch (YarnRemoteException e) {
} catch (IOException e) {
} }
return scheduler; return scheduler;
} }
Expand Down
Original file line number Original file line Diff line number Diff line change
Expand Up @@ -209,23 +209,15 @@ public String getTrackingURL() {
* completed. * completed.
*/ */
public float mapProgress() throws IOException { public float mapProgress() throws IOException {
try { return job.mapProgress();
return job.mapProgress();
} catch (InterruptedException ie) {
throw new IOException(ie);
}
} }


/** /**
* A float between 0.0 and 1.0, indicating the % of reduce work * A float between 0.0 and 1.0, indicating the % of reduce work
* completed. * completed.
*/ */
public float reduceProgress() throws IOException { public float reduceProgress() throws IOException {
try { return job.reduceProgress();
return job.reduceProgress();
} catch (InterruptedException ie) {
throw new IOException(ie);
}
} }


/** /**
Expand All @@ -245,33 +237,21 @@ public float cleanupProgress() throws IOException {
* completed. * completed.
*/ */
public float setupProgress() throws IOException { public float setupProgress() throws IOException {
try { return job.setupProgress();
return job.setupProgress();
} catch (InterruptedException ie) {
throw new IOException(ie);
}
} }


/** /**
* Returns immediately whether the whole job is done yet or not. * Returns immediately whether the whole job is done yet or not.
*/ */
public synchronized boolean isComplete() throws IOException { public synchronized boolean isComplete() throws IOException {
try { return job.isComplete();
return job.isComplete();
} catch (InterruptedException ie) {
throw new IOException(ie);
}
} }


/** /**
* True iff job completed successfully. * True iff job completed successfully.
*/ */
public synchronized boolean isSuccessful() throws IOException { public synchronized boolean isSuccessful() throws IOException {
try { return job.isSuccessful();
return job.isSuccessful();
} catch (InterruptedException ie) {
throw new IOException(ie);
}
} }


/** /**
Expand Down Expand Up @@ -302,11 +282,7 @@ public synchronized int getJobState() throws IOException {
* Tells the service to terminate the current job. * Tells the service to terminate the current job.
*/ */
public synchronized void killJob() throws IOException { public synchronized void killJob() throws IOException {
try { job.killJob();
job.killJob();
} catch (InterruptedException ie) {
throw new IOException(ie);
}
} }




Expand All @@ -331,14 +307,10 @@ public synchronized void setJobPriority(String priority)
*/ */
public synchronized void killTask(TaskAttemptID taskId, public synchronized void killTask(TaskAttemptID taskId,
boolean shouldFail) throws IOException { boolean shouldFail) throws IOException {
try { if (shouldFail) {
if (shouldFail) { job.failTask(taskId);
job.failTask(taskId); } else {
} else { job.killTask(taskId);
job.killTask(taskId);
}
} catch (InterruptedException ie) {
throw new IOException(ie);
} }
} }


Expand Down Expand Up @@ -378,16 +350,12 @@ public String toString() {
* Returns the counters for this job * Returns the counters for this job
*/ */
public Counters getCounters() throws IOException { public Counters getCounters() throws IOException {
try { Counters result = null;
Counters result = null; org.apache.hadoop.mapreduce.Counters temp = job.getCounters();
org.apache.hadoop.mapreduce.Counters temp = job.getCounters(); if(temp != null) {
if(temp != null) { result = Counters.downgrade(temp);
result = Counters.downgrade(temp);
}
return result;
} catch (InterruptedException ie) {
throw new IOException(ie);
} }
return result;
} }


@Override @Override
Expand Down
Loading

0 comments on commit 725c34b

Please sign in to comment.