Skip to content

Conversation

@justinuang
Copy link

No description provided.

@hadoop-yetus
Copy link

🎊 +1 overall

Vote Subsystem Runtime Comment
0 reexec 24 Docker mode activated.
_ Prechecks _
+1 @author 0 The patch does not contain any @author tags.
+1 test4tests 0 The patch appears to include 3 new or modified test files.
_ trunk Compile Tests _
+1 mvninstall 987 trunk passed
+1 compile 31 trunk passed
+1 checkstyle 23 trunk passed
+1 mvnsite 34 trunk passed
+1 shadedclient 713 branch has no errors when building and testing our client artifacts.
+1 findbugs 45 trunk passed
+1 javadoc 25 trunk passed
_ Patch Compile Tests _
+1 mvninstall 30 the patch passed
+1 compile 28 the patch passed
+1 javac 28 the patch passed
-0 checkstyle 18 hadoop-tools/hadoop-aws: The patch generated 8 new + 5 unchanged - 0 fixed = 13 total (was 5)
+1 mvnsite 32 the patch passed
+1 whitespace 0 The patch has no whitespace issues.
+1 shadedclient 735 patch has no errors when building and testing our client artifacts.
+1 findbugs 50 the patch passed
+1 javadoc 22 the patch passed
_ Other Tests _
+1 unit 272 hadoop-aws in the patch passed.
+1 asflicense 28 The patch does not generate ASF License warnings.
3186
Subsystem Report/Notes
Docker Client=17.05.0-ce Server=17.05.0-ce base: https://builds.apache.org/job/hadoop-multibranch/job/PR-645/1/artifact/out/Dockerfile
GITHUB PR #645
Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle
uname Linux 215678f1cf61 4.4.0-138-generic #164-Ubuntu SMP Tue Oct 2 17:16:02 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
Build tool maven
Personality personality/hadoop.sh
git revision trunk / ce4bafd
maven version: Apache Maven 3.3.9
Default Java 1.8.0_191
findbugs v3.1.0-RC1
checkstyle https://builds.apache.org/job/hadoop-multibranch/job/PR-645/1/artifact/out/diff-checkstyle-hadoop-tools_hadoop-aws.txt
Test Results https://builds.apache.org/job/hadoop-multibranch/job/PR-645/1/testReport/
Max. process+thread count 410 (vs. ulimit of 5500)
modules C: hadoop-tools/hadoop-aws U: hadoop-tools/hadoop-aws
Console output https://builds.apache.org/job/hadoop-multibranch/job/PR-645/1/console
Powered by Apache Yetus 0.9.0 http://yetus.apache.org

This message was automatically generated.

@hadoop-yetus
Copy link

🎊 +1 overall

Vote Subsystem Runtime Comment
0 reexec 23 Docker mode activated.
_ Prechecks _
+1 @author 0 The patch does not contain any @author tags.
+1 test4tests 0 The patch appears to include 3 new or modified test files.
_ trunk Compile Tests _
+1 mvninstall 1005 trunk passed
+1 compile 29 trunk passed
+1 checkstyle 19 trunk passed
+1 mvnsite 33 trunk passed
+1 shadedclient 706 branch has no errors when building and testing our client artifacts.
+1 findbugs 41 trunk passed
+1 javadoc 25 trunk passed
_ Patch Compile Tests _
+1 mvninstall 29 the patch passed
+1 compile 28 the patch passed
+1 javac 28 the patch passed
+1 checkstyle 17 the patch passed
+1 mvnsite 32 the patch passed
+1 whitespace 0 The patch has no whitespace issues.
+1 shadedclient 713 patch has no errors when building and testing our client artifacts.
+1 findbugs 48 the patch passed
+1 javadoc 19 the patch passed
_ Other Tests _
+1 unit 275 hadoop-aws in the patch passed.
+1 asflicense 24 The patch does not generate ASF License warnings.
3154
Subsystem Report/Notes
Docker Client=17.05.0-ce Server=17.05.0-ce base: https://builds.apache.org/job/hadoop-multibranch/job/PR-645/2/artifact/out/Dockerfile
GITHUB PR #645
Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle
uname Linux 61df0233a9b8 4.4.0-138-generic #164-Ubuntu SMP Tue Oct 2 17:16:02 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
Build tool maven
Personality personality/hadoop.sh
git revision trunk / ce4bafd
maven version: Apache Maven 3.3.9
Default Java 1.8.0_191
findbugs v3.1.0-RC1
Test Results https://builds.apache.org/job/hadoop-multibranch/job/PR-645/2/testReport/
Max. process+thread count 468 (vs. ulimit of 5500)
modules C: hadoop-tools/hadoop-aws U: hadoop-tools/hadoop-aws
Console output https://builds.apache.org/job/hadoop-multibranch/job/PR-645/2/console
Powered by Apache Yetus 0.9.0 http://yetus.apache.org

This message was automatically generated.

*/
public static final boolean CHANGE_DETECT_REQUIRE_VERSION_DEFAULT = true;

public static final String MULTIPART_DOWNLOAD_ENABLED =
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

should add javadoc for each of these, and add to core-default.xml

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Some documentation should be added to hadoop-aws/index.md and hadoop-aws/performance.md about this capability.


public static final String MULTIPART_DOWNLOAD_CHUNK_SIZE =
"fs.s3a.multipartdownload.chunk-size";
public static final long DEFAULT_MULTIPART_DOWNLOAD_CHUNK_SIZE = 262_144;
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

the javadoc should explain the difference between part size and chunk size, also why a new configuration property is needed instead of reusing fs.s3a.multipart.size


public static final String MULTIPART_DOWNLOAD_NUM_THREADS =
"fs.s3a.multipartdownload.num-threads";
public static final int DEFAULT_MULTIPART_DOWNLOAD_NUM_THREADS = 8;
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

why not reuse fs.s3a.threads.max (and ideally, the existing thread pool corresponding to that config)?

String text = String.format("%s %s at %d",
operation, uri, rangeStart);

S3Object object = Invoker.once(text, "s3a://" + key + "/" + bucket,
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Do the key and bucket need to be switched here? Also, use requestBucket instead of bucket?

*/
public abstract class AbortableInputStream extends InputStream {

public abstract void abort();
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

add method javadoc


private final S3ObjectInputStream s3ObjectInputStream;

public AbortableS3ObjectInputStream(S3ObjectInputStream s3ObjectInputStream) {
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

add constructor javadoc


public static final String MULTIPART_DOWNLOAD_BUFFER_SIZE =
"fs.s3a.multipartdownload.buffer-size";
public static final long DEFAULT_MULTIPART_DOWNLOAD_BUFFER_SIZE = 20_000_000;
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

How did you determine the optimal value for this? Javadoc for it should specify that the size is "(in bytes)".

@steveloughran steveloughran added the fs/s3 changes related to hadoop-aws; submitter must declare test endpoint label Mar 28, 2019
@bolkedebruin
Copy link

@justinuang are you still working on this?

@justinuang
Copy link
Author

justinuang commented Jul 10, 2019 via email

@justinuang justinuang closed this Jul 10, 2019
shanthoosh pushed a commit to shanthoosh/hadoop that referenced this pull request Oct 15, 2019
… instead of run-time objects

As per subject, changed caching table descriptor to take table descriptor instead of run-time objects
 - Added BaseHybridTableDescriptor, which models a hybrid table that may contain other tables
 - Modified StreamApplicationDescriptorImpl to also include tables contained within a hybrid table

Author: Wei Song <wsong@linkedin.com>

Reviewers: Jagadish Venkatraman <jvenkatraman@linkedin.com>

Closes apache#645 from weisong44/SAMZA-1854 and squashes the following commits:

2c0d1362 [Wei Song] Updated based on review comments
dd18bbee [Wei Song] Merge branch 'master' into SAMZA-1854
a6c94ad [Wei Song] Merge remote-tracking branch 'upstream/master'
41299b5 [Wei Song] Merge remote-tracking branch 'upstream/master'
239a095 [Wei Song] Merge remote-tracking branch 'upstream/master'
a87a9b04 [Wei Song] SAMZA-1854: Changed caching table descriptor to take table descriptor instead of run-time objects
eca0020 [Wei Song] Merge remote-tracking branch 'upstream/master'
5156239 [Wei Song] Merge remote-tracking branch 'upstream/master'
de708f5 [Wei Song] Merge remote-tracking branch 'upstream/master'
df2f8d7 [Wei Song] Merge remote-tracking branch 'upstream/master'
f28b491 [Wei Song] Merge remote-tracking branch 'upstream/master'
4782c61 [Wei Song] Merge remote-tracking branch 'upstream/master'
0440f75 [Wei Song] Merge remote-tracking branch 'upstream/master'
aae0f38 [Wei Song] Merge remote-tracking branch 'upstream/master'
a15a7c9 [Wei Song] Merge remote-tracking branch 'upstream/master'
5cbf9af [Wei Song] Merge remote-tracking branch 'upstream/master'
3f7ed71 [Wei Song] Added self to committer list
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

fs/s3 changes related to hadoop-aws; submitter must declare test endpoint

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants