Skip to content

0.10.0#4482

Closed
guoke111 wants to merge 132 commits intoapache:release-0.10.0from
guoke111:master
Closed

0.10.0#4482
guoke111 wants to merge 132 commits intoapache:release-0.10.0from
guoke111:master

Conversation

@guoke111
Copy link
Copy Markdown

Tips

What is the purpose of the pull request

(For example: This pull request adds quick-start document.)

Brief change log

(for example:)

  • Modify AnnotationLocation checkstyle rule in checkstyle.xml

Verify this pull request

(Please pick either of the following options)

This pull request is a trivial rework / code cleanup without any test coverage.

(or)

This pull request is already covered by existing tests, such as (please describe tests).

(or)

This change added tests and can be verified as follows:

(example:)

  • Added integration tests for end-to-end.
  • Added HoodieClientWriteTest to verify the change.
  • Manually verified the change by running a job locally.

Committer checklist

  • Has a corresponding JIRA in PR title & commit

  • Commit message is descriptive of the change

  • CI is green

  • Necessary doc changes done or have another open PR

  • For large changes, please consider breaking it into sub-tasks under an umbrella JIRA.

danny0405 and others added 30 commits November 27, 2021 17:22
…to 'DefaultHoodieRecordPayload' (#4115)" (#4169)

This reverts commit 88067f5.
* Fixing loading of props from default dir

* addressing comments
…writers and table services (#4186)

- Co-authored-by: Rajesh Mahindra <rmahindra@Rajeshs-MacBook-Pro.local>
- Co-authored-by: Sivabalan Narayanan <n.siva.b@gmail.com>
…ing virtual keys by default for metadata table (#4194)
…ulting in incorrect `KeyGenerator` configuration (#4195)
…on base files over S3 (#4185)

- Fetching partition files or all partitions from the metadata table is failing
   when run over S3. Metadata table uses HFile format for the base files and the
   record lookup uses HFile.Reader and HFileScanner interfaces to get records by
   partition keys. When the backing storage is S3, this record lookup from HFiles
   is failing with IOException, in turn failing the caller commit/update operations.

 - Metadata table looks up HFile records with positional read enabled so as to
   perform better for random lookups. But this positional read key lookup is
   returning with partial read sizes over S3 leading to HFile scanner throwing
   IOException. This doesn't happen over HDFS. Metadata table though uses the HFile
   for random key lookups, the positional read is not mandatory as we sort the keys
   when doing a lookup for multiple keys.

 - The fix is to disable HFile positional read for all HFile scanner based
   key lookups.
…bleFileSystemView, aiming to reduce unnecessary list/get requests"

Co-authored-by: yuezhang <yuezhang@freewheel.tv>
* skip shutdown zookeeper in `@AfterAll` in TestHBaseIndex

* rebalance CI tests
…concurrent operations (#4211)

* Fix kafka connect readme

* Fix handling of errors in write records for kafka connect

* By default, ensure we skip error records and keep the pipeline alive

* Fix indentation

Co-authored-by: Rajesh Mahindra <rmahindra@Rajeshs-MacBook-Pro.local>
…inflight (#4206)

* [HUDI-2923] Fixing metadata table reader when metadata compaction is inflight

* Fixing retry of pending compaction in metadata table and enhancing tests
…kpoint retrival (#4216)

- We now seek backwards to find the checkpoint
 - No need to return empty anymore
danny0405 and others added 27 commits December 22, 2021 11:10
Co-authored-by: yuzhaojing <yuzhaojing@bytedance.com>
[HUDI-3008] Fixing HoodieFileIndex partition column parsing for nested fields
…with empty checkpoint (#4334)

* Adding ability to read entire data with HoodieIncrSource with empty checkpoint

* Addressing comments
* [HUDI-3083] Support component data types for flink bulk_insert

* add nested row type test
Co-authored-by: yuzhaojing <yuzhaojing@bytedance.com>
…3173)

* [HUDI-2154] Add index key field to HoodieKey

* [HUDI-2157] Add the bucket index and its read/write implemention of Spark engine.
* revert HUDI-2154 add index key field to HoodieKey
* fix all comments and introduce a new tricky way to get index key at runtime
support double insert for bucket index
* revert spark read optimizer based on bucket index
* add the storage layout
* index tag, hash function and add ut
* fix ut
* address partial comments
* Code review feedback
* add layout config and docs
* fix ut
* rename hoodie.layout and rebase master

Co-authored-by: Vinoth Chandar <vinoth@apache.org>
Co-authored-by: yuzhaojing <yuzhaojing@bytedance.com>
* [HUDI-3095] abstract partition filter logic to enable code reuse

* [HUDI-3095] address reviews
…hms (#4453)

* constructDropPartitions when drop partitions using jdbc

* done

* done

* code style

* code review

Co-authored-by: yuezhang <yuezhang@freewheel.tv>
@hudi-bot
Copy link
Copy Markdown
Collaborator

CI report:

Bot commands @hudi-bot supports the following commands:
  • @hudi-bot run azure re-run the last Azure build

@guoke111 guoke111 changed the title 分支0.10.0 0.10.0 Dec 31, 2021
@guoke111 guoke111 closed this Dec 31, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.