Skip to content
Upserts And Incremental Processing on Big Data
Branch: master
Clone or download
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
deploy Add ossrh profile to publish maven artifacts to oss.sonatype.org (syn… Dec 21, 2016
docker Timeline Service with Incremental View Syncing support May 16, 2019
hoodie-cli Timeline Service with Incremental View Syncing support May 16, 2019
hoodie-client Timeline Service with Incremental View Syncing support May 16, 2019
hoodie-common Timeline Service with Incremental View Syncing support May 16, 2019
hoodie-hadoop-mr [maven-release-plugin] prepare for next development iteration May 14, 2019
hoodie-hive [maven-release-plugin] prepare for next development iteration May 14, 2019
hoodie-integ-test [maven-release-plugin] prepare for next development iteration May 14, 2019
hoodie-spark Support nested types for recordKey, partitionPath and combineKey May 18, 2019
hoodie-timeline-service Timeline Service with Incremental View Syncing support May 16, 2019
hoodie-utilities Support nested types for recordKey, partitionPath and combineKey May 18, 2019
packaging Timeline Service with Incremental View Syncing support May 16, 2019
style General enhancements Dec 18, 2018
.gitignore Bucketized Bloom Filter checking May 11, 2019
.travis.yml Add m2 directory to Travis cache Dec 31, 2018
CHANGELOG.md Added CHANGELOG.md and updated community contributions guideline Jun 16, 2017
KEYS HUDI-75: Add KEYS Mar 18, 2019
LICENSE.txt Importing Hoodie Client from internal repo Dec 16, 2016
README.md Update site url in README Feb 16, 2019
RELEASE_NOTES.md Update release notes for 0.4.6 release May 14, 2019
_config.yml Set theme jekyll-theme-minimal Dec 30, 2016
pom.xml Downgrading fasterxml jackson to 2.6.7 to be spark compatible May 16, 2019

README.md

Hudi

Hudi (pronounced Hoodie) stands for Hadoop Upserts anD Incrementals. Hudi manages storage of large analytical datasets on HDFS and serve them out via two types of tables

  • Read Optimized Table - Provides excellent query performance via purely columnar storage (e.g. Parquet)
  • Near-Real time Table (WIP) - Provides queries on real-time data, using a combination of columnar & row based storage (e.g Parquet + Avro)

For more, head over here

You can’t perform that action at this time.