Skip to content

ritchie46/Light_ETL_Challenge

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 

Repository files navigation

Light_ETL_Challenge

extract data fom a csv with different number of columns and export to Delta Lake I started with Duckdb and Polars, but I expect more engines like chdb, Raft etc

the script will download 60 files, around 4 GB uncompressed

Please keep the local Path as

for raw data : raw_landing='/lakehouse/default/Files/raw'

for Delta : '/lakehouse/default/Tables/'

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 100.0%