Skip to content
forked from rsyi/whale

🐳 The stupidly simple data discovery tool.

License

Notifications You must be signed in to change notification settings

peterthesling/whale

 
 

Repository files navigation

Whale is actively being built and maintained by Dataframe. For our full, collaborative, and just as simple data discovery platform, check out dataframe.ai.

The simplest way to keep track of your warehouse tables

whale is a lightweight data discovery, documentation, and quality engine for your data warehouse.

  • Automatically index all of the tables in your warehouse as plain markdown files -- so they're easily versionable, searchable, and editable either locally or through a remote git server.
  • Search for tables and documentation through the CLI or through a git remote server like Github.
  • Define and schedule basic metric calculations (in beta).
  • Run quality tests and systematically monitor anomalies (in roadmap).

😁 Join the discussion on slack.


codecov slack

For a live demo, check out dataframehq/whale-bigquery-public-data.

📔 Documentation

Read the docs for a full overview of whale's capabilities.

Installation

Mac OS

brew install dataframehq/tap/whale

All others

Make sure rust is installed on your local system. Then, clone this directory and run the following in the base directory of the repo:

make && make install

If you are running this multiple times, make sure ~/.whale/libexec does not exist, or your virtual environment may not rebuild. We don't explicitly add an alias for the whale binary, so you'll want to add the following alias to your .bash_profile or .zshrc file.

alias wh=~/.whale/bin/whale

Getting started

Setup

For individual use, run the following command to go through the onboarding process. It will (a) set up all necessary files in ~/.whale, (b) walk you through cron job scheduling to periodically scrape metadata, and (c) set up a warehouse:

wh init

The cron job will run as you schedule it (by default, every 6 hours). If you're feeling impatient, you can also manually run wh etl to pull down the latest data from your warehouse.

For team use, see the docs for instructions on how to set up and point your whale installation at a remote git server.

Seeding some sample data

If you just want to get a feel for how whale works, remove the ~/.whale directory and follow the instructions at dataframehq/whale-bigquery-public-data.

Go go go!

Run:

wh

to search over all metadata. Hitting enter will open the editable part of the docs in your default text editor, defined by the environmental variable $EDITOR (if no value is specified, whale will use the command open).

About

🐳 The stupidly simple data discovery tool.

Resources

License

Code of conduct

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Python 83.1%
  • Rust 16.6%
  • Makefile 0.3%