Skip to content
@opendatahub-io

Open Data Hub

Organization for Open Data Hub community

Welcome to Open Data Hub 👋

Open Data Hub (ODH) is an open source project that provides open source AI tools for running large and distributed AI workloads on the OpenShift Container Platform. Currently, the Open Data Hub project provides open source tools for distributed AI and Machine Learning (ML) workflows, Jupyter Notebook development environment and monitoring. The Open Data Hub project roadmap offers a view on new tools and integration the project developers are planning to add.

Included in the Open Data Hub core deployment are several open source components, which can be individually enabled. They include:

  • Jupyter Notebooks
  • ODH Dashboard
  • Data Science Pipelines Operator
  • Model Mesh Serving

For further information refer to the architecture and release notes.

Special Interest Groups and Working Groups

Open Data Hub (ODH) adopted an organization and governance model to foster transparent decision making and innovation. The project is organized into Special Interest Groups (SIG) and Working Groups (WG). The table below offers an overview of the current structure. For more information, read the Governance section, or the individual SIG/WG links:

Name Chairs Scope Repositories
SIG ML Dev Experience - Jay Koehler
- Kyle Walker
- Erwan Granger
- Romeo Kienzler
End-to-end experience for a data scientist, a data analyst or an application developer sig-ml-developer-experience
odh-dashboard
notebooks
SIG ML Ops - Prasanth Anbalagan
- Anish Asthana
- Nick Hill
- Giulio Frasca
Services and APIs responsible for providing ML Ops capabilities to users of the Open Data Hub modelmesh-serving
odh-model-controller
rest-proxy
modelmesh
modelmesh-runtime-adapter
data-science-pipelines-operator
data-science-pipelines
SIG Platform - Taneem Ibrahim
- Landon LaSmith
- Vaishinavi Hire
- Gabe Goodheart
ODH operator, odh-core components, ecosystem integration, authentication, and monitoring opendatahub-operator
odh-manifests
WG Distributed Workloads - Anish Asthana
- Michael Clifford
- Mustafa Eyceoz
Easy and user-friendly access to distributed workloads and workload orchestration on Open Data Hub distributed-workloads
WG On-prem - Landon LaSmith
- JooHo Lee
- Vaishnavi Hire
Guidance for on premises deployments of Open Data Hub
WG Explainable AI - Rebecca Whitworth
- Tommaso Teofili
- Rui Vieira
- Rob Geada
Explainability framework for ODH developers trustyai-explainability


Important Links

Contribution Guidelines: https://github.com/opendatahub-io/opendatahub-community/blob/master/contributing.md

Roadmap: https://opendatahub.io/docs/roadmap/future.html

Getting Started: https://opendatahub.io/docs/getting-started/quick-installation.html

Repositories

Core

  • opendatahub-operator: A repository that helps to deploy, monitor and manage the lifecycle of Kubeflow. Built using the Operator Framework which offers an open source toolkit to build, test, package operators and manage the lifecycle of operators.
  • odh-dashboard: A dashboard for Open Data Hub components. Shows what's installed and what's available for installation, as well as, links to component UIs and links to component documentation.
  • odh-manifests: A repository for Open Data Hub components Kustomize manifests.
  • odh-notebook-controller: This repository contains the code for the Open Data Hub notebook controller with support for OAuth authentication.
  • data-science-pipelines: Project bringing Kubeflow Pipelines and Tekton together. The current code allows you run Kubeflow Pipelines with Tekton backend end to end.
  • modelmesh: Distributed Model Serving Framework.
    • modelmesh-runtime-adapter: Unified runtime-adapter image of the sidecar containers which run in the modelmesh pods
    • rest-proxy: KServe V2 Protocol Rest API Implementation Proxy.
    • modelmesh-serving: The Controller for managing ModelMesh, a general-purpose model serving management/routing layer.

Utilities

  • odh-s2i-project-cookiecutter: A logical, reasonably standardized, but flexible project structure for doing and sharing data science work.
  • notebooks: These images were created to be used with Open Data Hub (ODH) using the ODH Notebook Controller as the launcher.
  • s2i-lab-elyra: JupyterHub enabled image deploying JupyterLab with Elyra.
  • odh-images:Source files for building container images that have been customized for Open Data Hub deployment.
  • manifests: A repository for Kustomize manifests.

Test

  • openshift-test-kit: Repos for helper module used with the Peak test framework.
  • peak: This test runner is based on the radanalyticsio openshift-test-kit repository.

Templates

  • odh-template-sig: Repository template for Open Data Hub Special Interest Groups.
  • odh-s2i-project-simple: Simple, unstructured, and unopinionated project for data science that is deployable as an OpenShift s2i Application.
  • odh-s2i-project-cds: A logical, reasonably standardized, but flexible project structure for doing and sharing data science work.

Contribution

  • opendatahub-community: This is the starting point for joining and contributing to the Open Data Hub community.

Other

Archived

  • data-engineering-and-machine-learning-workshop
  • Fraud-Detection-with-Business-Workflows-Tutorial
  • odh-model-controller

Pinned

  1. A repository for Open Data Hub Kustomize manifests extending upstream Kubeflow manifests

    Shell 56 176

  2. Forked from kubeflow/kfctl

    Open Data Hub operator implementation, based on Kubeflow Operator

    Go 37 31

  3. Dashboard for ODH

    TypeScript 17 87

Repositories

Top languages

Loading…

Most used topics

Loading…