Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
72 changes: 72 additions & 0 deletions .github/ISSUE_TEMPLATE/bug_report.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,72 @@
name: Bug Report
description: File a bug report
title: "[Bug]: "
labels: ["bug"]
body:
- type: markdown
attributes:
value: |
Thanks for taking the time to fill out this bug report!
- type: textarea
id: what-happened
attributes:
label: What happened?
description: A clear description of what the bug is
placeholder: Tell us what you see!
validations:
required: true
- type: textarea
id: reproduce
attributes:
label: Steps to reproduce
description: How do we reproduce this?
placeholder: |
1.
2.
3.
validations:
required: true
- type: textarea
id: expected
attributes:
label: Expected behavior
description: What did you expect to happen?
validations:
required: true
- type: input
id: version
attributes:
label: Version
description: What version of Artifacta are you running?
placeholder: "0.1.1"
validations:
required: true
- type: dropdown
id: os
attributes:
label: Operating System
options:
- macOS
- Linux
- Windows
- Other
validations:
required: true
- type: input
id: python-version
attributes:
label: Python Version
placeholder: "3.11"
validations:
required: true
- type: textarea
id: logs
attributes:
label: Relevant log output
description: Please copy and paste any relevant log output
render: shell
- type: textarea
id: additional
attributes:
label: Additional context
description: Add any other context about the problem here
5 changes: 5 additions & 0 deletions .github/ISSUE_TEMPLATE/config.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,5 @@
blank_issues_enabled: false
contact_links:
- name: Ask a Question
url: https://github.com/walkerbdev/artifacta/discussions
about: Ask questions and discuss with the community
39 changes: 39 additions & 0 deletions .github/ISSUE_TEMPLATE/feature_request.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,39 @@
name: Feature Request
description: Suggest an idea for this project
title: "[Feature]: "
labels: ["enhancement"]
body:
- type: markdown
attributes:
value: |
Thanks for suggesting a new feature!
- type: textarea
id: problem
attributes:
label: Is your feature request related to a problem?
description: A clear description of what the problem is
placeholder: I'm always frustrated when...
- type: textarea
id: solution
attributes:
label: Describe the solution you'd like
description: A clear description of what you want to happen
validations:
required: true
- type: textarea
id: alternatives
attributes:
label: Describe alternatives you've considered
description: Any alternative solutions or features you've considered
- type: textarea
id: use-case
attributes:
label: Use case
description: Describe your use case and how this feature would help
validations:
required: true
- type: textarea
id: additional
attributes:
label: Additional context
description: Add any other context or screenshots about the feature request
30 changes: 30 additions & 0 deletions .github/pull_request_template.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,30 @@
## Description

Please include a summary of the changes and the related issue.

Fixes # (issue)

## Type of change

- [ ] Bug fix (non-breaking change which fixes an issue)
- [ ] New feature (non-breaking change which adds functionality)
- [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
- [ ] Documentation update

## How Has This Been Tested?

Please describe the tests that you ran to verify your changes.

- [ ] Test A
- [ ] Test B

## Checklist

- [ ] My code follows the style guidelines of this project
- [ ] I have performed a self-review of my own code
- [ ] I have commented my code, particularly in hard-to-understand areas
- [ ] I have made corresponding changes to the documentation
- [ ] My changes generate no new warnings
- [ ] I have added tests that prove my fix is effective or that my feature works
- [ ] New and existing unit tests pass locally with my changes
- [ ] I have run `pre-commit run --all-files` and it passes
85 changes: 85 additions & 0 deletions CODE_OF_CONDUCT.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,85 @@

# Contributor Covenant Code of Conduct

## Our Pledge

We as members, contributors, and leaders pledge to make participation in our community a harassment-free experience for everyone, regardless of age, body size, visible or invisible disability, ethnicity, sex characteristics, gender identity and expression, level of experience, education, socio-economic status, nationality, personal appearance, race, caste, color, religion, or sexual identity and orientation.

We pledge to act and interact in ways that contribute to an open, welcoming, diverse, inclusive, and healthy community.

## Our Standards

Examples of behavior that contributes to a positive environment for our community include:

* Demonstrating empathy and kindness toward other people
* Being respectful of differing opinions, viewpoints, and experiences
* Giving and gracefully accepting constructive feedback
* Accepting responsibility and apologizing to those affected by our mistakes, and learning from the experience
* Focusing on what is best not just for us as individuals, but for the overall community

Examples of unacceptable behavior include:

* The use of sexualized language or imagery, and sexual attention or advances of any kind
* Trolling, insulting or derogatory comments, and personal or political attacks
* Public or private harassment
* Publishing others' private information, such as a physical or email address, without their explicit permission
* Other conduct which could reasonably be considered inappropriate in a professional setting

## Enforcement Responsibilities

Community leaders are responsible for clarifying and enforcing our standards of acceptable behavior and will take appropriate and fair corrective action in response to any behavior that they deem inappropriate, threatening, offensive, or harmful.

Community leaders have the right and responsibility to remove, edit, or reject comments, commits, code, wiki edits, issues, and other contributions that are not aligned to this Code of Conduct, and will communicate reasons for moderation decisions when appropriate.

## Scope

This Code of Conduct applies within all community spaces, and also applies when an individual is officially representing the community in public spaces. Examples of representing our community include using an official e-mail address, posting via an official social media account, or acting as an appointed representative at an online or offline event.

## Enforcement

Instances of abusive, harassing, or otherwise unacceptable behavior may be reported to the community leaders responsible for enforcement at [INSERT CONTACT METHOD]. All complaints will be reviewed and investigated promptly and fairly.

All community leaders are obligated to respect the privacy and security of the reporter of any incident.

## Enforcement Guidelines

Community leaders will follow these Community Impact Guidelines in determining the consequences for any action they deem in violation of this Code of Conduct:

### 1. Correction

**Community Impact**: Use of inappropriate language or other behavior deemed unprofessional or unwelcome in the community.

**Consequence**: A private, written warning from community leaders, providing clarity around the nature of the violation and an explanation of why the behavior was inappropriate. A public apology may be requested.

### 2. Warning

**Community Impact**: A violation through a single incident or series of actions.

**Consequence**: A warning with consequences for continued behavior. No interaction with the people involved, including unsolicited interaction with those enforcing the Code of Conduct, for a specified period of time. This includes avoiding interactions in community spaces as well as external channels like social media. Violating these terms may lead to a temporary or permanent ban.

### 3. Temporary Ban

**Community Impact**: A serious violation of community standards, including sustained inappropriate behavior.

**Consequence**: A temporary ban from any sort of interaction or public communication with the community for a specified period of time. No public or private interaction with the people involved, including unsolicited interaction with those enforcing the Code of Conduct, is allowed during this period. Violating these terms may lead to a permanent ban.

### 4. Permanent Ban

**Community Impact**: Demonstrating a pattern of violation of community standards, including sustained inappropriate behavior, harassment of an individual, or aggression toward or disparagement of classes of individuals.

**Consequence**: A permanent ban from any sort of public interaction within the community.

## Attribution

This Code of Conduct is adapted from the [Contributor Covenant][homepage], version 2.1, available at [https://www.contributor-covenant.org/version/2/1/code_of_conduct.html][v2.1].

Community Impact Guidelines were inspired by [Mozilla's code of conduct enforcement ladder][Mozilla CoC].

For answers to common questions about this code of conduct, see the FAQ at [https://www.contributor-covenant.org/faq][FAQ]. Translations are available at [https://www.contributor-covenant.org/translations][translations].

[homepage]: https://www.contributor-covenant.org
[v2.1]: https://www.contributor-covenant.org/version/2/1/code_of_conduct.html
[Mozilla CoC]: https://github.com/mozilla/diversity
[FAQ]: https://www.contributor-covenant.org/faq
[translations]: https://www.contributor-covenant.org/translations

43 changes: 33 additions & 10 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,34 @@

---

## 🎯 The Problem

Modern data science and machine learning workflows involve countless experiments—tweaking hyperparameters, adjusting data preprocessing, testing different architectures, updating dependencies, modifying code. **Every change produces different results**, but tracking and comparing these variations manually becomes overwhelming:

- 📋 Which parameters, environment, or code version led to that breakthrough result last week?
- 🔍 How does changing the learning rate affect convergence across multiple runs?
- 📊 What's the actual performance difference between model architectures?
- 🤔 Which preprocessing steps improved accuracy by 2%?
- 🔧 Did upgrading that dependency break model performance?
- 💻 What code changes caused the regression?

Without systematic tracking of **parameters, metrics, code changes, dependencies, and environment**, you're flying blind—relying on scattered notes, terminal output, and memory. **Artifacta solves this** by automatically capturing experiments, configurations, code versions, and artifacts in one place with intelligent visualization.

---

## 🌍 Ecosystem & Alternatives

Artifacta is part of a growing ecosystem of experiment tracking tools. Popular alternatives include:

- [**MLflow**](https://mlflow.org/) - Open-source platform from Databricks for ML lifecycle management
- [**Weights & Biases**](https://wandb.ai/) - Cloud-first experiment tracking with team collaboration features
- [**Neptune.ai**](https://neptune.ai/) - Metadata store for MLOps with extensive integrations
- [**Comet ML**](https://www.comet.com/) - ML platform with experiment tracking and model production monitoring

**Why Artifacta?** We focus on **automatic visualization discovery**, **domain-agnostic tracking** (not just ML), and **simple self-hosting** with a pre-built UI. No heavy dependencies, no mandatory cloud services—just install and start tracking.

---

## ✨ Key Features

- 🌐 **Domain-agnostic** - Track any experiment comparing parameters, data, and outcomes
Expand Down Expand Up @@ -43,23 +71,18 @@ Browse and preview datasets, models, code, images, videos, and documents with bu

### Installation

#### Standard Installation

**Prerequisites:** Python 3.9+

```bash
# Clone the repository
git clone https://github.com/walkerbdev/artifacta.git
cd artifacta

# Install Python package
pip install -e .
pip install artifacta
```

**Note:** The UI is pre-built and bundled. No Node.js required.
That's it! The UI is pre-built and bundled. No Node.js required.

#### Development Installation

If you want to contribute or modify the source:

**Prerequisites:** Python 3.9+, Node.js 16+

```bash
Expand All @@ -70,7 +93,7 @@ cd artifacta
# Build UI from source
npm install && npm run build

# Install Python package
# Install Python package in editable mode
pip install -e .
```

Expand Down
42 changes: 29 additions & 13 deletions docs/user-guide.rst
Original file line number Diff line number Diff line change
@@ -1,24 +1,40 @@
Quick Start Guide
=================

The Problem
-----------

Modern data science and machine learning workflows involve countless experiments—tweaking hyperparameters, adjusting data preprocessing, testing different architectures, updating dependencies, modifying code. **Every change produces different results**, but tracking and comparing these variations manually becomes overwhelming:

- 📋 Which parameters, environment, or code version led to that breakthrough result last week?
- 🔍 How does changing the learning rate affect convergence across multiple runs?
- 📊 What's the actual performance difference between model architectures?
- 🤔 Which preprocessing steps improved accuracy by 2%?
- 🔧 Did upgrading that dependency break model performance?
- 💻 What code changes caused the regression?

Without systematic tracking of **parameters, metrics, code changes, dependencies, and environment**, you're flying blind—relying on scattered notes, terminal output, and memory. **Artifacta solves this** by automatically capturing experiments, configurations, code versions, and artifacts in one place with intelligent visualization.

Ecosystem & Alternatives
------------------------

Artifacta is part of a growing ecosystem of experiment tracking tools. Popular alternatives include:

- `MLflow <https://mlflow.org/>`_ - Open-source platform from Databricks for ML lifecycle management
- `Weights & Biases <https://wandb.ai/>`_ - Cloud-first experiment tracking with team collaboration features
- `Neptune.ai <https://neptune.ai/>`_ - Metadata store for MLOps with extensive integrations
- `Comet ML <https://www.comet.com/>`_ - ML platform with experiment tracking and model production monitoring

**Why Artifacta?** We focus on **automatic visualization discovery**, **domain-agnostic tracking** (not just ML), and **simple self-hosting** with a pre-built UI. No heavy dependencies, no mandatory cloud services—just install and start tracking.

Installation
------------

Standard Installation
~~~~~~~~~~~~~~~~~~~~~

**Prerequisites:** Python 3.9+

For most users, this is the recommended installation method:

.. code-block:: bash

# Clone the repository
git clone https://github.com/walkerbdev/artifacta.git
cd artifacta

# Install Python package with bundled UI
pip install -e .
pip install artifacta

.. note::
The UI is pre-built and bundled with the package. No Node.js required.
Expand All @@ -28,7 +44,7 @@ Development Installation

**Prerequisites:** Python 3.9+, Node.js 16+

For contributors who want to modify the UI:
For contributors who want to modify the source code or UI:

.. code-block:: bash

Expand All @@ -39,7 +55,7 @@ For contributors who want to modify the UI:
# Build UI from source
npm install && npm run build

# Install Python package
# Install Python package in editable mode
pip install -e .

To run examples or tests, install with optional dependencies:
Expand Down