Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: Enabling automation of experiments running v2.0 #469

Open
wants to merge 24 commits into
base: main
Choose a base branch
from
Open
Changes from 1 commit
Commits
Show all changes
24 commits
Select commit Hold shift + click to select a range
8726ab8
Revising to enable automation of experiments running v1.0
xisen-w Nov 4, 2024
b44bef5
Any new updates
xisen-w Nov 15, 2024
c100876
Revising to enable automation of experiments running v1.0
xisen-w Nov 4, 2024
18370d4
Any new updates
xisen-w Nov 15, 2024
21a99d2
Add template
you-n-g Nov 15, 2024
86ae0b2
Stoping tracking additional env
xisen-w Nov 20, 2024
f94dbff
Merge branch 'automated-evaluation' of https://github.com/microsoft/R…
xisen-w Nov 20, 2024
66ffd6d
Uploading relevant envs
xisen-w Nov 20, 2024
0ef80a5
Adding tests
xisen-w Nov 20, 2024
907d980
Updating
xisen-w Nov 20, 2024
51388d1
Updated collect.py to extract result from trace
xisen-w Nov 23, 2024
af6220e
Update .gitignore to remove the unecessary ones
xisen-w Nov 23, 2024
54c3c6d
"Remove unnecessary files"
xisen-w Nov 23, 2024
78708e4
Merge branch 'automated-evaluation' of https://github.com/microsoft/R…
xisen-w Nov 25, 2024
3f131f3
Merge branch 'main' into automated-evaluation
xisen-w Nov 25, 2024
38bb9e6
Updated to enable automatic collection of experiment result information
xisen-w Nov 25, 2024
10b0053
Updating the env files & Upading test_system file
xisen-w Nov 25, 2024
238f492
Updated relevant env for better testing
xisen-w Nov 25, 2024
68ca63a
Updated README.md
xisen-w Nov 25, 2024
8b18fad
reverting gitignore back
xisen-w Nov 25, 2024
2395dc5
Updates
xisen-w Dec 3, 2024
b7cc98e
README update
xisen-w Dec 3, 2024
0b5a09d
Updates on env README
xisen-w Dec 3, 2024
24cd0c2
Updating collect.py
xisen-w Dec 3, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
Updating
  • Loading branch information
xisen-w committed Nov 20, 2024
commit 907d9805d848f0b398baaa70be2989e07a814898
60 changes: 50 additions & 10 deletions scripts/exp/tools/collect.py
Original file line number Diff line number Diff line change
@@ -1,24 +1,64 @@
import os
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Will the env name (e.g. basic, max, pro) displayed in the collected results?

import json
from pathlib import Path
from datetime import datetime

def collect_results(dir_path) -> list[dict]:
summary = []
for root, _, filies in os.walk(dir_path):
for file in filies:
if file.endswith(".json"):
for root, _, files in os.walk(dir_path):
for file in files:
if file.endswith("_result.json"):
config_name = file.replace("_result.json", "")
with open(os.path.join(root, file), "r") as f:
data = json.load(f)
summary.append(data)
# Extract both CV and Kaggle submission results
summary.append({
"config": config_name,
"cv_results": data.get("cv_score", None),
"kaggle_score": data.get("kaggle_score", None),
"trace": data.get("trace", {})
})
return summary

def generate_summary(results, output_path):
# First analyze the results and generate a summary
# For each experiment, we find the best result, the metric, and result trajectory
#TODO: Implement this

# Then write the summary to the output path
summary = {
"configs": {},
"best_cv_result": {"config": None, "score": None},
"best_kaggle_result": {"config": None, "score": None},
"timestamp": datetime.now().strftime("%Y%m%d_%H%M%S")
}

for result in results:
config = result["config"]
metrics = {
"cv_score": result["cv_results"],
"kaggle_score": result["kaggle_score"],
"iterations": len(result["trace"].get("steps", [])),
"best_model": result["trace"].get("best_model")
}

summary["configs"][config] = metrics

# Update best CV result
if (metrics["cv_score"] is not None and
(summary["best_cv_result"]["score"] is None or
metrics["cv_score"] > summary["best_cv_result"]["score"])):
summary["best_cv_result"].update({
"config": config,
"score": metrics["cv_score"]
})

# Update best Kaggle result
if (metrics["kaggle_score"] is not None and
(summary["best_kaggle_result"]["score"] is None or
metrics["kaggle_score"] > summary["best_kaggle_result"]["score"])):
summary["best_kaggle_result"].update({
"config": config,
"score": metrics["kaggle_score"]
})

with open(output_path, "w") as f:
json.dump(results, f, indent = 4)
json.dump(summary, f, indent=4)

if __name__ == "__main__":
result_dir = os.path.join(os.getenv("EXP_DIR"), "results")
16 changes: 11 additions & 5 deletions scripts/exp/tools/test_system.sh
Original file line number Diff line number Diff line change
@@ -3,19 +3,25 @@
# Test directory setup
TEST_DIR="test_run"
mkdir -p "$TEST_DIR/results"
mkdir -p "$TEST_DIR/logs"

# Test 1: Environment loading
# Test 1: Environment loading verification
echo "Testing environment loading..."
./scripts/exp/tools/run_envs.sh -d scripts/exp/ablation/env -j 1 -- env | grep "if_using"

# Test 2: Parallel execution
echo "Testing parallel execution..."
# Test 2: Run actual experiments
echo "Running experiments with different configurations..."
./scripts/exp/tools/run_envs.sh -d scripts/exp/ablation/env -j 4 -- \
echo "Processing env with RAG setting: $if_using_vector_rag"
python -m rdagent.app.kaggle.loop \
--competition "titanic" \
--result_path "${TEST_DIR}/results/$(basename {} .env)_result.json"

# Test 3: Result collection
echo "Testing result collection..."
echo "Collecting and analyzing results..."
EXP_DIR="$TEST_DIR" python scripts/exp/tools/collect.py

# Display results location
echo "Test results available at: $TEST_DIR"

# Cleanup
rm -rf "$TEST_DIR"
Loading
Oops, something went wrong.