Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

v1.0: Code Cleanup / Streamlining #2

Open
9 of 25 tasks
Innixma opened this issue Apr 23, 2023 · 0 comments
Open
9 of 25 tasks

v1.0: Code Cleanup / Streamlining #2

Innixma opened this issue Apr 23, 2023 · 0 comments

Comments

@Innixma
Copy link
Collaborator

Innixma commented Apr 23, 2023

We should ensure all primary components in the code-base are cleaned up to not be overly complicated / hacky, and to enhance ease of contribution and maintenance, while minimizing code-dupe.

  • Running new benchmarks / configs (AutoGluon-Benchmark, AutoMLBenchmark)
    • Simplify generation of BenchmarkContexts for new benchmarks.
    • Standardize S3 file location + naming
    • Automate retries on failed runs
    • Repeated cross-validation support for reducing simulation overfitting
    • Run configs in nested bagging mode to eliminate validation overfitting
    • Add extra model config results: KNN, NN, XGB, FT-Transformer, TabPFN, Linear, VowpalWabbit, etc.
    • Expand search space for key models (CAT, FASTAI)
    • Augment datasets for more information (subsample rows for example)
    • Collect more high quality datasets (Ex: Kaggle, MachineHack)
  • Downloading input files from benchmarks for simulation & loading them (Missing results in metrics #15)
  • Handling & Accessing Simulation Contexts (Code to reproduce paper figures #18)
  • Handling & Accessing Model Predictions (WIP: Add time_limit support #13)
  • Running & Scoring Portfolios (ConfigurationListScorer)
  • Conducting Zeroshot Simulations (ZeroshotConfigGenerator, ZeroshotConfigGeneratorCV)
  • Generating and evaluating baselines + comparing against simulation results
    • Oracle baselines
    • Simple baselines
  • Storing results, extracting analytics, and comparison results from simulations (PortfolioCV, AutoGluon-Benchmark integration)
    • Properly calculate inference speed per task for portfolios
    • Properly calculate scores for portfolios when time_limit is specified.
    • Properly calculate scores for portfolios when infer_limit is specified.
    • Enable extra logic such as temperature scaling to match what occurs in AutoGluon.
  • Automate running portfolios on the benchmark to verify results + test additional features like repeated bagging + stacking for learned portfolios.
    • Do so without overfitting (cross-validated portfolios)
@Innixma Innixma changed the title Code Cleanup / Streamlining v1.0: Code Cleanup / Streamlining Apr 24, 2023
@Innixma Innixma transferred this issue from Innixma/autogluon-zeroshot May 30, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant