{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"dsir","owner":"p-lambda","isFork":false,"description":"DSIR large-scale data selection framework for language model training","topicNames":["data","language-models","large-scale","data-filtering","data-selection","importance-resampling"],"topicsNotShown":0,"allTopics":["data","language-models","large-scale","data-filtering","data-selection","importance-resampling"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":4,"starsCount":195,"forksCount":16,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-07T09:15:33.286Z"}},{"type":"Public","name":"wilds","owner":"p-lambda","isFork":false,"description":"A machine learning benchmark of in-the-wild distribution shifts, with data loaders, evaluators, and default models.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":9,"starsCount":540,"forksCount":123,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-26T17:09:51.779Z"}},{"type":"Public","name":"incontext-learning","owner":"p-lambda","isFork":false,"description":"Experiments and code to generate the GINC small-scale in-context learning dataset from \"An Explanation for In-context Learning as Implicit Bayesian Inference\" ","topicNames":["language-modeling","dataset","few-shot-learning","gpt-3","in-context-learning"],"topicsNotShown":0,"allTopics":["language-modeling","dataset","few-shot-learning","gpt-3","in-context-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":91,"forksCount":12,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-10T17:17:05.874Z"}},{"type":"Public","name":"jukemir","owner":"p-lambda","isFork":false,"description":"Perform transfer learning for MIR using Jukebox!","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Shell","color":"#89e051"},"pullRequestCount":0,"issueCount":3,"starsCount":165,"forksCount":22,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-12T15:50:09.714Z"}},{"type":"Public","name":"verified_calibration","owner":"p-lambda","isFork":false,"description":"Calibration library and code for the paper: Verified Uncertainty Calibration. Ananya Kumar, Percy Liang, Tengyu Ma. NeurIPS 2019 (Spotlight).","topicNames":["metrics","toolbox","calibration","ece","uncertainty-calibration","calibration-library"],"topicsNotShown":0,"allTopics":["metrics","toolbox","calibration","ece","uncertainty-calibration","calibration-library"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":7,"starsCount":133,"forksCount":19,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-11-10T22:09:24.344Z"}},{"type":"Public","name":"dragon","owner":"p-lambda","isFork":true,"description":"[NeurIPS 2022] DRAGON 🐲: Deep Bidirectional Language-Knowledge Graph Pretraining ","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":45,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-10-19T01:28:04.283Z"}},{"type":"Public","name":"LinkBERT","owner":"p-lambda","isFork":true,"description":"[ACL 2022] LinkBERT: A Knowledgeable Language Model 😎 Pretrained with Document Links","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":40,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-04-05T16:55:46.398Z"}},{"type":"Public","name":"swords","owner":"p-lambda","isFork":false,"description":"The Stanford Word Substitution (Swords) Benchmark","topicNames":["nlp","lexical-substitution","writing-assistant","benchmark","swords"],"topicsNotShown":0,"allTopics":["nlp","lexical-substitution","writing-assistant","benchmark","swords"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":2,"starsCount":29,"forksCount":6,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-03-24T18:40:51.567Z"}},{"type":"Public","name":"composed_finetuning","owner":"p-lambda","isFork":false,"description":"Code for the ICML 2021 paper \"Composed Fine-Tuning: Freezing Pre-Trained Denoising Autoencoders for Improved Generalization\" by Sang Michael Xie, Tengyu Ma, Percy Liang","topicNames":["translation","semi-supervised-learning","image-generation","structured-prediction","ood-generalization"],"topicsNotShown":0,"allTopics":["translation","semi-supervised-learning","image-generation","structured-prediction","ood-generalization"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":5,"forksCount":1,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-12-28T20:26:16.799Z"}},{"type":"Public","name":"in-n-out","owner":"p-lambda","isFork":false,"description":"Code for the ICLR 2021 Paper \"In-N-Out: Pre-Training and Self-Training using Auxiliary Information for Out-of-Distribution Robustness\"","topicNames":["self-training","robustness","pre-training","out-of-distribution"],"topicsNotShown":0,"allTopics":["self-training","robustness","pre-training","out-of-distribution"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":12,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-10-23T21:27:56.292Z"}},{"type":"Public","name":"gradual_domain_adaptation","owner":"p-lambda","isFork":false,"description":"","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":36,"forksCount":12,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-04-16T20:57:36.785Z"}},{"type":"Public","name":"robust_tradeoff","owner":"p-lambda","isFork":false,"description":"Code for the ICML 2020 paper \"Understanding and Mitigating the Tradeoff Between Robustness and Accuracy\", Aditi Raghunathan, Sang Michael Xie, Fanny Yang, John Duchi, and Percy Liang. Paper available at https://arxiv.org/pdf/2002.10716.pdf. ","topicNames":["robustness","adversarial-examples","tradeoff"],"topicsNotShown":0,"allTopics":["robustness","adversarial-examples","tradeoff"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":8,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-01-04T18:36:54.588Z"}}],"repositoryCount":12,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}