{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"dreamitate","owner":"cvlab-columbia","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":26,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-25T22:16:57.417Z"}},{"type":"Public","name":"pix2gestalt","owner":"cvlab-columbia","isFork":false,"description":"Code for the paper \"pix2gestalt: Amodal Segmentation by Synthesizing Wholes\" (CVPR 2024)","allTopics":["recognition","3d","zero-shot","amodal-perception","amodal-instance-segmentation","stable-diffusion","cvpr2024"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":119,"forksCount":8,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-03T02:48:53.532Z"}},{"type":"Public","name":"paperbot","owner":"cvlab-columbia","isFork":false,"description":"PaperBot: Learning to Design Real-World Tools Using Paper","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":11,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-15T01:25:57.334Z"}},{"type":"Public","name":"viper","owner":"cvlab-columbia","isFork":false,"description":"Code for the paper \"ViperGPT: Visual Inference via Python Execution for Reasoning\"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":12,"starsCount":1635,"forksCount":117,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-29T21:29:37.427Z"}},{"type":"Public","name":"RaidarLLMDetect","owner":"cvlab-columbia","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":9,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-29T16:36:59.686Z"}},{"type":"Public","name":"DoubleRight","owner":"cvlab-columbia","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":25,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-25T15:36:14.647Z"}},{"type":"Public","name":"zero123","owner":"cvlab-columbia","isFork":false,"description":"Zero-1-to-3: Zero-shot One Image to 3D Object (ICCV 2023)","allTopics":["zero-shot","novel-view-synthesis","image-to-3d","single-view-reconstruction","stable-diffusion"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":50,"starsCount":2580,"forksCount":188,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-05T09:55:45.680Z"}},{"type":"Public","name":"expert","owner":"cvlab-columbia","isFork":false,"description":"Code for Learning to Learn Language from Narrated Video","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":1,"starsCount":33,"forksCount":6,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-03T22:35:05.453Z"}},{"type":"Public","name":"Equi4Rob","owner":"cvlab-columbia","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-08-20T18:59:34.740Z"}},{"type":"Public","name":"hyperfuture","owner":"cvlab-columbia","isFork":false,"description":"Code for the paper Learning the Predictability of the Future (CVPR 2021)","allTopics":["video-representation-learning","self-supervised-learning","hyperbolic-embeddings","uncertainty-modeling","future-predi"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":159,"forksCount":25,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-07-31T15:24:34.646Z"}},{"type":"Public","name":"ZSRobust4FoundationModel","owner":"cvlab-columbia","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":33,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-06-11T19:09:49.278Z"}},{"type":"Public","name":"surfsup","owner":"cvlab-columbia","isFork":false,"description":"Code for the paper- SURFSUP: Learning Fluid Simulation for Novel Surfaces","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":1,"starsCount":2,"forksCount":0,"license":"Creative Commons Zero v1.0 Universal","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-10T15:01:39.090Z"}},{"type":"Public","name":"trajectories","owner":"cvlab-columbia","isFork":false,"description":"Code for the paper \"Representing Spatial Trajectories as Distributions\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":10,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-01-17T16:09:20.602Z"}},{"type":"Public","name":"CT4Recognition","owner":"cvlab-columbia","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":6,"starsCount":21,"forksCount":8,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-06-08T18:28:33.164Z"}},{"type":"Public","name":"voicecamo","owner":"cvlab-columbia","isFork":false,"description":"Code for the paper Real-Time Neural Voice Camouflage","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":28,"forksCount":5,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-04-13T18:44:10.129Z"}},{"type":"Public","name":"globetrotter","owner":"cvlab-columbia","isFork":false,"description":"Code for the Globetrotter project","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":23,"forksCount":6,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-03-17T22:43:04.786Z"}},{"type":"Public","name":"SelfSupDefense","owner":"cvlab-columbia","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":28,"forksCount":5,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-02-17T22:04:28.585Z"}},{"type":"Public","name":"GenInt","owner":"cvlab-columbia","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":19,"forksCount":3,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-12-25T04:51:47.875Z"}},{"type":"Public","name":"aha","owner":"cvlab-columbia","isFork":false,"description":"Code for Video Representations of Goals Emerge from Watching Failure","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-10-22T04:39:39.063Z"}},{"type":"Public","name":"oops","owner":"cvlab-columbia","isFork":false,"description":"Code for Oops! Predicting Unintentional Action in Video","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":4,"starsCount":77,"forksCount":11,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-04-13T16:26:28.102Z"}},{"type":"Public","name":"CATER","owner":"cvlab-columbia","isFork":true,"description":"CATER: A diagnostic dataset for Compositional Actions and TEmporal Reasoning","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":1,"forksCount":19,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-03-19T21:56:20.068Z"}}],"repositoryCount":21,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"cvlab-columbia repositories"}