{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"HomeGPT","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-10T22:22:42.219Z"}},{"type":"Public","name":"FishNet2.0","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-09T09:09:07.409Z"}},{"type":"Public","name":"MiniGPT4-video","owner":"Vision-CAIR","isFork":false,"description":"Official code for MiniGPT4-video ","allTopics":["video-understanding","video-question-answering"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":23,"starsCount":420,"forksCount":44,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,55,19,19,25,14,11,1,0,2,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-19T20:03:46.516Z"}},{"type":"Public","name":"MiniGPT-4","owner":"Vision-CAIR","isFork":false,"description":"Open-sourced codes for MiniGPT-4 and MiniGPT-v2 (https://minigpt-4.github.io, https://minigpt-v2.github.io/)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":21,"issueCount":328,"starsCount":25076,"forksCount":2896,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-22T11:51:48.838Z"}},{"type":"Public","name":"MammalNet","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":23,"forksCount":3,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-15T14:25:15.906Z"}},{"type":"Public","name":"3DCoMPaT-v2","owner":"Vision-CAIR","isFork":false,"description":"3DCoMPaT++: An improved large-scale 3D vision dataset for compositional recognition","allTopics":["computer-vision","deep-learning","3d","multimodal-deep-learning","compositional-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":69,"forksCount":7,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-31T10:09:04.890Z"}},{"type":"Public","name":"LTVRR","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":8,"starsCount":35,"forksCount":6,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-21T04:07:49.676Z"}},{"type":"Public","name":"RelTransformer","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":29,"forksCount":6,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-04T19:38:03.735Z"}},{"type":"Public","name":"affectiveVisDial","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":6,"forksCount":2,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,3,1,4,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-09-27T08:14:18.797Z"}},{"type":"Public","name":"affective-vision-language","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-08-29T15:45:20.055Z"}},{"type":"Public","name":"VisualGPT","owner":"Vision-CAIR","isFork":false,"description":"VisualGPT, CVPR 2022 Proceeding, GPT as a decoder for vision-language models","allTopics":["image-caption","visualgpt","data-efficient-image-caption"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":3,"starsCount":307,"forksCount":48,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-05-16T06:13:12.854Z"}},{"type":"Public","name":"saai-factory-tutorial-creative-ai","owner":"Vision-CAIR","isFork":false,"description":"Creative AI for Visual Art and Music slides and demos.","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":11,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-05-02T11:48:44.393Z"}},{"type":"Public","name":"3DCoMPaT","owner":"Vision-CAIR","isFork":false,"description":"Official repository for the 3DCoMPaT dataset (ECCV2022 Oral)","allTopics":["computer-vision","deep-learning","dataset","compositional-data"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":1,"starsCount":16,"forksCount":1,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-19T20:40:30.692Z"}},{"type":"Public","name":"ChatCaptioner","owner":"Vision-CAIR","isFork":false,"description":"Official Repository of ChatCaptioner","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":5,"starsCount":446,"forksCount":25,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-13T20:05:09.456Z"}},{"type":"Public","name":"AF-Guide","owner":"Vision-CAIR","isFork":false,"description":"Official repository of Action-Free Guide","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":13,"forksCount":1,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-03-22T07:54:38.357Z"}},{"type":"Public","name":"artelingo","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":1,"starsCount":4,"forksCount":1,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-22T08:33:48.627Z"}},{"type":"Public","name":"artemis-v2","owner":"Vision-CAIR","isFork":false,"description":"Code for the paper: It is Okay to Not Be Okay: Overcoming Emotional Bias in Affective Image Captioning by Contrastive Data Collection","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":1,"issueCount":2,"starsCount":16,"forksCount":4,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-11-27T11:04:58.996Z"}},{"type":"Public","name":"CWAN","owner":"Vision-CAIR","isFork":false,"description":"Creative Walk Adversarial Networks: Novel Art Generation with Probabilistic Random Walk Deviation from Style Norms","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":7,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-06-27T09:55:45.661Z"}},{"type":"Public","name":"Zero-Shot-Learning","owner":"Vision-CAIR","isFork":false,"description":"VisionCAIR Zero-Shot Learning Research","allTopics":["transfer-learning","zero-shot-learning","continual-learning","domain-generalization"],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-12-10T09:45:30.083Z"}},{"type":"Public","name":"UnlikelihoodMotionForecasting","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-11-05T12:17:57.705Z"}},{"type":"Public","name":"Affective-and-Creative-AI","owner":"Vision-CAIR","isFork":false,"description":"VisionCAIR Affective and Creative AI Research","allTopics":[],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-10-17T20:36:46.983Z"}},{"type":"Public","name":"cs326-few-shot-classification","owner":"Vision-CAIR","isFork":false,"description":"CS326 Practical assignment #2: few-shot classification","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":5,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-10-07T12:03:11.039Z"}},{"type":"Public","name":"WAGA","owner":"Vision-CAIR","isFork":false,"description":"Code for Wölfflin Affective Generative Analysis paper published in ICCC 2021","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":6,"forksCount":1,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-09-17T17:41:17.596Z"}},{"type":"Public","name":"GRaWD","owner":"Vision-CAIR","isFork":false,"description":"Imaginative Walks: Generative Random Walk Deviation Loss for Improved Unseen Learning Representation. CVPR 2022 Workshop, ICCC 2022.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-09-16T14:40:29.777Z"}},{"type":"Public","name":"HalentNet","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":7,"forksCount":3,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-04-21T18:40:29.326Z"}},{"type":"Public","name":"CIZSLv2","owner":"Vision-CAIR","isFork":false,"description":"CIZSL++: Creativity Inspired Generative Zero-Shot Learning. T-PAMI under review.","allTopics":["attributes","creativity","zero-shot-learning","generative-models","vision-and-language","unseen-classes-understanding"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":6,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-02-08T07:10:01.625Z"}},{"type":"Public","name":"lifelong_fact_learning","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-08-21T17:05:18.646Z"}},{"type":"Public","name":"feelings","owner":"Vision-CAIR","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-04-22T18:20:17.506Z"}},{"type":"Public","name":"CIZSL","owner":"Vision-CAIR","isFork":true,"description":"Creativity Inspired Zero-Shot Learning","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":8,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-02-27T16:17:41.639Z"}},{"type":"Public","name":"GDPP","owner":"Vision-CAIR","isFork":true,"description":"Generator loss to reduce mode-collapse and to improve the generated samples quality.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":4,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2019-07-03T16:01:05.521Z"}}],"repositoryCount":30,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}