{"payload":{"pageCount":2,"repositories":[{"type":"Public","name":"NATTEN","owner":"SHI-Labs","isFork":false,"description":"Neighborhood Attention Extension. Bringing attention to a neighborhood near you!","allTopics":["cuda","pytorch","neighborhood-attention"],"primaryLanguage":{"name":"Cuda","color":"#3A4E3A"},"pullRequestCount":0,"issueCount":17,"starsCount":307,"forksCount":23,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-25T23:55:39.590Z"}},{"type":"Public","name":"CuMo","owner":"SHI-Labs","isFork":false,"description":"CuMo: Scaling Multimodal LLM with Co-Upcycled Mixture-of-Experts","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":115,"forksCount":8,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T06:04:21.828Z"}},{"type":"Public","name":"Diffusion-Driven-Test-Time-Adaptation-via-Synthetic-Domain-Alignment","owner":"SHI-Labs","isFork":false,"description":"Everything to the Synthetic: Diffusion-driven Test-time Adaptation via Synthetic-Domain Alignment","allTopics":["diffusion-models","test-time-adaptation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":12,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T17:04:27.355Z"}},{"type":"Public","name":"StyleNAT","owner":"SHI-Labs","isFork":false,"description":"New flexible and efficient image generation framework that sets new SOTA on FFHQ-256 with FID 2.05, 2022","allTopics":["gan","image-generation","neighborhood-attention"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":98,"forksCount":10,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T06:18:21.571Z"}},{"type":"Public","name":"Neighborhood-Attention-Transformer","owner":"SHI-Labs","isFork":false,"description":"Neighborhood Attention Transformer, arxiv 2022 / CVPR 2023. Dilated Neighborhood Attention Transformer, arxiv 2022","allTopics":["pytorch","neighborhood-attention"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":1012,"forksCount":82,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-15T01:12:12.916Z"}},{"type":"Public","name":"CompactNet","owner":"SHI-Labs","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":30,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-05T00:38:27.946Z"}},{"type":"Public","name":"VCoder","owner":"SHI-Labs","isFork":false,"description":"VCoder: Versatile Vision Encoders for Multimodal Large Language Models, arXiv 2023 / CVPR 2024","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":3,"starsCount":241,"forksCount":12,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-17T05:09:38.772Z"}},{"type":"Public","name":"FcF-Inpainting","owner":"SHI-Labs","isFork":false,"description":"[WACV 2023] Keys to Better Image Inpainting: Structure and Texture Go Hand in Hand","allTopics":["pytorch","image-inpainting","stylegan2","fast-fourier-convolution","fcfgan"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":1,"issueCount":7,"starsCount":166,"forksCount":13,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-03T10:49:30.737Z"}},{"type":"Public","name":"Smooth-Diffusion","owner":"SHI-Labs","isFork":false,"description":"Smooth Diffusion: Crafting Smooth Latent Spaces in Diffusion Models arXiv 2023 / CVPR 2024","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":4,"starsCount":275,"forksCount":7,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-25T11:51:53.288Z"}},{"type":"Public","name":"Rethinking-Text-Segmentation","owner":"SHI-Labs","isFork":false,"description":"[CVPR 2021] Rethinking Text Segmentation: A Novel Dataset and A Text-Specific Refinement Approach","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":7,"starsCount":240,"forksCount":27,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-02T20:46:36.762Z"}},{"type":"Public","name":"Matting-Anything","owner":"SHI-Labs","isFork":false,"description":"Matting Anything Model (MAM), an efficient and versatile framework for estimating the alpha matte of any instance in an image with flexible and interactive visual or linguistic user prompt guidance.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":560,"forksCount":43,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-18T04:27:11.070Z"}},{"type":"Public","name":"Prompt-Free-Diffusion","owner":"SHI-Labs","isFork":false,"description":"Prompt-Free Diffusion: Taking \"Text\" out of Text-to-Image Diffusion Models, arxiv 2023 / CVPR 2024","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":12,"starsCount":713,"forksCount":35,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-16T03:42:55.222Z"}},{"type":"Public","name":"VIM","owner":"SHI-Labs","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":4,"starsCount":51,"forksCount":4,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-08T01:48:33.705Z"}},{"type":"Public","name":"OneFormer","owner":"SHI-Labs","isFork":false,"description":"OneFormer: One Transformer to Rule Universal Image Segmentation, arxiv 2022 / CVPR 2023","allTopics":["transformer","coco","image-segmentation","semantic-segmentation","cityscapes","instance-segmentation","ade20k","panoptic-segmentation","universal-segmentation","oneformer"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":4,"issueCount":23,"starsCount":1381,"forksCount":128,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-25T16:43:29.538Z"}},{"type":"Public","name":"Versatile-Diffusion","owner":"SHI-Labs","isFork":false,"description":"Versatile Diffusion: Text, Images and Variations All in One Diffusion Model, arXiv 2022 / ICCV 2023","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":9,"starsCount":1301,"forksCount":80,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-08-10T10:17:01.938Z"}},{"type":"Public","name":"Forget-Me-Not","owner":"SHI-Labs","isFork":false,"description":"Forget-Me-Not: Learning to Forget in Text-to-Image Diffusion Models, 2023","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":6,"starsCount":105,"forksCount":6,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-07-17T20:58:45.977Z"}},{"type":"Public","name":"OneFormer-Colab","owner":"SHI-Labs","isFork":false,"description":"[Colab Demo Code] OneFormer: One Transformer to Rule Universal Image Segmentation.","allTopics":["transformer","coco","image-segmentation","semantic-segmentation","cityscapes","instance-segmentation","ade20k","panoptic-segmentation","universal-segmentation","oneformer"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":13,"forksCount":9,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-05-24T20:32:02.017Z"}},{"type":"Public","name":"PAIR-Diffusion","owner":"SHI-Labs","isFork":true,"description":"PAIR-Diffusion: Object-Level Image Editing with Structure-and-Appearance Paired Diffusion Models, 2023","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":24,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-05-19T20:24:43.350Z"}},{"type":"Public","name":"Text2Video-Zero","owner":"SHI-Labs","isFork":true,"description":"a copy of \"Text-to-Image Diffusion Models are Zero-Shot Video Generators\", ICCV 2023","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":335,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-05-06T22:35:28.641Z"}},{"type":"Public","name":"Text2Video-Zero-sd-webui","owner":"SHI-Labs","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":6,"starsCount":79,"forksCount":15,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-10T20:45:14.897Z"}},{"type":"Public","name":"SH-GAN","owner":"SHI-Labs","isFork":false,"description":"[WACV 2023] Image Completion with Heterogeneously Filtered Spectral Hints","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":62,"forksCount":4,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-03-28T17:37:11.780Z"}},{"type":"Public","name":"Compact-Transformers","owner":"SHI-Labs","isFork":false,"description":"Escaping the Big Data Paradigm with Compact Transformers, 2021 (Train your Vision Transformers in 30 mins on CIFAR-10 with a single GPU!)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":9,"starsCount":485,"forksCount":76,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-02-20T21:02:09.188Z"}},{"type":"Public","name":"Boosted-Dynamic-Networks","owner":"SHI-Labs","isFork":false,"description":" Boosted Dynamic Neural Networks, AAAI 2023","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":8,"forksCount":1,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-01T04:45:19.594Z"}},{"type":"Public","name":"VMFormer","owner":"SHI-Labs","isFork":false,"description":"[Preprint] VMFormer: End-to-End Video Matting with Transformer","allTopics":["video-matting","vision-transformer"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":8,"starsCount":103,"forksCount":10,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-11-30T05:09:29.723Z"}},{"type":"Public","name":"Unsupervised-Domain-Adaptation-with-Differential-Treatment","owner":"SHI-Labs","isFork":false,"description":"[CVPR 2020] Differential Treatment for Stuff and Things: A Simple Unsupervised Domain Adaptation Method for Semantic Segmentation","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":5,"starsCount":86,"forksCount":13,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-11-22T05:46:11.637Z"}},{"type":"Public","name":"Convolutional-MLPs","owner":"SHI-Labs","isFork":false,"description":"[Preprint] ConvMLP: Hierarchical Convolutional MLPs for Vision, 2021","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":161,"forksCount":18,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-10-11T19:20:01.326Z"}},{"type":"Public","name":"DiSparse-Multitask-Model-Compression","owner":"SHI-Labs","isFork":false,"description":"[CVPR 2022] DiSparse: Disentangled Sparsification for Multitask Model Compression","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":1,"starsCount":13,"forksCount":3,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-09-06T01:27:37.657Z"}},{"type":"Public","name":"Agriculture-Vision","owner":"SHI-Labs","isFork":false,"description":"[CVPR 2020 & 2021 & 2022 & 2023] Agriculture-Vision Dataset, Prize Challenge and Workshop: A joint effort with many great collaborators to bring Agriculture and Computer Vision/AI communities together to benefit humanity!","allTopics":[],"primaryLanguage":null,"pullRequestCount":1,"issueCount":1,"starsCount":193,"forksCount":34,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-06-12T04:32:34.977Z"}},{"type":"Public","name":"LIVE-Layerwise-Image-Vectorization","owner":"SHI-Labs","isFork":true,"description":"[CVPR 2022 Oral] Towards Layer-wise Image Vectorization","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":48,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-06-10T00:59:36.971Z"}},{"type":"Public","name":"VideoINR-Continuous-Space-Time-Super-Resolution","owner":"SHI-Labs","isFork":true,"description":"[CVPR 2022] VideoINR: Learning Video Implicit Neural Representation for Continuous Space-Time Super-Resolution","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":27,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-06-09T16:16:22.962Z"}}],"repositoryCount":57,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}