{"payload":{"pageCount":2,"repositories":[{"type":"Public","name":"stable-diffusion","owner":"CompVis","isFork":false,"description":"A latent text-to-image diffusion model","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":72,"issueCount":513,"starsCount":66309,"forksCount":9939,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-18T01:53:49.690Z"}},{"type":"Public","name":"zigma","owner":"CompVis","isFork":false,"description":"A PyTorch implementation of the paper \"ZigMa: A DiT-Style Mamba-based Diffusion Model\"","allTopics":["mamba","state-space-model","diffusion-models","flow-matching","zigma","stochastic-interpolant"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":193,"forksCount":9,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,9,27,3,0,0,0,0,0,17,5,0,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-14T09:41:39.975Z"}},{"type":"Public","name":"depth-fm","owner":"CompVis","isFork":false,"description":"DepthFM: Fast Monocular Depth Estimation with Flow Matching","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":9,"starsCount":295,"forksCount":17,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,10,0,1,0,0,0,0,0,0,0,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-13T19:48:55.787Z"}},{"type":"Public","name":"attribute-control","owner":"CompVis","isFork":false,"description":"Fine-Grained Subject-Specific Attribute Expression Control in T2I Models","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":1,"starsCount":96,"forksCount":7,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,3,2,3,1,0,0,0,0,0,0,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-13T10:33:42.646Z"}},{"type":"Public","name":"LoRAdapter","owner":"CompVis","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":1,"starsCount":72,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,1,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-16T13:13:16.961Z"}},{"type":"Public","name":"taming-transformers","owner":"CompVis","isFork":false,"description":"Taming Transformers for High-Resolution Image Synthesis","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":18,"issueCount":139,"starsCount":5503,"forksCount":1098,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-25T11:58:51.945Z"}},{"type":"Public","name":"fm-boosting","owner":"CompVis","isFork":false,"description":"Boosting Latent Diffusion with Flow Matching","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":3,"starsCount":95,"forksCount":0,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,21,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-28T17:45:55.221Z"}},{"type":"Public","name":"latent-diffusion","owner":"CompVis","isFork":false,"description":"High-Resolution Image Synthesis with Latent Diffusion Models","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":20,"issueCount":258,"starsCount":10944,"forksCount":1435,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-29T05:29:47.382Z"}},{"type":"Public","name":"instant-lora-composition","owner":"CompVis","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":30,"forksCount":0,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-08T17:16:01.587Z"}},{"type":"Public","name":"cuneiform-sign-detection-dataset","owner":"CompVis","isFork":false,"description":"Dataset provided with the article \"Deep learning for cuneiform sign detection with weak supervision using transliteration alignment\". It comprises image references, transliterations and sign annotations of clay tablets from the Neo-Assyrian epoch.","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":1,"starsCount":14,"forksCount":5,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-08-24T08:53:25.690Z"}},{"type":"Public","name":"cuneiform-sign-detection-code","owner":"CompVis","isFork":false,"description":"Code for the article \"Deep learning of cuneiform sign detection with weak supervision using transliteration alignment\"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":8,"forksCount":5,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-08-24T08:49:56.247Z"}},{"type":"Public","name":"geometry-free-view-synthesis","owner":"CompVis","isFork":false,"description":"Is a geometric model required to synthesize novel views from a single image?","allTopics":["novel-view-synthesis","transformers"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":9,"starsCount":367,"forksCount":34,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-16T21:13:47.367Z"}},{"type":"Public","name":"retrieval-augmented-diffusion-models","owner":"CompVis","isFork":false,"description":"Official codebase for the Paper “Retrieval-Augmented Diffusion Models”","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":2,"starsCount":100,"forksCount":6,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-05T14:31:43.702Z"}},{"type":"Public","name":"ipoke","owner":"CompVis","isFork":false,"description":"iPOKE: Poking a Still Image for Controlled Stochastic Video Synthesis","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":3,"starsCount":48,"forksCount":9,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-21T13:25:52.097Z"}},{"type":"Public","name":"net2net","owner":"CompVis","isFork":false,"description":"Network-to-Network Translation with Conditional Invertible Neural Networks","allTopics":["lightning","pytorch","generative-model","inn","gans","normalizing-flows","streamlit","autoencoders","pytorch-lightning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":6,"starsCount":219,"forksCount":20,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-20T12:11:14.400Z"}},{"type":"Public","name":"interactive-image2video-synthesis","owner":"CompVis","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":56,"forksCount":15,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-18T15:17:04.083Z"}},{"type":"Public","name":"behavior-driven-video-synthesis","owner":"CompVis","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":26,"forksCount":7,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-15T15:52:28.306Z"}},{"type":"Public","name":"imagebart","owner":"CompVis","isFork":false,"description":"ImageBART: Bidirectional Context with Multinomial Diffusion for Autoregressive Image Synthesis","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":4,"starsCount":119,"forksCount":12,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-03-14T20:03:30.195Z"}},{"type":"Public","name":"metric-learning-divide-and-conquer-improved","owner":"CompVis","isFork":false,"description":"Source code for the paper \"Improving Deep Metric Learning byDivide and Conquer\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":19,"forksCount":4,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-12-10T18:09:32.605Z"}},{"type":"Public","name":"visual-search","owner":"CompVis","isFork":false,"description":"Visual search interface","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":10,"forksCount":1,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-11-30T12:14:37.766Z"}},{"type":"Public","name":"image2video-synthesis-using-cINNs","owner":"CompVis","isFork":false,"description":"Implementation of Stochastic Image-to-Video Synthesis using cINNs.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":2,"starsCount":182,"forksCount":29,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-11-28T08:15:18.188Z"}},{"type":"Public","name":"Characterizing_Generalization_in_DML","owner":"CompVis","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-11-24T15:33:41.766Z"}},{"type":"Public","name":"brushstroke-parameterized-style-transfer","owner":"CompVis","isFork":false,"description":"TensorFlow implementation of our CVPR 2021 Paper \"Rethinking Style Transfer: From Pixels to Parameterized Brushstrokes\".","allTopics":["computer-vision","differentiable-rendering","deep-learning","style-transfer"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":160,"forksCount":19,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-07-16T16:27:38.484Z"}},{"type":"Public archive","name":"robust-disentangling","owner":"CompVis","isFork":false,"description":"Unsupervised Robust Disentangling of Latent Characteristics for Image Synthesis","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":23,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-07-01T10:04:21.785Z"}},{"type":"Public","name":"iin","owner":"CompVis","isFork":false,"description":"A Disentangling Invertible Interpretation Network","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":9,"starsCount":121,"forksCount":16,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-05-12T08:08:12.511Z"}},{"type":"Public","name":"content-style-disentangled-ST","owner":"CompVis","isFork":false,"description":"Content and Style Disentanglement for Artistic Style Transfer [ICCV19]","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":5,"starsCount":93,"forksCount":15,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-02-22T12:49:34.199Z"}},{"type":"Public","name":"cuneiform-sign-detection-webapp","owner":"CompVis","isFork":false,"description":"Code for demo web application of the article \"Deep learning for cuneiform sign detection with weak supervision using transliteration alignment\".","allTopics":[],"primaryLanguage":{"name":"JavaScript","color":"#f1e05a"},"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-01-27T15:08:42.399Z"}},{"type":"Public","name":"adaptive-style-transfer","owner":"CompVis","isFork":false,"description":"source code for the ECCV18 paper A Style-Aware Content Loss for Real-time HD Style Transfer","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":20,"starsCount":723,"forksCount":139,"license":"GNU General Public License v3.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-01-11T12:05:30.020Z"}},{"type":"Public","name":"invariances","owner":"CompVis","isFork":false,"description":"Making Sense of CNNs: Interpreting Deep Representations & Their Invariances with Invertible Neural Networks","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":52,"forksCount":6,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-12-18T13:21:38.271Z"}},{"type":"Public","name":"vunet","owner":"CompVis","isFork":false,"description":"A generative model conditioned on shape and appearance.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":10,"starsCount":497,"forksCount":104,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-12-18T12:26:43.175Z"}}],"repositoryCount":40,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}