{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"MicroDiffusion","owner":"UCSC-VLAA","isFork":false,"description":"[CVPR 2024] This repository includes the official implementation our paper \"MicroDiffusion: Implicit Representation-Guided Diffusion for 3D Reconstruction from Limited 2D Microscopy Projections\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":17,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,5,0,0,0,0,4,0,2,1,3,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-13T19:38:31.108Z"}},{"type":"Public","name":"UCSC-VLAA.github.io","owner":"UCSC-VLAA","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,2,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,2,0,0,0,2,0,1,2,0,0,0,0,1,0,0,0,1,0,0,1,0,0,2,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-13T05:58:24.200Z"}},{"type":"Public","name":"FedConv","owner":"UCSC-VLAA","isFork":false,"description":"[TMLR'24] This repository includes the official implementation our paper \"FedConv: Enhancing Convolutional Neural Networks for Handling Data Heterogeneity in Federated Learning\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":22,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-30T15:50:17.828Z"}},{"type":"Public","name":"EVP","owner":"UCSC-VLAA","isFork":false,"description":"[TMLR'24] This repository includes the official implementation our paper \"Unleashing the Power of Visual Prompting At the Pixel Level\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":34,"forksCount":3,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-30T15:49:39.767Z"}},{"type":"Public","name":"AdvXL","owner":"UCSC-VLAA","isFork":false,"description":"[CVPR 2024] This repository includes the official implementation our paper \"Revisiting Adversarial Training at Scale\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":14,"forksCount":1,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,5,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-21T18:10:15.821Z"}},{"type":"Public","name":"MixCon3D","owner":"UCSC-VLAA","isFork":false,"description":"[CVPR 2024] The official implementation of paper \"Sculpting Holistic 3D Representation in Contrastive Language-Image-3D Pre-training\"","allTopics":["pytorch","3d","multimodal-learning","contrastive-learning","zero-shot-classification","foundation-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":15,"forksCount":1,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,4,1,0,3,18,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-21T18:10:10.643Z"}},{"type":"Public","name":"HQ-Edit","owner":"UCSC-VLAA","isFork":false,"description":"HQ-Edit: A High-Quality and High-Coverage Dataset for General Image Editing","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":54,"forksCount":3,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,15,1,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-18T20:58:06.893Z"}},{"type":"Public","name":"Compress-Align","owner":"UCSC-VLAA","isFork":false,"description":"This repository includes the official implementation and dataset of our paper \"Compress & Align: Curating Image-Text Data with Human Knowledge\".","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":1,"starsCount":2,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-22T04:53:03.388Z"}},{"type":"Public","name":"AQA-Bench","owner":"UCSC-VLAA","isFork":false,"description":"Algorithmic-Q&A-Bench: An Interactive Benchmark for Evaluating LLMs’ Sequential Reasoning Ability","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":0,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-23T22:47:05.717Z"}},{"type":"Public","name":"vllm-safety-benchmark","owner":"UCSC-VLAA","isFork":false,"description":"Official PyTorch Implementation of \"How Many Unicorns Are in This Image? A Safety Evaluation Benchmark for Vision LLMs\"","allTopics":["benchmark","safety","datasets","robustness","adversarial-attacks","llm","vision-language-model","multimodal-llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":45,"forksCount":1,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,7,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-28T02:38:01.984Z"}},{"type":"Public","name":"Redteaming_Challenge","owner":"UCSC-VLAA","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":6,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-27T07:24:41.216Z"}},{"type":"Public","name":"CLIPA","owner":"UCSC-VLAA","isFork":false,"description":"[NeurIPS 2023] This repository includes the official implementation of our paper \"An Inverse Scaling Law for CLIP Training\"","allTopics":["zero-shot-learning","multimodal-learning","contrastive-learning","zero-shot-classification","foundation-models","neurips-2023","deep-learning","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":276,"forksCount":10,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-27T22:30:18.595Z"}},{"type":"Public","name":"SwinMM","owner":"UCSC-VLAA","isFork":false,"description":"[MICCAI 2023] This repository includes the official implementation our paper \"SwinMM: Masked Multi-view with Swin Transformers for 3D Medical Image Segmentation\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":90,"forksCount":5,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-13T19:01:28.921Z"}},{"type":"Public","name":"Sight-Beyond-Text","owner":"UCSC-VLAA","isFork":false,"description":"This repository includes the official implementation of our paper \"Sight Beyond Text: Multi-Modal Training Enhances LLMs in Truthfulness and Ethics\"","allTopics":["alignment","vlm","ai-alignment","vision-language","vicuna","llm","mllm","llava","llama2"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":19,"forksCount":1,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,4,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-09-15T02:03:49.064Z"}},{"type":"Public","name":"DMAE","owner":"UCSC-VLAA","isFork":false,"description":"[CVPR 2023] This repository includes the official implementation our paper \"Masked Autoencoders Enable Efficient Knowledge Distillers\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":95,"forksCount":3,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-07-24T16:37:20.279Z"}},{"type":"Public","name":"RobustCNN","owner":"UCSC-VLAA","isFork":false,"description":"[ICLR 2023] This repository includes the official implementation our paper \"Can CNNs Be More Robust Than Transformers?\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":141,"forksCount":12,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-01-23T01:49:12.940Z"}},{"type":"Public","name":"Image-Pretraining-for-Video","owner":"UCSC-VLAA","isFork":false,"description":"[ECCV 2022] This repository includes the official implementation our paper \"In Defense of Image Pre-Training for Spatiotemporal Recognition\".","allTopics":["image-classification","video-understanding","3d-convolutional-network","eccv2022","action-recognition"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":19,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-22T10:09:49.766Z"}},{"type":"Public","name":"vit_cert","owner":"UCSC-VLAA","isFork":false,"description":"[ECCV 2022] This repository includes the official implementation our paper \"ViP: Unified Certified Detection and Recovery for Patch Attack with Vision Transformers\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-07-21T18:06:39.593Z"}}],"repositoryCount":18,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}