{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"EmbodiedScan","owner":"OpenRobotLab","isFork":false,"description":"[CVPR 2024] EmbodiedScan: A Holistic Multi-Modal 3D Perception Suite Towards Embodied AI","allTopics":["computer-vision","robotics","3d-vision","multi-modal-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":3,"starsCount":414,"forksCount":26,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,2,11,3,1,3,48,0,1,1,3,0,6,1,0,4,1,0,3,2,2,0,1,0,1,0,0,1,0,0,0,0,0,0,0,0,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-19T02:27:27.840Z"}},{"type":"Public","name":"GRUtopia","owner":"OpenRobotLab","isFork":false,"description":"GRUtopia: Dream General Robots in a City at Scale","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":4,"starsCount":425,"forksCount":17,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,8,8,3,3,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-26T05:00:28.094Z"}},{"type":"Public","name":"OVExp","owner":"OpenRobotLab","isFork":false,"description":"OVExp: Open Vocabulary Exploration for Object-Oriented Navigation","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":23,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-15T04:12:39.800Z"}},{"type":"Public","name":"PointLLM","owner":"OpenRobotLab","isFork":false,"description":"[ECCV 2024 Oral] PointLLM: Empowering Large Language Models to Understand Point Clouds","allTopics":["chatbot","point-cloud","llama","representation-learning","3d","multimodal","vision-and-language","gpt-4","foundation-models","large-language-models","objaverse","pointllm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":479,"forksCount":22,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-08T23:58:49.493Z"}},{"type":"Public","name":"Grounded_3D-LLM","owner":"OpenRobotLab","isFork":false,"description":"Code&Data for Grounded 3D-LLM with Referent Tokens","allTopics":["3d","scene-understanding","llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":4,"starsCount":67,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-01T13:31:11.882Z"}},{"type":"Public","name":"OV_PARTS","owner":"OpenRobotLab","isFork":false,"description":"[NeurIPS 2023] OV-PARTS: Towards Open-Vocabulary Part Segmentation","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":69,"forksCount":6,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-24T07:21:51.073Z"}},{"type":"Public","name":"UniHSI","owner":"OpenRobotLab","isFork":false,"description":"[ICLR 2024 Spotlight] Unified Human-Scene Interaction via Prompted Chain-of-Contacts","allTopics":["unified","physics-simulation","human-scene-interaction","llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":4,"starsCount":150,"forksCount":5,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T03:33:49.267Z"}},{"type":"Public","name":"HIMLoco","owner":"OpenRobotLab","isFork":false,"description":"Learning-based locomotion control from OpenRobotLab, including Hybrid Internal Model & H-Infinity Locomotion Control","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":7,"starsCount":230,"forksCount":25,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-14T14:42:27.747Z"}},{"type":"Public","name":"P3Former","owner":"OpenRobotLab","isFork":false,"description":"[IJCV 2024] P3Former: Position-Guided Point Cloud Panoptic Segmentation Transformer","allTopics":["point-cloud","panoptic-segmentation","transformer-decoder","lidar-based"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":75,"forksCount":9,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-02T12:52:37.662Z"}},{"type":"Public","name":"DORT","owner":"OpenRobotLab","isFork":false,"description":"[CoRL 2023] DORT: Modeling Dynamic Objects in Recurrent for Multi-Camera 3D Object Detection and Tracking","allTopics":["structure-from-motion","3d-detection","3d-tracking","vision-based-perception"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":64,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-21T14:24:39.359Z"}}],"repositoryCount":10,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"OpenRobotLab repositories"}