{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"EXAONE-Atelier","owner":"LGAI-Research","isFork":false,"description":"Jupyter notebook examples for EXAONE Atelier in AWS Marketplace","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":12,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-08T20:18:54.326Z"}},{"type":"Public","name":"BITTERS","owner":"LGAI-Research","isFork":false,"description":"Large-Scale Bidirectional Training for Zero-Shot Image Captioning","allTopics":["deep-learning","pytorch","transformer","image-captioning","bitters","vqvae","pytorch-lightning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":21,"forksCount":2,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-02-14T05:52:52.148Z"}},{"type":"Public","name":"L-Verse","owner":"LGAI-Research","isFork":false,"description":"L-Verse: Bidirectional Generation Between Image and Text","allTopics":["transformer","image-captioning","image-to-text","text-to-image","vq-vae","pytorch-lightning","l-verse","deep-learning","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":108,"forksCount":6,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-11-15T05:51:27.121Z"}}],"repositoryCount":3,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}