{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"audio-ai-timeline","owner":"archinetai","isFork":false,"description":"A timeline of the latest AI models for audio generation, starting in 2023!","allTopics":["machine-learning","artificial-intelligence","audio-generation"],"primaryLanguage":null,"pullRequestCount":0,"issueCount":1,"starsCount":1872,"forksCount":66,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-04T01:15:07.756Z"}},{"type":"Public","name":"audio-diffusion-pytorch","owner":"archinetai","isFork":false,"description":"Audio generation using diffusion models, in PyTorch.","allTopics":["artificial-intelligence","denoising-diffusion","audio-generation","deep-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":15,"starsCount":1857,"forksCount":161,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-06-12T22:40:16.542Z"}},{"type":"Public","name":"a-unet","owner":"archinetai","isFork":false,"description":"A toolbox that provides hackable building blocks for generic 1D/2D/3D UNets, in PyTorch.","allTopics":["machine-learning","deep-learning","unet"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":71,"forksCount":9,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-06-12T22:27:06.044Z"}},{"type":"Public","name":"audio-encoders-pytorch","owner":"archinetai","isFork":false,"description":"A collection of audio autoencoders, in PyTorch.","allTopics":["audio","deep-learning","artificial-intelligence"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":1,"starsCount":36,"forksCount":6,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-03-07T00:08:24.820Z"}},{"type":"Public","name":"audio-data-pytorch","owner":"archinetai","isFork":false,"description":"A collection of useful audio datasets and transforms for PyTorch.","allTopics":["deep-learning","pytorch","datasets","artifical-intelligense","audio-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":125,"forksCount":22,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-02-11T23:30:31.620Z"}},{"type":"Public","name":"archisound","owner":"archinetai","isFork":false,"description":"A collection of pre-trained audio models, in PyTorch.","allTopics":["audio","deep-learning","artificial-intelligence"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":107,"forksCount":4,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-01-27T10:35:04.019Z"}},{"type":"Public","name":"audio-diffusion-pytorch-trainer","owner":"archinetai","isFork":false,"description":"Trainer for audio-diffusion-pytorch","allTopics":["deep-learning","artificial-intelligence","denoising-diffusion","audio-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":9,"starsCount":126,"forksCount":22,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-01-13T14:23:16.004Z"}},{"type":"Public","name":"cqt-pytorch","owner":"archinetai","isFork":false,"description":"An invertible and differentiable implementation of the Constant-Q Transform (CQT).","allTopics":["audio","deep-learning","pytorch","cqt"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":48,"forksCount":3,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-09T09:11:49.823Z"}},{"type":"Public","name":"aligner-pytorch","owner":"archinetai","isFork":false,"description":"Sequence alignement methods with helpers for PyTorch.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":24,"forksCount":3,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-11-30T07:18:49.256Z"}},{"type":"Public","name":"bitcodes-pytorch","owner":"archinetai","isFork":false,"description":"A vector quantization method with binary codes, in PyTorch.","allTopics":["machine-learning","deep-learning","artificial-intelligence","quantization"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":6,"forksCount":0,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-11-11T19:58:05.416Z"}},{"type":"Public","name":"a-transformers","owner":"archinetai","isFork":false,"description":"A collection of transformer models, in PyTorch.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":1,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-11-05T22:50:28.762Z"}},{"type":"Public","name":"quantizer-pytorch","owner":"archinetai","isFork":false,"description":"Different vector quantization methods, in PyTorch.","allTopics":["quantization","machine-learning","deep-learning","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-10-16T21:14:44.948Z"}},{"type":"Public archive","name":"a-transformers-pytorch","owner":"archinetai","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":1,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-10-15T14:20:33.077Z"}},{"type":"Public","name":"difformer-pytorch","owner":"archinetai","isFork":false,"description":"Diffusion based transformer, in PyTorch (Experimental).","allTopics":["deep-learning","artificial-intelligence","transformer","diffusion"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":25,"forksCount":2,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-09-13T10:01:06.499Z"}},{"type":"Public","name":"vat-pytorch","owner":"archinetai","isFork":false,"description":"Virtual Adversarial Training (VAT) techniques in PyTorch","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":16,"forksCount":2,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-07-19T20:58:15.883Z"}},{"type":"Public","name":"smart-pytorch","owner":"archinetai","isFork":false,"description":"PyTorch – SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models.","allTopics":["artificial-intelligence","fine-tuning","deep-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":58,"forksCount":6,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-06-28T12:43:48.301Z"}},{"type":"Public","name":"emergent-pytorch","owner":"archinetai","isFork":false,"description":"A new invertible and distributed training strategy for PyTorch. ","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":0,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-06-11T12:39:22.169Z"}},{"type":"Public","name":"surgeon-pytorch","owner":"archinetai","isFork":false,"description":"A library to inspect and extract intermediate layers of PyTorch models.","allTopics":["artificial-intelligence","deep-learning","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":466,"forksCount":16,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-05-12T21:55:03.421Z"}}],"repositoryCount":18,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}