{"payload":{"header_redesign_enabled":false,"results":[{"id":"473059611","archived":false,"color":"#3572A5","followers":1246,"has_funding_file":false,"hl_name":"MCG-NJU/VideoMAE","hl_trunc_description":"[NeurIPS 2022 Spotlight] VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training","language":"Python","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":473059611,"name":"VideoMAE","owner_id":62974475,"owner_login":"MCG-NJU","updated_at":"2023-12-08T13:44:48.183Z","has_issues":true}},"sponsorable":false,"topics":["pytorch","transformer","action-recognition","video-understanding","mae","video-analysis","video-representation-learning","self-supervised-learning","masked-autoencoder","vision-transformer","video-transformer","neurips-2022"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":92,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253AMCG-NJU%252FVideoMAE%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/MCG-NJU/VideoMAE/star":{"post":"Ny-lXrmSq1HTLNC1tb2O-Akhtfb5ejMnG-xA0vV2dSi6E2jBbjf-FfvOOLyz_AtVdtsimtpuiMUUpQx3SyezpQ"},"/MCG-NJU/VideoMAE/unstar":{"post":"rpxkh8Kc7VA4ZJ3p_jMxwIxb-uiZ0gZoBJsq4YjrQMKIpVL-vt2F5R6ppD-fFBEdHY69aPNNgQWTZ36woljfDg"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"4Gz4c3-JR9JNHD1OQHYIii_6xj-icxvzJcQ-n7w51fjyPKdkr8P2953ArWZacuGVIah3_VRWdDmx4TvSs9bzFw"}}},"title":"Repository search results"}