{"payload":{"header_redesign_enabled":false,"results":[{"id":"464909502","archived":false,"color":"#3572A5","followers":133,"has_funding_file":false,"hl_name":"huggingface/optimum-habana","hl_trunc_description":"Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)","language":"Python","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":464909502,"name":"optimum-habana","owner_id":25720743,"owner_login":"huggingface","updated_at":"2024-07-24T00:28:22.164Z","has_issues":true}},"sponsorable":false,"topics":["transformers","bert","fine-tuning","hpu","habana"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":1,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":74,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Ahuggingface%252Foptimum-habana%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/huggingface/optimum-habana/star":{"post":"-QPmIc_bvYu8uz_O_Git3jr2FxQtcG1MS2IcyxqVRo8D2TvtlkoDVG9wsfbtZaMoA-CjSDClsRwiHDk_WO2plg"},"/huggingface/optimum-habana/unstar":{"post":"Tae7k_lGzVh1BpG9EXbI8igNQVXWwMtAtX2Ou4TrPsZIovym9nZtiXGkNe1l17QYmLfw5vL8gXb3WLnM1QOLTQ"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"FirNW8s7metXAWtZTtngFnhgVC8XR6zSXzET_C-wSkeI0Oz2fuZxEPw1z9QTOwjVT9PAaHRcMKoZRpU_k6BFvg"}}},"title":"Repository search results"}