{"payload":{"header_redesign_enabled":false,"results":[{"id":"66823715","archived":false,"color":"#3572A5","followers":6488,"has_funding_file":false,"hl_name":"intel-analytics/ipex-llm","hl_trunc_description":"Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, MiniCPM, etc.) on Intel XPU …","language":"Python","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":66823715,"name":"ipex-llm","owner_id":10941215,"owner_login":"intel-analytics","updated_at":"2024-09-14T07:26:46.216Z","has_issues":true}},"sponsorable":false,"topics":["gpu","transformers","pytorch","llm"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":11,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":86,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Aintel-analytics%252Fipex-llm%2B%2Blanguage%253APython","metadata":null,"warn_limited_results":false,"csrf_tokens":{"/intel-analytics/ipex-llm/star":{"post":"vN7eTAqqvm5AIvqibVZw0x_h0ocAyxHt8s3Af8xRAvuhcmkZChvdY4cWLczNSAQ77LA_Vy4qBs0m98oSv26Y8g"},"/intel-analytics/ipex-llm/unstar":{"post":"eJBx3GJhiAJq55aidMjZVTszZU3ejOFBY7SFD7Z1Ro1trsGUlncczR7Kl_JUX9iFLrIr5BhAjo7IrHvhwConBg"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"jb8xwCThg90TC-t1hscjL9K9sYf4swKLHrT4ln0E-nsoq2W_JQBsxQ_FNZ8go2G-_jmYPzwllUq4-RjqIKkIZw"}}},"title":"Repository search results"}