{"payload":{"header_redesign_enabled":false,"results":[{"id":"738776129","archived":false,"color":"#3572A5","followers":90,"has_funding_file":false,"hl_name":"intel/auto-round","hl_trunc_description":"SOTA Weight-only Quantization Algorithm for LLMs. This is official implementation of \"Optimize Weight Rounding via Signed Gradient Descen…","language":"Python","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":738776129,"name":"auto-round","owner_id":17888862,"owner_login":"intel","updated_at":"2024-06-14T00:44:45.258Z","has_issues":true}},"sponsorable":false,"topics":["rounding","quantization","awq","int4","gptq","neural-compressor","weight-only"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":78,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Aintel%252Fauto-round%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/intel/auto-round/star":{"post":"TBuCI2gd2byb1f3bLuDHxQ1XgvuVClXysceyYifa-xvPqxrGXzn_di-JdB-h1w5si5RH3vcE4VJyWJY_f0xIBg"},"/intel/auto-round/unstar":{"post":"mEEEyD94LhxXvEmrb67ZDg6hSWQsQ5gCbVbX9ZOTNbkEbDaVKT0q21GfA9zjwIQ1QSsLqeoosmFMzIUCA908bg"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"7S2tO5tdmeErrsgL5gNctMFEgfksEBZbGzN9sREGE0KOnr69C-FaINnzcW2H4MiD-TbPVHsxfFBaaJ8hTr8iow"}}},"title":"Repository search results"}