{"payload":{"header_redesign_enabled":false,"results":[{"id":"648507291","archived":false,"color":"#DA5B0B","followers":5,"has_funding_file":false,"hl_name":"dasdristanta13/LLM-Lora-PEFT_accumulate","hl_trunc_description":"LLM-Lora-PEFT_accumulate explores optimizations for Large Language Models (LLMs) using PEFT, LORA, and QLORA. Contribute experiments and …","language":"Jupyter Notebook","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":648507291,"name":"LLM-Lora-PEFT_accumulate","owner_id":70366198,"owner_login":"dasdristanta13","updated_at":"2023-06-16T06:19:26.496Z","has_issues":true}},"sponsorable":false,"topics":["falcon","llama","lora","alpaca","int8","peft","llm","qlora","bitsandbytes"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":67,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Adasdristanta13%252FLLM-Lora-PEFT_accumulate%2B%2Blanguage%253A%2522Jupyter%2BNotebook%2522","metadata":null,"warn_limited_results":false,"csrf_tokens":{"/dasdristanta13/LLM-Lora-PEFT_accumulate/star":{"post":"Ev2T6Xl1wJxzzp6gVSX4zTtmc5iSt4liIjf34xnUwqaaOyF4rPhrCcBxdjfZ-2C2-BPyOO_iH56zqhka1VXN1g"},"/dasdristanta13/LLM-Lora-PEFT_accumulate/unstar":{"post":"AyRCwZVepb2ePcsFWuIey7KsNyewPJIXBtfiyMi5D0zPZyPQ6itCAua5967vy38oRub6zrd5_QgSg35WxuNQgg"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"zSjkgpIQnwehhF7UKPvlFxrcxu-mhxlMLY2VzRfXYd-VJn_vzCZcHG_lyoEgpicQo1Dleit-JWYkhof913ECwQ"}}},"title":"Repository search results"}