{"payload":{"header_redesign_enabled":false,"results":[{"id":"351783073","archived":false,"color":"#3572A5","followers":423,"has_funding_file":false,"hl_name":"Xirider/finetune-gpt2xl","hl_trunc_description":"Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpeed","language":"Python","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":351783073,"name":"finetune-gpt2xl","owner_id":37597043,"owner_login":"Xirider","updated_at":"2023-06-14T02:34:46.365Z","has_issues":true}},"sponsorable":false,"topics":["finetuning","gpt2","huggingface","huggingface-transformers","gpt3","deepspeed","gpt-neo","gpt-neo-fine-tuning"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":78,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253AXirider%252Ffinetune-gpt2xl%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/Xirider/finetune-gpt2xl/star":{"post":"lH_s_CS7MHg1AXBweE_w21AwI5oomMX72hAaxlL-LRu3r-aG_BHwSiYM-5TJZR7O1zPKx2RKSOciti5Z7nT2UQ"},"/Xirider/finetune-gpt2xl/unstar":{"post":"h6ROymjRHO89rpj_jycJHwY0sWydj8hEnhWrXbORCGR_N3En9j9KYnEogStjw1teFx6p1ZZQGyUngMPcP2AhFA"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"Y0wpqkvnqI5gyqutbVb06UUb8MdWCsL1jMFgu-zizeUbKM8wvZ16hVwn0NOgX1hs5JxaJLhVjqgt3lKNREkdJQ"}}},"title":"Repository search results"}