{"payload":{"header_redesign_enabled":false,"results":[{"id":"690511522","archived":false,"color":"#f1e05a","followers":668,"has_funding_file":false,"hl_name":"RahulSChand/gpu_poor","hl_trunc_description":"Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization","language":"JavaScript","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":690511522,"name":"gpu_poor","owner_id":16897807,"owner_login":"RahulSChand","updated_at":"2023-11-04T23:17:31.643Z","has_issues":true}},"sponsorable":false,"topics":["gpu","pytorch","llama","quantization","language-model","huggingface","llm","llamacpp","ggml","llama2"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":115,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253ARahulSChand%252Fgpu_poor%2B%2Blanguage%253AJavaScript","metadata":null,"csrf_tokens":{"/RahulSChand/gpu_poor/star":{"post":"KDzzF3DPhCzzJL-4keO120pWFGiBTkD-STeg5xTqdVzMYMM-ejxKXRabhaJ6Qd4xtk6t0lguLPFZOtewiYgAlA"},"/RahulSChand/gpu_poor/unstar":{"post":"LBwhTeBdSCywhWEqlhglFj3ZTqWDPHgv9-V-ZoUk-AmmOsO1aMUmfle8rYM6knprQbMJLx5QCNJ2Bim0UmfWcQ"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"8PGfztXO6TR9xQ8MgvAFj5heYAoFxeD3Zjq3DG22u7W7-LAXjcY3RkT0FfZlsw1j9etdYKWg2KoIe1SGMhwrjA"}}},"title":"Repository search results"}