{"payload":{"header_redesign_enabled":false,"results":[{"id":"741810306","archived":false,"color":"#3572A5","followers":536,"has_funding_file":false,"hl_name":"ymcui/Chinese-Mixtral","hl_trunc_description":"中文Mixtral混合专家大模型(Chinese Mixtral MoE LLMs)","language":"Python","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":741810306,"name":"Chinese-Mixtral","owner_id":16095339,"owner_login":"ymcui","updated_at":"2024-04-30T04:29:06.644Z","has_issues":true}},"sponsorable":false,"topics":["nlp","moe","64k","mixture-of-experts","32k","large-language-models","llm","mixtral"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":60,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Aymcui%252FChinese-Mixtral%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/ymcui/Chinese-Mixtral/star":{"post":"X334d_8B-I-pNqGzJeW7hb-Ev-nIwO-QXEX1Yk0qTN3s81eOAv6mPNV4KtOaHWvJfOzAyDZbhMGHlGVFm0FsNg"},"/ymcui/Chinese-Mixtral/unstar":{"post":"JjCv41N65CZavDn5FhMeO-tO7GLlwL6jz8xGzgg0pjv-NDvste-8oqUhErJiC-lb8suy5j9vZJiLS-QCAfY56w"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"pizVZrAlgRvIfrkbyBxUZBKE4agMmDGCg-uPLo_WF0vT_9wt-yDNaJm5T7dOO4DaM9Rj-y5lijBxPqXj68FxdQ"}}},"title":"Repository search results"}