{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"llama-moe","owner":"pjlab-sys4nlp","isFork":false,"description":"⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training","allTopics":["moe","llama","mixture-of-experts","llm","continual-pre-training","expert-partition"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":753,"forksCount":32,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-26T04:58:35.308Z"}},{"type":"Public","name":"opencompass-Huawei","owner":"pjlab-sys4nlp","isFork":true,"description":"OpenCompass is an LLM evaluation platform, supporting a wide range of models (LLaMA, LLaMa2, ChatGLM2, ChatGPT, Claude, etc) over 50+ datasets.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":0,"forksCount":308,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-03T04:09:10.148Z"}},{"type":"Public","name":"NeMo","owner":"pjlab-sys4nlp","isFork":true,"description":"NeMo: a toolkit for conversational AI","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":2202,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-31T04:05:21.049Z"}},{"type":"Public","name":"NeMo-Megatron-Launcher","owner":"pjlab-sys4nlp","isFork":true,"description":"NeMo Megatron launcher and tools","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":117,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-31T04:05:09.713Z"}},{"type":"Public","name":"DeepSpeed","owner":"pjlab-sys4nlp","isFork":true,"description":"DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":0,"starsCount":0,"forksCount":3908,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-29T16:10:55.704Z"}},{"type":"Public","name":"Megatron-LM","owner":"pjlab-sys4nlp","isFork":true,"description":"Ongoing research training transformer models at scale","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":2006,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-19T08:33:44.475Z"}},{"type":"Public","name":"Megatron-DeepSpeed","owner":"pjlab-sys4nlp","isFork":true,"description":"Ongoing research training transformer language models at scale, including: BERT & GPT-2","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":0,"forksCount":2006,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-13T03:33:26.454Z"}},{"type":"Public","name":"llama","owner":"pjlab-sys4nlp","isFork":true,"description":"Inference code for LLaMA models","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":9258,"license":"GNU General Public License v3.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-15T06:17:27.169Z"}}],"repositoryCount":8,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}