{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"OpenRLHF","owner":"OpenLLMAI","isFork":false,"description":"An Easy-to-use, Scalable and High-performance RLHF Framework (Support 70B+ full tuning & LoRA & Mixtral & KTO)","allTopics":["reinforcement-learning","raylib","transformers","deepspeed","large-language-models","reinforcement-learning-from-human-feedback","vllm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":47,"starsCount":1529,"forksCount":131,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T11:19:40.342Z"}},{"type":"Public","name":"OpenLLMDE","owner":"OpenLLMAI","isFork":false,"description":"OpenLLMDE: An open source data engineering framework for LLMs","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":9,"starsCount":16,"forksCount":3,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,2,0,0,1,3,2,4,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-09-09T10:45:34.411Z"}}],"repositoryCount":2,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}