{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"konduktor","owner":"Trainy-ai","isFork":false,"description":"cluster/scheduler health monitoring for GPU jobs on k8s","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":15,"forksCount":0,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T20:19:25.392Z"}},{"type":"Public archive","name":"llm-atc","owner":"Trainy-ai","isFork":false,"description":"Fine-tuning and serving LLMs on any cloud","allTopics":["finetuning","llms","vllm","llama2"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":83,"forksCount":2,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-02T10:52:19.015Z"}},{"type":"Public","name":"training","owner":"Trainy-ai","isFork":true,"description":"Reference implementations of MLPerf™ training benchmarks","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":0,"forksCount":545,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-21T02:49:43.784Z"}},{"type":"Public","name":"dashboard","owner":"Trainy-ai","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-18T01:55:37.418Z"}},{"type":"Public","name":"vllm","owner":"Trainy-ai","isFork":true,"description":"A high-throughput and memory-efficient inference and serving engine for LLMs","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":0,"forksCount":2852,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-15T22:01:20.093Z"}},{"type":"Public","name":"airoboros","owner":"Trainy-ai","isFork":true,"description":"Customizable implementation of the self-instruct paper.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":0,"forksCount":65,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-15T22:00:58.922Z"}},{"type":"Public","name":"FastChat","owner":"Trainy-ai","isFork":true,"description":"An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":0,"forksCount":4315,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-14T23:26:38.000Z"}},{"type":"Public","name":"RWKV-LM","owner":"Trainy-ai","isFork":true,"description":"RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, \"infinite\" ctx_len, and free sentence embedding.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":814,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-02T04:05:19.595Z"}},{"type":"Public","name":"nodify","owner":"Trainy-ai","isFork":false,"description":"Profiling tools for distributed training","allTopics":[],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":1,"starsCount":37,"forksCount":3,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-31T05:25:29.840Z"}},{"type":"Public","name":"trainy","owner":"Trainy-ai","isFork":false,"description":"A simple Pure Python/PyTorch performance daemon for training workloads","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":13,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-08-02T07:36:46.036Z"}},{"type":"Public","name":"dynolog","owner":"Trainy-ai","isFork":true,"description":"Dynolog is a telemetry daemon for performance monitoring and tracing. It exports metrics from different components in the system like the linux kernel, CPU, disks, Intel PT, GPUs etc. Dynolog also integrates with pytorch and can trigger traces for distributed training applications.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":33,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-06-29T22:44:49.371Z"}}],"repositoryCount":11,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}