{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":664913876,"defaultBranch":"main","name":"xtuner","ownerLogin":"InternLM","currentUserCanPush":false,"isFork":false,"isEmpty":false,"createdAt":"2023-07-11T03:18:13.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/135356492?v=4","public":true,"private":false,"isOrgOwned":true},"refInfo":{"name":"","listCacheKey":"v0:1715421014.0","currentOid":""},"activityList":{"items":[{"before":"ed844be1fe00107c40ff5ec4712acc7481b69f36","after":"b1099fecdd8843bad05b0056ab2c82050a037db0","ref":"refs/heads/main","pushedAt":"2024-05-17T07:13:14.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Bug] fix internlm2 flash attn (#693)\n\n* fix internlm2 flash attn\r\n\r\n* fix SUPPORT_FLASH2","shortMessageHtmlLink":"[Bug] fix internlm2 flash attn (#693)"}},{"before":"3b14f4859904da6ff7fc30ce1dd957d8ed603a46","after":"ed844be1fe00107c40ff5ec4712acc7481b69f36","ref":"refs/heads/main","pushedAt":"2024-05-17T07:12:21.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Feature] support auto saving tokenizer (#696)\n\nsupport auto saving tokenizer","shortMessageHtmlLink":"[Feature] support auto saving tokenizer (#696)"}},{"before":"e745a0e0befa841dce002f4ade6164e57d9ee3db","after":"3b14f4859904da6ff7fc30ce1dd957d8ed603a46","ref":"refs/heads/main","pushedAt":"2024-05-16T06:45:17.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Fix] ZeRO2 Checkpoint Convert Bug (#684)\n\nfix z2 convert","shortMessageHtmlLink":"[Fix] ZeRO2 Checkpoint Convert Bug (#684)"}},{"before":"aab528c33a9418a2ce760a667cac34918ec88da6","after":"e745a0e0befa841dce002f4ade6164e57d9ee3db","ref":"refs/heads/main","pushedAt":"2024-05-12T13:58:08.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Enhancement] Optimizing Memory Usage during ZeRO Checkpoint Convert (#582)\n\noptimize memory usage\r\n\r\nCo-authored-by: Zhihao Lin <36994684+LZHgrla@users.noreply.github.com>","shortMessageHtmlLink":"[Enhancement] Optimizing Memory Usage during ZeRO Checkpoint Convert (#…"}},{"before":"193f614ffbb2463010808ebb2e689331a9c5e4f6","after":"aab528c33a9418a2ce760a667cac34918ec88da6","ref":"refs/heads/main","pushedAt":"2024-05-11T09:49:12.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"bump version to 0.1.19 (#675)","shortMessageHtmlLink":"bump version to 0.1.19 (#675)"}},{"before":"eb647cab41c2df2832f0620aa1c42a47b99e14a8","after":"193f614ffbb2463010808ebb2e689331a9c5e4f6","ref":"refs/heads/main","pushedAt":"2024-05-10T15:30:00.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Improve] Support `safe_serialization` saving (#648)\n\n* Update merge.py\r\n\r\n* Update pth_to_hf.py","shortMessageHtmlLink":"[Improve] Support safe_serialization saving (#648)"}},{"before":"fc419431f303aa19abe14226f0e003545d84dd35","after":"eb647cab41c2df2832f0620aa1c42a47b99e14a8","ref":"refs/heads/main","pushedAt":"2024-05-10T15:29:41.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"Remove test info in DatasetInfoHook (#622)\n\nremove test info in DatasetInfoHook","shortMessageHtmlLink":"Remove test info in DatasetInfoHook (#622)"}},{"before":"648d63a44be018e240fe690880e62da5a2dbe5dd","after":"fc419431f303aa19abe14226f0e003545d84dd35","ref":"refs/heads/main","pushedAt":"2024-05-10T15:28:17.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Feature] add HFCheckpointHook to auto save hf model after the whole training phase (#621)\n\n* add HFCheckpointHook to auto save hf model after the whole training phase\r\n\r\n* refinie HFCheckpointHook\r\n\r\n* fix lint\r\n\r\n* delete useless codes\r\n\r\n* fix bugs\r\n\r\n* support non-dist training","shortMessageHtmlLink":"[Feature] add HFCheckpointHook to auto save hf model after the whole …"}},{"before":"117b4c129d3eaea306c0c1b7302020e8e07de1af","after":"648d63a44be018e240fe690880e62da5a2dbe5dd","ref":"refs/heads/main","pushedAt":"2024-05-10T15:26:44.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Fix] Fix batch-size setting of single-card LLaVA-Llama-3-8B configs (#598)\n\n* Update llava_llama3_8b_instruct_qlora_clip_vit_large_p14_336_e1_gpu1_finetune.py\r\n\r\n* Update llava_llama3_8b_instruct_quant_clip_vit_large_p14_336_e1_gpu1_pretrain.py","shortMessageHtmlLink":"[Fix] Fix batch-size setting of single-card LLaVA-Llama-3-8B configs (#…"}},{"before":"b283b998356801f6648859a941f60960970b32a9","after":"117b4c129d3eaea306c0c1b7302020e8e07de1af","ref":"refs/heads/main","pushedAt":"2024-05-10T15:19:27.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Fix] `convert_xtuner_weights_to_hf` with frozen ViT (#661)\n\n* Update convert_xtuner_weights_to_hf.py\r\n\r\n* fix pre-commit","shortMessageHtmlLink":"[Fix] convert_xtuner_weights_to_hf with frozen ViT (#661)"}},{"before":"d12bc05c1b35a57d18345b8f7f5bee93d23e2f4d","after":"b283b998356801f6648859a941f60960970b32a9","ref":"refs/heads/main","pushedAt":"2024-05-10T15:18:34.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"check transformers version before dispatch (#672)","shortMessageHtmlLink":"check transformers version before dispatch (#672)"}},{"before":"086df69e99888e1735d6f1e2b63eb3792693a36a","after":"d12bc05c1b35a57d18345b8f7f5bee93d23e2f4d","ref":"refs/heads/main","pushedAt":"2024-05-10T14:49:25.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Feature] Add qwen1.5 110b cfgs (#632)\n\n* add qwen1.5 110b cfgs\r\n\r\n* add qwen 110b readme and cfg","shortMessageHtmlLink":"[Feature] Add qwen1.5 110b cfgs (#632)"}},{"before":"aac7f578e0a9e64129abb4c9d06659bb04e7eb19","after":"086df69e99888e1735d6f1e2b63eb3792693a36a","ref":"refs/heads/main","pushedAt":"2024-05-06T05:35:50.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Feature] Support dsz3+qlora (#600)\n\n* support qlora zero3\r\n\r\n* add llama3 70B sft config and llava llama3 70B pretrain config\r\n\r\n* rename\r\n\r\n* change llama3 70b base to llama3 70b instruct","shortMessageHtmlLink":"[Feature] Support dsz3+qlora (#600)"}},{"before":"df83c6c0e4843db8a269cd1e6016109712fd1d14","after":"aac7f578e0a9e64129abb4c9d06659bb04e7eb19","ref":"refs/heads/main","pushedAt":"2024-04-29T11:39:21.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Docs] Delete colab and add speed benchmark (#617)\n\n* delete colab and add speed benchmark\r\n\r\n* change speed benchmark figures\r\n\r\n* fix en readme","shortMessageHtmlLink":"[Docs] Delete colab and add speed benchmark (#617)"}},{"before":"81d66e6906bd39140b2f3ae0fe4a5439f6638fa0","after":"df83c6c0e4843db8a269cd1e6016109712fd1d14","ref":"refs/heads/main","pushedAt":"2024-04-28T08:38:47.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Fix] Convert nan to 0 just for logging (#625)\n\nconvert nan to 0 just for logging","shortMessageHtmlLink":"[Fix] Convert nan to 0 just for logging (#625)"}},{"before":"1cd36282eeac3ca1aad933ae6c427d4ef064eb3d","after":"81d66e6906bd39140b2f3ae0fe4a5439f6638fa0","ref":"refs/heads/main","pushedAt":"2024-04-28T04:48:27.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Feature] Add conversion scripts for LLaVA-Llama-3-8B (#618)\n\n* update\r\n\r\n* update\r\n\r\n* fix typo\r\n\r\n* Update README.md\r\n\r\n* Update README.md","shortMessageHtmlLink":"[Feature] Add conversion scripts for LLaVA-Llama-3-8B (#618)"}},{"before":"e446753cbacaca29894e23b28f57e7dc26359d4c","after":"1cd36282eeac3ca1aad933ae6c427d4ef064eb3d","ref":"refs/heads/main","pushedAt":"2024-04-26T08:08:31.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Feature] Refine sp api (#619)\n\n* refine pad_for_sequence_parallel\r\n\r\n* refine sp api","shortMessageHtmlLink":"[Feature] Refine sp api (#619)"}},{"before":"75703c39b95ea792fa13df3b0baf70aec0832c15","after":"e446753cbacaca29894e23b28f57e7dc26359d4c","ref":"refs/heads/main","pushedAt":"2024-04-26T08:06:39.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"Update README.md (#608)\n\nHuggingFace -> Hugging Face","shortMessageHtmlLink":"Update README.md (#608)"}},{"before":"60e0cc9675cb3cdd630adfc7081b59c69084ce1f","after":"75703c39b95ea792fa13df3b0baf70aec0832c15","ref":"refs/heads/main","pushedAt":"2024-04-25T14:39:27.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"LZHgrla","name":"Zhihao Lin","path":"/LZHgrla","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/36994684?s=80&v=4"},"commit":{"message":"[Release] LLaVA-Phi-3-mini (#615)\n\n* update\r\n\r\n* add readme\r\n\r\n* update\r\n\r\n* update\r\n\r\n* fix typo\r\n\r\n* add links\r\n\r\n* update\r\n\r\n* add llava-llama-3-8b\r\n\r\n* update\r\n\r\n* update\r\n\r\n* add enter\r\n\r\n* fix typo\r\n\r\n* Update README.md\r\n\r\n* update README\r\n\r\n* update\r\n\r\n* update\r\n\r\n* update\r\n\r\n* Update README.md\r\n\r\n* Update README.md","shortMessageHtmlLink":"[Release] LLaVA-Phi-3-mini (#615)"}},{"before":"fc4225aaa26b330e064d8a6425aebdafa4851444","after":"60e0cc9675cb3cdd630adfc7081b59c69084ce1f","ref":"refs/heads/main","pushedAt":"2024-04-25T07:22:08.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Fix] Do not set attn_implementation to flash_attention_2 or sdpa if users already set it in XTuner configs. (#609)\n\n* do not set attn_implementation to flash_attention_2 or sdpa if users already set it\r\n\r\n* check cfg: If we want to use varlen attn or sequence parallel, we should set attn_implementation to flash_attention_2 or do not set this attribute.","shortMessageHtmlLink":"[Fix] Do not set attn_implementation to flash_attention_2 or sdpa if …"}},{"before":"ffc4ea8001aa6b9f4866a6e6f3af9f08c94eabb8","after":"fc4225aaa26b330e064d8a6425aebdafa4851444","ref":"refs/heads/main","pushedAt":"2024-04-24T09:49:24.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Fix] set `dataloader_num_workers=4` for llava training (#611)\n\nfix","shortMessageHtmlLink":"[Fix] set dataloader_num_workers=4 for llava training (#611)"}},{"before":"649cab995d5240ba9e4506fcdce1e3e27cd6aab4","after":"ffc4ea8001aa6b9f4866a6e6f3af9f08c94eabb8","ref":"refs/heads/main","pushedAt":"2024-04-24T09:48:54.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Feature]Support Mircosoft Phi3 4K&128K Instruct Models (#603)\n\n* support phi3\r\n\r\n* dispatch sft\r\n\r\n* rename configs\r\n\r\n* add phi3 llava configs\r\n\r\n* dispatch llava\r\n\r\n* fix phi3 dispatch (#3)\r\n\r\n* remove readme; fix ckpt name\r\n\r\n* remove unused file\r\n\r\n* add comma\r\n\r\n* fix typo\r\n\r\n* rename\r\n\r\n* set dataloader_num_workers = 0\r\n\r\n---------\r\n\r\nCo-authored-by: whcao <41630003+HIT-cwh@users.noreply.github.com>\r\nCo-authored-by: linzhihao ","shortMessageHtmlLink":"[Feature]Support Mircosoft Phi3 4K&128K Instruct Models (#603)"}},{"before":"0e6241fa3c64df1928af33f7facf0bc22066c639","after":"649cab995d5240ba9e4506fcdce1e3e27cd6aab4","ref":"refs/heads/main","pushedAt":"2024-04-24T09:38:22.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Feature] Support load_json_file with json.load (#610)\n\nsupport load_json_file with json.load","shortMessageHtmlLink":"[Feature] Support load_json_file with json.load (#610)"}},{"before":"bdafd58b20739dd3f8368a5e8d84475048d1188f","after":"0e6241fa3c64df1928af33f7facf0bc22066c639","ref":"refs/heads/main","pushedAt":"2024-04-23T05:10:10.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Docs] Add wisemodel badge (#597)\n\n* Update README.md\r\n\r\n* Update README_zh-CN.md","shortMessageHtmlLink":"[Docs] Add wisemodel badge (#597)"}},{"before":"4b6d5c9ede1f1f81d1cfd5f472dcca02f4ba82a5","after":"bdafd58b20739dd3f8368a5e8d84475048d1188f","ref":"refs/heads/main","pushedAt":"2024-04-22T17:51:23.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Improve] Add single-gpu configs for LLaVA-Llama-3-8B (#596)\n\n* update\r\n\r\n* update\r\n\r\n* update readme","shortMessageHtmlLink":"[Improve] Add single-gpu configs for LLaVA-Llama-3-8B (#596)"}},{"before":"36a1f263509102243f91bb50b9ed98e2e2b89a56","after":"4b6d5c9ede1f1f81d1cfd5f472dcca02f4ba82a5","ref":"refs/heads/main","pushedAt":"2024-04-22T11:24:35.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Feature] Release LLaVA-Llama-3-8B (#595)\n\n* update\r\n\r\n* 2048\r\n\r\n* readme\r\n\r\n* update\r\n\r\n* Update README.md\r\n\r\n* Update README.md\r\n\r\n* Update README.md\r\n\r\n* Update README.md\r\n\r\n* update\r\n\r\n* update\r\n\r\n* update figure\r\n\r\n* update figure\r\n\r\n* add fig\r\n\r\n* fix\r\n\r\n* fix\r\n\r\n* fix\r\n\r\n* fix\r\n\r\n* fix\r\n\r\n* fix\r\n\r\n* fix\r\n\r\n* fix\r\n\r\n* Update README_zh-CN.md\r\n\r\nCo-authored-by: Haian Huang(深度眸) <1286304229@qq.com>\r\n\r\n---------\r\n\r\nCo-authored-by: Haian Huang(深度眸) <1286304229@qq.com>","shortMessageHtmlLink":"[Feature] Release LLaVA-Llama-3-8B (#595)"}},{"before":"ae1d9811471b9ba8dea69cac52a03e3c37e34eff","after":"36a1f263509102243f91bb50b9ed98e2e2b89a56","ref":"refs/heads/main","pushedAt":"2024-04-22T03:13:58.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"LZHgrla","name":"Zhihao Lin","path":"/LZHgrla","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/36994684?s=80&v=4"},"commit":{"message":"[Fix] LLaVA-v1.5 official settings (#594)\n\n* Update llava_v15_7b_finetune_lora.py\r\n\r\n* Update llava_v15_13b_finetune_lora.py","shortMessageHtmlLink":"[Fix] LLaVA-v1.5 official settings (#594)"}},{"before":"2eb3ac64d287c135b3ff94f47705ab2543574b60","after":"ae1d9811471b9ba8dea69cac52a03e3c37e34eff","ref":"refs/heads/main","pushedAt":"2024-04-19T11:20:12.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"HIT-cwh","name":"whcao","path":"/HIT-cwh","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/41630003?s=80&v=4"},"commit":{"message":"bump version to 0.1.18 (#590)","shortMessageHtmlLink":"bump version to 0.1.18 (#590)"}},{"before":"d722775b27c42e81570b4806f26e9483fb8ac5ab","after":"2eb3ac64d287c135b3ff94f47705ab2543574b60","ref":"refs/heads/main","pushedAt":"2024-04-19T11:16:00.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Docs] Add tokenizer to sft in Case 2 (#583)\n\nadd tokenizer to sft in Case 2","shortMessageHtmlLink":"[Docs] Add tokenizer to sft in Case 2 (#583)"}},{"before":"f125a896ba106c558eba047794d82b0ae0decfb0","after":"582a3987e7d2adba809d62452a2539394002e4a2","ref":"refs/heads/docs","pushedAt":"2024-04-19T11:15:39.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"pppppM","name":null,"path":"/pppppM","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/67539920?s=80&v=4"},"commit":{"message":"[Docs] Add tokenizer to sft in Case 2 (#584)\n\nadd tokenizer to sft in Case 2","shortMessageHtmlLink":"[Docs] Add tokenizer to sft in Case 2 (#584)"}}],"hasNextPage":true,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAAETJ8nqgA","startCursor":null,"endCursor":null}},"title":"Activity · InternLM/xtuner"}