{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":651337680,"defaultBranch":"master","name":"llama.cpp","ownerLogin":"nomic-ai","currentUserCanPush":false,"isFork":true,"isEmpty":false,"createdAt":"2023-06-09T03:03:13.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/102670180?v=4","public":true,"private":false,"isOrgOwned":true},"refInfo":{"name":"","listCacheKey":"v0:1717526480.0","currentOid":""},"activityList":{"items":[{"before":"ed12631213e1069d27d6a88913d489301ae9b1a1","after":"b2db03acf299111885af2921a4230de07623eaf8","ref":"refs/heads/master","pushedAt":"2024-06-04T18:41:37.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"llama: replace ngl=0 hack with llama_model_using_gpu\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"llama: replace ngl=0 hack with llama_model_using_gpu"}},{"before":null,"after":"b2db03acf299111885af2921a4230de07623eaf8","ref":"refs/heads/cuda-early-alloc","pushedAt":"2024-06-04T18:41:20.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"llama: replace ngl=0 hack with llama_model_using_gpu\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"llama: replace ngl=0 hack with llama_model_using_gpu"}},{"before":"f67f4651fac0b2f377dc53fe853b1dafa96f9aa9","after":"ed12631213e1069d27d6a88913d489301ae9b1a1","ref":"refs/heads/master","pushedAt":"2024-05-30T20:32:44.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"llama : allow skipping CPU graph with llama_set_skip_cpu\n\nThis allows us to exercise the device graph in order to allocate VRAM\nwithout wasting cycles on the CPU graph.\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"llama : allow skipping CPU graph with llama_set_skip_cpu"}},{"before":"fadf1135a54e80188d644df42ad6a53bf986e8b0","after":"f67f4651fac0b2f377dc53fe853b1dafa96f9aa9","ref":"refs/heads/master","pushedAt":"2024-05-28T17:32:14.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"llama : work around load crash caused by out-of-vocab token IDs\n\nUpstream issue: ggerganov/llama.cpp#2378","shortMessageHtmlLink":"llama : work around load crash caused by out-of-vocab token IDs"}},{"before":"40bac11e427f2307305b86c322cb366bb95fcb8a","after":"fadf1135a54e80188d644df42ad6a53bf986e8b0","ref":"refs/heads/master","pushedAt":"2024-05-20T15:52:58.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"kompute : fix use-after-free in ggml_vk_get_device\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"kompute : fix use-after-free in ggml_vk_get_device"}},{"before":"a3f03b7e793ee611c4918235d4532ee535a9530d","after":null,"ref":"refs/heads/master-old","pushedAt":"2024-05-14T22:18:21.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"}},{"before":null,"after":"a3f03b7e793ee611c4918235d4532ee535a9530d","ref":"refs/heads/gpt4all-v2.7.5","pushedAt":"2024-05-14T22:18:04.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"kompute : put device with most VRAM first, not least\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"kompute : put device with most VRAM first, not least"}},{"before":null,"after":"2a086f71f5b570a0f047f88d88cf5704aae7ec7c","ref":"refs/heads/gpt4all-v2.7.3","pushedAt":"2024-05-14T22:16:57.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"mpt : remove output tensor name to satisfy quantize check\n\n(cherry picked from commit daf88100eb22c620a691228b200282156771e737)\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"mpt : remove output tensor name to satisfy quantize check"}},{"before":null,"after":"cfb5f030658966ebdba9bf170655d97450dda50d","ref":"refs/heads/gpt4all-v2.7.2","pushedAt":"2024-05-14T22:15:56.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"llama : fix -Wunused-const-variable warning for non-Kompute build\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"llama : fix -Wunused-const-variable warning for non-Kompute build"}},{"before":null,"after":"7162b64190f0bf6118336098d66b0d08566e2ce5","ref":"refs/heads/gpt4all-v2.7.2-88e330ef","pushedAt":"2024-05-14T22:15:17.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"kompute : enable GPU support for 10 more model architectures\n\nThese are Baichuan, Bert and Nomic Bert, CodeShell, GPT-2, InternLM,\nMiniCPM, Orion, Qwen, and StarCoder.\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"kompute : enable GPU support for 10 more model architectures"}},{"before":null,"after":"12c7910ec32550fb5a5520b9b1a105071431576d","ref":"refs/heads/gpt4all-v2.7.2-fc6c5ea0","pushedAt":"2024-05-14T22:14:21.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"llama : fix loading models with shared tok_embd and output (#5651)\n\nggml-ci","shortMessageHtmlLink":"llama : fix loading models with shared tok_embd and output (ggerganov…"}},{"before":null,"after":"b61ee89fca2038e9937317a794e28e08391b7888","ref":"refs/heads/gpt4all-v2.7.1","pushedAt":"2024-05-14T22:13:29.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"llama : disable space prefix in tokenize\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"llama : disable space prefix in tokenize"}},{"before":null,"after":"7d4ced850548642b9a1740fa25ecdef249fbf47f","ref":"refs/heads/gpt4all-v2.7.1-7810b757","pushedAt":"2024-05-14T22:12:48.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"kompute : add gemma, phi-2, qwen2, and stablelm to whitelist\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"kompute : add gemma, phi-2, qwen2, and stablelm to whitelist"}},{"before":null,"after":"822a9c894eb3770c65f0b4a724aae34605c90029","ref":"refs/heads/gpt4all-v2.7.1-6fdec808","pushedAt":"2024-05-14T22:11:40.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"Early return for zero size calls to get_tensor. (#5482)\n\n* Early return for zero size calls to get_tensor.\r\n\r\nSigned-off-by: Adam Treat \r\n\r\n* Update ggml-kompute.cpp\r\n\r\nCo-authored-by: Georgi Gerganov \r\n\r\n* Update ggml-kompute.cpp\r\n\r\nCo-authored-by: Georgi Gerganov \r\n\r\n* Add an early return to the get/set tensor when the size is null.\r\n\r\nSigned-off-by: Adam Treat \r\n\r\n* Early return after the assertions.\r\n\r\nSigned-off-by: Adam Treat \r\n\r\n* Since we do the early return in the generic backend now no reason to do so here as well.\r\n\r\nSigned-off-by: Adam Treat \r\n\r\n---------\r\n\r\nSigned-off-by: Adam Treat \r\nCo-authored-by: Georgi Gerganov ","shortMessageHtmlLink":"Early return for zero size calls to get_tensor. (ggerganov#5482)"}},{"before":null,"after":"74f977c196286e937fc3a40af9f1638f018761a8","ref":"refs/heads/gpt4all-v2.5.3","pushedAt":"2024-05-14T22:09:16.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"llama : fix Vulkan whitelist (#11)","shortMessageHtmlLink":"llama : fix Vulkan whitelist (#11)"}},{"before":null,"after":"2b7b3f37f9d0193943c3d31e3e617bc6351838c8","ref":"refs/heads/gpt4all-v2.5.0-f605a5b6","pushedAt":"2024-05-14T22:06:33.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"Delete TODO now that we have q8_0.","shortMessageHtmlLink":"Delete TODO now that we have q8_0."}},{"before":null,"after":"4458a8eaf443e7fa0e764682d22213fa4fef90c3","ref":"refs/heads/gpt4all-v2.4.7","pushedAt":"2024-05-14T22:00:55.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"metal kernels: cpy f16->f16, alibi","shortMessageHtmlLink":"metal kernels: cpy f16->f16, alibi"}},{"before":null,"after":"47aec1bcc09e090f0b8f196dc0a4e43b89507e4a","ref":"refs/heads/gpt4all-v2.6.2","pushedAt":"2024-05-14T21:50:40.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"kompute : add missing call to ggml_backend_kompute_device_unref\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"kompute : add missing call to ggml_backend_kompute_device_unref"}},{"before":"e21d0b32e9324b1df40107d9004cc6b67cdd1567","after":"40bac11e427f2307305b86c322cb366bb95fcb8a","ref":"refs/heads/master","pushedAt":"2024-05-13T16:07:15.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"kompute: fix leaks in ggml_vk_current_device\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"kompute: fix leaks in ggml_vk_current_device"}},{"before":"acafa7ffe6adcb45b4e2f229c688d1c25c1bbdb7","after":"e21d0b32e9324b1df40107d9004cc6b67cdd1567","ref":"refs/heads/master","pushedAt":"2024-05-09T16:01:54.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"kompute: update submodule for install fix\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"kompute: update submodule for install fix"}},{"before":"a3f03b7e793ee611c4918235d4532ee535a9530d","after":"acafa7ffe6adcb45b4e2f229c688d1c25c1bbdb7","ref":"refs/heads/master","pushedAt":"2024-05-08T19:46:24.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"kompute: update submodule for install fix\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"kompute: update submodule for install fix"}},{"before":null,"after":"a3f03b7e793ee611c4918235d4532ee535a9530d","ref":"refs/heads/master-old","pushedAt":"2024-05-08T19:45:20.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"kompute : put device with most VRAM first, not least\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"kompute : put device with most VRAM first, not least"}},{"before":"9fb04f5818b0adf15caf563150bfba06e429af85","after":"b627b8dd2ea97985283a5b4506f2796500850cc8","ref":"refs/heads/gpt4all-cuda","pushedAt":"2024-05-08T18:30:30.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"kompute: update submodule for install fix\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"kompute: update submodule for install fix"}},{"before":"64b3aed560a572a4aeda488466d340e116031361","after":"9fb04f5818b0adf15caf563150bfba06e429af85","ref":"refs/heads/gpt4all-cuda","pushedAt":"2024-05-08T15:25:06.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"kompute: update submodule for install fix\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"kompute: update submodule for install fix"}},{"before":"43f7cf5044022f9ec12db3bb0706b42f5217e7bc","after":"64b3aed560a572a4aeda488466d340e116031361","ref":"refs/heads/gpt4all-cuda","pushedAt":"2024-05-07T17:52:05.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"cuda: remove unconditional stderr prints on init\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"cuda: remove unconditional stderr prints on init"}},{"before":null,"after":"43f7cf5044022f9ec12db3bb0706b42f5217e7bc","ref":"refs/heads/gpt4all-cuda","pushedAt":"2024-05-07T17:45:42.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"cuda: remove unconditional stderr prints on init\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"cuda: remove unconditional stderr prints on init"}},{"before":"e3c4f65d786d26f1daa7aebfb1b67cd6c31ea082","after":"a3f03b7e793ee611c4918235d4532ee535a9530d","ref":"refs/heads/master","pushedAt":"2024-05-01T18:48:02.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"kompute : put device with most VRAM first, not least\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"kompute : put device with most VRAM first, not least"}},{"before":"ee32d48b1422ea6effb92f4bf16f2dbc7abaf7fa","after":"e3c4f65d786d26f1daa7aebfb1b67cd6c31ea082","ref":"refs/heads/master","pushedAt":"2024-03-25T20:47:59.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"wpm : portable unicode tolower (ggerganov/llama.cpp#6305)\n\nexcludes unicodedata.cpp split\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"wpm : portable unicode tolower (ggerganov#6305)"}},{"before":"43c20ce8004a4eac25ffe89e52bdf94bc7c47c02","after":"ee32d48b1422ea6effb92f4bf16f2dbc7abaf7fa","ref":"refs/heads/master","pushedAt":"2024-03-25T16:19:51.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"Fix heap corruption from wmode out-of-bound writes on windows (#6272)\n\n* would throw error on VS2022 on GGML_FREE(wmode)\n* wchar_t is usually 2 bytes, but malloc wants bytes\n * therefore `*wmode_p++ = (wchar_t)*mode;` could write off the end of the allocation\n* Fixes error possibly introduced by https://github.com/ggerganov/llama.cpp/pull/6248\n\n(cherry picked from commit a32b77c4b2c1808654d0b952f26c37d73d2e746b)\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"Fix heap corruption from wmode out-of-bound writes on windows (ggerga…"}},{"before":"2a086f71f5b570a0f047f88d88cf5704aae7ec7c","after":"43c20ce8004a4eac25ffe89e52bdf94bc7c47c02","ref":"refs/heads/master","pushedAt":"2024-03-25T16:17:15.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"cebtenzzre","name":"Jared Van Bortel","path":"/cebtenzzre","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14168726?s=80&v=4"},"commit":{"message":"llama : expose model name and architecture via API\n\nSigned-off-by: Jared Van Bortel ","shortMessageHtmlLink":"llama : expose model name and architecture via API"}}],"hasNextPage":true,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAAEXFsfdwA","startCursor":null,"endCursor":null}},"title":"Activity · nomic-ai/llama.cpp"}