Releases: nomic-ai/llama.cpp
Releases · nomic-ai/llama.cpp
b2245
b2023
Early return for zero size calls to get_tensor. (#5482) * Early return for zero size calls to get_tensor. Signed-off-by: Adam Treat <treat.adam@gmail.com> * Update ggml-kompute.cpp Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> * Update ggml-kompute.cpp Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> * Add an early return to the get/set tensor when the size is null. Signed-off-by: Adam Treat <treat.adam@gmail.com> * Early return after the assertions. Signed-off-by: Adam Treat <treat.adam@gmail.com> * Since we do the early return in the generic backend now no reason to do so here as well. Signed-off-by: Adam Treat <treat.adam@gmail.com> --------- Signed-off-by: Adam Treat <treat.adam@gmail.com> Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
b2022
kompute : make partial tensor copies faster by syncing less data (#15) Signed-off-by: Jared Van Bortel <jared@nomic.ai>
b2021
kompute : do not list Intel GPUs as they are unsupported (#14) Signed-off-by: Jared Van Bortel <jared@nomic.ai>
b2020
kompute : disable GPU offload for Mixtral We haven't implemented the necessary GPU kernels yet. Fixes this crash: ggml_vk_graph_compute: error: unsupported op 'ARGSORT' GGML_ASSERT: /home/jared/src/forks/gpt4all/gpt4all-backend/llama.cpp-mainline/ggml-kompute.cpp:1508: !"unsupported op" Signed-off-by: Jared Van Bortel <jared@nomic.ai>
b2019
common : remove llama_token_to_piece for compatibility with hack Signed-off-by: Jared Van Bortel <jared@nomic.ai>
b1782
Merge branch 'ceb/nomic-vulkan' into nomic
b1780
Merge branch 'ceb/nomic-vulkan' into nomic
b1720
Merge branch 'ceb/nomic-vulkan' into nomic
b1641
kompute : fix -Wunused-private-field warnings from clang Fixes nomic-ai/gpt4all#1722