18 Commits (dc5a645434f0ea6364c426c6ba112da1afa40cb2)

Author SHA1 Message Date
Daniel Hiltgen 1c6669e64c
Re-remove cuda v11 (#10694) 10 months ago
Michael Yang 139f84cf21
fix cmakelists (#10804) 11 months ago
Michael Yang 375839ea2d
chore: disable debug in binary libraries (#10788) 11 months ago
Jeffrey Morgan 09bb2e30f6
ml/backend/ggml: use default CUDA compression mode (#10314) 12 months ago
Jeffrey Morgan dc264be6ff
ml: add missing cmake property and remove additional CMakeLists.txt (#10310) 12 months ago
Jeffrey Morgan 943464ccb8
llama: update to commit 71e90e88 (#10192) 12 months ago
saman-amd ead27aa9fe
Add gfx1200 & gfx1201 support on linux (#9878) 1 year ago
Jeffrey Morgan 96a97adf9b
build: use correct GGML_HIP_NO_VMM compiler definition for ggml-hip (#9451) 1 year ago
Jeffrey Morgan e75c6126e9
build: set GGML_CUDA_NO_VMM for ggml-hip target (#9449) 1 year ago
Michael Yang b42aba40ed cuda: enable flash attention 1 year ago
Michael Yang 08a299e1d0 cmake: avoid building intel backends on linux 1 year ago
Jeffrey Morgan a4f69a0191
build: add -DGGML_CUDA_NO_PEER_COPY=ON for rocm builds on windows (#9060) 1 year ago
Michael Yang abb8dd57f8
add gfx instinct gpus (#8933) 1 year ago
Michael Yang ae7e368f75
build(rocm): add numa, elf (#8900) 1 year ago
Michael Yang 9a4757ae66
build(rocm): add tinfo (#8899) 1 year ago
Michael Yang b698f9a0d8
build: add missing dependencies (#8896) 1 year ago
Jeffrey Morgan 4759ecae19
ml/backend/ggml: fix library loading on macOS amd64 (#8827) 1 year ago
Michael Yang dcfb7a105c
next build (#8539) 1 year ago
Michael Yang 442dec1c6f vendor llama.cpp 3 years ago
Michael Yang fd4792ec56 call llama.cpp directly from go 3 years ago