..
ggml-alloc.h
ggml : upgrade init_tensor API to return a ggml_status ( #11854 )
2025-02-28 14:41:47 +01:00
ggml-backend.h
Add `--no-op-offload` to improve `-ot` pp perf in MoE models like llama4 400B ( #13386 )
2025-05-11 14:18:39 +02:00
ggml-blas.h
ggml : build backends as libraries ( #10256 )
2024-11-14 18:04:35 +01:00
ggml-cann.h
ggml : build backends as libraries ( #10256 )
2024-11-14 18:04:35 +01:00
ggml-cpp.h
ggml : fix ggml_gallocr_ptr type (ggml/1205)
2025-05-01 09:58:44 +03:00
ggml-cpu.h
ggml: move fp16/bf16 conversion optimizations to CPU backend + export conversion APIs ( #13107 )
2025-04-26 16:05:31 +02:00
ggml-cuda.h
ggml : build backends as libraries ( #10256 )
2024-11-14 18:04:35 +01:00
ggml-kompute.h
ggml : build backends as libraries ( #10256 )
2024-11-14 18:04:35 +01:00
ggml-metal.h
repo : update links to new url ( #11886 )
2025-02-15 16:40:57 +02:00
ggml-opencl.h
Introducing experimental OpenCL backend with support for Qualcomm Adreno GPUs ( #10693 )
2024-12-13 12:23:52 -08:00
ggml-opt.h
llama/ggml: add LLM training support ( #10544 )
2025-05-12 14:44:49 +02:00
ggml-rpc.h
rpc : do not wait for response when sending RPC_CMD_SET_TENSOR ( #12943 )
2025-04-25 10:08:08 +03:00
ggml-sycl.h
ggml : build backends as libraries ( #10256 )
2024-11-14 18:04:35 +01:00
ggml-vulkan.h
vulkan: Make Vulkan optional at runtime ( #11493 ). ( #11494 )
2025-02-10 07:17:21 +01:00
ggml.h
llama/ggml: add LLM training support ( #10544 )
2025-05-12 14:44:49 +02:00
gguf.h
GGUF: C++ refactor, backend support, misc fixes ( #11030 )
2025-01-07 18:01:58 +01:00