b8340
📦 llama-cppView on GitHub →
✨ 2 features🔧 1 symbols
Summary
This release introduces native AVX512-FP16 support for F16 operations within ggml, alongside the release of numerous pre-compiled binaries for various platforms and hardware configurations.
Migration Steps
- If building from source, note that the AVX512-FP16 path is only enabled for native builds or when custom flags are used.
✨ New Features
- Added native AVX512-FP16 support for F16 operations in ggml.
- New pre-built binaries available for macOS (Apple Silicon and Intel), Linux (various configurations including Vulkan, ROCm 7.2, OpenVINO), Windows (CPU, CUDA 12.4, CUDA 13.1, Vulkan, SYCL, HIP), and openEuler.