Change8

b8340

📦 llama-cppView on GitHub →
2 features🔧 1 symbols

Summary

This release introduces native AVX512-FP16 support for F16 operations within ggml, alongside the release of numerous pre-compiled binaries for various platforms and hardware configurations.

Migration Steps

  1. If building from source, note that the AVX512-FP16 path is only enabled for native builds or when custom flags are used.

✨ New Features

  • Added native AVX512-FP16 support for F16 operations in ggml.
  • New pre-built binaries available for macOS (Apple Silicon and Intel), Linux (various configurations including Vulkan, ROCm 7.2, OpenVINO), Windows (CPU, CUDA 12.4, CUDA 13.1, Vulkan, SYCL, HIP), and openEuler.

Affected Symbols