Skip to content

Releases: JamePeng/llama-cpp-python

v0.3.14-cu128-AVX2-win-20250801

01 Aug 17:24

Choose a tag to compare

  1. Compile with cuda12.8.1 for Blackwell architecture (sm_100 and sm_120) computing cards
  2. Sync llama.cpp API 20250801
  3. Remove sm_70 in cuda12.8.1 action flow

v0.3.14-cu128-AVX2-linux-20250801

01 Aug 14:34

Choose a tag to compare

  1. Compile with cuda12.8.1 for Blackwell architecture (sm_100 and sm_120) computing cards
  2. Sync llama.cpp API 20250801
  3. Remove sm_70 in cuda12.8.1 action flow

v0.3.14-cu126-AVX2-win-20250801

01 Aug 16:26

Choose a tag to compare

Sync llama.cpp API 20250801

v0.3.14-cu126-AVX2-linux-20250801

01 Aug 13:24

Choose a tag to compare

Sync llama.cpp API 20250801

v0.3.14-cu124-AVX2-win-20250801

01 Aug 16:25

Choose a tag to compare

Sync llama.cpp API 20250801

v0.3.14-cu124-AVX2-linux-20250801

01 Aug 13:24

Choose a tag to compare

Sync llama.cpp API 20250801

v0.3.13-cu126-AVX2-win-20250717

17 Jul 16:16

Choose a tag to compare

fix memory_seq_rm crash bug

v0.3.13-cu126-AVX2-linux-20250717

17 Jul 13:08

Choose a tag to compare

fix memory_seq_rm crash bug

v0.3.13-cu124-AVX2-win-20250717

17 Jul 16:00

Choose a tag to compare

fix memory_seq_rm crash bug

v0.3.13-cu124-AVX2-linux-20250717

17 Jul 13:07

Choose a tag to compare

fix memory_seq_rm crash bug