Package Details: whisper.cpp-cuda 1.8.2-1

Git Clone URL: https://aur.archlinux.org/whisper.cpp-cuda.git (read-only, click to copy)
Package Base: whisper.cpp-cuda
Description: Port of OpenAI's Whisper model in C/C++ (with NVIDIA CUDA optimizations)
Upstream URL: https://github.com/ggerganov/whisper.cpp
Licenses: MIT
Conflicts: whisper.cpp
Provides: whisper.cpp
Submitter: robertfoster
Maintainer: robertfoster
Last Packager: robertfoster
Votes: 2
Popularity: 0.46
First Submitted: 2024-12-11 23:08 (UTC)
Last Updated: 2025-10-16 07:57 (UTC)

Latest Comments

1 2 Next › Last »

luiscastro193 commented on 2025-10-16 09:33 (UTC)

It compiled fine for me

Kcchouette commented on 2025-10-16 09:29 (UTC) (edited on 2025-10-16 09:44 (UTC) by Kcchouette)

[ 86%] Linking CXX executable ../../../bin/wchess
[ 87%] Building CXX object examples/talk-llama/CMakeFiles/whisper-talk-llama.dir/llama-model-saver.cpp.o
[ 89%] Building CXX object examples/talk-llama/CMakeFiles/whisper-talk-llama.dir/llama-model.cpp.o
/var/tmp/pamac-build/whisper.cpp-cuda/src/whisper.cpp-1.8.2/examples/talk-llama/llama-model.cpp: In constructor ‘llm_build_apertus::llm_build_apertus(const llama_model&, const llm_graph_params&)’:
/var/tmp/pamac-build/whisper.cpp-cuda/src/whisper.cpp-1.8.2/examples/talk-llama/llama-model.cpp:19330:43: error: ‘ggml_xielu’ was not declared in this scope; did you mean ‘ggml_silu’?
19330 |                 ggml_tensor * activated = ggml_xielu(ctx0, up, alpha_n_val, alpha_p_val, beta_val, eps_val);
      |                                           ^~~~~~~~~~
      |                                           ggml_silu
make[2]: *** [examples/talk-llama/CMakeFiles/whisper-talk-llama.dir/build.make:373: examples/talk-llama/CMakeFiles/whisper-talk-llama.dir/llama-model.cpp.o] Error 1
make[1]: *** [CMakeFiles/Makefile2:733: examples/talk-llama/CMakeFiles/whisper-talk-llama.dir/all] Error 2
make[1]: *** Waiting for unfinished jobs....
[ 89%] Built target wchess
make: *** [Makefile:136: all] Error 2
==> ERROR: A failure occurred in build().
    Aborting...

Edit: building again libggml-cuda-git fix the problem

evorster commented on 2025-08-18 14:24 (UTC)

So close!

-- Installing: /home/evert/Aur/whisper.cpp-cuda/pkg/whisper.cpp-cuda/usr/lib/pkgconfig/whisper.pc
/home/evert/Aur/whisper.cpp-cuda/PKGBUILD: line 40: cd: /home/evert/Aur/whisper.cpp-cuda/src/build/bin: No such file or directory
==> ERROR: A failure occurred in package().
    Aborting...
 -> error making: whisper.cpp-cuda-exit status 4
 -> Failed to install the following packages. Manual intervention is required:
whisper.cpp-cuda - exit status 4
evert@Evert ~ [1]> 

ashs commented on 2025-08-11 16:24 (UTC) (edited on 2025-08-11 16:26 (UTC) by ashs)

I get this error at 95%:

/home/ashish/.cache/yay/whisper.cpp-cuda/src/whisper.cpp-1.7.6/examples/talk-llama/llama-model.cpp:224:42: error: too few arguments to function ‘ggml_tensor* ggml_ssm_scan(ggml_context*, ggml_tensor*, ggml_tensor*, ggml_tensor*, ggml_tensor*, ggml_tensor*, ggml_tensor*, ggml_tensor*)’
  224 |                 op_tensor = ggml_ssm_scan(ctx, s, x, dt, w, B, C);
      |                             ~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~

dbb commented on 2025-07-06 11:53 (UTC)

Fails to build in a clean chroot with:

/usr/bin/ld: warning: libcuda.so.1, needed by /usr/lib/libggml-cuda.so, not found (try using -rpath or -rpath-link)
/usr/bin/ld: /usr/lib/libggml-cuda.so: undefined reference to `cuMemCreate'
/usr/bin/ld: /usr/lib/libggml-cuda.so: undefined reference to `cuMemAddressReserve'
/usr/bin/ld: /usr/lib/libggml-cuda.so: undefined reference to `cuMemUnmap'
/usr/bin/ld: /usr/lib/libggml-cuda.so: undefined reference to `cuMemSetAccess'
/usr/bin/ld: /usr/lib/libggml-cuda.so: undefined reference to `cuDeviceGet'
/usr/bin/ld: /usr/lib/libggml-cuda.so: undefined reference to `cuMemAddressFree'
/usr/bin/ld: /usr/lib/libggml-cuda.so: undefined reference to `cuGetErrorString'
/usr/bin/ld: /usr/lib/libggml-cuda.so: undefined reference to `cuDeviceGetAttribute'
/usr/bin/ld: /usr/lib/libggml-cuda.so: undefined reference to `cuMemMap'
/usr/bin/ld: /usr/lib/libggml-cuda.so: undefined reference to `cuMemRelease'
/usr/bin/ld: /usr/lib/libggml-cuda.so: undefined reference to `cuMemGetAllocationGranularity'
collect2: error: ld returned 1 exit status
make[2]: *** [tests/CMakeFiles/test-vad.dir/build.make:107: bin/test-vad] Error 1
make[1]: *** [CMakeFiles/Makefile2:1418: tests/CMakeFiles/test-vad.dir/all] Error 2
make: *** [Makefile:146: all] Error 2

Had to add nvidia-utils to dependencies to get it to build.

evorster commented on 2025-06-27 17:18 (UTC) (edited on 2025-06-27 17:18 (UTC) by evorster)

I now get this compile error:

[ 84%] Building CXX object examples/talk-llama/CMakeFiles/whisper-talk-llama.dir/unicode-data.cpp.o
/home/evert/Aur/whisper.cpp-cuda/src/whisper.cpp-1.7.6/examples/talk-llama/llama-graph.cpp: In member function ‘ggml_tensor* llm_graph_context::build_moe_ffn(ggml_tensor*, ggml_tensor*, ggml_tensor*, ggml_tensor*, ggml_tensor*, ggml_tensor*, int64_t, int64_t, llm_ffn_op_type, bool, bool, float, llama_expert_gating_func_type, int) const’:
/home/evert/Aur/whisper.cpp-cuda/src/whisper.cpp-1.7.6/examples/talk-llama/llama-graph.cpp:722:34: error: ‘ggml_repeat_4d’ was not declared in this scope; did you mean ‘ggml_repeat’?
  722 |         ggml_tensor * repeated = ggml_repeat_4d(ctx0, cur, n_embd, n_expert_used, n_tokens, 1);
      |                                  ^~~~~~~~~~~~~~
      |                                  ggml_repeat
[ 85%] Linking CXX executable ../../bin/vad-speech-segments
make[2]: *** [examples/talk-llama/CMakeFiles/whisper-talk-llama.dir/build.make:205: examples/talk-llama/CMakeFiles/whisper-talk-llama.dir/llama-graph.cpp.o] Error 1
make[2]: *** Waiting for unfinished jobs....

brauliobo commented on 2025-05-30 21:37 (UTC)

it the maintainer active? last update is more than a month ago...

mikecr commented on 2025-05-28 07:32 (UTC)

I had to add this to make it build cos ggml_backend_sched_new() seemingly needs an extra arg now:

prepare() {
  cd "${srcdir}/${_pkgbase}-${pkgver}"
  sed -i '602s~.*~    sched = ggml_backend_sched_new(backends.data(), nullptr, backends.size(), WHISPER_MAX_NODES, false, false);~' src/whisper.cpp
  sed -i '9833s~.*~            ctx->sched.reset(ggml_backend_sched_new(backend_ptrs.data(), backend_buft.data(), backend_ptrs.size(), max_nodes, pipeline_parallel, false));~' examples/talk-llama/llama.cpp
  sed -i '4i #include <cstdint>' examples/talk-llama/llama-mmap.h
}

hsantanna commented on 2025-05-10 06:27 (UTC) (edited on 2025-05-10 06:32 (UTC) by hsantanna)

If you get this error:

[ 84%] Building CXX object examples/talk-llama/CMakeFiles/whisper-talk-llama.dir/llama-mmap.cpp.o
In file included from /home/henrique/.cache/pikaur/build/whisper.cpp-cuda/src/whisper.cpp-1.7.5/examples/talk-llama/llama-mmap.cpp:1:
/home/henrique/.cache/pikaur/build/whisper.cpp-cuda/src/whisper.cpp-1.7.5/examples/talk-llama/llama-mmap.h:26:5: error: ‘uint32_t’ does not name a type
26 |     uint32_t read_u32() const;
|     ^~~~~~~~
/home/henrique/.cache/pikaur/build/whisper.cpp-cuda/src/whisper.cpp-1.7.5/examples/talk-llama/llama-mmap.h:5:1: note: ‘uint32_t’ is defined in header ‘<cstdint>’; this is probably fixable by adding ‘#include <cstdint>’
4 | #include <vector>
+++ |+#include <cstdint>
5 |

Just add this to PKGBUILD:

prepare() {
  cd "${srcdir}/${_pkgbase}-${pkgver}"
  sed -i '4i #include <cstdint>' examples/talk-llama/llama-mmap.h
}

lhy7889678 commented on 2025-03-15 10:21 (UTC)

Adding -DCMAKE_CUDA_ARCHITECTURES=89-real is not always a good idea, because not every GPU supports such a high compute capability. My RTX2060 fails during runtime (not compilation) with this line. As for me, changing it to -DCMAKE_CUDA_ARCHITECTURES=75 (compute capability 7.5) fixes the problem. You may need to check https://developer.nvidia.com/cuda-gpus for the compute capability of a given GPU model.