the build is taking 45gb! just the build/bin folder takes 38gb:
braulio @ whitebeast ➜ bin git:(master) pwd
/home/braulio/.cache/yay/llama.cpp-cuda/src/build/bin
braulio @ whitebeast ➜ bin git:(master) du -h --max-depth=1
38G .
braulio @ whitebeast ➜ bin git:(master) ls
llama-batched llama-gbnf-validator llama-lookup llama-quantize llama-vdot test-grammar-integration test-sampling
llama-batched-bench llama-gguf llama-lookup-create llama-quantize-stats rpc-server test-grammar-parser test-tokenizer-0
llama-bench llama-gguf-hash llama-lookup-merge llama-retrieval test-arg-parser test-json-schema-to-grammar test-tokenizer-1-bpe
llama-cli llama-gguf-split llama-lookup-stats llama-save-load-state test-autorelease test-llama-grammar test-tokenizer-1-spm
llama-convert-llama2c-to-ggml llama-gritlm llama-minicpmv-cli llama-server test-backend-ops test-log
llama-cvector-generator llama-imatrix llama-parallel llama-simple test-barrier test-model-load-cancel
llama-embedding llama-infill llama-passkey llama-simple-chat test-c test-quantize-fns
llama-eval-callback llama-llava-cli llama-perplexity llama-speculative test-chat-template test-quantize-perf
llama-export-lora llama-lookahead llama-q8dot llama-tokenize test-grad0 test-rope
Pinned Comments
txtsd commented on 2024-10-26 20:17 (UTC) (edited on 2024-12-06 14:15 (UTC) by txtsd)
Alternate versions
llama.cpp
llama.cpp-vulkan
llama.cpp-sycl-fp16
llama.cpp-sycl-fp32
llama.cpp-cuda
llama.cpp-cuda-f16
llama.cpp-hip