diff options
author | Felix Erkinger | 2023-12-08 18:47:18 +0100 |
---|---|---|
committer | Felix Erkinger | 2023-12-08 18:47:18 +0100 |
commit | ff79985edbfc667077f0e2592af5077fe6750842 (patch) | |
tree | 1c8e006e317f303e0aa25b2ff7b61a2aa92bb890 | |
parent | 7801efa2a1d549f427b3a497591badc1c0e63da3 (diff) | |
download | aur-ff79985edbfc667077f0e2592af5077fe6750842.tar.gz |
make buildtype changeable from env, upgpkg: localai-git v2.0.0.14.gb181503-1
-rw-r--r-- | .SRCINFO | 10 | ||||
-rw-r--r-- | PKGBUILD | 21 |
2 files changed, 13 insertions, 18 deletions
@@ -1,7 +1,7 @@ pkgbase = localai-git - pkgdesc = The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. - pkgver = v2.0.0.6.g997119c - pkgrel = 2 + pkgdesc = Self-hosted OpenAI API alternative - Open Source, community-driven and local-first. + pkgver = v2.0.0.14.gb181503 + pkgrel = 1 url = https://github.com/mudler/LocalAI arch = x86_64 license = MIT @@ -29,9 +29,9 @@ pkgbase = localai-git pkgname = localai-git pkgname = localai-git-cuda - pkgdesc = The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. (with CUDA support) + pkgdesc = Self-hosted OpenAI API alternative - Open Source, community-driven and local-first. (with CUDA support) depends = cuda pkgname = localai-git-rocm - pkgdesc = The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. (with ROCM support) + pkgdesc = Self-hosted OpenAI API alternative - Open Source, community-driven and local-first. (with ROCM support) depends = rocm-hip-runtime @@ -1,9 +1,9 @@ # Maintainer: wuxxin <wuxxin@gmail.com> # to only build for cpu, set ENABLE_CUDA and ENABLE_ROCM to 0 -_ENABLE_CUDA=1 -_ENABLE_ROCM=1 -_SKIP_CPU=0 +_ENABLE_CUDA=${_ENABLE_CUDA:-1} +_ENABLE_ROCM=${_ENABLE_ROCM:-1} +_SKIP_CPU=${_SKIP_CPU:-0} _GO_TAGS="" # _GO_TAGS="tts stablediffusion" _OPTIONAL_BACKENDS="" @@ -19,9 +19,9 @@ _pkgname="localai" pkgbase="${_pkgname}-git" pkgname=("${pkgbase}") -pkgver=v2.0.0.6.g997119c -pkgrel=2 -pkgdesc="The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first." +pkgver=v2.0.0.14.gb181503 +pkgrel=1 +pkgdesc="Self-hosted OpenAI API alternative - Open Source, community-driven and local-first." url="https://github.com/mudler/LocalAI" license=('MIT') arch=('x86_64') @@ -45,8 +45,8 @@ makedepends=( if test "$(echo "$_GO_TAGS" | grep -o "tts")" = "tts"; then makedepends+=( 'onnxruntime' - 'piper-phonemize' ) + # 'piper-phonemize' is build from piper fi if [[ $_ENABLE_CUDA = 1 ]]; then @@ -140,7 +140,7 @@ build() { else _AMDGPU_TARGETS="${AMDGPU_TARGETS:-gfx900;gfx906;gfx908;gfx90a;gfx1030;gfx1100;gfx1101;gfx1102}" fi - # XXX workaround build error on ROCM by removing cf-procetion from CXX_FLAGS + # XXX workaround build error on ROCM by removing unsupported cf-protection from CMAKE_CXX_FLAGS sed -i '1s/^/string(REPLACE "-fcf-protection" "" CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS}")\n/' backend/cpp/llama/llama.cpp/CMakeLists.txt MAGMA_HOME="$ROCM_HOME" AMDGPU_TARGETS="$_AMDGPU_TARGETS" GPU_TARGETS="$_AMDGPU_TARGETS" \ _build hipblas @@ -149,11 +149,6 @@ build() { _package_install() { install -Dm755 "local-ai" "${pkgdir}/usr/bin/local-ai" - # sources/go-piper/piper/build/pi/lib/* /usr/lib/ - - # add 1-2 7b high performing models yaml configs based on mistral as gpt-3.5 - # prefer chatml, add example working preload-models.yaml, - install -Dm644 README.md -t "${pkgdir}/usr/share/doc/${_pkgname}" } |