## START: Set by rpmautospec ## (rpmautospec version 0.6.0) ## RPMAUTOSPEC: autorelease, autochangelog %define autorelease(e:s:pb:n) %{?-p:0.}%{lua: release_number = 1; base_release_number = tonumber(rpm.expand("%{?-b*}%{!?-b:1}")); print(release_number + base_release_number - 1); }%{?-e:.%{-e*}}%{?-s:.%{-s*}}%{!?-n:%{?dist}} ## END: Set by rpmautospec %global pypi_name llama-cpp-python %global pypi_version 0.2.87 # it's all python code %global debug_package %{nil} Name: python-%{pypi_name} Version: %{pypi_version} Release: %autorelease License: MIT Summary: Simple Python bindings for @ggerganov's llama.cpp library URL: https://github.com/abetlen/llama-cpp-python Source: %{url}/archive/v%{version}/%{pypi_name}-%{version}.tar.gz Patch1: 0001-don-t-build-llama.cpp-and-llava.patch Patch2: 0002-search-for-libllama-so-in-usr-lib64.patch %bcond_without test # this is what llama-cpp is on # and this library is by default installed in /usr/lib64/python3.12/site-packages/llama_cpp/__init__.py ExclusiveArch: x86_64 aarch64 BuildRequires: git-core BuildRequires: gcc BuildRequires: gcc-c++ BuildRequires: ninja-build BuildRequires: python3-devel BuildRequires: llama-cpp-devel %if %{with test} BuildRequires: python3-pytest BuildRequires: python3-scipy %endif %generate_buildrequires %pyproject_buildrequires %description %{pypi_name} provides: Low-level access to C API via `ctypes` interface. High-level Python API for text completion. OpenAI compatible web server %package -n python3-%{pypi_name} Summary: %{summary} # -devel has the unversioned libllama.so Requires: llama-cpp-devel %description -n python3-%{pypi_name} %{pypi_name} provides: Low-level access to C API via `ctypes` interface. High-level Python API for text completion. OpenAI compatible web server %prep %autosetup -p1 -n %{pypi_name}-%{version} -Sgit %build %pyproject_wheel %if %{with test} %check # most test_llama.py tests utilize model ggml-vocab-llama.gguf from vendored llama.cpp %pytest -vs tests/test_llama.py::test_llama_cpp_version tests/test_llama.py::test_logits_to_logprobs tests/test_llama_speculative.py tests/test_llama_chat_format.py %endif %install %pyproject_install %pyproject_save_files -l llama_cpp -L %files -n python3-%{pypi_name} -f %{pyproject_files} %license LICENSE.md %doc README.md %changelog ## START: Generated by rpmautospec * Fri Nov 15 2024 Tomas Tomecek - 0.2.87-1 - update to 0.2.87 to match llama-cpp-b3561 * Mon Jun 24 2024 Tomas Tomecek - 0.2.75-5 - build the package in f40 * Wed Jun 19 2024 Python Maint - 0.2.75-4 - Rebuilt for Python 3.13 * Thu May 23 2024 Tomas Tomecek - 0.2.75-3 - build with tests on * Thu May 23 2024 Tomas Tomecek - 0.2.75-2 - use %%%%pyproject_save_files -L * Thu May 23 2024 Mohammadreza Hendiani - 0.2.75-1 - update source to llama-cpp-python-0.2.75.tar.gz * Wed Apr 17 2024 Tomas Tomecek - 0.2.60-1 - initial import, version 0.2.60 ## END: Generated by rpmautospec