# test ovc with NumPy 2.x on Ubuntu 24 with default Python 3.12
# test against NumPy 1.x with older Python versions
# optimum still requires numpy<2.0.0
numpy==1.26.4; python_version < "3.12"
numpy==2.1.1; python_version >= "3.12"
torch==2.9.0
--extra-index-url https://download.pytorch.org/whl/cpu

torchvision==0.24.0
torchaudio==2.9.0
# before updating transformers version, make sure no tests (esp. sdpa2pa) are failing
transformers==4.47.1
pytest==7.0.1; python_version < '3.10'
pytest==7.2.0; python_version >= '3.10'
pytest-html==4.2.0
pytest-xdist[psutil]==3.6.1
defusedxml==0.7.1

autoawq==0.2.7; platform_system == "Linux" and platform_machine == "x86_64"
# triton is a dependency of autoawq, newer versions lead to TorchFX test failures
triton==3.5.1; platform_system == "Linux" and platform_machine == "x86_64"
auto-gptq==0.7.1; platform_system == "Linux" and platform_machine == "x86_64" and python_version < "3.12"
# peft is a dependency of auto-gptq, newer versions lead to gptq model loading failures
peft==0.17.1; platform_system == "Linux" and platform_machine == "x86_64" and python_version < "3.12"
av==13.0.0
basicsr==1.4.2; python_version < "3.12"
datasets==3.0.1
easyocr==1.7.2
facexlib==0.3.0; python_version < "3.12"
librosa==0.10.2; python_version < "3.12"
packaging==24.1
pandas==2.2.3
protobuf==5.28.3
pyctcdecode==0.5.0; python_version < "3.12"
sacremoses==0.1.1
sentencepiece==0.2.1
soundfile==0.12.1
super-image==0.1.7; python_version < "3.12"
timm==1.0.24
wheel==0.44.0
PyYAML==6.0.2
kornia==0.7.3
super-image==0.1.7
# huggingface-hub required for super-image
huggingface-hub==0.25.2

# For now, we decided to pin a specific working version of optimum-intel.
# It will be discussed in the future how to manage versioning of the components properly.
git+https://github.com/huggingface/optimum-intel.git@8ba536cd0a2bf93e9e88408b0048a7695db5be0b; python_version < "3.12"
# set 'export HF_HUB_ENABLE_HF_TRANSFER=1' to benefits from hf_transfer
hf_transfer==0.1.9

# requirements for specific models
# - hf-tiny-model-private/tiny-random-RoFormerForCausalLM
rjieba==0.1.11

# - katuni4ka/tiny-random-qwen
# - katuni4ka/tiny-random-internlm2
transformers_stream_generator==0.0.5
einops==0.8.0
