llm_topic_modelling / requirements_cpu.txt
seanpedrickcase's picture
Added model compatibility for OpenAI and Azure endpoints. Added some Bedrock models, now compatible with thinking models
3085585
pandas==2.3.3
gradio==5.49.1
transformers==4.56.0
spaces==0.42.1
boto3==1.40.48
pyarrow==21.0.0
openpyxl==3.1.5
markdown==3.7
tabulate==0.9.0
lxml==5.3.0
google-genai==1.33.0
openai==2.2.0
html5lib==1.1
beautifulsoup4==4.12.3
rapidfuzz==3.13.0
python-dotenv==1.1.0
torch==2.8.0 --extra-index-url https://download.pytorch.org/whl/cpu
llama-cpp-python==0.3.16 -C cmake.args="-DGGML_BLAS=ON;-DGGML_BLAS_VENDOR=OpenBLAS"
# Direct wheel links if above doesn't work
# I have created CPU Linux, Python 3.11 compatible wheels:
# https://github.com/seanpedrick-case/llama-cpp-python-whl-builder/releases/download/v0.1.0/llama_cpp_python-0.3.16-cp311-cp311-linux_x86_64.whl
# Windows, Python 3.11 compatible CPU wheels available:
# https://github.com/seanpedrick-case/llama-cpp-python-whl-builder/releases/download/v0.1.0/llama_cpp_python-0.3.16-cp311-cp311-win_amd64_cpu_openblas.whl
# If above doesn't work for Windows, try looking at'windows_install_llama-cpp-python.txt' for instructions on how to build from source