|
# This file was autogenerated by uv via the following command: |
|
# uv pip compile pyproject.toml -o requirements.txt |
|
accelerate==1.7.0 |
|
# via |
|
# flagembedding |
|
# peft |
|
# transformers |
|
aiofiles==24.1.0 |
|
# via gradio |
|
aiohappyeyeballs==2.6.1 |
|
# via aiohttp |
|
aiohttp==3.12.11 |
|
# via |
|
# fsspec |
|
# huggingface-hub |
|
# llama-index-core |
|
aiosignal==1.3.2 |
|
# via aiohttp |
|
aiosqlite==0.21.0 |
|
# via llama-index-core |
|
annotated-types==0.7.0 |
|
# via pydantic |
|
anyio==4.9.0 |
|
# via |
|
# gradio |
|
# httpx |
|
# openai |
|
# starlette |
|
arxiv==2.2.0 |
|
# via agent-hackathon (pyproject.toml) |
|
attrs==25.3.0 |
|
# via aiohttp |
|
banks==2.1.2 |
|
# via llama-index-core |
|
beautifulsoup4==4.13.4 |
|
# via |
|
# ir-datasets |
|
# llama-index-readers-file |
|
cbor==1.0.0 |
|
# via trec-car-tools |
|
certifi==2025.4.26 |
|
# via |
|
# httpcore |
|
# httpx |
|
# llama-cloud |
|
# requests |
|
charset-normalizer==3.4.2 |
|
# via requests |
|
click==8.2.1 |
|
# via |
|
# duckduckgo-search |
|
# llama-cloud-services |
|
# nltk |
|
# typer |
|
# uvicorn |
|
colorama==0.4.6 |
|
# via griffe |
|
dataclasses-json==0.6.7 |
|
# via llama-index-core |
|
datasets==3.6.0 |
|
# via flagembedding |
|
deprecated==1.2.18 |
|
# via |
|
# banks |
|
# llama-index-core |
|
dill==0.3.8 |
|
# via |
|
# datasets |
|
# multiprocess |
|
dirtyjson==1.0.8 |
|
# via llama-index-core |
|
distro==1.9.0 |
|
# via openai |
|
duckduckgo-search==6.4.2 |
|
# via llama-index-tools-duckduckgo |
|
fastapi==0.115.12 |
|
# via gradio |
|
feedparser==6.0.11 |
|
# via arxiv |
|
ffmpy==0.6.0 |
|
# via gradio |
|
filelock==3.18.0 |
|
# via |
|
# datasets |
|
# huggingface-hub |
|
# torch |
|
# transformers |
|
filetype==1.2.0 |
|
# via llama-index-core |
|
flagembedding==1.3.5 |
|
# via agent-hackathon (pyproject.toml) |
|
frozenlist==1.6.2 |
|
# via |
|
# aiohttp |
|
# aiosignal |
|
fsspec==2025.3.0 |
|
# via |
|
# datasets |
|
# gradio-client |
|
# huggingface-hub |
|
# llama-index-core |
|
# torch |
|
gradio==5.33.1 |
|
# via agent-hackathon (pyproject.toml) |
|
gradio-client==1.10.3 |
|
# via gradio |
|
greenlet==3.2.3 |
|
# via sqlalchemy |
|
griffe==1.7.3 |
|
# via banks |
|
groovy==0.1.2 |
|
# via gradio |
|
grpcio==1.67.1 |
|
# via pymilvus |
|
h11==0.16.0 |
|
# via |
|
# httpcore |
|
# uvicorn |
|
hf-xet==1.1.3 |
|
# via huggingface-hub |
|
html2text==2025.4.15 |
|
# via llama-hub |
|
httpcore==1.0.9 |
|
# via httpx |
|
httpx==0.28.1 |
|
# via |
|
# agent-hackathon (pyproject.toml) |
|
# gradio |
|
# gradio-client |
|
# llama-cloud |
|
# llama-index-core |
|
# openai |
|
# safehttpx |
|
huggingface-hub==0.32.4 |
|
# via |
|
# agent-hackathon (pyproject.toml) |
|
# accelerate |
|
# datasets |
|
# gradio |
|
# gradio-client |
|
# llama-index-embeddings-huggingface |
|
# llama-index-embeddings-huggingface-api |
|
# llama-index-llms-huggingface-api |
|
# llama-index-utils-huggingface |
|
# peft |
|
# sentence-transformers |
|
# smolagents |
|
# tokenizers |
|
# transformers |
|
idna==3.10 |
|
# via |
|
# anyio |
|
# httpx |
|
# requests |
|
# yarl |
|
ijson==3.4.0 |
|
# via ir-datasets |
|
inquirerpy==0.3.4 |
|
# via huggingface-hub |
|
inscriptis==2.6.0 |
|
# via ir-datasets |
|
ir-datasets==0.5.10 |
|
# via flagembedding |
|
jinja2==3.1.6 |
|
# via |
|
# banks |
|
# gradio |
|
# smolagents |
|
# torch |
|
jiter==0.10.0 |
|
# via openai |
|
joblib==1.5.1 |
|
# via |
|
# nltk |
|
# scikit-learn |
|
llama-cloud==0.1.23 |
|
# via |
|
# llama-cloud-services |
|
# llama-index-indices-managed-llama-cloud |
|
llama-cloud-services==0.6.30 |
|
# via llama-parse |
|
llama-hub==0.0.79.post1 |
|
# via agent-hackathon (pyproject.toml) |
|
llama-index==0.12.41 |
|
# via llama-hub |
|
llama-index-agent-openai==0.4.9 |
|
# via |
|
# llama-index |
|
# llama-index-program-openai |
|
llama-index-cli==0.4.3 |
|
# via llama-index |
|
llama-index-core==0.12.41 |
|
# via |
|
# llama-cloud-services |
|
# llama-index |
|
# llama-index-agent-openai |
|
# llama-index-cli |
|
# llama-index-embeddings-huggingface |
|
# llama-index-embeddings-huggingface-api |
|
# llama-index-embeddings-openai |
|
# llama-index-indices-managed-llama-cloud |
|
# llama-index-llms-huggingface |
|
# llama-index-llms-huggingface-api |
|
# llama-index-llms-nebius |
|
# llama-index-llms-openai |
|
# llama-index-llms-openai-like |
|
# llama-index-multi-modal-llms-openai |
|
# llama-index-program-openai |
|
# llama-index-question-gen-openai |
|
# llama-index-readers-file |
|
# llama-index-readers-llama-parse |
|
# llama-index-tools-duckduckgo |
|
# llama-index-utils-huggingface |
|
# llama-index-vector-stores-milvus |
|
llama-index-embeddings-huggingface==0.5.4 |
|
# via agent-hackathon (pyproject.toml) |
|
llama-index-embeddings-huggingface-api==0.3.1 |
|
# via agent-hackathon (pyproject.toml) |
|
llama-index-embeddings-openai==0.3.1 |
|
# via |
|
# llama-index |
|
# llama-index-cli |
|
llama-index-indices-managed-llama-cloud==0.7.4 |
|
# via llama-index |
|
llama-index-llms-huggingface==0.5.0 |
|
# via agent-hackathon (pyproject.toml) |
|
llama-index-llms-huggingface-api==0.5.0 |
|
# via agent-hackathon (pyproject.toml) |
|
llama-index-llms-nebius==0.1.2 |
|
# via agent-hackathon (pyproject.toml) |
|
llama-index-llms-openai==0.4.4 |
|
# via |
|
# llama-index |
|
# llama-index-agent-openai |
|
# llama-index-cli |
|
# llama-index-llms-openai-like |
|
# llama-index-multi-modal-llms-openai |
|
# llama-index-program-openai |
|
# llama-index-question-gen-openai |
|
llama-index-llms-openai-like==0.4.0 |
|
# via llama-index-llms-nebius |
|
llama-index-multi-modal-llms-openai==0.5.1 |
|
# via llama-index |
|
llama-index-program-openai==0.3.2 |
|
# via |
|
# llama-index |
|
# llama-index-question-gen-openai |
|
llama-index-question-gen-openai==0.3.1 |
|
# via llama-index |
|
llama-index-readers-file==0.4.8 |
|
# via llama-index |
|
llama-index-readers-llama-parse==0.4.0 |
|
# via llama-index |
|
llama-index-tools-duckduckgo==0.3.0 |
|
# via agent-hackathon (pyproject.toml) |
|
llama-index-utils-huggingface==0.3.0 |
|
# via llama-index-embeddings-huggingface-api |
|
llama-index-vector-stores-milvus==0.8.4 |
|
# via agent-hackathon (pyproject.toml) |
|
llama-parse==0.6.30 |
|
# via llama-index-readers-llama-parse |
|
lxml==5.4.0 |
|
# via |
|
# inscriptis |
|
# ir-datasets |
|
lz4==4.4.4 |
|
# via ir-datasets |
|
markdown-it-py==3.0.0 |
|
# via rich |
|
markupsafe==3.0.2 |
|
# via |
|
# gradio |
|
# jinja2 |
|
marshmallow==3.26.1 |
|
# via dataclasses-json |
|
mdurl==0.1.2 |
|
# via markdown-it-py |
|
milvus-lite==2.4.12 |
|
# via pymilvus |
|
mpmath==1.3.0 |
|
# via sympy |
|
multidict==6.4.4 |
|
# via |
|
# aiohttp |
|
# yarl |
|
multiprocess==0.70.16 |
|
# via datasets |
|
mypy-extensions==1.1.0 |
|
# via typing-inspect |
|
nest-asyncio==1.6.0 |
|
# via llama-index-core |
|
networkx==3.4.2 |
|
# via |
|
# llama-index-core |
|
# torch |
|
nltk==3.9.1 |
|
# via |
|
# llama-index |
|
# llama-index-core |
|
numpy==2.2.6 |
|
# via |
|
# accelerate |
|
# datasets |
|
# gradio |
|
# ir-datasets |
|
# llama-index-core |
|
# pandas |
|
# peft |
|
# scikit-learn |
|
# scipy |
|
# transformers |
|
# trec-car-tools |
|
openai==1.84.0 |
|
# via |
|
# agent-hackathon (pyproject.toml) |
|
# llama-index-agent-openai |
|
# llama-index-embeddings-openai |
|
# llama-index-llms-openai |
|
orjson==3.10.18 |
|
# via gradio |
|
packaging==25.0 |
|
# via |
|
# accelerate |
|
# datasets |
|
# gradio |
|
# gradio-client |
|
# huggingface-hub |
|
# marshmallow |
|
# peft |
|
# transformers |
|
pandas==2.3.0 |
|
# via |
|
# datasets |
|
# gradio |
|
# llama-index-readers-file |
|
# pymilvus |
|
peft==0.15.2 |
|
# via flagembedding |
|
pfzy==0.3.4 |
|
# via inquirerpy |
|
pillow==11.2.1 |
|
# via |
|
# gradio |
|
# llama-index-core |
|
# sentence-transformers |
|
# smolagents |
|
pip==25.1.1 |
|
# via agent-hackathon (pyproject.toml) |
|
platformdirs==4.3.8 |
|
# via |
|
# banks |
|
# llama-cloud-services |
|
primp==0.15.0 |
|
# via duckduckgo-search |
|
prompt-toolkit==3.0.51 |
|
# via inquirerpy |
|
propcache==0.3.1 |
|
# via |
|
# aiohttp |
|
# yarl |
|
protobuf==6.31.1 |
|
# via |
|
# flagembedding |
|
# pymilvus |
|
psutil==7.0.0 |
|
# via |
|
# accelerate |
|
# llama-hub |
|
# peft |
|
pyaml==23.12.0 |
|
# via llama-hub |
|
pyarrow==20.0.0 |
|
# via |
|
# datasets |
|
# ir-datasets |
|
pydantic==2.11.5 |
|
# via |
|
# banks |
|
# fastapi |
|
# gradio |
|
# llama-cloud |
|
# llama-cloud-services |
|
# llama-index-core |
|
# openai |
|
pydantic-core==2.33.2 |
|
# via pydantic |
|
pydub==0.25.1 |
|
# via gradio |
|
pygments==2.19.1 |
|
# via rich |
|
pymilvus==2.5.10 |
|
# via llama-index-vector-stores-milvus |
|
pypdf==5.6.0 |
|
# via llama-index-readers-file |
|
pyprojroot==0.3.0 |
|
# via agent-hackathon (pyproject.toml) |
|
python-dateutil==2.9.0.post0 |
|
# via pandas |
|
python-dotenv==1.1.0 |
|
# via |
|
# agent-hackathon (pyproject.toml) |
|
# llama-cloud-services |
|
# pymilvus |
|
# smolagents |
|
python-multipart==0.0.20 |
|
# via gradio |
|
pytz==2025.2 |
|
# via pandas |
|
pyyaml==6.0.2 |
|
# via |
|
# accelerate |
|
# datasets |
|
# gradio |
|
# huggingface-hub |
|
# ir-datasets |
|
# llama-index-core |
|
# peft |
|
# pyaml |
|
# transformers |
|
regex==2024.11.6 |
|
# via |
|
# nltk |
|
# tiktoken |
|
# transformers |
|
requests==2.32.3 |
|
# via |
|
# arxiv |
|
# datasets |
|
# huggingface-hub |
|
# inscriptis |
|
# ir-datasets |
|
# llama-index-core |
|
# smolagents |
|
# tiktoken |
|
# transformers |
|
retrying==1.3.4 |
|
# via llama-hub |
|
rich==14.0.0 |
|
# via |
|
# smolagents |
|
# typer |
|
ruff==0.11.13 |
|
# via gradio |
|
safehttpx==0.1.6 |
|
# via gradio |
|
safetensors==0.5.3 |
|
# via |
|
# accelerate |
|
# peft |
|
# transformers |
|
scikit-learn==1.7.0 |
|
# via sentence-transformers |
|
scipy==1.15.3 |
|
# via |
|
# scikit-learn |
|
# sentence-transformers |
|
semantic-version==2.10.0 |
|
# via gradio |
|
sentence-transformers==4.1.0 |
|
# via |
|
# flagembedding |
|
# llama-index-embeddings-huggingface |
|
sentencepiece==0.2.0 |
|
# via flagembedding |
|
setuptools==80.9.0 |
|
# via pymilvus |
|
sgmllib3k==1.0.0 |
|
# via feedparser |
|
shellingham==1.5.4 |
|
# via typer |
|
six==1.17.0 |
|
# via |
|
# python-dateutil |
|
# retrying |
|
smolagents==1.17.0 |
|
# via agent-hackathon (pyproject.toml) |
|
sniffio==1.3.1 |
|
# via |
|
# anyio |
|
# openai |
|
soupsieve==2.7 |
|
# via beautifulsoup4 |
|
sqlalchemy==2.0.41 |
|
# via llama-index-core |
|
starlette==0.46.2 |
|
# via |
|
# fastapi |
|
# gradio |
|
striprtf==0.0.26 |
|
# via llama-index-readers-file |
|
sympy==1.13.1 |
|
# via torch |
|
tenacity==9.1.2 |
|
# via llama-index-core |
|
threadpoolctl==3.6.0 |
|
# via scikit-learn |
|
tiktoken==0.9.0 |
|
# via llama-index-core |
|
tokenizers==0.21.1 |
|
# via transformers |
|
tomlkit==0.13.3 |
|
# via gradio |
|
torch==2.6.0 |
|
# via |
|
# accelerate |
|
# flagembedding |
|
# llama-index-llms-huggingface |
|
# peft |
|
# sentence-transformers |
|
# transformers |
|
tqdm==4.67.1 |
|
# via |
|
# datasets |
|
# huggingface-hub |
|
# ir-datasets |
|
# llama-index-core |
|
# milvus-lite |
|
# nltk |
|
# openai |
|
# peft |
|
# sentence-transformers |
|
# transformers |
|
transformers==4.52.4 |
|
# via |
|
# flagembedding |
|
# llama-index-llms-huggingface |
|
# llama-index-llms-openai-like |
|
# peft |
|
# sentence-transformers |
|
trec-car-tools==2.6 |
|
# via ir-datasets |
|
typer==0.16.0 |
|
# via gradio |
|
typing-extensions==4.14.0 |
|
# via |
|
# aiosqlite |
|
# anyio |
|
# beautifulsoup4 |
|
# fastapi |
|
# gradio |
|
# gradio-client |
|
# huggingface-hub |
|
# llama-index-core |
|
# openai |
|
# pydantic |
|
# pydantic-core |
|
# pyprojroot |
|
# sentence-transformers |
|
# sqlalchemy |
|
# torch |
|
# typer |
|
# typing-inspect |
|
# typing-inspection |
|
typing-inspect==0.9.0 |
|
# via |
|
# dataclasses-json |
|
# llama-index-core |
|
typing-inspection==0.4.1 |
|
# via pydantic |
|
tzdata==2025.2 |
|
# via pandas |
|
ujson==5.10.0 |
|
# via pymilvus |
|
unlzw3==0.2.3 |
|
# via ir-datasets |
|
urllib3==2.4.0 |
|
# via requests |
|
uvicorn==0.34.3 |
|
# via gradio |
|
warc3-wet==0.2.5 |
|
# via ir-datasets |
|
warc3-wet-clueweb09==0.2.5 |
|
# via ir-datasets |
|
wcwidth==0.2.13 |
|
# via prompt-toolkit |
|
websockets==15.0.1 |
|
# via gradio-client |
|
wrapt==1.17.2 |
|
# via |
|
# deprecated |
|
# llama-index-core |
|
xxhash==3.5.0 |
|
# via datasets |
|
yarl==1.20.0 |
|
# via aiohttp |
|
zlib-state==0.1.9 |
|
# via ir-datasets |
|
|