Spaces:
Running
on
Zero
Running
on
Zero
Commit History
Added topic validation and llm deduplication functionality. Now checks for input token count before LLM calls.
8af499b
Added model compatibility for OpenAI and Azure endpoints. Added some Bedrock models, now compatible with thinking models
3085585
Changed config defaults to enable successful app run without local model inference on first run
5fed34d
Added deduplication with LLM functionality. Minor package updates. Updated installation documentation.
6f3d42c
Added examples for structured summaries and groups. Adapted functions for structured summaries. Simplified front tab GUI
5ed844b
Optimised prompts. Updated Gradio. Added example for zero shot topics. Added support for Granite 4 local model
9e8c029
No longer str capitalize summaries
3ee11fd
Corrected blank KV_QUANT_LEVEL values. Removed erroneous extract topics output
7ae3b47
Corrected KV quant value when not specified (set to None)
4cd2443
Corrected KV quantisation definition in config. Moved examples to top of screen under intro.
73188ff
Generally improved inference for low vram systems, unsloth usage improvements, updated packages, switched default local model to Qwen 3 4b
bd1a015
Updated output logging files. Updated examples and readme. Minor prompt adjustments and package updates
fff212b
Corrected reasoning setting
4998b3c
Revised intro and readme. Reasoning suffix setting simplified. All in one xlsx returns correct tables.
8161b79
tracking model state through all_in_one_function
1a1a845
Corrected usage of stop strings, streaming
6154c1e
Added stop strings, optimised llama-cpp-python inference for streaming
6eaced0
Trying out inference with unsloth vs transformers
4d01a46
Added 'all-in-one' function. Corrected local model load when not loaded initially. Environment variables for max data rows and topics.
d6ff533
Added speculative decoding to transformers calls for gemma 3
7cadb40
It should now be possible to load the local model globally at the start to avoid repeated loading throughout the stages of topic extraction
fd02514
Added framework of support for Azure models (although untested)
138286c
Corrected misplaced example file reference. Added AWS Nova models to model list
5fa40a6
Minor corrections to transformers inference
6797022
Further optimisations to transformers inference
0895a36
Enabled GPU-based local model inference with the transformers package
72d517c
Added possibility of adding examples quickly to the input files
8c54223
Enhanced app functionality by adding new UI elements for summary file management, put in Bedrock model toggle, and refined logging messages. Updated Dockerfile and requirements for better compatibility and added install guide to readme. Removed deprecated code and unnecessary comments.
ba1a951
Removed unnecessary print statements
6da6ac6
Enhanced app functionality by adding new logging variables, refining file input options, and updating prompts for better user experience. Updated Dockerfile for improved environment setup and adjusted requirements for compatibility. Removed unnecessary print statements and added error handling in data loading functions.
714810a
Corrected full_prompt save
d25e491
Minor fixes for Gemini, model calls. Updated Dockerfile for non-GPU systems
8ec0f3d
Merge branch 'dev' into gpt-oss-compat
7497349
unverified
Sean Pedrick-Case
commited on