llm_topic_modelling / tools /helper_functions.py

Commit History

Added model compatibility for OpenAI and Azure endpoints. Added some Bedrock models, now compatible with thinking models
3085585

seanpedrickcase commited on

Added deduplication with LLM functionality. Minor package updates. Updated installation documentation.
6f3d42c

seanpedrickcase commited on

Added stop strings, optimised llama-cpp-python inference for streaming
6eaced0

seanpedrickcase commited on

Added possibility of adding examples quickly to the input files
8c54223

seanpedrickcase commited on

Allowed possibility to run all analysis steps in one click
2e33e29

seanpedrickcase commited on

Enhanced app functionality by adding new logging variables, refining file input options, and updating prompts for better user experience. Updated Dockerfile for improved environment setup and adjusted requirements for compatibility. Removed unnecessary print statements and added error handling in data loading functions.
714810a

seanpedrickcase commited on

Minor fixes for Gemini, model calls. Updated Dockerfile for non-GPU systems
8ec0f3d

seanpedrickcase commited on

GPT-OSS 20b should now work correctly
12c4a40

Sonnyjim commited on

Added cost code logging functionality. More customised logging output, can now save logs to DynamoDB.
9bd035b

seanpedrickcase commited on

Allowed for save to xlsx file. Some package updates
11004c5

seanpedrickcase commited on

You can now create structured summaries by group by changing a setting under 'I have my own list of topics'
4753a16

seanpedrickcase commited on

Improved on implementation of group-based analysis. Now should be possible all the way to summarisation
f2d85f1

seanpedrickcase commited on

Wrapper function added to add possibility of analysis by group. Some bug fixes, package changes
1937645

seanpedrickcase commited on

Downgraded Hugging Face space llama-cpp-python requirements. Updated to new Gemini SDK. File output and overall summarisation fixes.
ce10c9a

seanpedrickcase commited on

Fixed some issues with output folder references. Minor changes to column names, summary output
ad90f21

seanpedrickcase commited on

Code reorganisation to better use config files. Adapted code to use Gemma 3 as local model. Minor package updates
0c0a08a

seanpedrickcase commited on

Package updates. Can now ask the model to only assign specified topics
99b54b3

seanpedrickcase commited on

Upgraded Gradio. More resilient to cases where LLM calls do not return valid markdown tables (will reattempt with different temperature). Minor fixes
b9301bd

seanpedrickcase commited on

Allowed for manual modification of output topic table. Did some fixes to deduplication and Excel file input. Allowed for General topic specification in zero shot topics
75d1651

seanpedrickcase commited on

Changed default requirements to CPU version of llama cpp. Added Gemini Flash 2.0 to model list. Output files should contain only final files.
b0e08c8

seanpedrickcase commited on

Added presentation of summary table outputs
cc6683a

seanpedrickcase commited on

Added support for using local models (specifically Gemma 2b) for topic extraction and summary. Generally improved output format safeguards.
b7f4700

seanpedrickcase commited on

Added more guidance in Readme. Now wipes variables on click to create or summarise topics
f8f34c2

seanpedrickcase commited on