diff options
| author | S. Solomon Darnell | 2025-03-28 21:52:21 -0500 |
|---|---|---|
| committer | S. Solomon Darnell | 2025-03-28 21:52:21 -0500 |
| commit | 4a52a71956a8d46fcb7294ac71734504bb09bcc2 (patch) | |
| tree | ee3dc5af3b6313e921cd920906356f5d4febc4ed /.venv/lib/python3.12/site-packages/core/configs/full_lm_studio.toml | |
| parent | cc961e04ba734dd72309fb548a2f97d67d578813 (diff) | |
| download | gn-ai-master.tar.gz | |
Diffstat (limited to '.venv/lib/python3.12/site-packages/core/configs/full_lm_studio.toml')
| -rw-r--r-- | .venv/lib/python3.12/site-packages/core/configs/full_lm_studio.toml | 57 |
1 files changed, 57 insertions, 0 deletions
diff --git a/.venv/lib/python3.12/site-packages/core/configs/full_lm_studio.toml b/.venv/lib/python3.12/site-packages/core/configs/full_lm_studio.toml new file mode 100644 index 00000000..daae73a1 --- /dev/null +++ b/.venv/lib/python3.12/site-packages/core/configs/full_lm_studio.toml @@ -0,0 +1,57 @@ +[app] +# LLM used for internal operations, like deriving conversation names +fast_llm = "lm_studio/llama-3.2-3b-instruct" + +# LLM used for user-facing output, like RAG replies +quality_llm = "lm_studio/llama-3.2-3b-instruct" + +# LLM used for ingesting visual inputs +vlm = "lm_studio/llama3.2-vision" # TODO - Replace with viable candidate + +# LLM used for transcription +audio_lm = "lm_studio/llama-3.2-3b-instruct" # TODO - Replace with viable candidate + +[embedding] +provider = "litellm" +base_model = "lm_studio/text-embedding-nomic-embed-text-v1.5" +base_dimension = nan +batch_size = 128 +add_title_as_prefix = true +concurrent_request_limit = 2 + +[completion_embedding] +# Generally this should be the same as the embedding config, but advanced users may want to run with a different provider to reduce latency +provider = "litellm" +base_model = "lm_studio/text-embedding-nomic-embed-text-v1.5" +base_dimension = nan +batch_size = 128 +add_title_as_prefix = true +concurrent_request_limit = 2 + +[agent] +tools = ["search_file_knowledge"] + +[completion] +provider = "litellm" +concurrent_request_limit = 1 + + [completion.generation_config] + temperature = 0.1 + top_p = 1 + max_tokens_to_sample = 1_024 + stream = false + +[ingestion] +provider = "unstructured_local" +strategy = "auto" +chunking_strategy = "by_title" +new_after_n_chars = 512 +max_characters = 1_024 +combine_under_n_chars = 128 +overlap = 20 +chunks_for_document_summary = 16 +document_summary_model = "lm_studio/llama-3.2-3b-instruct" +automatic_extraction = false + +[orchestration] +provider = "hatchet" |
