From 3bc99af2e2d0cc89d473ddba9b07baf8c8a8123e Mon Sep 17 00:00:00 2001 From: OrPaynter Date: Sun, 13 Jul 2025 17:54:34 -0500 Subject: [PATCH 1/2] Configures default provider and adds models Updates the default provider to minimax, and adds several local transformer models including Llama2, Llama3, Mistral, Falcon, and Vicuna. Also introduces minimax model configuration and modifies the ollama configuration including updated temperature and base URL. Finally, configures a chromadb vector database with default settings. --- trae_config.json | 91 +++++++++++++++++++++++++++++++++++++++++--- trae_config.json.bak | 6 +++ 2 files changed, 92 insertions(+), 5 deletions(-) create mode 100644 trae_config.json.bak diff --git a/trae_config.json b/trae_config.json index 570e2135..284ddfe8 100644 --- a/trae_config.json +++ b/trae_config.json @@ -1,5 +1,5 @@ { - "default_provider": "anthropic", + "default_provider": "minimax", "max_steps": 20, "enable_lakeview": true, "model_providers": { @@ -43,11 +43,11 @@ "max_retries": 10 }, "ollama": { - "api_key": "ollama", - "base_url": "http://localhost:11434/v1", - "model": "model_name", + "provider": "ollama", + "model": "llama3", + "base_url": "http://localhost:11434", "max_tokens": 4096, - "temperature": 0.5, + "temperature": 0.9, "top_p": 1, "top_k": 0, "max_retries": 10 @@ -70,10 +70,91 @@ "temperature": 0.5, "top_p": 1, "max_retries": 20 + }, + "local_transformers": { + "provider": "local_transformers", + "model": "TheBloke/Llama-2-7B-GGUF", + "model_path": "./models/llama-2-7b.gguf", + "framework": "transformers", + "max_tokens": 4096, + "temperature": 0.5, + "top_p": 1, + "top_k": 0, + "max_retries": 10 + }, + "minimax": { + "provider": "minimax", + "api_key": "your_minimax_api_key", + "base_url": "https://api.minimax.chat/v1", + "model": "abab5.5-chat", + "max_tokens": 4096, + "temperature": 0.5, + "top_p": 1, + "max_retries": 10 + }, + "llama2": { + "provider": "local_transformers", + "model": "meta-llama/Llama-2-7b-hf", + "model_path": "./models/llama-2-7b-hf", + "framework": "transformers", + "max_tokens": 4096, + "temperature": 0.5, + "top_p": 1, + "top_k": 0, + "max_retries": 10 + }, + "llama3": { + "provider": "local_transformers", + "model": "meta-llama/Meta-Llama-3-8B", + "model_path": "./models/llama-3-8b", + "framework": "transformers", + "max_tokens": 8192, + "temperature": 0.5, + "top_p": 1, + "top_k": 0, + "max_retries": 10 + }, + "mistral": { + "provider": "local_transformers", + "model": "mistralai/Mistral-7B-v0.1", + "model_path": "./models/mistral-7b-v0.1", + "framework": "transformers", + "max_tokens": 8192, + "temperature": 0.5, + "top_p": 1, + "top_k": 0, + "max_retries": 10 + }, + "falcon": { + "provider": "local_transformers", + "model": "tiiuae/falcon-7b", + "model_path": "./models/falcon-7b", + "framework": "transformers", + "max_tokens": 8192, + "temperature": 0.5, + "top_p": 1, + "top_k": 0, + "max_retries": 10 + }, + "vicuna": { + "provider": "local_transformers", + "model": "lmsys/vicuna-7b-v1.5", + "model_path": "./models/vicuna-7b-v1.5", + "framework": "transformers", + "max_tokens": 4096, + "temperature": 0.5, + "top_p": 1, + "top_k": 0, + "max_retries": 10 } }, "lakeview_config": { "model_provider": "anthropic", "model_name": "claude-sonnet-4-20250514" + }, + "vector_database": { + "type": "chromadb", + "path": "./vector_db/", + "collection": "default" } } diff --git a/trae_config.json.bak b/trae_config.json.bak new file mode 100644 index 00000000..a7314189 --- /dev/null +++ b/trae_config.json.bak @@ -0,0 +1,6 @@ +{ + "default_provider": "llama3", + "max_steps": 20, + "enable_lakeview": true, + "model_providers": { + "openai": { From cf62e6d66797268a9a1092dcefbbff9bc46afe72 Mon Sep 17 00:00:00 2001 From: OrPaynter Date: Wed, 16 Jul 2025 17:21:11 -0500 Subject: [PATCH 2/2] add vscode pytest configuration settings --- .vscode/settings.json | 7 +++++++ 1 file changed, 7 insertions(+) create mode 100644 .vscode/settings.json diff --git a/.vscode/settings.json b/.vscode/settings.json new file mode 100644 index 00000000..3e99ede3 --- /dev/null +++ b/.vscode/settings.json @@ -0,0 +1,7 @@ +{ + "python.testing.pytestArgs": [ + "." + ], + "python.testing.unittestEnabled": false, + "python.testing.pytestEnabled": true +} \ No newline at end of file