From fdbb9fd9c3ab40838b982275a5d55533c4b60d77 Mon Sep 17 00:00:00 2001 From: hkr04 Date: Sun, 13 Apr 2025 00:07:59 +0800 Subject: [PATCH] remove personal configs --- config/config.toml | 15 -------------- config/config_embd.toml | 17 ---------------- config/config_llm.toml | 45 ----------------------------------------- config/config_mcp.toml | 23 --------------------- config/config_mem.toml | 19 ----------------- config/config_vec.toml | 8 -------- 6 files changed, 127 deletions(-) delete mode 100644 config/config.toml delete mode 100644 config/config_embd.toml delete mode 100644 config/config_llm.toml delete mode 100644 config/config_mcp.toml delete mode 100644 config/config_mem.toml delete mode 100644 config/config_vec.toml diff --git a/config/config.toml b/config/config.toml deleted file mode 100644 index 4ebd83e..0000000 --- a/config/config.toml +++ /dev/null @@ -1,15 +0,0 @@ -[humanus_cli] -llm = "qwen-max-latest" # Key in config_llm.toml -memory = "long-context" # Key in config_mem.toml -tools = ["filesystem", "playwright", "image_loader"] # Builtin tools configuration -mcp_servers = ["python_execute"] # Key in config_mcp.toml, all MCP tools provided by servers will be added -max_steps = 30 # Maximum automatic steps without user's check -duplicate_threshold = 2 # Used to detect repeating condition (will be checked by LCS) - -[humanus_plan] -llm = "deepseek-chat" -memory = "long-context" -tools = ["filesystem", "playwright", "image_loader"] -mcp_servers = ["python_execute"] -max_steps = 30 -duplicate_threshold = 2 \ No newline at end of file diff --git a/config/config_embd.toml b/config/config_embd.toml deleted file mode 100644 index dc4b8bc..0000000 --- a/config/config_embd.toml +++ /dev/null @@ -1,17 +0,0 @@ -["nomic-embed-text-v1.5"] -provider = "oai" # Only support OAI-Compatible style for now -base_url = "http://localhost:8080" # Base url. Note: Don't add any endpoint behind -endpoint = "/v1/embeddings" # Endpoint of embeddings -model = "nomic-embed-text-v1.5.f16.gguf" # Model name -api_key = "" # Your API Key -embeddings_dim = 768 # Dimension of embeddings (refer to API docs) -max_retries = 3 # Maximum retry count - -[qwen-text-embedding-v3] -provider = "oai" -base_url = "https://dashscope.aliyuncs.com" -endpoint = "/compatible-mode/v1/embeddings" -model = "text-embedding-v3" -api_key = "sk-cb1bb2a240d84182bb93f6dd0fe03600" -embeddings_dim = 1024 -max_retries = 3 \ No newline at end of file diff --git a/config/config_llm.toml b/config/config_llm.toml deleted file mode 100644 index d58a2e6..0000000 --- a/config/config_llm.toml +++ /dev/null @@ -1,45 +0,0 @@ -[qwen-max] -model = "qwen-max" # Model name -base_url = "https://dashscope.aliyuncs.com" # Base url. Note: Don't add any endpoint behind -endpoint = "/compatible-mode/v1/chat/completions" # Endpoint of chat completions -api_key = "sk-cb1bb2a240d84182bb93f6dd0fe03600" # Your API Key - -[qwen-max-latest] -model = "qwen-max-latest" -base_url = "https://dashscope.aliyuncs.com" -endpoint = "/compatible-mode/v1/chat/completions" -api_key = "sk-cb1bb2a240d84182bb93f6dd0fe03600" - -[qwen-vl-max-latest] -model = "qwen-vl-max-latest" -base_url = "https://dashscope.aliyuncs.com" -endpoint = "/compatible-mode/v1/chat/completions" -api_key = "sk-cb1bb2a240d84182bb93f6dd0fe03600" -enable_vision = true # This means the model could accept content item like {"image_url", {"url", "xxx"}} - -["claude-3.5-sonnet"] -model = "anthropic/claude-3.5-sonnet" -base_url = "https://openrouter.ai" -endpoint = "/api/v1/chat/completions" -api_key = "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" -enable_vision = true - -["claude-3.7-sonnet"] -model = "anthropic/claude-3.7-sonnet" -base_url = "https://openrouter.ai" -endpoint = "/api/v1/chat/completions" -api_key = "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" -enable_vision = true - -[deepseek-chat] -model = "deepseek-chat" -base_url = "https://api.deepseek.com" -endpoint = "/v1/chat/completions" -api_key = "sk-93c5bfcb920c4a8aa345791d429b8536" - -[deepseek-r1] -model = "deepseek-reasoner" -base_url = "https://api.deepseek.com" -endpoint = "/v1/chat/completions" -api_key = "sk-93c5bfcb920c4a8aa345791d429b8536" -enable_tool = false diff --git a/config/config_mcp.toml b/config/config_mcp.toml deleted file mode 100644 index 51130ec..0000000 --- a/config/config_mcp.toml +++ /dev/null @@ -1,23 +0,0 @@ -[python_execute] -type = "sse" -host = "localhost" -port = 8895 -sse_endpoint = "/sse" -message_enpoint = "/message" - -[puppeteer] -type = "stdio" -command = "npx" -args = ["-y", "@modelcontextprotocol/server-puppeteer"] - -[playwright] -type = "stdio" -command = "npx" -args = ["-y", "@executeautomation/playwright-mcp-server"] - -[filesystem] -type = "stdio" -command = "npx" -args = ["-y", - "@modelcontextprotocol/server-filesystem", - "/Users/hyde/Desktop"] # Allowed paths \ No newline at end of file diff --git a/config/config_mem.toml b/config/config_mem.toml deleted file mode 100644 index f4e6513..0000000 --- a/config/config_mem.toml +++ /dev/null @@ -1,19 +0,0 @@ -[default] -max_messages = 16 # Maximum number of messages in short-term memory -max_tokens_message = 32768 # Maximum number of tokens in single message -max_tokens_messages = 65536 # Maximum number of tokens in short-term memory -max_tokens_context = 131072 # Maximum number of tokens in context (used by `get_messages`) -retrieval_limit = 32 # Maximum number of results to retrive from long-term memory -embedding_model = "qwen-text-embedding-v3" # Key in config_embd.toml -vector_store = "hnswlib" # Key in config_vec.toml -llm = "qwen-max-latest" # Key in config_llm.toml - -[long-context] -max_messages = 32 -max_tokens_message = 64000 -max_tokens_messages = 128000 -max_tokens_context = 128000 -retrieval_limit = 32 -embedding_model = "qwen-text-embedding-v3" -vector_store = "hnswlib" -llm = "qwen-max-latest" \ No newline at end of file diff --git a/config/config_vec.toml b/config/config_vec.toml deleted file mode 100644 index 4b14f07..0000000 --- a/config/config_vec.toml +++ /dev/null @@ -1,8 +0,0 @@ -[hnswlib] -provider = "hnswlib" -dim = 768 # Dimension of the elements -max_elements = 100 # Maximum number of elements, should be known beforehand -M = 16 # Tightly connected with internal dimensionality of the data - # strongly affects the memory consumption -ef_construction = 200 # Controls index search speed/build speed tradeoff -metric = "L2" # Distance metric to use, can be L2 or IP \ No newline at end of file