mirrored 10 minutes ago
0
HiroidAdd multiple new modules and tools to enhance the functionality and extensibility of the Maestro project (#333) * Added a **pyproject.toml** file to define project metadata and dependencies. * Added **run\_maestro.py** and **osworld\_run\_maestro.py** to provide the main execution logic. * Introduced multiple new modules, including **Evaluator**, **Controller**, **Manager**, and **Sub-Worker**, supporting task planning, state management, and data analysis. * Added a **tools module** containing utility functions and tool configurations to improve code reusability. * Updated the **README** and documentation with usage examples and module descriptions. These changes lay the foundation for expanding the Maestro project’s functionality and improving the user experience. Co-authored-by: Hiroid <guoliangxuan@deepmatrix.com>3a4b673
# Supported Model Providers and Model Lists

## LLM Model Providers

### 1. OpenAI

**Provider**

- `openai`

**Supported Models:**

- `gpt-5` Window: 400,000  Max Output Tokens: 128,000
- `gpt-5-mini` Window: 400,000  Max Output Tokens: 128,000
- `gpt-4.1-nano` Window: 400,000  Max Output Tokens: 128,000
- `gpt-4.1` Window: 1,047,576   Max Output Tokens: 32,768
- `gpt-4.1-mini`  Window: 1,047,576   Max Output Tokens: 32,768
- `gpt-4.1-nano`   Window: 1,047,576   Max Output Tokens: 32,768
- `gpt-4o`   Window: 128,000   Max Output Tokens: 16,384 
- `gpt-4o-mini`   Window: 128,000   Max Output Tokens: 16,384 
- `o1`   Window: 200,000   Max Output Tokens: 100,000 
- `o1-pro`   Window: 200,000   Max Output Tokens: 100,000 
- `o1-mini`   Window: 200,000   Max Output Tokens: 100,000 
- `o3`   Window: 200,000   Max Output Tokens: 100,000 
- `o3-pro`   Window: 200,000   Max Output Tokens: 100,000 
- `o3-mini`    Window: 200,000   Max Output Tokens: 100,000 
- `o4-mini`    Window: 200,000   Max Output Tokens: 100,000 

**Embedding Models:**

- `text-embedding-3-small`
- `text-embedding-3-large`
- `text-embedding-ada-002`

📚 **Reference Link:** <https://platform.openai.com/docs/pricing>

---

### 2. Anthropic Claude

**Provider**

- `anthropic`

**Supported Models:**

- `claude-opus-4-1-20250805`  Context window: 200K  Max output: 32000
- `claude-opus-4-20250514`   Context window: 200K  Max output: 32000
- `claude-sonnet-4-20250514`  Context window: 200K  Max output: 64000
- `claude-3-7-sonnet-20250219`   Context window: 200K  Max output: 64000
- - `claude-3-5-sonnet-20240620`   Context window: 200K  Max output: 64000
- `claude-3-5-haiku-20241022`    Context window: 200K  Max output: 8192

📚 **Reference Link:** <https://www.anthropic.com/api>

---

### 3. AWS Bedrock

**Provider**

- `bedrock`


**Supported Claude Models:**

- `Claude-Opus-4`
- `Claude-Sonnet-4`
- `Claude-Sonnet-3.7`
- `Claude-Sonnet-3.5`

📚 **Reference Link:** <https://aws.amazon.com/bedrock/>

---

### 4. Google Gemini

**Provider**

- `gemini`

**Supported Models:**

- `gemini-2.5-pro` in: 1,048,576  out: 65536
- `gemini-2.5-flash`  in: 1,048,576  out: 65536
- `gemini-2.0-flash` in: 1,048,576  out: 8192
- `gemini-1.5-pro`  in: 2,097,152  out: 8192
- `gemini-1.5-flash` in: 1,048,576  out: 8192

**Embedding Models:**

- `gemini-embedding-001`

📚 **Reference Link:** <https://ai.google.dev/gemini-api/docs/pricing>

---

### 5. Groq

**Provider**

- `groq`

**Supported Models:**

- `Kimi-K2-Instruct`
- `Llama-4-Scout-17B-16E-Instruct`
- `Llama-4-Maverick-17B-128E-Instruct`
- `Llama-Guard-4-12B`
- `DeepSeek-R1-Distill-Llama-70B`
- `Qwen3-32B`
- `Llama-3.3-70B-Instruct`

📚 **Reference Link:** <https://groq.com/pricing>

---

### 6. Monica (Proxy Platform)

**Provider**

- `monica`

**OpenAI Models:**

- `gpt-4.1`
- `gpt-4.1-mini`
- `gpt-4.1-nano`
- `gpt-4o-2024-11-20`
- `gpt-4o-mini-2024-07-18`
- `o4-mini`
- `o3`

**Anthropic Claude Models:**

- `claude-opus-4-20250514`
- `claude-sonnet-4-20250514`
- `claude-3-7-sonnet-latest`
- `claude-3-5-sonnet-20241022`
- `claude-3-5-sonnet-20240620`
- `claude-3-5-haiku-20241022`


**Google Gemini Models:**

- `gemini-2.5-pro-preview-03-25`
- `gemini-2.5-flash-lite`
- `gemini-2.5-flash-preview-05-20`
- `gemini-2.0-flash-001`
- `gemini-1.5-pro-002`
- `gemini-1.5-flash-002`

**DeepSeek Models:**

- `deepseek-reasoner`
- `deepseek-chat`

**Meta Llama Models:**

- `Llama-4-Scout-17B-16E-Instruct`   Context length: 10M tokens
- `Llama-4-Maverick-17B-128E-Instruct `   Context length: 1M tokens
- `llama-3.3-70b-instruct`
- `llama-3-70b-instruct`
- `llama-3.1-405b-instruct`

**xAI Grok Models:**

- `grok-3-beta`
- `grok-beta`

📚 **Reference Link:** <https://platform.monica.im/docs/en/models-and-pricing>

---

### 7. OpenRouter (Proxy Platform)

**Provider**

- `openrouter`

**OpenAI Models:**

- `gpt-4.1`
- `gpt-4.1-mini`
- `o1`
- `o1-pro`
- `o1-mini`
- `o3`
- `o3-pro`
- `o3-mini`
- `o4-mini`

**xAI Grok Models:**

- `grok-4`  Total Context: 256K   Max Output: 256K
- `grok-3`
- `grok-3-mini`

**Anthropic Claude Models:**

- `claude-opus-4`
- `claude-sonnet-4`

**Google Gemini Models:**

- `gemini-2.5-flash`
- `gemini-2.5-pro`

📚 **Reference Link:** <https://openrouter.ai/models>

---

### 8. Azure OpenAI

**Provider**

- `azure`


**Supported Models:**

- `gpt-4.1`
- `gpt-4.1-mini`
- `gpt-4.1-nano`
- `o1`
- `o3`
- `o4-mini`

📚 **Reference Link:** <https://azure.microsoft.com/en-us/pricing/details/cognitive-services/openai-service/>

---

### 9. Lybic AI

**Provider:**

- `lybic`

**Supported Models:**

- `gpt-5`
- `gpt-4.1`
- `gpt-4.1-mini`
- `gpt-4.1-nano`
- `gpt-4.5-preview`
- `gpt-4o`
- `gpt-4o-realtime-preview`
- `gpt-4o-mini`
- `o1`
- `o1-pro`
- `o1-mini`
- `o3`
- `o3-pro`
- `o3-mini`
- `o4-mini`

**Note:** Lybic AI provides OpenAI-compatible API endpoints with the same model names and pricing structure.

📚 **Reference Link:** <https://aigw.lybicai.com/>

---

### 10. DeepSeek

**Provider**

- `deepseek`

**Supported Models:**

- `deepseek-chat`  Context length: 128K, Output length: Default 4K, Max 8K
- `deepseek-reasoner`  Context length: 128K, Output length: Default 32K, Max 64K

📚 **Reference Link:** <https://platform.deepseek.com/>

---

### 11. Alibaba Cloud Qwen

**Supported Models:**

- `qwen-max-latest`  Context window: 32,768  Max input token length: 30,720  Max generation token length: 8,192
- `qwen-plus-latest`  Context window: 131,072  Max input token length: 98,304 (thinking)  Max generation token length: 129,024  Max output: 16,384
- `qwen-turbo-latest`  Context window: 1,000,000  Max input token length: 1,000,000  Max generation token length: 16,384
- `qwen-vl-max-latest` (Grounding)  Context window: 131,072  Max input token length: 129,024  Max generation token length: 8,192
- `qwen-vl-plus-latest` (Grounding)  Context window: 131,072  Max input token length: 129,024  Max generation token length: 8,192

**Embedding Models:**

- `text-embedding-v4`
- `text-embedding-v3`

📚 **Reference Link:** <https://bailian.console.aliyun.com/?tab=doc#/doc/?type=model&url=https%3A%2F%2Fhelp.aliyun.com%2Fdocument_detail%2F2840914.html&renderType=iframe>

---

### 12. ByteDance Doubao

**Supported Models:**

- `doubao-seed-1-6-flash-250615`  Context window: 256k  Max input token length: 224k  Max generation token length: 32k  Max thinking content token length: 32k
- `doubao-seed-1-6-thinking-250715`  Context window: 256k  Max input token length: 224k  Max generation token length: 32k  Max thinking content token length: 32k
- `doubao-seed-1-6-250615` Context window: 256k  Max input token length: 224k  Max generation token length: 32k  Max thinking content token length: 32k
- `doubao-1.5-vision-pro-250328` (Grounding)  Context window: 128k  Max input token length: 96k   Max generation token length: 16k  Max thinking content token length: 32k
- `doubao-1-5-thinking-vision-pro-250428` (Grounding)  Context window: 128k  Max input token length: 96k  Max generation token length: 16k  Max thinking content token length: 32k
- `doubao-1-5-ui-tars-250428` (Grounding)  Context window: 128k  Max input token length: 96k  Max generation token length: 16k  Max thinking content token length: 32k

**Embedding Models:**

- `doubao-embedding-large-text-250515`
- `doubao-embedding-text-240715`

📚 **Reference Link:** <https://console.volcengine.com/ark/region:ark+cn-beijing/model?vendor=Bytedance&view=LIST_VIEW>

---

### 13. Zhipu GLM

**Supported Models:**

- `GLM-4.5`  Max in: 128k Max output: 0.2K
- `GLM-4.5-X`  Max in: 128k  Max output: 0.2K
- `GLM-4.5-Air` Max in: 128k  Max output: 0.2K
- `GLM-4-Plus`
- `GLM-4-Air-250414`
- `GLM-4-AirX` (Grounding)
- `GLM-4V-Plus-0111` (Grounding)

**Embedding Models:**

- `Embedding-3`
- `Embedding-2`

📚 **Reference Link:** <https://open.bigmodel.cn/pricing>

---

### 14. SiliconFlow

**Supported Models:**

- `Kimi-K2-Instruct`   Context Length: 128K
- `DeepSeek-V3`
- `DeepSeek-R1`
- `Qwen3-32B`

📚 **Reference Link:** <https://cloud.siliconflow.cn/sft-d1pi8rbk20jc73c62gm0/models>

---

## 🔤 Dedicated Embedding Providers

### 15. Jina AI

**Embedding Models:**

- `jina-embeddings-v4`
- `jina-embeddings-v3`

📚 **Reference Link:** <https://jina.ai/embeddings>

---

## 🔍 AI Search Engines

### 16. Bocha AI

**Service Type:** AI Research & Search

📚 **Reference Link:** <https://open.bochaai.com/overview>

---

### 17. Exa

**Service Type:** AI Research & Search

**Pricing Model:**

- $5.00 / 1k agent searches
- $5.00 / 1k exa-research agent page reads
- $10.00 / 1k exa-research-pro agent page reads
- $5.00 / 1M reasoning tokens

📚 **Reference Link:** <https://dashboard.exa.ai/home>