--- myst: html_meta: title: AutoRAG - MiniMax LLM description: Use MiniMax LLM in AutoRAG. Generate answers using MiniMax M2.7 and M2.5 models. keywords: AutoRAG,RAG,LLM,generator,MiniMax,M2.7,M2.5 --- # MiniMax LLM The `minimax_llm` module integrates [MiniMax](https://www.minimax.io/) models into AutoRAG via the OpenAI-compatible API. ## Supported Models | Model | Context Window | |-------|---------------| | MiniMax-M2.7 | 1,048,576 tokens | | MiniMax-M2.7-highspeed | 1,048,576 tokens | | MiniMax-M2.5 | 1,048,576 tokens | | MiniMax-M2.5-highspeed | 204,800 tokens | ## Features ### Auto-truncate prompt Prompts that exceed the model's token limit are automatically truncated to prevent API errors. ### Temperature clamping MiniMax models accept temperature values between 0 and 1. Values above 1.0 are automatically clamped to 1.0. ### Think-tag stripping MiniMax M2.5+ models may include `...` reasoning tags in their output. These are automatically stripped from the generated text. ## **Module Parameters** - **llm**: The MiniMax model name. For example, `MiniMax-M2.7` or `MiniMax-M2.5-highspeed`. - **batch**: The batch size for API calls. Default is 16. - **truncate**: Whether to truncate input prompts to the model's max length. Default is True. - **api_key**: MiniMax API key. You can also set this to env variable `MINIMAX_API_KEY`. - And all parameters from the [OpenAI Chat Completion API](https://platform.openai.com/docs/api-reference/chat/create) (MiniMax uses an OpenAI-compatible endpoint). ## **Example config.yaml** ```yaml modules: - module_type: minimax_llm llm: [MiniMax-M2.7, MiniMax-M2.5-highspeed] temperature: [0.1, 0.5] max_tokens: 512 api_key: ${MINIMAX_API_KEY} ```