llama_index LLM¶
The llama_index_llm
module is generator based
on llama_index. It gets the LLM instance from llama
index, and returns generated text by the input prompt.
It does not generate log probs.
Support chat prompt¶
From v0.3.19, you can use chat prompt with llama_index_llm
module.
For using chat prompt, you have to use chat_fstring
module for prompt maker.
Module Parameters¶
llm: The generator node requires setting parameters related to the Large Language Model (LLM) being used. This includes specifying the LLM provider (e.g., openai
or a list of providers like [openai, huggingfacellm]
) and the model configuration. By default, if only openai
is specified without a model, the system uses the default model set in llama_index
, which is gpt-3.5-turbo
.
Tip
Information about the LLM model can be found Supporting LLM models.
Additional Parameters:
batch: How many calls to make at once. Default is 16.
Other LLM-related parameters such as
model
,temperature
, andmax_token
can be set. These are passed as keyword arguments (kwargs
) to the LLM object, allowing for further customization of the LLM’s behavior.
Example config.yaml¶
modules:
- module_type: llama_index_llm
llm: [openai]
model: [gpt-3.5-turbo-16k, gpt-3.5-turbo-1106]
temperature: [0.5, 1.0, 1.5]