StandardOpenAILLMService¶
Description¶
A Controller Service that provides integration with OpenAI’s Chat Completion API. Supports configurable parameters including model selection, temperature, top_p, max tokens, and retry behavior. Handles API authentication, request retries with exponential backoff, and error handling.
Tags¶
ai, chat completion, chatgpt, large language model, llm, openai, openflow
Properties¶
In the list below required Properties are shown with an asterisk (*). Other properties are considered optional. The table also indicates any default values, and whether a property supports the NiFi Expression Language.
| Display Name | API Name | Default Value | Allowable Values | Description |
|---|---|---|---|---|
| Backoff Base Delay (ms) * | Backoff Base Delay (ms) | 1000 | The base delay in milliseconds for exponential backoff between retries | |
| Max Response Tokens | Max Response Tokens | The maximum number of tokens to generate in the response. | ||
| Max Retries * | Max Retries | 3 | The maximum number of retry attempts for API calls | |
| Model Name * | Model Name | gpt-4o-mini | The name of the OpenAI model. | |
| OpenAI API Key * | OpenAI API Key | The API Key for authenticating to OpenAI. | ||
| Seed | Seed | The seed to use for generating the response | ||
| Temperature | Temperature | The temperature to use for generating the response. | ||
| Top P | Top P | The top_p value for nucleus sampling. It controls the diversity of the generated responses. | ||
| User | User | Your end user, sent to OpenAI for monitoring and detection of abuse | ||
| Web Client Service * | Web Client Service | The Web Client Service to use for communicating with the LLM provider. |
State management¶
This component does not store state.
Restricted¶
This component is not restricted.
System Resource Considerations¶
This component does not specify system resource considerations.