I am working on enhancing the capabilities of my agent powered by Azure OpenAI in PhiData. One of the key improvements I’m aiming for is increasing the max_tokens
limit to allow for processing longer inputs and generating more comprehensive responses. Here’s how I’ve set up the Azure OpenAI model in my code:
azure_model: AzureOpenAIChat = AzureOpenAIChat(
id=os.getenv("AZURE_OPENAI_MODEL_NAME"),
api_key=os.getenv("AZURE_OPENAI_API_KEY"),
azure_endpoint=os.getenv("AZURE_OPENAI_ENDPOINT"),
azure_deployment=os.getenv("AZURE_OPENAI_DEPLOYMENT"),
)
Could anyone provide guidance or best practices on how to modify or configure the max_tokens
parameter in PhiData when using Azure OpenAI’s LLM? Any suggestions on optimizing token usage for better performance would be greatly appreciated!