Hi there prabhu
Thanks for using QandA platform
try increasing max_tokens
to allow for extended replies and adjusting temperature
(e.g., 0.5) and top_p
for better control over randomness. Providing explicit instructions in your prompt, such as "Give a detailed explanation with examples," can also improve response length. Ensure that stop_sequences
aren’t unintentionally cutting off responses. If context is lacking, use Azure AI Search to retrieve relevant information before querying the model.
If this helps kindly accept the answr thanks.