I have noticed that my deployment of gpt-35-turbo on "Azure AI Studio" is not giving consistent responses to my chat completion prompts even when I set the temperature to 0. The longer the prompt, the more inconsistency I see.
I thought the idea with setting temperature to 0 meant consistent (deterministic) responses (given the same model). Is that not the case?