Output truncated when calling the OpenAI API

I had some Python code written that called the OpenAI API. For some reason, during execution of this code I noticed that the output was truncated.

This was the snippet of the Python code that I had written that called the OpenAI API:

response = await openai.ChatCompletion.acreate (
        model=model,
        messages=messages,
        temperature=0.7,
        max_tokens=200,
        n=1,
        stop=None,
    )

The truncation of the output was due to the max_tokens parameter in my OpenAI API call. The max_tokens parameter limits the number of tokens (words and punctuation marks) in the response.

  • The maximum allowed tokens for gpt-3.5-turbo is 4096.
  • The maximum allowed tokens for gpt-4o is 124,000.

I was using gpt-4o and upped my max tokens to 4000 and my issue was resolved.

It is possible to look into the stream=True option if the response was too large, which was unnecessary in my case.