Migration Guide - LiteLLM v1.0.0+
When we have breaking changes (i.e. going from 1.x.x to 2.x.x), we will document those changes here.
1.0.0​
Last Release before breaking change: 0.14.0
What changed?
- Requires openai>=1.0.0
- openai.InvalidRequestError →Â- openai.BadRequestError
- openai.ServiceUnavailableError→- openai.APIStatusError
- NEW litellm client, allow users to pass api_key- litellm.Litellm(api_key="sk-123")
 
- response objects now inherit from BaseModel(prev.OpenAIObject)
- NEW default exception - APIConnectionError(prev.APIError)
- litellm.get_max_tokens() now returns an int not a dictmax_tokens = litellm.get_max_tokens("gpt-3.5-turbo") # returns an int not a dict
 assert max_tokens==4097
- Streaming - OpenAI Chunks now return Nonefor empty stream chunks. This is how to process stream chunks with contentresponse = litellm.completion(model="gpt-3.5-turbo", messages=messages, stream=True)
 for part in response:
 print(part.choices[0].delta.content or "")
How can we communicate changes better? Tell us
- Discord
- Email (krrish@berri.ai/ishaan@berri.ai)
- Text us (+17708783106)