Skip to content

Commit

Permalink
max_file_size_mb in float
Browse files Browse the repository at this point in the history
  • Loading branch information
ishaan-jaff committed Jul 16, 2024
1 parent 38cef1c commit 0bd747e
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions litellm/types/router.py
Original file line number Diff line number Diff line change
Expand Up @@ -154,7 +154,7 @@ class GenericLiteLLMParams(BaseModel):
input_cost_per_second: Optional[float] = None
output_cost_per_second: Optional[float] = None

max_file_size_mb: Optional[int] = None
max_file_size_mb: Optional[float] = None

model_config = ConfigDict(extra="allow", arbitrary_types_allowed=True)

Expand Down Expand Up @@ -187,7 +187,7 @@ def __init__(
output_cost_per_token: Optional[float] = None,
input_cost_per_second: Optional[float] = None,
output_cost_per_second: Optional[float] = None,
max_file_size_mb: Optional[int] = None,
max_file_size_mb: Optional[float] = None,
**params,
):
args = locals()
Expand Down Expand Up @@ -248,7 +248,7 @@ def __init__(
aws_region_name: Optional[str] = None,
# OpenAI / Azure Whisper
# set a max-size of file that can be passed to litellm proxy
max_file_size_mb: Optional[int] = None,
max_file_size_mb: Optional[float] = None,
**params,
):
args = locals()
Expand Down

0 comments on commit 0bd747e

Please sign in to comment.