diff --git a/src/konoha/api/v1/batch_tokenization.py b/src/konoha/api/v1/batch_tokenization.py index fd6755d..e164810 100644 --- a/src/konoha/api/v1/batch_tokenization.py +++ b/src/konoha/api/v1/batch_tokenization.py @@ -26,7 +26,7 @@ class TokenizeParameter(BaseModel): def generate_cache_key(params): - params = params.dict(exclude={"text", "texts"}) + params = params.model_dump(exclude={"text", "texts"}) return ".".join(f"{k}-{v}" for k, v in params.items()) diff --git a/src/konoha/api/v1/tokenization.py b/src/konoha/api/v1/tokenization.py index f3e0a2b..55cefb4 100644 --- a/src/konoha/api/v1/tokenization.py +++ b/src/konoha/api/v1/tokenization.py @@ -26,7 +26,7 @@ class TokenizeParameter(BaseModel): def generate_cache_key(params): - params = params.dict(exclude={"text", "texts"}) + params = params.model_dump(exclude={"text", "texts"}) return ".".join(f"{k}-{v}" for k, v in params.items())