diff --git a/llama_cpp/llama_chat_format.py b/llama_cpp/llama_chat_format.py index f738ab9bb..fce66ef9e 100644 --- a/llama_cpp/llama_chat_format.py +++ b/llama_cpp/llama_chat_format.py @@ -3171,7 +3171,7 @@ def from_pretrained( class ObsidianChatHandler(Llava15ChatHandler): # Prompt Format - # The model followed ChatML format. However, with ### as the seperator + # The model followed ChatML format. However, with ### as the separator # <|im_start|>user # What is this sign about?\n diff --git a/llama_cpp/server/settings.py b/llama_cpp/server/settings.py index 13c951241..0cf98c343 100644 --- a/llama_cpp/server/settings.py +++ b/llama_cpp/server/settings.py @@ -60,7 +60,7 @@ class ModelSettings(BaseSettings): ) rpc_servers: Optional[str] = Field( default=None, - description="comma seperated list of rpc servers for offloading", + description="comma separated list of rpc servers for offloading", ) # Context Params seed: int = Field(