diff --git a/llms/mlx_lm/chat.py b/llms/mlx_lm/chat.py index 7968a868..ea1a99c7 100644 --- a/llms/mlx_lm/chat.py +++ b/llms/mlx_lm/chat.py @@ -56,7 +56,7 @@ def main(): tokenizer_config={"trust_remote_code": True}, ) - print(f"[INFO] Starting chat sessiong with {args.model}. To exit, enter 'q'.") + print(f"[INFO] Starting chat session with {args.model}. To exit, enter 'q'.") prompt_cache = make_prompt_cache(model, args.max_kv_size) while True: query = input(">> ") diff --git a/llms/mlx_lm/requirements.txt b/llms/mlx_lm/requirements.txt index 4c9874ed..48012863 100644 --- a/llms/mlx_lm/requirements.txt +++ b/llms/mlx_lm/requirements.txt @@ -1,4 +1,4 @@ -mlx>=0.19.3 +mlx>=0.19.2 numpy transformers[sentencepiece]>=4.39.3 protobuf