-
Dear mlx community, I am working on MAC with apple silicon, and I ran into an issue running local LLMs using the mlx server. (the same code below works on llama-cpp server) Below is the code I used: openai_client = openai.OpenAI(api_key = "placeholder-api", base_url="http://localhost:8080") response = openai_client.chat.completions.create( I confirm that mlx server opens successfully, and running below on command line works curl localhost:8080/v1/chat/completions However, when I send query via the openai API, it gives me the NotFoundError: Not Found Any help would be sincerely appreciated! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
Fix is in #877 |
Beta Was this translation helpful? Give feedback.
Fix is in #877