See various READMEs:
Build the llm-tool
scheme in Xcode.
To run this in Xcode simply press cmd-opt-r to set the scheme arguments. For example:
--model mlx-community/Mistral-7B-v0.1-hf-4bit-mlx
--prompt "swift programming language"
--max-tokens 50
Then cmd-r to run.
Note: you may be prompted for access to your Documents directory -- this is where the Hugging Face HubApi stores the downloaded files.
The model should be a path in the Hugging Face repository, e.g.:
mlx-community/Mistral-7B-v0.1-hf-4bit-mlx
mlx-community/phi-2-hf-4bit-mlx
See LLM for more info.
llm-tool
can also be run from the command line if built from Xcode, but
the DYLD_FRAMEWORK_PATH
must be set so that the frameworks and bundles can be found:
The easiest way to do this is drag the Products/llm-tool into Terminal to get the path:
DYLD_FRAMEWORK_PATH=~/Library/Developer/Xcode/DerivedData/mlx-examples-swift-ceuohnhzsownvsbbleukxoksddja/Build/Products/Debug ~/Library/Developer/Xcode/DerivedData/mlx-examples-swift-ceuohnhzsownvsbbleukxoksddja/Build/Products/Debug/llm-tool --prompt "swift programming language"
If the program crashes with a very deep stack trace you may need to build in Release configuration. This seems to depend on the size of the model.
There are a couple options:
- build Release
- force the model evaluation to run on the main thread, e.g. using @MainActor
- build
Cmlx
with optimizations by modifyingmlx/Package.swift
and adding.unsafeFlags(["-O"]),
around line 87
Building in Release / optimizations will remove a lot of tail calls in the C++ layer. These lead to the stack overflows.
See discussion here: ml-explore#3