Within 24 hours of the release, community members began porting the algorithm to popular local AI libraries like MLX for Apple Silicon and llama.cpp.
Running a large language model is expensive, and a surprising amount of that cost comes down to memory, not computation.