Can I run an LLM on the device itself

Generated on 8/11/2024

1 search

Yes, you can run a large language model (LLM) on the device itself using Apple's platforms. Apple provides tools and frameworks such as Core ML to facilitate the deployment and execution of machine learning models on their devices. You can start with a model developed in PyTorch, convert it to the Core ML format using Core ML tools, and then integrate it into your app. Core ML optimizes the model for hardware-accelerated execution across the CPU, GPU, and neural engine, making it possible to run models like Whisper, Stable Diffusion, and Mistral on Apple devices.

For more details on running models on device, you can refer to the session Explore machine learning on Apple platforms (07:16).