
In business, this becomes an on-premises AI that can be accessed remotely by authorized endpoints (you, your iPhone, your employees’ devices). The beauty of this arrangement is that whatever data you share or requests you might make are handled only by the devices and software you control.
How it might work
You might be running an open-source Llama large language model (LLM) to analyze your business documents and databases — combined with data (privately) found on the web — to give your field operatives access to up-to-the minute analysis relevant to them.
In this model, you might have a couple of high-memory Macs (even an M1 Max Mac Studio, which you can get second-hand for around $1,000) securely hosted at your offices, with access managed by your choice of secure remote access solutions and your own endpoint security profiling/MDM tools. You might use Apple’s ML framework, MLX, installing models you choose, or turn to other solutions, including Ollama.
This story originally appeared on Computerworld
