latent arranges your images, text, code, and PDFs in 3D based on what's in them. Things that are about the same topic cluster together. Runs entirely on your Mac, nothing leaves your device.
macOS 15+ · Apple Silicon · Free tier available
You give it files. It figures out what each one is about, then places them in a 3D space accordingly.
MobileCLIP runs locally on your Mac's neural engine. Your files never touch a server. No accounts, works offline.
Drop images, text files, source code, and PDFs into the same space. The model understands all of them and places them relative to each other.
Embeddings are reduced to 3D and rendered with Metal. Rotate, zoom, pan around. Handles hundreds of items without breaking a sweat.
Drop files onto the window. They show up in the visualization as embeddings are computed. No import wizard, no project setup.
SwiftUI app. Small download, opens fast. Looks and feels like the rest of your Mac.
Files about similar things end up near each other. Useful for spotting patterns you wouldn't notice otherwise, or just getting a feel for what's in a folder.
Drag anything onto the window — images, text, code, PDFs. Mix types freely.
MobileCLIP processes each file on-device and generates a 512-dimensional vector. This takes a few seconds depending on how many files you add.
Vectors are reduced to 3D and rendered in real time. Orbit around, zoom in on clusters, click items to inspect them.
Free tier has no time limit. Pro is a one-time purchase. No subscription, because there's nothing running on our end.
An embedding is a list of numbers that represents what a file "means." A photo of a dog and the text "golden retriever" would have similar numbers, so they'd appear near each other in the visualization. The model we use (MobileCLIP) understands both images and text.
Yes. The ML model runs on your Mac. Nothing is sent anywhere. There's no server, no analytics, no telemetry. Works fully offline.
The free tier supports 300 items. Pro is unlimited, though performance depends on your hardware. We've tested with 1,500+ items without issues.
Images (JPEG, PNG, WebP), plain text, source code, and PDFs. You can mix them all in the same space. MobileCLIP handles both visual and text content.
Everything runs on your Mac, so there's no ongoing cost on our side. If we add cloud features later, that'd be separate.
macOS 15 (Sequoia) or later, on Apple Silicon. Intel Macs aren't supported. The neural engine is what makes on-device inference fast enough to actually use.