Description of App
Run powerful local AI on iPhone, Mac and visionOS with Noema 2.0, now with Relay, multimodal vision, richer tools, global languages and full offline privacy.
Noema brings large-language-model intelligence to all your devices, fully offline. Download lightweight models directly from Hugging Face and pair them with curated textbooks and your own PDFs. The privacy-first design means your data never leaves your device, whether you are on iPhone, Mac, or visionOS.
- Native macOS app: Run the full Noema experience on your desktop with a rebuilt interface that feels at home on macOS.
- VisionOS support: Use Noema in spatial computing environments, with windows you can place around your workspace.
- Noema Relay: Connect your iPhone to your Mac via CloudKit, with no local Wi-Fi required, so one device can host a model while another becomes the client.
- Vision support for models: Attach photographs to your prompts and use multimodal models for on-device image understanding and analysis.
- Open Textbook Library integration: Browse and import entire textbooks from OTL through the built-in Explore view; Noema indexes them locally so you can search and retrieve relevant passages on demand.
- Bring your own data: Add personal documents in PDF or EPUB formats, which are embedded and indexed on-device to power retrieval-augmented generation.
- Integrated Hugging Face search: Discover and install any quantized model from the Hugging Face hub, with one-tap installation, automatic dependency management, and real-time download progress.
- RAM check and model size helper: A built-in advisor estimates each modelβs memory footprint and shows a badge when it fits your deviceβs budget; it can also compute the maximum context length that fits in RAM.
- Triple-backend support: Run models in GGUF, MLX or Liquid AIβs Leap format, so you can choose between high-performance quantized models, Apple-optimised MLX models, and Liquid AIβs lightweight SLMs.
- Low-RAM, high-knowledge advantage: Noema shifts knowledge into compact datasets rather than bloated weights, allowing larger knowledge bases on low-memory devices.
- Advanced settings for power users: Fine-tune context length, quantization and GPU acceleration; enable tool-calling for built-in search and other functions; and customise model parameters for optimal performance.
- Built-in tool calling and RAG: Use integrated search tools and retrieval-augmented generation to query your data without hitting context limits.
- Localization upgrades: Experience Noema in 10 languages, so international teams can work in the interface that suits them best.
- Private and offline: All processing happens locally, and your conversations and files never leave the device.
Accessibility Comments
- The More Actions button lets you attach images to messages and do other stuff, but it brings up a screen with a bunch of unlabeled buttons. You can still navigate this screen though.
- The Drag button brings up a menu with two Add buttons, which is confusing.
The developer is informed of both these issues and willing to improve the app further per user feedback. Version 1.4 had serious issues like every word and even spaces in between being treated as individual elements, meaning you would have to flick right or left with one finger to navigate to the next or previous word, including the spaces between them. I'm mentioning this to clarify how my accessibility and general suggestions and the developer's cooperation have been transforming the app. Several of my suggestions are still waiting to be implemented, but version 2.0 already introduces various major fixes and improvements.