Noema brings large‑language‑model intelligence to your iPhone and iPad, fully offline. Download lightweight models directly from Hugging Face and pair them with curated textbooks and your own PDFs. The privacy‑first design means your data never leaves your device.
- Open Textbook Library integration: Browse and import entire textbooks from OTL through the built‑in Explore view; Noema indexes them locally so you can search and retrieve relevant passages on demand.
- Bring your own data: Add personal documents in PDF or EPUB formats, which are embedded and indexed on‑device to power retrieval‑augmented generation.
- Integrated Hugging Face search: Discover and install any quantized model from the Hugging Face hub, no pre‑set list, with one‑tap installation, automatic dependency management and real‑time download progress.
- RAM check and model size helper: A built‑in advisor estimates each model’s memory footprint and shows a badge when it fits your device’s budget; it can also compute the maximum context length that fits in RAM.
- Triple‑backend support: Run models in GGUF, MLX or Liquid AI’s SLM format: a first for mobile LLM apps. This broad compatibility lets you choose between high‑performance quantized models, Apple‑optimised MLX models and Liquid AI’s lightweight SLMs.
- Low‑RAM, high‑knowledge advantage: Noema shifts knowledge into compact datasets rather than bloated weights, allowing bigger knowledge bases on low‑memory devices.
- Advanced settings for power users: Fine‑tune context length, quantization and GPU acceleration; enable tool‑calling for built‑in search and other functions; and customise model parameters for optimal performance.
- Built‑in tool calling and RAG: Use integrated search tools and retrieval‑augmented generation to query your data without hitting context limits.
- Private and offline: All processing happens locally, and your conversations and files never leave the device.