DEV Community

Cover image for One Open Source Project a Day (No. 65): OpenHuman - A Local-First Personal AI Super Intelligence That Actually Knows You
WonderLab
WonderLab

Posted on

One Open Source Project a Day (No. 65): OpenHuman - A Local-First Personal AI Super Intelligence That Actually Knows You

Introduction

"Private, Simple and extremely powerful."

This is the No.65 article in the "One Open Source Project a Day" series. Today, we are exploring OpenHuman.

Most AI assistants share a fundamental limitation: they have no memory. Every conversation starts from zero. They don't know what project you're working on, what's in your Gmail inbox, or what happened in your GitHub repository last week.

OpenHuman exists to solve exactly that. Its goal is not to build a better chatbot, but to create an AI super intelligence that truly integrates into your daily life—pulling fresh data from all your connected apps every 20 minutes, compressing it into a local SQLite knowledge tree, giving the AI access to your complete, up-to-date work context at all times. 5.6k Stars, built in Rust + Tauri, GPL-3.0 licensed—early Beta, but already presenting a distinctly original technical direction.

What You Will Learn

  • How OpenHuman's Memory Tree achieves genuine persistent memory (not just conversation history)
  • How 118+ OAuth integrations + auto-sync every 20 minutes actually works
  • How TokenJuice compression reduces LLM API costs by up to 80%
  • How Model Routing intelligently directs tasks to reasoning, fast, or vision models
  • Why the Desktop Mascot is a meaningful product design choice, not a gimmick
  • Why choosing Rust + Tauri over Electron is a significant architectural decision

Prerequisites

  • Basic familiarity with AI assistants and agents
  • Understanding of OAuth authorization at a conceptual level
  • Rust development background is helpful for appreciating the technical design but not required

Project Background

Project Introduction

OpenHuman is an open-source personal AI agent assistant, positioning itself as a "Personal AI Super Intelligence." Its core differentiation rests on three keywords:

  • Private: All workflow data is stored locally, encrypted—never uploaded to any cloud
  • Simple: From install to a working agent in just a few clicks, no terminal setup required
  • Powerful: 118+ app integrations + persistent memory + intelligent compression + multi-model routing

It is not just a chat window. It is an actively running background agent: pulling data on a schedule, continuously updating its knowledge tree, ready to provide full context whenever you need it.

Author/Team Introduction

  • Organization: tinyhumansai
  • Creator: @senamakel
  • Project Status: Early Beta, actively developed
  • Technology choice: Built in Rust (69.7%) + TypeScript (26.1%) + Tauri for the desktop app, rather than the mainstream Electron—a deliberate statement about performance and memory overhead priorities

Project Data

  • ⭐ GitHub Stars: 5,600+
  • 🍴 Forks: 459
  • 🔧 Primary Languages: Rust 69.7% + TypeScript 26.1%
  • 📄 License: GPL-3.0
  • 🖥️ Supported Platforms: macOS, Linux (x64), Windows
  • 🌐 Repository: tinyhumansai/openhuman
  • 🔗 Website: tinyhumans.ai/openhuman

Main Features

Core Utility

OpenHuman's essence: an AI agent that actively perceives your work context, rather than a chatbot that passively waits to be asked.

The fundamental difference from traditional AI assistants:

Traditional AI assistant:
  User asks → AI answers from training data → Conversation ends (memory resets)

OpenHuman:
  Every 20 minutes in background:
    Pull latest data from Gmail / GitHub / Notion / Slack / ...
        ↓
  Memory Tree: Compress and archive into local knowledge tree (SQLite + Obsidian Vault)
        ↓
  User asks: AI answers based on your complete, current work context
        ↓
  Conversation ends: Context is preserved — picks up next time
Enter fullscreen mode Exit fullscreen mode

Use Cases

  1. Cross-Application Project Context

    • "Summarize the progress on this GitHub PR and compare it with the latest comments on the related Linear issue"—the AI has already pulled both, and answers directly.
  2. Email and Task Correlation

    • "Did I receive any emails about this project today?" The AI scans the synced Gmail data and delivers a summary and list of important messages.
  3. Meeting Assistant

    • The desktop mascot joins Google Meet as a participant, recording discussion in real time and surfacing relevant background information.
  4. Code and Documentation Q&A

    • Answer questions about code logic, historical changes, and PR comments based on synced GitHub repository data.
  5. Obsidian Knowledge Base Enhancement

    • All synced data is simultaneously written to an Obsidian-compatible Vault, allowing users to browse and edit the AI-maintained knowledge in a familiar note-taking interface.

Quick Start

Method 1: Download the Installer (Recommended)

# macOS / Linux (one-command install script)
curl -fsSL https://raw.githubusercontent.com/tinyhumansai/openhuman/main/install.sh | bash

# Windows (PowerShell)
irm https://raw.githubusercontent.com/tinyhumansai/openhuman/main/install.ps1 | iex

# Or download directly from the website:
# https://tinyhumans.ai/openhuman
Enter fullscreen mode Exit fullscreen mode

After installation, complete the UI wizard:

  1. Choose an AI model provider (OpenAI / Anthropic / local Ollama)
  2. Add OAuth integrations (pick what you need from the 118 available apps)
  3. Start using

Method 2: Developer Source Build

# Requirements:
# - Node.js 24+
# - pnpm 10.10.0
# - Rust 1.93.0 (with rustfmt + clippy)
# - CMake

git clone https://github.com/tinyhumansai/openhuman.git
cd openhuman

pnpm install

# Development mode
pnpm tauri dev

# Production build
pnpm tauri build
Enter fullscreen mode Exit fullscreen mode

Run Fully Offline with Local Ollama:

# Install and start Ollama
ollama serve
ollama pull llama3.2  # or any other model

# In OpenHuman settings, select "Ollama" as model provider
# Point to local endpoint: http://localhost:11434
Enter fullscreen mode Exit fullscreen mode

Core Characteristics

1. Memory Tree — The Technical Engine of Persistent Memory

This is OpenHuman's most central technical innovation. It doesn't just save conversation history—it builds a genuine knowledge tree:

Raw data (Gmail messages / GitHub PRs / Notion pages / Slack messages / ...)
        ↓
Content normalization (HTML → Markdown, URL shortening, remove non-ASCII)
        ↓
Chunking (each chunk ≤ 3k tokens)
        ↓
Importance scoring (based on recency, relevance, frequency)
        ↓
Hierarchical summary tree (parent node = summary of child summaries)
        ↓
Dual write:
    → SQLite local database (for AI queries)
    → Obsidian-compatible Vault (for user browsing)
Enter fullscreen mode Exit fullscreen mode

The advantage of a hierarchical summary tree: when the AI needs to answer "what's the overall status of project X?", it reads the high-level summary node directly; when it needs specifics, it drills down into concrete data chunks. This is more structured than simple vector retrieval—it mirrors how human memory actually organizes information.

2. 118+ OAuth Integrations + Auto-Sync

Covers the full ecosystem of mainstream productivity tools:

Category Representative Apps
Email / Messaging Gmail, Outlook, Slack
Project Management Notion, Linear, Jira, Asana, Trello
Code Hosting GitHub, GitLab, Bitbucket
Docs / Files Google Drive, Dropbox, OneDrive, Confluence
Calendar / Meetings Google Calendar, Outlook Calendar
CRM / Finance Stripe, HubSpot, Salesforce
Other Airtable, Figma, Zapier, Webhooks...

The 20-minute auto-sync means: you don't need to manually "tell" the AI what happened—it goes and fetches it itself.

3. TokenJuice — LLM Cost Compression Technology

TokenJuice is an internal module that compresses all content before it reaches the LLM:

Raw tool output / web scrape / API response
        ↓
TokenJuice processing pipeline:
  1. HTML → plain Markdown (strip all HTML tags)
  2. URL shortening (replace long URLs with short identifiers)
  3. Remove non-ASCII characters (emoji, special symbols)
  4. Deduplicate redundant content (nav bars, footers, ads)
  5. Extract key information (headings, body text, metadata)
        ↓
  Result: cost and latency reduced by up to 80%
Enter fullscreen mode Exit fullscreen mode

For an agent that calls LLMs frequently, this compression layer matters enormously. Monthly API costs can easily be cut in half or more.

4. Intelligent Model Routing

Different tasks fit different models. OpenHuman routes automatically:

Task Type Route Target Reason
Complex reasoning (code analysis, architecture design) Reasoning models (o3, Claude Opus) Accuracy first
Simple lookups (find data, format conversion) Fast models (GPT-4o-mini, Haiku) Cost and speed first
Image/screenshot analysis Vision models (GPT-4V, Claude Vision) Multimodal requirements
Fully offline scenarios Local Ollama model Privacy first

5. Desktop Mascot

This is not a pure UI gimmick—it is a functional background agent interface:

  • Meeting participation: Joins Google Meet as a participant, recording discussion in real time
  • Background processing: Continuously runs scheduled sync tasks while you work
  • Proactive reminders: Surfaces upcoming deadlines based on calendar and task data
  • Personalized interaction: Has personality and memory—not a stateless "help bot"

6. Local-First Privacy Architecture

All workflow data → Local SQLite (AES encrypted)
AI inference → Optional local Ollama (fully offline)
OAuth tokens → Locally encrypted, never routed through OpenHuman servers
Third-party data → Lives only on your device
Enter fullscreen mode Exit fullscreen mode

This is fundamentally different from most AI assistants that send your data to the cloud for indexing.

Project Advantages

Feature OpenHuman Notion AI / Copilot ChatGPT / Claude.ai Mem.ai
Persistent Memory ✅ Memory Tree Platform content only ❌ Resets each conversation ✅ But cloud storage
Cross-App Integration ✅ 118+ OAuth apps Limited Limited
Local / Privacy ✅ Local SQLite, encrypted ❌ Cloud ❌ Cloud ❌ Cloud
Auto-Sync ✅ Every 20 minutes None None Yes
Open Source ✅ GPL-3.0
Native Desktop ✅ Rust + Tauri Web plugin Web Web
Local AI Models ✅ Ollama support

Detailed Analysis

1. Why Rust + Tauri Instead of Electron?

This is one of OpenHuman's most deliberate architectural decisions:

The problem with Electron:

  • Every Electron app bundles a full Chromium engine
  • Baseline memory usage is typically 200–500MB
  • Higher CPU overhead, noticeable battery drain when running in the background

The advantages of Tauri + Rust:

  • Tauri uses the system's native WebView (WKWebView on macOS, WebView2 on Windows)
  • Core logic written in Rust: memory-safe, zero-cost abstractions, extremely low memory footprint (typically < 50MB)
  • Much smaller builds: a Tauri app is typically 3–10MB vs 100MB+ for Electron

For an application that needs to run constantly in the background and execute sync tasks every 20 minutes, this architectural choice directly determines user experience. OpenHuman's resource footprint feels like a native system utility, not a heavy web app.

2. Memory Tree vs Vector Retrieval: Two Memory Philosophies

Most AI tools with memory use a vector database: chunk content, vectorize it, retrieve the most similar chunks at query time. OpenHuman chose a different path—a hierarchical summary tree:

Vector retrieval approach:
  Query: "What's the current status of project X?"
  → Vector search finds Top-K similar chunks (from various times and perspectives)
  → Concatenate as context → LLM answers
  Problem: Fragmented, lacks a holistic view

Memory Tree approach:
  Query: "What's the current status of project X?"
  → Directly read the high-level summary node for "project X"
  → Drill down into sub-nodes for details when needed
  → Answer has hierarchy and coherence
Enter fullscreen mode Exit fullscreen mode

Both approaches have their place. OpenHuman's choice is better suited for questions like "understand the overall state of a long-running project"—which is precisely the use case it targets.


Project Links & Resources

Official Resources

Target Audience

  • Knowledge workers: Using multiple SaaS tools simultaneously (Gmail + Notion + GitHub + Slack) who need AI to understand context across all of them
  • Solo developers / one-person companies: Managing projects, code, and email alone—who want an AI assistant that genuinely understands their project state
  • Privacy-conscious users: Who don't want work data uploaded to an AI company's cloud
  • AI tool researchers: Interested in the architectural design of local-first AI assistants

Summary

Key Takeaways

  1. Memory Tree: Hierarchical summary tree + local SQLite storage—genuine cross-session persistent memory
  2. 118+ OAuth + auto-sync every 20 min: The AI actively perceives your work context instead of waiting to be told
  3. TokenJuice: Intelligent pre-LLM compression, up to 80% cost reduction
  4. Rust + Tauri: Native desktop architecture, minimal background resource usage
  5. Local-first privacy: All data encrypted on-device, with support for fully offline Ollama local models

One-Line Review

OpenHuman is tackling the hardest problem in AI assistant design: making the AI genuinely know you—not because you told it, but because it actively observes your work world.


Find more useful knowledge and interesting products on my Homepage

Top comments (0)