Everything GENAPA does.
Browse by area of concern, or jump to a section from the sidebar.
Find anything by describing what you need.
Three ways to search your archive -- all driven by meaning, not keywords.
Technical detail
Seeker uses a manual tool-call loop with SearchMemory chaining across iterations. Vector search powered by pgvector with HNSW indexes. Cross-field scoring with configurable per-field weighting. Results are relevancy-ranked by the LLM, not just by vector distance.
You talk naturally and GENAPA finds what you need across everything -- even when you do not remember the right terms or where you saved something.
Optional layers of understanding. You choose the depth.
Every processing layer is independently configurable. Start with the baseline and add depth when your use case calls for it.
Technical detail
Version-based change tracking per operation ensures only modified content triggers re-processing. Domain-specific extraction prompts defined in YAML. Two-stage entity disambiguation: vector similarity candidates, then LLM evaluation with domain-specific rules.
Start with fast, powerful search. Add depth when you need it. Every layer is independently configurable, and you control which AI model handles each task.
Purpose-built AI agents, each with a specific role.
Four specialized agents that handle different kinds of knowledge work -- and compose together.
Beyond the core agents, GENAPA includes a composable skills system with 19+ skills that can be loaded dynamically. Custom agent identities and personas let you tailor how agents interact with your content.
Technical detail
Tool-calling loop pattern with persistent session memory. Cross-agent memory linking preserves the chain of discovery. Sub-agent spawning with configurable recursion depth. Skill discovery and composition via semantic search.
These are specialized knowledge workers, not generic chatbots. Each handles a specific kind of knowledge work, and they compose -- Omni can use any other agent as a tool.
Your knowledge archive stays on your machine. Always.
Data sovereignty is built into the architecture, not bolted on as an afterthought.
- Your entire knowledge archive — every extracted entity, relationship, synthesis, conversation, and memory — lives in a local database on your machine. GENAPA the company never has access to it.
- The database stores metadata and descriptions -- not your actual file content. Content is retrieved on-demand from your filesystem when needed.
- Choose from 10 AI providers and connect with your own API key — GENAPA never proxies through its servers. For fully air-gapped operation, run Ollama locally with zero external network calls.
- All outputs stay on your machine and belong to you. Extracted knowledge, synthesized insights, embeddings, and search indexes are generated and stored locally. Nothing is sent back to GENAPA.
- Docker security hardening: read-only containers, capability restrictions, no-new-privileges.
- Optional usage telemetry (aggregate source counts, token usage, and version) can be sent to genapa.com for license management. No file content, knowledge, or archive data is ever included. Disable it entirely with a single configuration setting.
- Encrypted credential storage (DPAPI on Windows). API key authentication with scoped access.
- Passphrase-encrypted configuration export and import for secure portability.
Technical detail
Docker containers run with read-only filesystem, cap_drop ALL, no-new-privileges, and tmpfs for temp directories. DPAPI-encrypted LINK credentials. Fixed-time secret comparison prevents timing attacks. Scoped API keys (Admin, Mcp, Portal) restrict access per endpoint group.
Most AI tools require your data on their servers to function. GENAPA keeps your knowledge archive entirely local — only the AI provider you choose sees the content you send for processing, and you can go fully offline at any time. During extraction and chat, relevant content from your files is sent directly to your provider's API — it never passes through GENAPA's infrastructure. Regulatory alignment (GDPR, EU AI Act) is built into the architecture: no user content is stored or processed on GENAPA's infrastructure.
Built for speed and scale.
Process thousands of files without re-analyzing your entire archive every time something changes.
- Incremental processing: Version-based change tracking means only modified content is re-processed. Edit a file and only the changed parts are re-analyzed -- the rest of your archive is untouched.
- Lightweight footprint: The database stores metadata and descriptions, not file content. Content is retrieved on-demand via GENAPA Link. This keeps the database compact enough to support large-scale archives.
- Smart accumulation: File watchers batch changes with configurable debounce timers that reset when you are actively editing, preventing processing during active work sessions.
- Write coalescing: Rapid memory updates are merged to prevent filesystem thrashing.
- Connection pooling: Configurable PostgreSQL connection pooling with lifetime management.
- Distributed locking: Redis-based locking with automatic renewal for safe concurrent processing.
Technical detail
Per-operation version counters on source nodes. Content-addressed chunks via SHA256 for change detection. Accumulation timers with reset-on-edit. Redis distributed locks with configurable expiration and renewal intervals.
Process thousands of files without re-analyzing your entire archive every time something changes. GENAPA is designed to be efficient with both your time and your AI costs.
Every aspect is configurable. Any knowledge domain.
No vendor lock-in. No one-size-fits-all processing. You control exactly how your content is analyzed.
- 10 LLM providers: Azure OpenAI, OpenAI, Ollama (local), Anthropic, Google Vertex/Gemini, Amazon Bedrock, Hugging Face, Mistral, DeepSeek, xAI (Grok).
- Per-feature model assignment: Use a powerful model for synthesis and a fast, cheap model for chunking. Each processing feature can use a different provider and model.
- Domain-configurable extraction: Built-in domains for source code, technical documentation, and general content. Add new domains by writing a YAML configuration -- no code changes required.
- Any knowledge domain: Source code, research papers, legal documents, personal notes, creative projects -- customize extraction instructions to pull out what matters for your domain.
- Per-job configuration: Each processing job can have different settings, operations, and domain types.
- Spend controls: Usage tracking and configurable spend caps with warning thresholds.
Technical detail
LlmFeature enum maps each feature to an independent provider/model configuration. Domain types carry extraction guidance, disambiguation rules, worked examples, field specifications, and synth type declarations. All defined in YAML.
No vendor lock-in. No one-size-fits-all processing. You control exactly how your content is analyzed and which AI models do the work.
Connects to your existing workflow.
Your archive fits into your tools -- editors, scripts, automation -- not the other way around.
Technical detail
MCP over Streamable HTTP at /mcp. SignalR hubs for chat streaming and archive updates. LINK uses FileSystemWatcher with System.Reactive coalescing. Redis-backed job queuing with distributed locking.
GENAPA fits into your existing tools -- editors, scripts, CI/CD. You do not have to switch to the portal to use your knowledge archive.
The full GENAPA platform, available as a shell command.
The GENAPA CLI (genapa) exposes everything the portal does -- plus things it doesn't. Submit ingestion jobs from a JSON file. Run the Seeker agent from a bash script. Retrieve any memory node by key from any machine. Cross-platform native binaries ship in the release bundle. No separate install. No package manager. If you prefer the portal, see the full web interface →
--json for machine-readable, pipeline-composable output.genapa seek <query> runs the full multi-pass SEEKER discovery agent. genapa explore <question> runs ORACLE for written synthesis. Full agent intelligence, not a simplified API wrapper.genapa auth pair uses a challenge/response pairing flow that binds credentials to your deployment. Credentials are locally encrypted; all subsequent commands authenticate automatically. The --api-key option is also available for scripted or CI environments where interactive pairing isn't possible.genapa memory content <key> retrieves any memory node by GUID key from any authenticated machine -- search results, agent session records, or task context, in any environment with CLI credentials.--environment dev|test|prod routes to the right deployment with a single flag. In server-only or containerized environments where no browser is available, the CLI is the full operational surface.Technical detail
Commands: auth pair, seek, explore, search, memory content, memory relations, collections list, jobs list/get/create/cancel/stop/resume. All data commands support --json for machine-readable output. Global options: --environment, --api-url, --api-key. Auth pairing supports single-step, split begin/complete, and resume workflows. Native binaries per platform -- Windows, Linux, macOS x64, macOS ARM -- ship in the release bundle without requiring .NET SDK on the target machine.
Everything the portal does through the browser is available as a composable shell command. For teams running GENAPA in CI/CD, server environments, or automation workflows, the CLI is not a convenience -- it is the primary interface. Memory nodes are accessible from the CLI -- retrieve any saved context by key with genapa memory content <key>. See the Memory section.
Context that travels. Memory that's searchable.
GENAPA memory is not preference tracking -- it is structured context you can save at runtime, address by key, and retrieve from any session, any machine, any agent role. Every memory enters the same embedding and indexing pipeline as your documents, so Seeker surfaces relevant memories during discovery alongside actual sources. No other major AI platform makes its memory searchable in the same knowledge base. Pass a single key and unlock any amount of typed, formatted context.
genapa memory content <key>. No separate database. No external storage.Technical detail
Typed memory hierarchy: GenericMemory, SearchMemory, NodeMappingMemory, SeekerMemory, OracleMemory, WeaverMemory, OmniMemory. Every write goes to a YAML file via an atomic temp-file-then-rename operation with 500ms write coalescing. On retrieval, cache is checked first; misses load from disk and re-cache. Startup rehydration loads all non-expired files in parallel. MemoryToMemory graph relations are materialized in the knowledge store after each disk write -- memory chains are navigable as graph edges. Bulk save, retrieve, and delete in single calls.
Most AI memory systems remember facts about you. GENAPA memory carries structured work -- task packages, search results, agent session records, synthesis outputs -- across any number of sessions, agents, and machines. The key-based handoff pattern is what makes multi-agent orchestration tractable: no shared state, no context window limits, no coordination infrastructure beyond a GUID string.
See the shape of your knowledge.
Understanding your content is not just about text results -- it is about seeing how everything connects.
Technical detail
Two persistent Sigma.js v3 WebGL renderers using graphology. ForceAtlas2 physics layout. Diff-based graph updates via SignalR. Incremental layout for smooth additions without restarting the simulation.
Understanding your knowledge is not just about text results -- it is about seeing how everything connects and exploring relationships visually.
