Skip to content

girishlade111/research-assistant

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

368 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

πŸ”¬ ResAgent β€” Advanced Multi-Agent Research Orchestrator

Next.js React TypeScript Tailwind CSS NVIDIA NIM

Next-Generation Multi-Agent Research Engine
Transforming raw queries into exhaustive, structured, and fact-checked intelligence reports using a fleet of specialized AI experts.

Project Overview β€’ Key Features β€’ System Architecture β€’ Dev Stack β€’ Installation β€’ Configuration β€’ Project Stats β€’ Usage Guide β€’ Maintainer


πŸ“‹ Project Overview

ResAgent is a production-grade, multi-agent AI research system engineered for depth, accuracy, and scale. It orchestrates a fleet of specialized AI agents across a multi-phase pipeline to deliver exhaustive, citation-rich research reports in real-time.

Important

ResAgent features Dynamic Model Routing with automatic fallback to high-capacity context models (up to 131,072 tokens). A unique race-condition fallback mechanism ensures zero downtime by firing concurrent requests to OpenRouter if primary endpoints stall.


✨ Key Features

🌐 Intelligent Data Retrieval

  • Targeted Augmentation: Concurrent web searches triggered by refined research blueprints rather than raw user input.
  • Multi-Modal Intake: Seamlessly ingest and parse complex local files:
    • PDF Parsing: High-fidelity text extraction via pdfjs-dist.
    • Word Documents: Comprehensive DOCX processing via mammoth.
    • Structured Data: CSV and datasheet handling with PapaParse.
    • Image OCR: WebAssembly-powered text extraction from images via Tesseract.js.

πŸ€– Specialized Agent Fleet

The system dynamically assigns models based on task complexity and domain expertise.

Agent Purpose Primary Model (NVIDIA NIM) Fallback (OpenRouter)
Query Intelligence Refines queries & builds research plans mistral-large-3 gpt-oss-120b:free
Web Search Concurrent real-time data retrieval dracarys-70b llama-3.3-70b:free
Financial Analysis Market trends & fiscal data correlation deepseek-v3.2 gpt-oss-120b:free
Deep Reasoning Risk assessment & complex logic kimi-k2-thinking gpt-oss-120b:free
Code Generation Technical snippet & algorithm generation qwen3-coder-480b qwen3-coder:free
Summarization High-speed overview generation minimax-m2.7 glm-4.5-air:free
Report Synthesis Final markdown assembly & QC nemotron-3-super nemotron-3:free

πŸ—οΈ System Architecture

ResAgent utilizes a sophisticated Control Plane vs. Data Plane architecture to manage high-concurrency multi-agent workflows.

🧩 Granular Orchestration Workflow

flowchart TB
    %% 1. Presentation Layer
    subgraph Client [Presentation Layer - React 19]
        UI[Main Chat UI]
        State[Zustand State Manager]
        SSE_Rec[SSE Stream Consumer]
        MD_Render[Progressive MD Renderer]
    end

    %% 2. Service Gateway
    subgraph API [Service Gateway - Next.js]
        Route[POST /api/research]
        Cache[SHA-256 Cache Check]
        SSE_Push[SSE Stream Controller]
    end

    %% 3. Control Plane
    subgraph Control_Plane [Control Plane - Logic]
        QI[Query Intel Agent]
        BP[Research Blueprint]
        MS[Model Selector Agent]
        Health[NIM Health Check]
    end

    %% 4. Data Plane
    subgraph Data_Plane [Data Plane - Parallel Fleet]
        Parallel[Parallel Execution Engine]
        A1[Analysis Agent]
        A2[Coding Agent]
        A3[Fact-Check Agent]
        A4[Summary Agent]
    end

    %% 5. Grounding Layer
    subgraph Grounding [Grounding Layer - RAG]
        WS[Web Search - Perplexity]
        OCR[File Parser - WASM]
        Semantic[Semantic Scoring]
    end

    %% 6. Resilience Layer
    subgraph Resilience [Resilience - Auto-Fallback]
        Race[Fallback Race Condition]
        NIM[Primary: NVIDIA NIM]
        OR[Fallback: OpenRouter]
    end

    %% 7. Finalization Layer
    subgraph Assembly [Finalization Layer]
        RS[Report Synthesis Agent]
        Export[MD/PDF Export Engine]
    end

    %% Connections
    UI --> Route
    Route --> Cache
    Cache --> QI
    QI --> BP
    BP --> MS
    MS --> Health
    Health --> Parallel
    
    Parallel --> A1 & A2 & A3 & A4
    WS & OCR & Semantic --> A1 & A2 & A3 & A4
    
    A1 & A2 & A3 & A4 --> Race
    Race --> NIM
    Race -.-> OR
    
    NIM & OR --> RS
    RS --> SSE_Push
    SSE_Push --> SSE_Rec
    SSE_Rec --> State
    State --> MD_Render
    MD_Render --> UI

    %% Styling
    classDef client fill:#111,stroke:#38bdf8,color:#fff
    classDef api fill:#111,stroke:#818cf8,color:#fff
    classDef control fill:#111,stroke:#a5b4fc,color:#fff
    classDef data fill:#111,stroke:#34d399,color:#fff
    classDef grounding fill:#111,stroke:#fbbf24,color:#fff
    classDef resilience fill:#111,stroke:#f0abfc,color:#fff
    classDef assembly fill:#111,stroke:#f59e0b,color:#fff

    class UI,State,SSE_Rec,MD_Render client
    class Route,Cache,SSE_Push api
    class QI,BP,MS,Health control
    class Parallel,A1,A2,A3,A4 data
    class WS,OCR,Semantic grounding
    class Race,NIM,OR resilience
    class RS,Export assembly
Loading

πŸ›‘οΈ Technical Deep Dives

1. The Model Routing & Health Plane

Unlike static LLM implementations, ResAgent employs a Health-Aware Control Plane (model-selector-agent.ts):

  • Dynamic Task Classification: Every research section is classified into 8 specialized task types (e.g., web_search, financial_analysis, deep_reasoning).
  • Pre-emptive Health Checks: The system pings the NVIDIA NIM health endpoint with a 4s timeout. If latency exceeds this threshold or the service returns a non-200 status, the Control Plane automatically swaps primary assignments to OpenRouter fallbacks before execution begins.

2. Resilient Parallelization Framework

The engine leverages Node.js asynchronous primitives and Promise.allSettled to manage the agent fleet:

  • Non-Blocking Aggregation: Web searching and local document parsing run concurrently. Document parsing is handled via WebAssembly (WASM) threads, offloading CPU-intensive OCR tasks from the main event loop.
  • Graceful Degradation: Each agent is wrapped in a withGracefulTimeout wrapper. If a sub-agent stalls (150s ceiling), the system returns a partial result with a clear "Data Limitations" notice, ensuring the final report is delivered even if one "expert" fails.

3. WASM-Powered Grounding Layer (RAG)

To ensure zero hallucinations, the system uses a Semantic Blackboard Architecture (context-builder.ts):

  • Semantic Scoring: Text extracted from files and web results is chunked and scored against the research query using keyword density and relevance proximity.
  • Token Budgeting (70/30 Split): The system intelligently allocates context window space, prioritizing local file content (70% budget) over web search results (30% budget) to ensure groundedness in user-provided data.

πŸ› οΈ Development Stack

Frontend Core

  • Framework: Next.js 16.2.4 (App Router, Turbopack)
  • Library: React 19.2.4 (Concurrent Rendering)
  • Styling: Tailwind CSS v4 + tw-animate-css
  • Animations: Framer Motion 12.38.0
  • Components: shadcn/ui + Base UI 1.4.0
  • State Management: Zustand (Local UI state)

AI & Orchestration

  • Inference: NVIDIA NIM (Primary), OpenRouter (Fallback)
  • Web Search: Perplexity Sonar API
  • Data Parsing:
    • PDF: pdfjs-dist (High-fidelity extraction)
    • DOCX: Mammoth (Semantic HTML conversion)
    • CSV: PapaParse (Stream parsing)
    • Images: Tesseract.js (WASM OCR)
  • Streaming: Native Server-Sent Events (SSE) for real-time progress.

Export Engine

  • PDF Export: jspdf + jspdf-autotable
  • Visuals: html-to-image for capturing UI elements.

πŸš€ Installation & Setup

1. Prerequisites

  • Node.js 20+
  • NPM 10+
  • API Keys for NVIDIA NIM, OpenRouter, and Perplexity.

2. Clone & Install

git clone https://github.com/girishlade111/research-assistant.git
cd research-assistant
npm install

3. Configure Environment

Create a .env.local in the root and add your keys:

# Primary platform (NVIDIA NIM)
NVIDIA_API_KEY=nvapi-your-key-here

# Fallback platform (OpenRouter)
OPENROUTER_API_KEY=sk-or-your-key-here

# Web Search (Perplexity)
SONAR_API_KEY=pplx-your-key-here

4. Launch Development

npm run dev
# Open http://localhost:3000 to see the application.

βš™οΈ Configuration & Stats

Token Governance

The system uses a tiered token budgeting strategy based on task priority.

Parameter Value Description
Global Context 131,072 Maximum supported context for massive document sets.
Max Response 32,768 Total budget for the final synthesized report.
Per-Agent Cap 16,384 Individual context budget for specialized sub-agents.
Agent Timeout 150,000ms Maximum duration before a sub-agent triggers graceful failure.
Health Check 4,000ms Maximum latency allowed for primary NVIDIA NIM endpoints.

Project Metrics

  • 8+ Specialized AI Agents working in parallel.
  • 15+ State-of-the-art LLMs integrated into the model registry.
  • 3 Research Tiers: Corpus (Document-only), Deep (4 sources), Pro (8+ sources).
  • 4 File Types Supported: PDF, DOCX, CSV, Image (OCR).
  • 0.2s Stagger delay for concurrent agent launches to prevent rate limiting.
  • 100% SSE-based real-time streaming for all agent activities.

πŸ“– Usage Guide

  1. Select Research Mode:
    • Corpus: Focuses exclusively on your uploaded files.
    • Deep: Combines files with targeted web search (4 sources).
    • Pro: Exhaustive research using 8+ sources and deep reasoning agents.
  2. Toggle Specialized Agents: Customize your pipeline by enabling/disabling specific agents (e.g., enable Coding Agent for technical queries).
  3. Context Injection: Upload PDFs, DOCX, or Images to ground the AI in your local data.
  4. Real-Time Tracking: Watch the Thinking Panel as agents progress through Initialization, Research, and Synthesis.
  5. Interactive Exploration: Use the Citation Graph to visualize the connections between findings and sources.
  6. High-Fidelity Export: Download your findings as professional PDF or Markdown reports.

πŸ‘€ Maintainer

Girish Lade

Full-Stack AI Solutions Architect & UI/UX Expert

Specializing in high-performance multi-agent systems and immersive AI-driven interfaces.


Website LinkedIn GitHub Email


πŸ“„ License

Private and Proprietary. Powered by the Lade Stack ecosystem. All rights reserved Β© 2026.

About

Next-Generation Multi-Agent Research Engine Transforming raw queries into structured, verified intelligence reports.research assistant

Topics

Resources

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors