Skip to main content

Feature Guides

Comprehensive guides for all NeuroLink features organized by category. Each guide includes setup, usage patterns, configuration, and troubleshooting.


Latest Features (Q1 2026)

FeatureDescription
MCP EnhancementsAdvanced MCP features: ToolRouter, ToolCache, RequestBatcher, tool annotations, elicitation protocol, and custom MCP server creation.
PPT GenerationGenerate professional PowerPoint presentations from text prompts with 35 slide types, 5 themes, and optional AI images.
Video GenerationGenerate videos from text prompts using RunwayML (ML5, ML6 Turbo models).
Image Generation with GeminiNative image generation using Gemini 2.0 Flash Experimental with imagen-3.0-generate-002 model.
HTTP/Streamable HTTP Transport for MCPConnect to remote MCP servers via HTTP with authentication, rate limiting, retry support, and session management.
Audio InputReal-time voice conversations with Gemini Live and audio streaming capabilities.
Server AdaptersExpose NeuroLink AI agents as HTTP APIs with Hono, Express, Fastify, and Koa. Production-ready with auth, rate limiting, and streaming.
RAG Document ProcessingComprehensive document chunking (10 strategies), hybrid search (BM25 + vector), and reranking (5 types) for retrieval-augmented generation.
Context Compaction4-stage context compaction pipeline with automatic budget management, per-provider token estimation, and non-destructive message tagging.
MemoryPer-user condensed memory that persists across conversations. LLM-powered condensation with S3, Redis, or SQLite storage backends.
Claude Subscription SupportMultiple authentication methods for Claude (API key, OAuth) with support for Free, Pro, Max, and API tiers.
Client SDKType-safe HTTP, SSE, and WebSocket clients with React hooks and Vercel AI SDK adapter.
Claude ProxyMulti-account Claude proxy with OAuth pooling, rate-limit failover, token refresh, and launchd daemon for crash recovery.
Authentication ProvidersSecure AI endpoints with 11 auth providers (Auth0, Clerk, Firebase, Supabase, Cognito, Keycloak, WorkOS, Better Auth, OAuth2, JWT, Custom) with RBAC, session management, and rate limiting.

Q1 2026 Highlights:

  • Authentication Providers: Pluggable multi-provider authentication system with 11 providers, per-call token validation in generate/stream, composable middleware (auth + RBAC + rate limiting), in-memory and Redis session storage, AsyncLocalStorage context propagation, and CLI management commands
  • MCP Enhancements: Advanced MCP features including intelligent tool routing (round-robin, least-loaded, capability-based), result caching with LRU/FIFO/LFU eviction, request batching, tool annotations with safety levels, elicitation protocol for interactive input, MCPServerBase for custom servers, and multi-server management with load balancing
  • PPT Generation: Create AI-powered PowerPoint presentations with 35 slide types (title, content, charts, timelines, dashboards, composite layouts), 5 built-in themes, optional AI-generated images, and multi-provider support (Vertex, OpenAI, Anthropic, Google AI, Azure, Bedrock)
  • Video Generation: Create AI-generated videos with RunwayML integration supporting ML5 and ML6 Turbo models, customizable duration (5-10s), and watermark control
  • Gemini Image Generation: Native support for Google's imagen-3.0-generate-002 model through Gemini 2.0 Flash Experimental for high-quality image synthesis
  • Remote MCP Servers: HTTP/Streamable HTTP transport enables connecting to cloud-hosted MCP servers with Bearer token authentication, configurable rate limits, automatic retry with exponential backoff, and session management via Mcp-Session-Id header
  • Audio Input: Real-time voice conversations with Gemini Live API enabling bidirectional audio streaming for interactive voice-based AI experiences
  • Server Adapters: Deploy NeuroLink as production HTTP APIs with support for Hono (recommended), Express, Fastify, and Koa frameworks. Includes built-in authentication, rate limiting, caching, validation middleware, and SSE streaming support.
  • RAG Document Processing: Full-featured retrieval-augmented generation with 10 chunking strategies (character, recursive, sentence, token, markdown, html, json, latex, semantic, semantic-markdown), hybrid search combining BM25 and vector similarity, 5 reranking types (simple, LLM, batch, cross-encoder, Cohere), and integration with Pinecone, Weaviate, and Chroma vector stores.
  • Claude Subscription Support: Flexible authentication supporting API keys and OAuth for Claude Pro/Max subscribers, with model availability tracking and quota management

Core Features (Q4 2025)

FeatureDescription
Image GenerationGenerate images from text prompts using Gemini models via Vertex AI or Google AI Studio.
Enterprise HITLProduction-ready HITL with approval workflows, confidence thresholds, and enterprise patterns.
Interactive CLIAI development environment with loop mode, session variables, and conversation memory.
MCP Tools ShowcaseComplete guide to 6 built-in tools and 58+ external MCP servers across 6 categories.
Human-in-the-Loop (HITL)Pause AI tool execution for user approval before risky operations like file deletion or API calls.
Guardrails MiddlewareContent filtering, PII detection, and safety checks for AI outputs with zero configuration.
Redis Conversation ExportExport complete session history as JSON for analytics, debugging, and compliance auditing.
Context CompactionAutomatic conversation compression for long-running sessions to stay within token limits.
LiteLLM IntegrationAccess 100+ AI models from all major providers through unified LiteLLM routing interface.
SageMaker IntegrationDeploy and use custom-trained models on AWS SageMaker infrastructure with full control.

Core Features (Q3 2025)

FeatureDescription
Multimodal Chat ExperiencesStream text and images together with automatic provider fallbacks and format conversion.
CSV File SupportProcess CSV files for data analysis with automatic format conversion. Works with all providers.
PDF File SupportProcess PDF documents for visual analysis and content extraction. Native provider support.
Office DocumentsProcess DOCX, PPTX, XLSX files for document analysis. Native Bedrock, Vertex, Anthropic support.
Auto Evaluation EngineAutomated quality scoring and metrics export for AI response validation using LLM-as-judge.
CLI Loop SessionsPersistent interactive mode with conversation memory and session state for prompt engineering.
Regional Streaming ControlsRegion-specific model deployment and routing for compliance and latency optimization.
Provider Orchestration BrainAdaptive provider and model selection with intelligent fallbacks based on task classification.

Platform Capabilities at a Glance

CategoryFeaturesDocumentation
Provider unification14+ providers with automatic failover, cost-aware routing, provider orchestration (Q3)Provider Setup
Multimodal pipelineStream images + CSV data + PDF documents + Office files across providers with auto-detection for mixed file types.Multimodal Guide, CSV Support, PDF Support, Office Docs
Quality & governanceAuto-evaluation engine (Q3), guardrails middleware (Q4), HITL workflows (Q4), audit loggingAuto Evaluation, Guardrails, HITL
Memory & contextConversation memory, per-user memory, Mem0 integration, Redis history export (Q4), context summarization (Q4)Conversation Memory, Memory, Redis Export
CLI toolingLoop sessions (Q3), setup wizard, config validation, Redis auto-detect, JSON outputCLI Loop, CLI Commands
Enterprise opsProxy support, regional routing (Q3), telemetry hooks, configuration managementEnterprise Proxy, Observability
Tool ecosystemMCP auto discovery, LiteLLM hub access, SageMaker custom deployment, web searchMCP Integration, MCP Catalog

AI Provider Integration

NeuroLink supports 13 AI providers with unified API access:

ProviderKey FeaturesFree TierTool SupportStatusDocumentation
OpenAIGPT-4o, GPT-4o-mini, o1 modelsNoFullProductionSetup Guide
AnthropicClaude 4.6, 4.5/4.0 Sonnet, Opus, HaikuNoFullProductionSetup Guide, Subscription Guide
Google AIGemini 2.5 Flash/ProFree TierFullProductionSetup Guide
AWS BedrockClaude, Titan, Llama, NovaNoFullProductionSetup Guide
Google VertexGemini via GCPNoFullProductionSetup Guide
Azure OpenAIGPT-4, GPT-4o, o1NoFullProductionSetup Guide
LiteLLM100+ models unifiedVariesFullProductionIntegration Guide
AWS SageMakerCustom deployed modelsNoFullProductionIntegration Guide
Mistral AIMistral Large, SmallFree TierFullProductionSetup Guide
Hugging Face100,000+ modelsFreePartialProductionSetup Guide
OllamaLocal modelsFree (Local)PartialProductionSetup Guide
OpenAI CompatibleAny compatible endpointVariesFullProductionSetup Guide
OpenRouter300+ models via unified APIFree TierFullProductionSetup Guide

Provider Comparison Guide - Full feature matrix


Advanced CLI Capabilities

Interactive Setup Wizard

NeuroLink includes a revolutionary interactive setup wizard that guides users through provider configuration in 2-3 minutes:

# Launch interactive setup wizard
npx @juspay/neurolink setup

# Provider-specific guided setup
npx @juspay/neurolink setup --provider openai
npx @juspay/neurolink setup --provider bedrock

Wizard Features:

  • Secure credential collection with validation
  • Real-time authentication testing
  • Automatic .env file creation
  • Recommended model selection
  • Quick-start command examples
  • Interactive provider discovery

15+ CLI Commands

Complete command-line toolkit for every workflow:

CommandDescriptionKey Features
generate/genText generationMultimodal input, tool support, streaming
streamReal-time streamingLive token output, evaluation
loopInteractive sessionPersistent variables, conversation memory
setupGuided configurationProvider wizard, validation
statusHealth monitoringProvider health, latency checks
models listModel discoveryCapability filtering, availability
configConfiguration managementInit, validate, export, reset
memoryConversation managementExport, import, stats, clear
mcpMCP server managementList, discover, connect, status
providerProvider operationsList, test, health dashboard
ollamaOllama managementModel download, list, remove
sagemakerSageMaker operationsStatus, endpoint management
vertexVertex AI operationsAuth status, quota checks
completionShell completionBash and Zsh support
validateConfig validationEnvironment verification

Shell Integration

Bash and Zsh completions for faster command-line workflows:

# Install Bash completion
neurolink completion bash >> ~/.bashrc

# Install Zsh completion
neurolink completion zsh >> ~/.zshrc

Learn more: Complete CLI Reference


Built-in Tools & MCP Integration

8 Core Built-in Agent Tools

Complete autonomous agent foundation with security and validation:

ToolFunctionCapabilitiesSecurityStatus
getCurrentTimeTime accessDate/time with timezone supportSafeActive
readFileFile readingSecure file system access with path validationSandboxedActive
writeFileFile writingFile creation and modification with safety checksHITLActive
listFilesDirectory listingDirectory navigation and listingRestrictedActive
createDirectoryDirectory creationDirectory creation with permission checksValidatedActive
deleteFileFile deletionFile and directory deletion with confirmationHITLActive
executeCommandCommand executionSystem command execution with safety limitsHITLActive
websearchGroundingWeb searchGoogle Vertex web search integrationAPI-basedActive

Tool Management System:

  • Dynamic tool registration and validation
  • Secure execution with sandboxing
  • Result processing and error recovery
  • Tool discovery and availability tracking

Custom Tools Guide - Create your own tools


Model Context Protocol (MCP) - Enterprise-Grade Ecosystem

5 Built-in MCP Servers

NeuroLink includes 5 production-ready MCP servers for enterprise agent deployment:

ServerPurposeTools ProvidedStatus
AI CoreProvider orchestrationgenerate, select-provider, check-statusOperational
AI AnalysisAnalytics capabilitiesanalyze-usage, performance-metricsOperational
AI WorkflowWorkflow automationexecute-workflow, batch-processOperational
Direct ToolsAgent integrationfile-ops, web-search, executeOperational
UtilitiesGeneral utilitiestime, calculations, formattingOperational

Advanced MCP Infrastructure

ComponentCapabilitiesStatus
Tool RegistryTool registration, execution, statisticsActive
External Server ManagerLifecycle management, health monitoringActive
Tool Discovery ServiceAutomatic tool discovery and registrationActive
MCP FactoryLighthouse-compatible server creationActive
Flexible Tool ValidatorUniversal safety validationActive
Context ManagerRich context with 15+ fieldsActive
Tool OrchestratorSequential pipelines, error handlingActive

Lighthouse MCP Compatibility

  • Factory Pattern: createMCPServer() fully compatible with Lighthouse architecture
  • Transport Mechanisms: stdio, HTTP/Streamable HTTP, SSE, WebSocket support (99% compatibility)
  • Tool Standards: Full MCP specification compliance
  • Context Passing: Rich context with sessionId, userId, permissions (15+ fields)

58+ External MCP Servers

Supported for extended functionality:

Categories:

  • Development: GitHub, GitLab, filesystem access
  • Databases: PostgreSQL, MySQL, SQLite
  • Cloud Storage: Google Drive, AWS S3
  • Communication: Slack, email
  • And many more...

Quick Example:

// Add any MCP server dynamically
await neurolink.addExternalMCPServer("github", {
command: "npx",
args: ["-y", "@modelcontextprotocol/server-github"],
transport: "stdio",
env: { GITHUB_TOKEN: process.env.GITHUB_TOKEN },
});

// Tools automatically available to AI
const result = await neurolink.generate({
input: { text: 'Create a GitHub issue titled "Bug in auth flow"' },
});

MCP Integration Guide - Setup and usage MCP Server Catalog - Complete server list (58+)


Developer Experience Features

SDK Features

FeatureDescriptionDocumentation
Auto Provider SelectionIntelligent provider fallbackSDK Guide
Streaming ResponsesReal-time token streamingStreaming Guide
Conversation MemoryAutomatic context managementMemory Guide
Full Type SafetyComplete TypeScript typesType Reference
Error HandlingGraceful provider fallbackError Guide
Analytics & EvaluationUsage tracking, quality scoresAnalytics Guide
Middleware SystemRequest/response hooksMiddleware Guide
Framework IntegrationNext.js, SvelteKit, ExpressFramework Guides

CLI Features

FeatureDescriptionDocumentation
Interactive SetupGuided provider configurationSetup Guide
Text GenerationCLI-based generationGenerate Command
StreamingReal-time streaming outputStream Command
Loop SessionsPersistent interactive modeLoop Sessions
Provider ManagementHealth checks and statusCLI Guide
Model EvaluationAutomated testingEval Command
MCP ManagementServer discovery and installationMCP CLI

15+ Commands for every workflow - see Complete CLI Reference


Smart Model Selection & Cost Optimization

Cost Optimization Features

  • Automatic Cost Optimization: Selects cheapest models for simple tasks
  • LiteLLM Model Routing: Access 100+ models with automatic load balancing
  • Capability-Based Selection: Find models with specific features (vision, function calling)
  • Intelligent Fallback: Seamless switching when providers fail

CLI Examples:

# Cost optimization - automatically use cheapest model
npx @juspay/neurolink generate "Hello" --optimize-cost

# LiteLLM specific model selection
npx @juspay/neurolink generate "Complex analysis" --provider litellm --model "anthropic/claude-3-5-sonnet"

# Auto-select best available provider
npx @juspay/neurolink generate "Write code" # Automatically chooses optimal provider

Learn more: Provider Orchestration Guide


Interactive Loop Mode

NeuroLink features a powerful interactive loop mode that transforms the CLI into a persistent, stateful session.

Key Capabilities

  • Run any CLI command without restarting session
  • Persistent session variables: set provider openai, set temperature 0.9
  • Conversation memory: AI remembers previous turns within session
  • Redis auto-detection: Automatically connects if REDIS_URL is set
  • Export session history as JSON for analytics

Quick Start

# Start loop with Redis-backed conversation memory
npx @juspay/neurolink loop --enable-conversation-memory --auto-redis

# Start loop without Redis auto-detection
npx @juspay/neurolink loop --enable-conversation-memory --no-auto-redis

Example Session

# Start the interactive session
$ npx @juspay/neurolink loop

neurolink » set provider google-ai
✓ provider set to google-ai

neurolink » set temperature 0.8
✓ temperature set to 0.8

neurolink » generate "Tell me a fun fact about space"
The quietest place on Earth is an anechoic chamber at Microsoft's headquarters...

# Exit the session
neurolink » exit

Complete Loop Guide - Full documentation with all commands


Enterprise & Production Features

Production Capabilities

FeatureDescriptionUse CaseDocumentation
Enterprise ProxyCorporate proxy supportBehind firewallsProxy Setup
Redis MemoryDistributed conversation stateMulti-instance deploymentRedis Guide
Cost OptimizationAutomatic cheapest model selectionBudget controlCost Guide
Multi-Provider FailoverAutomatic provider switchingHigh availabilityFailover Guide
Telemetry & MonitoringOpenTelemetry integrationObservabilityObservability Guide
Security HardeningCredential management, auditingComplianceSecurity Guide
Custom Model HostingSageMaker integrationPrivate modelsSageMaker Guide
Load BalancingLiteLLM proxy integrationScale & routingLoad Balancing Guide
Audit TrailsComprehensive loggingComplianceAudit Guide
Configuration ManagementEnvironment & credential managementMulti-environment deploymentConfig Guide

Advanced Security Features

Human-in-the-Loop (HITL) Policy Engine

Enterprise-grade approval system for sensitive operations:

// HITL Policy Configuration
type HITLPolicy = {
requireApprovalFor: string[]; // Tool-specific policies
autoApprove: string[]; // Safe operation whitelist
alwaysDeny: string[]; // Blacklist operations
timeoutBehavior: "deny" | "approve"; // Timeout handling
};

HITL Capabilities:

  • User consent for dangerous operations
  • Configurable policy engine
  • Comprehensive audit trail logging
  • Timeout handling
  • Bulk approval for batch operations

Advanced Proxy Support

Corporate network compatibility:

Proxy TypeSupportFeatures
AWS ProxyFullAWS-specific proxy configuration
HTTP/HTTPS ProxyFullUniversal proxy across all providers
No-Proxy BypassFullBypass configuration and utilities

Enhanced Guardrails

AI-powered content security:

  • Content Filtering: Automatic content screening
  • Toxicity Detection: Toxic content filtering
  • PII Redaction: Privacy protection and PII detection
  • Custom Rules: Configurable policy rules
  • Security Reporting: Detailed security event reporting

Security & Compliance Certifications

  • SOC2 Type II compliant deployments
  • ISO 27001 certified infrastructure compatible
  • GDPR-compliant data handling (EU providers available)
  • HIPAA compatible (with proper configuration)
  • Hardened OS verified (SELinux, AppArmor)
  • Zero credential logging
  • Encrypted configuration storage

Enterprise Deployment Guide - Complete production patterns


Middleware & Extension System

Advanced Middleware Architecture

Pluggable request/response processing for custom workflows:

Built-in Middleware

MiddlewarePurposeFeaturesStatus
AnalyticsUsage tracking & monitoringToken counting, timing, performance metricsActive
GuardrailsContent securityContent policies, toxicity detection, PII filteringActive
Auto EvaluationQuality scoringLLM-as-judge, accuracy metrics, safety validationActive

Middleware System Capabilities

// Middleware Configuration
type MiddlewareFactoryOptions = {
middleware?: NeuroLinkMiddleware[]; // Custom middleware registration
enabledMiddleware?: string[]; // Selective activation
disabledMiddleware?: string[]; // Selective deactivation
middlewareConfig?: Record<string, MiddlewareConfig>; // Per-middleware configuration
preset?: string; // Preset configurations
global?: {
// Global settings
maxExecutionTime?: number;
continueOnError?: boolean;
};
};

Middleware Features:

  • Dynamic middleware registration
  • Pipeline execution with performance tracking
  • Runtime configuration changes
  • Error handling and graceful recovery
  • Priority-based execution order
  • Detailed execution statistics

Custom Middleware Guide - Build your own middleware


Performance & Optimization

Intelligent Cost Optimization

  • Model Resolver: Cost optimization algorithms and intelligent routing
  • Performance Routing: Speed-optimized provider selection
  • Concurrent Initialization: Reduced latency through parallel loading
  • Caching Strategies: Intelligent response and configuration caching

Advanced SageMaker Features

Beyond basic integration - enterprise-grade custom model deployment:

FeatureDescriptionStatus
Adaptive SemaphoreDynamic concurrency control for optimal throughputImplemented
Structured Output ParserComplex response parsing and validationImplemented
Capability DetectionAutomatic endpoint capability discoveryImplemented
Batch InferenceEfficient batch processing for high-volume workloadsImplemented
Diagnostics SystemReal-time endpoint monitoring and debuggingImplemented

Error Handling & Resilience

Production-grade fault tolerance:

  • MCP Circuit Breaker: Fault tolerance with state management
  • Error Hierarchies: Comprehensive error types for HITL, providers, and MCP
  • Graceful Degradation: Intelligent fallback strategies
  • Retry Logic: Configurable retry with exponential backoff

Performance Optimization Guide - Complete optimization strategies


Advanced Integrations

IntegrationDescription
LiteLLM IntegrationAccess 100+ models from all major providers via LiteLLM routing with unified interface.
SageMaker IntegrationDeploy and call custom endpoints directly from NeuroLink CLI/SDK with full control.
MemoryPer-user condensed memory with S3/Redis/SQLite storage and LLM-powered condensation.
Enterprise ProxyConfigure outbound policies and compliance posture for corporate environments.
Configuration ManagementManage environments, regions, and credentials safely across deployments.

Advanced Features

FeatureDescription
🏭 Factory Pattern ArchitectureUnified provider interface with automatic fallbacks and type-safe implementations.
🗄️ Conversation MemoryDeep dive into memory management and Redis integration.
🔧 Custom MiddlewareBuild request/response hooks for logging, filtering, and custom processing.
Performance OptimizationCaching, connection pooling, and latency optimization strategies.
📊 Telemetry & ObservabilityOpenTelemetry integration for distributed tracing and monitoring.
🧪 Testing GuideProvider-agnostic testing, mocking, and quality assurance strategies.
📊 Analytics & EvaluationUsage tracking, cost monitoring, and quality scoring for AI responses.
StreamingReal-time token streaming with provider-specific optimizations.
Thinking ConfigurationConfigure extended thinking levels for supported models (Anthropic, Gemini 2.5+).
Structured OutputJSON schema-based structured output with provider-specific formatting.
Text-to-Speech (TTS)Basic TTS support via Google Cloud TTS (Neural2, Wavenet, Standard voices).

See Also