๐ง Environment Variables Configuration Guide
This guide provides comprehensive setup instructions for all AI providers supported by NeuroLink. The CLI automatically loads environment variables from .env files, making configuration seamless.
๐ Quick Setupโ
Automatic .env Loading โจ NEW!โ
NeuroLink CLI automatically loads environment variables from .env files in your project directory:
# Create .env file (automatically loaded)
echo 'OPENAI_API_KEY="sk-your-key"' > .env
echo 'AWS_ACCESS_KEY_ID="your-key"' >> .env
# Test configuration
npx @juspay/neurolink status
Manual Export (Also Supported)โ
export OPENAI_API_KEY="sk-your-key"
export AWS_ACCESS_KEY_ID="your-key"
npx @juspay/neurolink status
๐๏ธ Enterprise Configuration Managementโ
โจ NEW: Automatic Backup Systemโ
# Configure backup settings
NEUROLINK_BACKUP_ENABLED=true # Enable automatic backups (default: true)
NEUROLINK_BACKUP_RETENTION=30 # Days to keep backups (default: 30)
NEUROLINK_BACKUP_DIRECTORY=.neurolink.backups # Backup directory (default: .neurolink.backups)
# Config validation settings
NEUROLINK_VALIDATION_STRICT=false # Strict validation mode (default: false)
NEUROLINK_VALIDATION_WARNINGS=true # Show validation warnings (default: true)
# Provider status monitoring
NEUROLINK_PROVIDER_STATUS_CHECK=true # Monitor provider availability (default: true)
NEUROLINK_PROVIDER_TIMEOUT=30000 # Provider timeout in ms (default: 30000)
Interface Configurationโ
# MCP Registry settings
NEUROLINK_REGISTRY_CACHE_TTL=300 # Cache TTL in seconds (default: 300)
NEUROLINK_REGISTRY_AUTO_DISCOVERY=true # Auto-discover MCP servers (default: true)
NEUROLINK_REGISTRY_STATS_ENABLED=true # Enable registry statistics (default: true)
# Execution context settings
NEUROLINK_DEFAULT_TIMEOUT=30000 # Default execution timeout (default: 30000)
NEUROLINK_DEFAULT_RETRIES=3 # Default retry count (default: 3)
NEUROLINK_CONTEXT_LOGGING=info # Context logging level (default: info)
Performance & Optimizationโ
# Tool execution settings
NEUROLINK_TOOL_EXECUTION_TIMEOUT=1000 # Tool execution timeout in ms (default: 1000)
NEUROLINK_PIPELINE_TIMEOUT=22000 # Pipeline execution timeout (default: 22000)
NEUROLINK_CACHE_ENABLED=true # Enable execution caching (default: true)
# Error handling
NEUROLINK_AUTO_RESTORE_ENABLED=true # Enable auto-restore on config failures (default: true)
NEUROLINK_ERROR_RECOVERY_ATTEMPTS=3 # Error recovery attempts (default: 3)
NEUROLINK_GRACEFUL_DEGRADATION=true # Enable graceful degradation (default: true)
๐ AI Enhancement Featuresโ
Basic Enhancement Configurationโ
# AI response quality evaluation model (optional)
NEUROLINK_EVALUATION_MODEL="gemini-2.5-flash"
Description: Configures the AI model used for response quality evaluation when --enable-evaluation flag is used. Uses Google AI's fast Gemini 2.5 Flash model for quick quality assessment.
Supported Models:
gemini-2.5-flash(default) - Fast evaluation processinggemini-2.5-pro- More detailed evaluation (slower)
Usage:
# Enable evaluation with default model
npx @juspay/neurolink generate "prompt" --enable-evaluation
# Enable both analytics and evaluation
npx @juspay/neurolink generate "prompt" --enable-analytics --enable-evaluation
๐ Universal Evaluation System (Advanced)โ
Primary Configurationโ
# Primary evaluation provider
NEUROLINK_EVALUATION_PROVIDER="google-ai" # Default: google-ai
# Evaluation performance mode
NEUROLINK_EVALUATION_MODE="fast" # Options: fast, balanced, quality
NEUROLINK_EVALUATION_PROVIDER: Primary AI provider for evaluation
- Options:
google-ai,openai,anthropic,vertex,bedrock,azure,ollama,huggingface,mistral - Default:
google-ai - Usage: Determines which AI provider performs the quality evaluation
NEUROLINK_EVALUATION_MODE: Performance vs quality trade-off
- Options:
fast(cost-effective),balanced(optimal),quality(highest accuracy) - Default:
fast - Usage: Selects appropriate model for the provider (e.g., gemini-2.5-flash vs gemini-2.5-pro)
Fallback Configurationโ
# Enable automatic fallback when primary provider fails
NEUROLINK_EVALUATION_FALLBACK_ENABLED="true" # Default: true
# Fallback provider order (comma-separated)
NEUROLINK_EVALUATION_FALLBACK_PROVIDERS="openai,anthropic,vertex,bedrock"
NEUROLINK_EVALUATION_FALLBACK_ENABLED: Enable intelligent fallback system
- Options:
true,false - Default:
true - Usage: When enabled, automatically tries backup providers if primary fails
NEUROLINK_EVALUATION_FALLBACK_PROVIDERS: Backup provider order
- Format: Comma-separated provider names
- Default:
openai,anthropic,vertex,bedrock - Usage: Defines the order of providers to try if primary fails
Performance Tuningโ
# Evaluation timeout (milliseconds)
NEUROLINK_EVALUATION_TIMEOUT="10000" # Default: 10000 (10 seconds)
# Maximum tokens for evaluation response
NEUROLINK_EVALUATION_MAX_TOKENS="500" # Default: 500
# Temperature for consistent evaluation
NEUROLINK_EVALUATION_TEMPERATURE="0.1" # Default: 0.1 (low for consistency)
# Retry attempts for failed evaluations
NEUROLINK_EVALUATION_RETRY_ATTEMPTS="2" # Default: 2
Performance Variables:
- TIMEOUT: Maximum time to wait for evaluation (prevents hanging)
- MAX_TOKENS: Limits evaluation response length (controls cost)
- TEMPERATURE: Lower values = more consistent scoring
- RETRY_ATTEMPTS: Number of retry attempts for transient failures
Cost Optimizationโ
# Prefer cost-effective models and providers
NEUROLINK_EVALUATION_PREFER_CHEAP="true" # Default: true
# Maximum cost per evaluation (USD)
NEUROLINK_EVALUATION_MAX_COST_PER_EVAL="0.01" # Default: $0.01
NEUROLINK_EVALUATION_PREFER_CHEAP: Cost optimization preference
- Options:
true,false - Default:
true - Usage: When enabled, prioritizes cheaper providers and models
NEUROLINK_EVALUATION_MAX_COST_PER_EVAL: Cost limit per evaluation
- Format: Decimal number (USD)
- Default:
0.01($0.01) - Usage: Prevents expensive evaluations, switches to cheaper providers if needed
Complete Universal Evaluation Exampleโ
# Comprehensive evaluation configuration
NEUROLINK_EVALUATION_PROVIDER="google-ai"
NEUROLINK_EVALUATION_MODEL="gemini-2.5-flash"
NEUROLINK_EVALUATION_MODE="balanced"
NEUROLINK_EVALUATION_FALLBACK_ENABLED="true"
NEUROLINK_EVALUATION_FALLBACK_PROVIDERS="openai,anthropic,vertex"
NEUROLINK_EVALUATION_TIMEOUT="15000"
NEUROLINK_EVALUATION_MAX_TOKENS="750"
NEUROLINK_EVALUATION_TEMPERATURE="0.2"
NEUROLINK_EVALUATION_PREFER_CHEAP="false"
NEUROLINK_EVALUATION_MAX_COST_PER_EVAL="0.05"
NEUROLINK_EVALUATION_RETRY_ATTEMPTS="3"
Testing Universal Evaluationโ
# Test primary provider
npx @juspay/neurolink generate "What is AI?" --enable-evaluation --debug
# Test with custom domain
npx @juspay/neurolink generate "Fix this Python code" --enable-evaluation --evaluation-domain "Python expert"
# Test Lighthouse-style evaluation
npx @juspay/neurolink generate "Business analysis" --lighthouse-style --evaluation-domain "Business consultant"
๐ข Enterprise Proxy Configurationโ
Proxy Environment Variablesโ
# Corporate proxy support (automatic detection)
HTTPS_PROXY="http://proxy.company.com:8080"
HTTP_PROXY="http://proxy.company.com:8080"
NO_PROXY="localhost,127.0.0.1,.company.com"
| Variable | Description | Example |
|---|---|---|
HTTPS_PROXY | Proxy server for HTTPS requests | http://proxy.company.com:8080 |
HTTP_PROXY | Proxy server for HTTP requests | http://proxy.company.com:8080 |
NO_PROXY | Domains to bypass proxy | localhost,127.0.0.1,.company.com |
Authenticated Proxyโ
# Proxy with username/password authentication
HTTPS_PROXY="http://username:[email protected]:8080"
HTTP_PROXY="http://username:[email protected]:8080"
All NeuroLink providers automatically use proxy settings when configured.
For detailed proxy setup โ See Enterprise & Proxy Setup Guide
๐ค Provider Configurationโ
1. OpenAIโ
Required Variablesโ
OPENAI_API_KEY="sk-proj-your-openai-api-key"
Optional Variablesโ
OPENAI_MODEL="gpt-4o" # Default: gpt-4o
OPENAI_BASE_URL="https://api.openai.com" # Default: OpenAI API
How to Get OpenAI API Keyโ
- Visit OpenAI Platform
- Sign up or log in to your account
- Navigate to API Keys section
- Click Create new secret key
- Copy the key (starts with
sk-proj-orsk-) - Add billing information if required
Supported Modelsโ
gpt-4o(default) - Latest GPT-4 Optimizedgpt-4o-mini- Faster, cost-effective optiongpt-4-turbo- High-performance modelgpt-3.5-turbo- Legacy cost-effective option
2. Amazon Bedrockโ
Required Variablesโ
AWS_ACCESS_KEY_ID="AKIA..."
AWS_SECRET_ACCESS_KEY="your-secret-key"
AWS_REGION="us-east-1"
Model Configuration (โ ๏ธ Critical)โ
# Use full inference profile ARN for Anthropic models
BEDROCK_MODEL="arn:aws:bedrock:us-east-2:<account_id>:inference-profile/us.anthropic.claude-3-7-sonnet-20250219-v1:0"
# OR use simple model names for non-Anthropic models
BEDROCK_MODEL="amazon.titan-text-express-v1"
Optional Variablesโ
AWS_SESSION_TOKEN="IQoJb3..." # For temporary credentials
How to Get AWS Credentialsโ
- Sign up for AWS Account
- Navigate to IAM Console
- Create new user with programmatic access
- Attach policy:
AmazonBedrockFullAccess - Download access key and secret key
- Important: Request model access in Bedrock console
Bedrock Model Access Setupโ
- Go to AWS Bedrock Console
- Navigate to Model access
- Click Request model access
- Select desired models (Claude, Titan, etc.)
- Submit request and wait for approval
Supported Modelsโ
- Anthropic Claude:
arn:aws:bedrock:<region>:<account_id>:inference-profile/us.anthropic.claude-3-7-sonnet-20250219-v1:0arn:aws:bedrock:<region>:<account_id>:inference-profile/us.anthropic.claude-3-5-sonnet-20241022-v2:0
- Amazon Titan:
amazon.titan-text-express-v1amazon.titan-text-lite-v1
3. Google Vertex AIโ
Google Vertex AI supports three authentication methods. Choose the one that fits your deployment:
Method 1: Service Account File (Recommended)โ
GOOGLE_APPLICATION_CREDENTIALS="/absolute/path/to/service-account.json"
GOOGLE_VERTEX_PROJECT="your-gcp-project-id"
GOOGLE_VERTEX_LOCATION="us-central1"
Method 2: Service Account JSON Stringโ
GOOGLE_SERVICE_ACCOUNT_KEY='{"type":"service_account","project_id":"your-project",...}'
GOOGLE_VERTEX_PROJECT="your-gcp-project-id"
GOOGLE_VERTEX_LOCATION="us-central1"
Method 3: Individual Environment Variablesโ
GOOGLE_AUTH_CLIENT_EMAIL="[email protected]"
GOOGLE_AUTH_PRIVATE_KEY="-----BEGIN PRIVATE KEY-----\nMIIEvQIBADANBgkqhkiG9w0B..."
GOOGLE_VERTEX_PROJECT="your-gcp-project-id"
GOOGLE_VERTEX_LOCATION="us-central1"
Optional Variablesโ
VERTEX_MODEL="gemini-2.5-pro" # Default: gemini-2.5-pro
How to Set Up Google Vertex AIโ
- Create Google Cloud Project
- Enable Vertex AI API
- Create Service Account:
- Go to IAM & Admin > Service Accounts
- Click Create Service Account
- Grant Vertex AI User role
- Generate and download JSON key file
- Set
GOOGLE_APPLICATION_CREDENTIALSto the JSON file path
Supported Modelsโ
gemini-2.5-pro(default) - Most capable modelgemini-2.5-flash- Faster responsesclaude-3-5-sonnet@20241022- Claude via Vertex AI
4. Anthropic (Direct)โ
Anthropic supports two authentication methods: API key (traditional) and OAuth token (for Claude subscription users).
Method 1: API Key (Traditional)โ
Required Variablesโ
ANTHROPIC_API_KEY="sk-ant-api03-your-anthropic-key"
Optional Variablesโ
ANTHROPIC_MODEL="claude-3-5-sonnet-20241022" # Default model
How to Get Anthropic API Keyโ
- Visit Anthropic Console
- Sign up or log in
- Navigate to API Keys
- Click Create Key
- Copy the key (starts with
sk-ant-api03-) - Add billing information for usage
Method 2: OAuth Token (Claude Subscription)โ
Use OAuth authentication to access Claude models through a Claude Pro, Max, or Team subscription instead of pay-per-token API billing.
Required Variablesโ
# Either of these (ANTHROPIC_OAUTH_TOKEN takes precedence)
ANTHROPIC_OAUTH_TOKEN="your-oauth-access-token"
CLAUDE_OAUTH_TOKEN="your-oauth-access-token"
The OAuth token value can be a plain access token string or a JSON object with the following fields:
{
"accessToken": "your-access-token",
"refreshToken": "your-refresh-token",
"expiresAt": 1735689600000
}
Where expiresAt is the token expiry time in Unix milliseconds.
Optional Variablesโ
ANTHROPIC_MODEL="claude-3-5-sonnet-20241022" # Default model
ANTHROPIC_SUBSCRIPTION_TIER="pro" # Subscription tier override
ANTHROPIC_SUBSCRIPTION_TIER controls which models and rate limits are available. Valid values:
| Tier | Description |
|---|---|
free | Free tier with limited access |
pro | Claude Pro subscription (default for OAuth) |
max | Claude Max subscription |
max_5 | Claude Max with 5x usage |
max_20 | Claude Max with 20x usage |
api | Standard API key access (default without OAuth) |
If ANTHROPIC_SUBSCRIPTION_TIER is not set, the tier is auto-detected: pro when using OAuth, api when using an API key.
Environment Variables Referenceโ
| Variable | Required | Default | Description |
|---|---|---|---|
ANTHROPIC_API_KEY | * | - | Anthropic API key (required if not using OAuth) |
ANTHROPIC_OAUTH_TOKEN | * | - | OAuth access token, plain string or JSON (required if not using API key) |
CLAUDE_OAUTH_TOKEN | * | - | Alternative OAuth token env var (same format as ANTHROPIC_OAUTH_TOKEN) |
ANTHROPIC_MODEL | No | claude-3-5-sonnet-20241022 | Default model to use |
ANTHROPIC_SUBSCRIPTION_TIER | No | Auto-detected (pro or api) | Subscription tier override: free, pro, max, max_5, max_20, api |
ANTHROPIC_ENABLE_BETA_FEATURES | No | true (OAuth) / false (API key) | Enable Anthropic beta headers (OAuth beta, extended thinking) |
ANTHROPIC_OAUTH_REFRESH_TOKEN | No | - | OAuth refresh token (used for automatic token renewal) |
ANTHROPIC_AUTH_METHOD | No | Auto-detected | Force auth method: api_key or oauth |
* One of ANTHROPIC_API_KEY, ANTHROPIC_OAUTH_TOKEN, or CLAUDE_OAUTH_TOKEN must be set.
Supported Modelsโ
claude-3-5-sonnet-20241022(default) - Latest Claudeclaude-3-haiku-20240307- Fast, cost-effectiveclaude-3-opus-20240229- Most capable (if available)
5. Google AI Studioโ
Required Variablesโ
GOOGLE_AI_API_KEY="AIza-your-google-ai-api-key"
Optional Variablesโ
GOOGLE_AI_MODEL="gemini-2.5-pro" # Default model
How to Get Google AI Studio API Keyโ
- Visit Google AI Studio
- Sign in with your Google account
- Navigate to API Keys section
- Click Create API Key
- Copy the key (starts with
AIza) - Note: Google AI Studio provides free tier with generous limits
Supported Modelsโ
gemini-2.5-pro(default) - Latest Gemini Progemini-2.0-flash- Fast, efficient responses
6. Azure OpenAIโ
Required Variablesโ
AZURE_OPENAI_API_KEY="your-azureOpenai-key"
AZURE_OPENAI_ENDPOINT="https://your-resource.openai.azure.com/"
AZURE_OPENAI_DEPLOYMENT_ID="your-deployment-name"
Optional Variablesโ
AZURE_MODEL="gpt-4o" # Default: gpt-4o
AZURE_API_VERSION="2024-02-15-preview" # Default API version
How to Set Up Azure OpenAIโ
- Create Azure Account
- Apply for Azure OpenAI Service access
- Create Azure OpenAI Resource:
- Go to Azure Portal
- Search "OpenAI"
- Create new OpenAI resource
- Deploy Model:
- Go to Azure OpenAI Studio
- Navigate to Deployments
- Create deployment with desired model
- Get credentials from Keys and Endpoint section
Supported Modelsโ
gpt-4o(default) - Latest GPT-4 Optimizedgpt-4- Standard GPT-4gpt-35-turbo- Cost-effective option
7. Hugging Faceโ
Required Variablesโ
HUGGINGFACE_API_KEY="hf_your_huggingface_token"
Optional Variablesโ
HUGGINGFACE_MODEL="microsoft/DialoGPT-medium" # Default model
HUGGINGFACE_ENDPOINT="https://api-inference.huggingface.co" # Default endpoint
How to Get Hugging Face API Tokenโ
- Visit Hugging Face
- Sign up or log in
- Go to Settings โ Access Tokens
- Create new token with "read" scope
- Copy token (starts with
hf_)
Supported Modelsโ
- Open Source: Access to 100,000+ community models
microsoft/DialoGPT-medium(default) - Conversational AIgpt2- Classic GPT-2EleutherAI/gpt-neo-2.7B- Large open model- Any model from Hugging Face Hub
8. Ollama (Local AI)โ
Required Variablesโ
None! Ollama runs locally.
Optional Variablesโ
OLLAMA_BASE_URL="http://localhost:11434" # Default local server
OLLAMA_MODEL="llama2" # Default model
How to Set Up Ollamaโ
-
Install Ollama:
-
Start Ollama Service:
ollama serve # Usually auto-startsTip: To keep Ollama running in the background:
- macOS:
brew services start ollama - Linux (user):
systemctl --user enable --now ollama - Linux (system):
sudo systemctl enable --now ollama
- macOS:
-
Pull Models:
ollama pull llama2
ollama pull codellama
ollama pull mistral
Supported Modelsโ
llama2(default) - Meta's Llama 2codellama- Code-specialized Llamamistral- Mistral 7Bvicuna- Fine-tuned Llama- Any model from Ollama Library
9. Mistral AIโ
Required Variablesโ
MISTRAL_API_KEY="your_mistral_api_key"
Optional Variablesโ
MISTRAL_MODEL="mistral-small" # Default model
MISTRAL_ENDPOINT="https://api.mistral.ai" # Default endpoint
How to Get Mistral AI API Keyโ
- Visit Mistral AI Platform
- Sign up for an account
- Navigate to API Keys section
- Generate new API key
- Add billing information
Supported Modelsโ
mistral-tiny- Fastest, most cost-effectivemistral-small(default) - Balanced performancemistral-medium- Enhanced capabilitiesmistral-large- Most capable model
10. LiteLLM ๐โ
Required Variablesโ
LITELLM_BASE_URL="http://localhost:4000" # Local LiteLLM proxy (default)
LITELLM_API_KEY="sk-anything" # API key for local proxy (any value works)
Optional Variablesโ
LITELLM_MODEL="gemini-2.5-pro" # Default model
LITELLM_TIMEOUT="60000" # Request timeout (ms)
How to Use LiteLLMโ
LiteLLM provides access to 100+ AI models through a unified proxy interface:
- Local Setup: Run LiteLLM locally with your API keys (recommended)
- Self-Hosted: Deploy your own LiteLLM proxy server
- Cloud Deployment: Use cloud-hosted LiteLLM instances
Available Models (Example Configuration)โ
openai/gpt-4o- OpenAI GPT-4 Optimizedanthropic/claude-3-5-sonnet- Anthropic Claude Sonnetgoogle/gemini-2.0-flash- Google Gemini Flashmistral/mistral-large- Mistral Large model- Many more via LiteLLM Providers
Benefitsโ
- 100+ Models: Access to all major AI providers through one interface
- Cost Optimization: Automatic routing to cost-effective models
- Unified API: OpenAI-compatible API for all models
- Load Balancing: Automatic failover and load distribution
- Analytics: Built-in usage tracking and monitoring
11. Amazon SageMaker ๐โ
Required Variablesโ
AWS_ACCESS_KEY_ID="AKIA..."
AWS_SECRET_ACCESS_KEY="your-aws-secret-key"
AWS_REGION="us-east-1"
SAGEMAKER_DEFAULT_ENDPOINT="your-endpoint-name"
Optional Variablesโ
SAGEMAKER_MODEL="custom-model-name" # Model identifier (default: sagemaker-model)
SAGEMAKER_TIMEOUT="30000" # Request timeout in ms (default: 30000)
SAGEMAKER_MAX_RETRIES="3" # Retry attempts (default: 3)
AWS_SESSION_TOKEN="IQoJb3..." # For temporary credentials
SAGEMAKER_CONTENT_TYPE="application/json" # Request content type (default: application/json)
SAGEMAKER_ACCEPT="application/json" # Response accept type (default: application/json)
How to Set Up Amazon SageMakerโ
Amazon SageMaker allows you to deploy and use your own custom trained models:
-
Deploy Your Model to SageMaker:
- Train your model using SageMaker Training Jobs
- Deploy model to a SageMaker Real-time Endpoint
- Note the endpoint name for configuration
-
Set Up AWS Credentials:
- Use IAM user with
sagemaker:InvokeEndpointpermission - Or use IAM role for EC2/Lambda/ECS deployments
- Configure AWS CLI:
aws configure
- Use IAM user with
-
Configure NeuroLink:
export AWS_ACCESS_KEY_ID="your-access-key"
export AWS_SECRET_ACCESS_KEY="your-secret-key"
export AWS_REGION="us-east-1"
export SAGEMAKER_DEFAULT_ENDPOINT="my-model-endpoint" -
Test Connection:
npx @juspay/neurolink sagemaker status
npx @juspay/neurolink sagemaker test my-endpoint
How to Get AWS Credentials for SageMakerโ
-
Create IAM User:
- Go to AWS IAM Console
- Create new user with Programmatic access
- Attach the following policy:
{
"Version": "2012-10-17",
"Statement": [
{
"Effect": "Allow",
"Action": ["sagemaker:InvokeEndpoint"],
"Resource": "arn:aws:sagemaker:*:*:endpoint/*"
}
]
} -
Download Credentials:
- Save Access Key ID and Secret Access Key
- Set as environment variables
Supported Modelsโ
SageMaker supports any custom model you deploy:
- Custom Fine-tuned Models - Your domain-specific models
- Foundation Model Endpoints - Large language models deployed via SageMaker
- Multi-model Endpoints - Multiple models behind single endpoint
- Serverless Endpoints - Auto-scaling model deployments
Model Deployment Typesโ
- Real-time Inference - Low-latency model serving (recommended)
- Batch Transform - Batch processing (not supported by NeuroLink)
- Serverless Inference - Pay-per-request model serving
- Multi-model Endpoints - Host multiple models efficiently
Benefitsโ
- ๐๏ธ Custom Models - Deploy and use your own trained models
- ๐ฐ Cost Control - Pay only for inference usage, auto-scaling available
- ๐ Enterprise Security - Full control over model infrastructure and data
- โก Performance - Dedicated compute resources with predictable latency
- ๐ Global Deployment - Available in all major AWS regions
- ๐ Monitoring - Built-in CloudWatch metrics and logging
CLI Commandsโ
# Check SageMaker configuration and endpoint status
npx @juspay/neurolink sagemaker status
# Validate connection to specific endpoint
npx @juspay/neurolink sagemaker validate
# Test inference with specific endpoint
npx @juspay/neurolink sagemaker test my-endpoint
# Show current configuration
npx @juspay/neurolink sagemaker config
# Performance benchmark
npx @juspay/neurolink sagemaker benchmark my-endpoint
# List available endpoints (requires AWS CLI)
npx @juspay/neurolink sagemaker list-endpoints
# Interactive setup wizard
npx @juspay/neurolink sagemaker setup
Environment Variables Referenceโ
| Variable | Required | Default | Description |
|---|---|---|---|
AWS_ACCESS_KEY_ID | โ | - | AWS access key for authentication |
AWS_SECRET_ACCESS_KEY | โ | - | AWS secret key for authentication |
AWS_REGION | โ | us-east-1 | AWS region where endpoint is deployed |
SAGEMAKER_DEFAULT_ENDPOINT | โ | - | SageMaker endpoint name |
SAGEMAKER_TIMEOUT | โ | 30000 | Request timeout in milliseconds |
SAGEMAKER_MAX_RETRIES | โ | 3 | Number of retry attempts for failed requests |
AWS_SESSION_TOKEN | โ | - | Session token for temporary credentials |
SAGEMAKER_MODEL | โ | sagemaker-model | Model identifier for logging |
SAGEMAKER_CONTENT_TYPE | โ | application/json | Request content type |
SAGEMAKER_ACCEPT | โ | application/json | Response accept type |
Production Considerationsโ
- ๐ Security: Use IAM roles instead of access keys when possible
- ๐ Monitoring: Enable CloudWatch logging for your endpoints
- ๐ฐ Cost Optimization: Use auto-scaling and serverless options
- ๐ Multi-Region: Deploy endpoints in multiple regions for redundancy
- โก Performance: Choose appropriate instance types for your workload
๐ง Configuration Examplesโ
Complete .env File Exampleโ
# NeuroLink Environment Configuration - All 11 Providers
# OpenAI Configuration
OPENAI_API_KEY="sk-proj-your-openai-key"
OPENAI_MODEL="gpt-4o"
# Amazon Bedrock Configuration
AWS_ACCESS_KEY_ID="AKIA..."
AWS_SECRET_ACCESS_KEY="your-aws-secret"
AWS_REGION="us-east-1"
BEDROCK_MODEL="arn:aws:bedrock:us-east-1:<account_id>:inference-profile/us.anthropic.claude-3-5-sonnet-20241022-v2:0"
# Amazon SageMaker Configuration
AWS_ACCESS_KEY_ID="AKIA..."
AWS_SECRET_ACCESS_KEY="your-aws-secret"
AWS_REGION="us-east-1"
SAGEMAKER_DEFAULT_ENDPOINT="my-model-endpoint"
SAGEMAKER_TIMEOUT="30000"
SAGEMAKER_MAX_RETRIES="3"
# Google Vertex AI Configuration
GOOGLE_APPLICATION_CREDENTIALS="/path/to/service-account.json"
GOOGLE_VERTEX_PROJECT="your-gcp-project"
GOOGLE_VERTEX_LOCATION="us-central1"
VERTEX_MODEL="gemini-2.5-pro"
# Anthropic Configuration (API key or OAuth token)
ANTHROPIC_API_KEY="sk-ant-api03-your-key"
# ANTHROPIC_OAUTH_TOKEN="your-oauth-token" # Alternative: OAuth token for Claude subscription
# ANTHROPIC_SUBSCRIPTION_TIER="pro" # Optional: free, pro, max, max_5, max_20, api
# ANTHROPIC_ENABLE_BETA_FEATURES="true" # Optional: enable beta headers (default: true for OAuth)
# ANTHROPIC_OAUTH_REFRESH_TOKEN="" # Optional: OAuth refresh token for auto-renewal
# ANTHROPIC_AUTH_METHOD="oauth" # Optional: force auth method (api_key or oauth)
# Google AI Studio Configuration
GOOGLE_AI_API_KEY="AIza-your-google-ai-key"
GOOGLE_AI_MODEL="gemini-2.5-pro"
# Azure OpenAI Configuration
AZURE_OPENAI_API_KEY="your-azure-key"
AZURE_OPENAI_ENDPOINT="https://your-resource.openai.azure.com/"
AZURE_OPENAI_DEPLOYMENT_ID="gpt-4o-deployment"
AZURE_MODEL="gpt-4o"
# Hugging Face Configuration
HUGGINGFACE_API_KEY="hf_your_huggingface_token"
HUGGINGFACE_MODEL="microsoft/DialoGPT-medium"
# Ollama Configuration (Local AI - No API Key Required)
OLLAMA_BASE_URL="http://localhost:11434"
OLLAMA_MODEL="llama2"
# Mistral AI Configuration
MISTRAL_API_KEY="your_mistral_api_key"
MISTRAL_MODEL="mistral-small"
# LiteLLM Configuration
LITELLM_BASE_URL="http://localhost:4000"
LITELLM_API_KEY="sk-anything"
LITELLM_MODEL="openai/gpt-4o-mini"
Docker/Container Configurationโ
# Use environment variables in containers
docker run -e OPENAI_API_KEY="sk-..." \
-e AWS_ACCESS_KEY_ID="AKIA..." \
-e AWS_SECRET_ACCESS_KEY="..." \
your-app
CI/CD Configurationโ
# GitHub Actions example
env:
OPENAI_API_KEY: ${{ secrets.OPENAI_API_KEY }}
AWS_ACCESS_KEY_ID: ${{ secrets.AWS_ACCESS_KEY_ID }}
AWS_SECRET_ACCESS_KEY: ${{ secrets.AWS_SECRET_ACCESS_KEY }}
๐งช Testing Configurationโ
Test All Providersโ
# Check provider status
npx @juspay/neurolink status --verbose
# Test specific provider
npx @juspay/neurolink generate "Hello" --provider openai
# Get best available provider
npx @juspay/neurolink get-best-provider
Expected Outputโ
โ
openai: Working (1245ms)
โ
bedrock: Working (2103ms)
โ
vertex: Working (1876ms)
โ
anthropic: Working (1654ms)
โ
azure: Working (987ms)
๐ Summary: 5/5 providers working
๐ Security Best Practicesโ
API Key Managementโ
- โ Use .env files for local development
- โ Use environment variables in production
- โ Rotate keys regularly (every 90 days)
- โ Never commit keys to version control
- โ Never hardcode keys in source code
.gitignore Configurationโ
# Add to .gitignore
.env
.env.local
.env.production
*.pem
service-account*.json
Production Deploymentโ
- Use secret management systems (AWS Secrets Manager, Azure Key Vault)
- Implement key rotation policies
- Monitor API usage and rate limits
- Use least privilege access policies
๐จ Troubleshootingโ
Common Issuesโ
1. "Missing API Key" Errorโ
# Check if environment is loaded
npx @juspay/neurolink status
# Verify .env file exists and has correct format
cat .env
2. AWS Bedrock "Not Authorized" Errorโ
- โ Verify account has model access in Bedrock console
- โ Use full inference profile ARN for Anthropic models
- โ Check IAM permissions include Bedrock access
3. Google Vertex AI Import Issuesโ
- โ Ensure Vertex AI API is enabled
- โ Verify service account has correct permissions
- โ Check JSON file path is absolute and accessible
4. CLI Not Loading .envโ
- โ
Ensure
.envfile is in current directory - โ Check file has correct format (no spaces around =)
- โ Verify CLI version supports automatic loading
Debug Commandsโ
# Verbose status check
npx @juspay/neurolink status --verbose
# Test specific provider
npx @juspay/neurolink generate "test" --provider openai --verbose
# Check environment loading
node -e "require('dotenv').config(); console.log(process.env.OPENAI_API_KEY)"
๐ Related Documentationโ
- Provider Configuration Guide - Detailed provider setup
- CLI Guide - Complete CLI command reference
- API Reference - Programmatic usage examples
- Framework Integration - Next.js, SvelteKit, React
๐ค Need Help?โ
- ๐ Check the troubleshooting section above
- ๐ Report issues in our GitHub repository
- ๐ฌ Join our Discord for community support
- ๐ง Contact us for enterprise support
Next Steps: Once configured, test your setup with npx @juspay/neurolink status and start generating AI content!