Frequently Asked Questions
Common questions and answers about NeuroLink usage, configuration, and troubleshooting.
🚀 Getting Started
Q: What is NeuroLink?
A: NeuroLink is an enterprise AI development platform that provides unified access to multiple AI providers (OpenAI, Google AI, Anthropic, AWS Bedrock, etc.) through a single SDK and CLI. It includes built-in tools, analytics, evaluation capabilities, and supports the Model Context Protocol (MCP) for extended functionality.
Q: Which AI providers does NeuroLink support?
A: NeuroLink supports 21+ AI providers:
- OpenAI (GPT-4o, GPT-4.1, o3, o4-mini)
- Google AI Studio (Gemini 3 Flash/Pro, Gemini 2.5 Pro/Flash)
- Google Vertex AI (Gemini 3, Claude via Vertex)
- Anthropic (Claude Opus 4.7, Sonnet 4.6, 4.5 Opus/Sonnet/Haiku)
- AWS Bedrock (Claude, Titan, Nova models)
- Azure OpenAI (GPT models)
- Hugging Face (Open source models)
- Ollama (Local AI models)
- Mistral AI (Mistral models)
- LiteLLM (100+ models via proxy)
- AWS SageMaker (Custom endpoints)
- OpenAI-compatible (Any OpenAI-API-compatible endpoint)
- OpenRouter (300+ models via OpenRouter)
- DeepSeek (DeepSeek V3, R1)
- NVIDIA NIM (Llama 3.3 70B, 400+ catalog models)
- LM Studio (Local models loaded in LM Studio)
- llama.cpp (Local GGUF models via llama-server)
Voice providers:
- OpenAI TTS (TTS-1, TTS-1-HD, GPT-4o Audio)
- ElevenLabs (Multilingual v2, Turbo v2.5, Flash v2.5)
- Deepgram (Nova-3, Nova-2, Enhanced — STT)
- Azure Speech (Azure Cognitive Services TTS + STT)
- Google TTS / STT (Google Cloud Speech)
- Whisper (OpenAI Whisper — STT)
- OpenAI Realtime + Gemini Live (realtime voice APIs)
Q: Do I need to install anything?
A: No installation required! You can use NeuroLink directly with npx:
npx @juspay/neurolink generate "Hello, AI!"
npx @juspay/neurolink status
For frequent use, you can install globally: npm install -g @juspay/neurolink