Models and Providers
HelseCLI is designed to be model-agnostic. You can switch between top-tier models from various providers seamlessly.
HelseCLI is engineered for model-agnosticism, enabling seamless transitions between leading Large Language Models from diverse providers.
Integrated Providers
1. OpenAI Integration
Standard for software engineering and general computational reasoning.
- Configuration: Requires
OPENAI_API_KEYin the environment configuration. - Primary Model:
gpt-4o(Optimized for balance of latency and intelligence). - Reasoning Models:
o1-preview,o3-mini(Specifically for complex logical synthesis).
2. Anthropic Integration
Recognized for high-fidelity technical writing and nuanced contextual analysis.
- Configuration: Requires
ANTHROPIC_API_KEYin the environment configuration. - Primary Model:
claude-3-5-sonnet(Industry-leading performance in coding and research).
3. Google AI Integration
Optimized for high-throughput tasks and extensive context window requirements.
- Configuration: Requires
GOOGLE_API_KEYin the environment configuration. - Primary Model:
gemini-2.0-flash(Optimized for low-latency response).
4. Local Inference via Ollama
Enables private, air-gapped execution with zero token-based costs.
- Configuration: Requires an active local Ollama instance. Default connection via
localhost:11434. - Supported Models:
llama3,codellama,qwen2.5-coder.
Model selection Protocols
The active inference model can be dynamically reconfigured during any active session:
| Command | Action |
|---|---|
!m | Access the indexed model selection interface. |
!model <name> | Explicitly transition to a defined model (e.g., !model gpt-4o). |
!default <name> | Persist a model choice as the default for future initializations. |
Custom Model Configuration
HelseCLI supports third-party OpenAI-compatible endpoints (e.g., LM Studio, LocalAI, Groq).
To integrate a custom endpoint, modify the custom_models.json manifest in the application root:
{
"custom-inference-endpoint": {
"api_base": "https://api.example-provider.com/v1",
"provider": "openai",
"is_custom": true
}
}
Inference Optimization Strategies
- Software Development:
gpt-4oandclaude-3-5-sonnetare recommended for code synthesis. - Large-Scale Document Analysis:
gemini-1.5-prois recommended for its extensive context window. - Iterative Research:
gpt-4ois optimized for complex tool orchestration and web data synthesis. - Privacy-Restricted Tasks: Local execution via
llama3(Ollama) is recommended for data sensitive workflows.