AI Engines (LLM Providers/Models)
AI Engines define which Large Language Model (LLM) provider/model MiaRec uses to run AI Tasks (summaries, sentiment, CSAT, Auto QA, and custom insights).
This chapter describes how platform operators should configure and govern AI Engines in a multi-tenant environment.
What an AI Engine is
An AI Engine is a named configuration that typically includes: - provider (e.g., OpenAI/Azure OpenAI/Anthropic/etc.) - model name/version - authentication (API key, endpoint, deployment name) - safety and policy settings (if applicable) - limits (rate limits, quotas) - default parameters (timeouts, max tokens, temperature) if the platform supports them
AI Tasks reference an AI Engine when executing.
Recommended engine strategy
Start with a small set of standard engines
For example: - Standard – balanced cost/quality - High accuracy – for complex scoring/categorization - Low latency – for near-real-time needs (if applicable)
Keep the catalog small to reduce operational complexity.
Decide how engine selection works
Common patterns: - Single global default engine for all tasks - Per-task engine selection (your UI appears to support this) - Per-tenant engine policy (if required for cost or compliance)
Credential and security management
- Store credentials in a secrets manager (recommended).
- Rotate credentials regularly and document rotation steps.
- Limit engine access by environment (dev/stage/prod).
- Maintain audit logs for engine configuration changes.
Validation / smoke tests
When adding an engine: 1. Run a minimal test prompt (health check) to verify credentials and connectivity. 2. Run a representative AI Task (e.g., summarization) on a test transcript. 3. Verify: - latency - successful JSON responses (if using JSON output) - stable output quality
Operational considerations
Cost management
- Track usage by tenant and by task (requests, tokens, cost).
- Use task filters to avoid running tasks on ineligible conversations.
- Consider “preview mode” for new tasks before broad enablement.
Reliability and failover (if supported)
- Define a fallback engine if the primary provider is down.
- Document expected behavior:
- automatic failover vs manual switch
- per-task vs global failover
Data handling
- Document data residency considerations (especially in partner-hosted deployments).
- Document whether transcripts are sent to external providers and what is included (metadata vs transcript only).
Where to configure
Menu path: Administration > Speech Analytics > AI Assistant > Engines
Figure: AI Engine configuration showing name, status, visibility settings, and model configuration options.
Engine settings
When configuring an AI Engine, you specify:
- Name – A descriptive name for the engine
- Status – Enabled or Disabled
- Visibility – Global (available to all tenants) or Tenant-specific
- Model settings – Provider-specific configuration (API endpoint, model name, authentication)
Per-task engine selection
Each AI Task can select which engine to use. This allows you to:
- Use different models for different task types (e.g., a faster model for simple tasks, a more capable model for complex analysis)
- Test new models on specific tasks before broader rollout
- Manage costs by using appropriate models for each use case
