Overview
Debugging voice agents requires systematic investigation of multiple components working together—transcription, LLM reasoning, voice synthesis, and action execution. The itellicoAI dashboard provides detailed logs and tools to help you quickly identify the root cause of issues.Dashboard Debugging Tools
Conversation Logs
Complete history of every conversation with transcripts, actions, and metadata
Real-Time Transcript
Live view of transcription and agent responses during test calls
Action Payloads
Detailed JSON of every API call, tool execution, and webhook
Error Messages
Specific error details when components fail
Systematic Debugging Approach
When something goes wrong, follow this systematic process:Reproduce the issue
Test again to confirm the problem is consistentNote exact conditions when it occurs
Identify the component
Determine which part of the pipeline failed:
- Transcriber (speech → text)
- LLM (understanding → response)
- TTS (text → speech)
- Action/Tool execution
- Knowledge retrieval
Review logs
Open Conversations and find the problematic callExamine transcripts, action payloads, and errors
Test components individually
Isolate the failing component:
- Try different transcriber
- Test LLM with simpler prompts
- Try different voice
- Test actions directly via API
Component-Level Debugging
Transcriber Issues
Transcriber Issues
How to identify:
- Check transcript in conversation logs
- Compare what was said vs what was transcribed
- Look for missing words, incorrect words, or gibberish
- Background noise
- Accent or language mismatch
- Audio quality problems
- Wrong transcriber model selected
- Navigate to Models → Transcriber
- Try different transcriber provider (Deepgram ↔ Azure)
- Try different model (e.g., Nova-2 ↔ Nova-3)
- Verify language setting matches speaker
- Test in quieter environment
- Check audio input quality
- Transcript accuracy
- Timing of transcription (delays?)
- Empty or partial transcriptions
- Language detection issues
LLM Response Issues
LLM Response Issues
How to identify:
- Agent gives wrong answers
- Agent goes off-topic
- Agent repeats itself
- Agent refuses to answer valid questions
- Agent hallucinates information
- Instructions too vague or conflicting
- Knowledge base missing information
- Context window overflow
- Model not suitable for task
- Temperature too high/low
- Review agent instructions in Abilities → Instructions
- Simplify instructions to isolate the issue
- Check knowledge base for missing information
- Try different LLM model (Claude Haiku 4.5 ↔ GPT-4.1 mini)
- Adjust temperature in model settings
- Review conversation logs to see full context
- Full conversation history leading to bad response
- Knowledge items retrieved (if using RAG)
- System prompts and context injection
Voice/TTS Issues
Voice/TTS Issues
How to identify:
- Unnatural speech patterns
- Mispronunciations
- Wrong emphasis or intonation
- Robotic sound
- Speed too fast/slow
- Voice not suited to content type
- Punctuation affecting pacing
- Numbers or acronyms not handled well
- Voice provider limitations
- Navigate to Models → Voice
- Try different voice from same provider
- Try different voice provider entirely
- Add custom pronunciations for problem words
- Adjust stability/clarity settings (ElevenLabs)
- Adjust speaking rate
- Modify text output to improve TTS
- Listen to audio recording
- Compare text vs how it was spoken
- Check for SSML tags (if used)
- Verify voice settings applied
Action/Tool Execution Issues
Action/Tool Execution Issues
How to identify:
- Action doesn’t trigger when expected
- Action triggers but fails
- Wrong data sent to action
- Action returns error
- Action not properly configured
- API endpoint down or slow
- Authentication failure
- Incorrect parameter extraction
- Network timeout
- Check if action was triggered in conversation logs
- Review action payload (JSON sent to API)
- Check API response and status code
- Test API endpoint directly (Postman, curl)
- Verify authentication credentials
- Check parameter extraction from conversation
- Review action instructions in agent prompt
custom_data.actionsor similar fields- API request payload
- API response body
- Error messages and stack traces
- Timestamp (did it timeout?)
Conversation logs show complete action payloads including request/response data.
Knowledge Retrieval Issues
Knowledge Retrieval Issues
How to identify:
- Agent can’t answer questions it should know
- Agent retrieves wrong knowledge
- Agent mixes irrelevant information into answers
- Knowledge not indexed yet
- RAG retrieval not finding relevant items
- Knowledge base not assigned to agent
- Verify knowledge base assigned to agent
- Check knowledge items are INDEXED (not just COMPLETED)
- Review knowledge item titles—make them descriptive
- Test with smaller knowledge base
- Try Context mode vs RAG mode
- Check conversation logs for retrieved knowledge
Using Conversation Logs for Debugging
Every test call creates a detailed log accessible in Conversations.What’s in the logs:
Basic information:- Call date, time, duration
- Agent used
- Phone number (if phone test)
- Call status (completed, failed, etc.)
- Full transcript (user + agent)
- Timestamps for each message
- Audio recording (if available)
- Actions triggered with payloads
- DTMF inputs captured
- Goal analysis results
- Post-call analysis responses
- Custom data fields
- Error messages
- Filter by agent name to find test calls
- Open specific call to see full details
- Read transcript to identify where it went wrong
- Check action payloads if actions failed
- Listen to audio if transcript looks correct but audio was wrong
- Review timestamps to identify latency issues
Getting Help
When you need additional support:Review Documentation
Check specific feature docs for configuration details
Check Provider Status
Visit status pages for OpenAI, Deepgram, ElevenLabs, Azure
Contact Support
Email support@itellico.ai with call logs and error details
- Agent ID or name
- Conversation ID from logs
- Specific error messages
- Steps to reproduce
- Screenshots if applicable
Next Steps
Launch Checklist
Once you’ve debugged your agent, review the launch checklist to prepare for production