Understanding Debug Tools
Beam provides multiple debugging approaches for different scenarios: Test Tool - Isolate and test individual nodes with mock input data to debug specific workflow steps Error Handling - Configure how agents respond to failures, including continue-on-error and retry logic Execution Logs - Detailed step-by-step traces showing inputs, outputs, and errors for each node LangFuse Integration - Advanced observability platform for tracing agent behavior and identifying issuesTest Tool for Node Debugging
Test individual workflow nodes in isolation without running the entire agent workflow. Configure nodes in Flow Builder.1
Select Node to Test
Open Flow builder and click the node experiencing issues or requiring validation.
2
Access Test Tool
In node configuration panel, locate “Test” or “Run Test” option to open test interface.
3
Provide Mock Input
Enter sample input data matching expected node input schema. Use real examples from failed tasks (see Task Executions) or create test cases in Test Datasets.
4
Execute Test
Run node test. System executes only this node with provided input, bypassing earlier workflow steps.
5
Inspect Output
Review node output, validation results, and any error messages. Compare against expected behavior defined in Evaluation Framework.
When to Use Test Tool
When to Use Test Tool
Node Producing Unexpected Output:
- Data extraction returning wrong values
- Classification categorizing incorrectly
- Integration call failing
- Field mapping errors
- Test new prompts before full workflow run
- Compare outputs with different prompt versions
- Verify edge case handling
- Test API authentication
- Validate request/response format
- Check integration configuration
- Verify data transformation logic
Test Tool Best Practices
Test Tool Best Practices
Use Real Production Data:
- Test with actual failed task inputs
- Include edge cases from production
- Cover variety of input formats
- Test → Adjust → Test cycle
- Small changes between iterations
- Document what works
- Test with old configuration
- Make single change
- Test with new configuration
- Measure improvement
- Save successful test inputs
- Reuse for regression testing
- Share with team for consistency
Error Handling Configuration
Configure how agents respond to errors at workflow and node levels.Continue on Error
Allow workflow to proceed even when specific nodes fail. Node-Level Setting:- Enable “Continue on error” in node configuration
- Workflow proceeds to next step even if this node fails
- Failed node output marked as error, available to downstream nodes
- Useful for non-critical steps where failure shouldn’t block entire workflow
- Optional data enrichment that shouldn’t block processing
- Logging or notification steps
- Non-essential validation checks
- Backup or fallback operations
Configuring Continue on Error
Configuring Continue on Error
When to Enable:
- Step is optional for workflow completion
- Downstream nodes handle missing data gracefully
- Failure rate is acceptable (e.g., external API occasionally unavailable)
- Manual review will catch issues later
- Critical data extraction or validation
- Security or compliance checks
- Payment or transaction processing
- Data accuracy is paramount
- Open node configuration in Flow builder
- Locate “Error Handling” section
- Toggle “Continue on error” option
- Test with failing scenarios to verify behavior
- Publish workflow changes
Handling Missing Data
Handling Missing Data
When upstream node fails with “Continue on error” enabled:Downstream Node Strategy:
- Check if required input exists before processing
- Use conditional branches to route around missing data
- Provide default values for missing fields
- Flag task for human review when critical data missing
- Node 1: Extract invoice data (Continue on error: enabled)
- Node 2: Check if invoice_amount exists
- Branch A: Amount exists → Process payment
- Branch B: Amount missing → Route to inbox for manual review
Auto-Retry on Error
Automatically retry failed steps to handle transient errors. Retry Configuration:- Set maximum retry attempts (typically 1-3)
- Configure retry delay between attempts
- Specify error types that trigger retry vs immediate failure
- API timeout errors (network issues)
- Rate limiting (429 errors)
- Temporary service unavailability (503 errors)
- Database connection failures
Retry Best Practices
Retry Best Practices
Appropriate for Retries:
- Network timeouts
- Temporary service outages
- Rate limiting
- Resource temporarily unavailable
- Authentication errors (won’t resolve with retry)
- Invalid request format (will fail again)
- Missing required data
- Business logic errors
- Start with 2 retries maximum
- Use exponential backoff for delays
- Monitor retry frequency
- High retry rates indicate underlying issue needing fix
Execution Logs and Tracing
Analyze detailed execution logs to identify root causes of failures.Accessing Execution Logs
View step-by-step execution details for any task. See Task Executions for comprehensive monitoring guidance.1
Navigate to Task
Go to Tasks page and select failed or problematic task execution.
2
View Execution Timeline
See complete workflow execution with each node’s status, duration, and output.
3
Inspect Failed Step
Click failed node to view error message, input data received, and attempted output.
4
Review Error Details
Examine error type, stack trace (for integration errors), and contextual information.
Common Error Patterns
Common Error Patterns
Data Format Errors:
- Symptom: Node expecting JSON receives text
- Solution: Check upstream node output format, update extraction prompt
- Symptom: Node fails with “field X is required”
- Solution: Verify upstream extraction captured all fields, check input data completeness
- Symptom: 401 Unauthorized from API calls
- Solution: Verify integration credentials in Integrations page, refresh tokens if expired
- Symptom: Node exceeds execution time limit
- Solution: Optimize prompt complexity, reduce input size, or enable auto-retry
- Symptom: Silent failure or truncation errors
- Solution: Reduce input size, summarize before processing, or split into smaller chunks
Debugging Workflow
Debugging Workflow
Systematic Debugging Approach:
-
Reproduce Error:
- Identify exact input causing failure
- Rerun task to confirm error persists
- Note if error is consistent or intermittent
-
Isolate Failed Step:
- Identify which node failed
- Review input that node received
- Check if input matches expected format
-
Test in Isolation:
- Use Test Tool with same input
- Verify error reproduces
- Try variations to find working input
-
Identify Root Cause:
- Prompt issue (unclear instructions)
- Integration issue (API error, auth failure)
- Data issue (malformed input, missing fields)
- Configuration issue (wrong tool selected, invalid parameters)
-
Apply Fix:
- Update prompt, configuration, or integration
- Test fix with same input
- Verify resolution
-
Validate:
- Rerun original failed task (see Rerunning Tasks)
- Run test dataset with similar cases (see Test Datasets)
- Monitor for recurrence in Task Executions
Advanced Debugging with LangFuse
Integrated observability platform for deep execution tracing and performance analysis. LangFuse Capabilities:- Complete execution traces across all nodes
- Token usage and cost tracking
- Latency analysis per step
- Model behavior insights
- Error pattern identification
- Navigate to agent configuration (see Agent Settings)
- Select “Learning” or “Observability” section
- View LangFuse dashboard for agent executions
- Filter by timeframe, status, or specific runs
Using LangFuse for Debugging
Using LangFuse for Debugging
Performance Debugging:
- Identify slow nodes causing delays
- Analyze token usage for cost optimization
- Compare execution times across test runs
- Review exact prompts sent to model
- See model responses and reasoning
- Identify prompt inconsistencies
- Track how model handles edge cases
- Trace error through full execution chain
- See context available at failure point
- Identify data quality issues
- Verify integration call details
LangFuse Best Practices
LangFuse Best Practices
Regular Monitoring:
- Review LangFuse traces for critical agents weekly
- Check for anomalies in execution patterns
- Monitor cost trends
- Always check LangFuse for persistent failures
- Compare successful vs failed execution traces
- Look for subtle input differences causing issues
- Identify nodes with excessive token usage
- Find opportunities to reduce latency
- Optimize prompt efficiency
Sandbox and Testing Environments
Isolated testing environments for safe experimentation without affecting production. Sandbox Benefits:- Test integration changes safely
- Validate workflow modifications
- Experiment with prompts
- Train team members
- Demo features to stakeholders
- E2E Environment: Full workflow testing with test data before production deployment (configure in Agent Settings)
- Production Environment: Live agent processing real tasks
Using Sandbox Effectively
Using Sandbox Effectively
Pre-Deployment Testing:
- Make changes in E2E environment
- Run test datasets to validate
- Review execution logs for errors
- Compare metrics to production baseline
- Deploy to production if tests pass
- Set up test API credentials for sandbox
- Use test accounts for external services
- Validate end-to-end data flow
- Check error handling with simulated failures
- Safe environment for team training
- Demo new features to stakeholders
- Experiment with new capabilities
- No risk to production data
Best Practices
Debug Systematically
Debug Systematically
Start Simple:
- Test individual nodes before full workflow
- Use Test Tool to isolate issues
- Verify one change at a time
- Record what worked and what didn’t
- Note error messages and resolutions
- Build knowledge base for team
- Use diverse test cases
- Include edge cases
- Validate with production-like data
Leverage Error Handling Strategically
Leverage Error Handling Strategically
Continue on Error:
- Only for truly optional steps
- Ensure downstream handles missing data
- Monitor failure rates
- For transient errors only
- Limit retry attempts to 2-3
- Track retry frequency to spot systematic issues
Monitor and Improve
Monitor and Improve
Regular Reviews:
- Check execution logs weekly
- Review LangFuse traces for patterns
- Identify recurring error types
- Fix high-frequency errors
- Optimize slow nodes
- Update prompts based on failures
- Improve error handling coverage
- Add error handling to critical paths
- Create fallback workflows
- Implement human review for edge cases
- Test disaster recovery scenarios