| layout | page |
|---|---|
| title | Direct Framework Guide |
| description | Simple, fast API calls for basic AI agent interactions |
The Direct Framework is KubeAgentic's default execution mode, designed for fast, straightforward interactions with AI agents. It provides simple API calls directly to the LLM provider without complex workflow orchestration.
✅ Perfect for:
- Chat bots and conversational agents
- Simple Q&A systems
- Basic tool usage scenarios
- High-throughput applications
- Lightweight agents with minimal resource requirements
- Straightforward request-response patterns
❌ Not ideal for:
- Complex multi-step reasoning
- Stateful conversation workflows
- Advanced tool orchestration
- Conditional logic between tools
- Long-running task workflows
- Response Time: ~100-500ms
- Resource Usage: Low CPU and memory footprint
- Scalability: Excellent horizontal scaling
- Complexity: Minimal configuration required
Here's a basic Direct Framework agent configuration:
apiVersion: ai.example.com/v1
kind: Agent
metadata:
name: simple-chatbot
spec:
framework: direct
provider: openai
model: gpt-4
systemPrompt: "You are a helpful customer service assistant."
apiSecretRef:
name: openai-secret
key: api-key
replicas: 2
resources:
requests:
memory: "128Mi"
cpu: "100m"
limits:
memory: "256Mi"
cpu: "200m"The Direct Framework requires minimal configuration. Just specify your provider, model, and system prompt.
Direct API calls to LLM providers ensure the fastest possible response times for your applications.
Minimal overhead means your agents use fewer resources, allowing for better cost optimization.
Simple request-response patterns make it easy to debug and monitor your agents.
Even with the Direct Framework, you can still integrate tools:
spec:
framework: direct
provider: openai
model: gpt-4
tools:
- name: calculator
description: "Basic math operations"
endpoint: "http://calculator-service:8080/calculate"
- name: weather
description: "Get current weather information"
endpoint: "http://weather-service:8080/weather"The Direct Framework provides built-in monitoring capabilities:
- Health Checks: Automatic health monitoring
- Metrics: Request/response metrics via Prometheus
- Logging: Structured logging for debugging
- Tracing: Request tracing for performance analysis
- Keep it Simple: Use Direct Framework for straightforward use cases
- Optimize Prompts: Well-crafted system prompts improve response quality
- Monitor Performance: Track response times and resource usage
- Scale Horizontally: Add more replicas for high-throughput scenarios
- Use Appropriate Models: Choose models based on your performance and cost requirements
If you're currently using a more complex framework and want to simplify:
- Evaluate Complexity: Determine if you really need complex workflows
- Simplify Logic: Move complex logic to your application layer
- Test Performance: Ensure Direct Framework meets your performance needs
- Update Configuration: Modify your agent specs to use
framework: direct
Slow Response Times
- Check your network connectivity to the LLM provider
- Verify your API key has sufficient quota
- Consider using a faster model or region
High Resource Usage
- Review your resource limits and requests
- Check for memory leaks in your application
- Monitor CPU usage patterns
Tool Integration Issues
- Verify tool endpoints are accessible
- Check tool response formats
- Ensure proper error handling
- View Examples - See real-world Direct Framework implementations
- API Reference - Detailed configuration options
- LangGraph Framework - For complex workflows
- Local Testing - Test your agents locally