We actively maintain and provide security updates for the following versions:
| Version | Supported |
|---|---|
| Latest | ✅ |
| < Latest | ❌ |
We take the security of the MultiAgentEval seriously. If you discover a security vulnerability, please follow these steps:
- Do NOT create a public GitHub issue for security vulnerabilities
- Send an email to [email protected] with the subject line: "Security Vulnerability Report"
- Include the following information:
- Description of the vulnerability
- Steps to reproduce the issue
- Potential impact assessment
- Any suggested fixes (if available)
- Your contact information for follow-up
- Acknowledgment: We will acknowledge receipt of your report within 48 hours
- Initial Assessment: We will provide an initial assessment within 5 business days
- Regular Updates: We will keep you informed of our progress every 7 days until resolution
- Resolution Timeline: We aim to resolve critical vulnerabilities within 30 days
We follow responsible disclosure practices:
- We will work with you to understand and resolve the issue
- We will credit you in our security advisory (unless you prefer to remain anonymous)
- We ask that you do not publicly disclose the vulnerability until we have released a fix
- Input Validation: Always validate and sanitize inputs, especially when executing AI agent code
- Sandboxing: Ensure proper isolation when running untrusted agent code
- Dependency Management:
- Keep dependencies up to date
- Use
pip-auditor similar tools to check for known vulnerabilities - Pin dependency versions in requirements files
- Code Execution:
- Use secure sandboxes for agent code execution
- Implement proper timeouts and resource limits
- Never execute agent code with elevated privileges
- Data Handling:
- Don't log sensitive information from agent interactions
- Sanitize outputs before storing or displaying
- Be cautious with file system access in evaluations
- API Keys and Secrets:
- Never commit API keys, tokens, or secrets to the repository
- Use environment variables or secure secret management
- Provide clear documentation on required environment variables
- Configuration:
- Use secure defaults in configuration files
- Validate configuration parameters
- Document security implications of configuration options
- Isolated Environment: Run evaluations in isolated environments (containers, VMs)
- Network Access: Be cautious about agent network access during evaluations
- File System: Limit agent file system access to necessary directories only
- Resource Limits: Set appropriate CPU, memory, and time limits
- API Security: Use secure connections (HTTPS) for AI model APIs
- Authentication: Properly secure API keys and authentication tokens
- Rate Limiting: Implement appropriate rate limiting to prevent abuse
- Data Privacy: Be aware of data sent to external AI services
- This framework executes AI-generated code, which inherently carries security risks
- Always run in isolated, sandboxed environments
- Review agent outputs before execution in production environments
- AI agents may attempt to install or import external packages
- Use virtual environments and dependency isolation
- Monitor and control package installation during evaluations
- Evaluation results may contain sensitive information
- Implement proper access controls for evaluation data
- Consider data retention and deletion policies
- Assessment: We assess the severity and impact of reported vulnerabilities
- Fix Development: We develop and test fixes in private branches
- Coordinated Release: We coordinate the release of fixes with security advisories
- Communication: We communicate security updates through:
- GitHub Security Advisories
- Release notes
- Email notifications to maintainers
We use the following severity levels:
- Critical: Immediate threat to user security or data
- High: Significant security risk requiring prompt attention
- Medium: Important security issue with moderate impact
- Low: Minor security concern with limited impact
This project aims to follow security best practices including:
- OWASP security guidelines
- Secure coding practices for Python
- Container security best practices (when applicable)
- API security standards
For security-related questions or concerns:
- Security issues: [email protected]
- General questions: [email protected]
- Maintainer: [email protected]
We appreciate the security research community's efforts in keeping open source projects secure. Contributors who responsibly disclose security vulnerabilities will be acknowledged in our security advisories.
This security policy is reviewed and updated regularly to ensure it remains current with best practices and project needs.