Llama Guard
AIAgentStore 未评级 fingerprint: aiagentstore:llama-guard
3.5 /5.0 B
Llama Guard is a Large Language Model (LLM)-based safeguard developed to ensure safe and appropriate human-AI interactions. It functions by classifying both user inputs and AI-generated outputs to identify and mitigate potential safety risks, such as prompt injections or inappropriate content. The model is instruction-tuned to handle various safety categories and can be customized to align with specific use cases. Llama Guard supports multi-class classification and generates binary decision scores to effectively moderate AI conversations.
🔗 前往使用
📊 六维信任评分
真实性 Authenticity
3.60
赌注 Commitment
3.30
一致性 Consistency
3.80
透明度 Transparency
3.30
📈 近90天评分趋势
🏷️ Badge预览
AgentRisk Badge AgentRisk Badge Flat
[![AgentRisk](https://api.agentrisk.app/v1/badge/aiagentstore:llama-guard?style=for-the-badge)](https://agentrisk.app/a/aiagentstore:llama-guard)

查看完整评分历史与深度分析

在 AgentRisk Dashboard 中查看完整的评分变化、风险事件和详细报告

在Dashboard中查看 →