LogoLogo
👨‍💻 API Reference📣 Release Notes📺 Request a Demo
  • Introduction to Fiddler
    • Monitor, Analyze, and Protect your ML Models and Gen AI Applications
  • Fiddler Doc Chatbot
  • First Steps
    • Getting Started With Fiddler Guardrails
    • Getting Started with LLM Monitoring
    • Getting Started with ML Model Observability
  • Tutorials & Quick Starts
    • LLM and GenAI
      • LLM Evaluation - Compare Outputs
      • LLM Monitoring - Simple
    • Fiddler Free Guardrails
      • Guardrails - Quick Start Guide
      • Guardrails - Faithfulness
      • Guardrails - Safety
      • Guardrails FAQ
    • ML Observability
      • ML Monitoring - Simple
      • ML Monitoring - NLP Inputs
      • ML Monitoring - Class Imbalance
      • ML Monitoring - Model Versions
      • ML Monitoring - Ranking
      • ML Monitoring - Regression
      • ML Monitoring - Feature Impact
      • ML Monitoring - CV Inputs
  • Glossary
    • Product Concepts
      • Baseline
      • Custom Metric
      • Data Drift
      • Embedding Visualization
      • Fiddler Guardrails
      • Fiddler Trust Service
      • LLM and GenAI Observability
      • Metric
      • Model Drift
      • Model Performance
      • ML Observability
      • Trust Score
  • Product Guide
    • LLM Application Monitoring & Protection
      • LLM-Based Metrics
      • Embedding Visualizations for LLM Monitoring and Analysis
      • Selecting Enrichments
      • Enrichments (Private Preview)
      • Guardrails for Proactive Application Protection
    • Optimize Your ML Models and LLMs with Fiddler's Comprehensive Monitoring
      • Alerts
      • Package-Based Alerts (Private Preview)
      • Class Imbalanced Data
      • Enhance ML and LLM Insights with Custom Metrics
      • Data Drift: Monitor Model Performance Changes with Fiddler's Insights
      • Ensuring Data Integrity in ML Models And LLMs
      • Embedding Visualization With UMAP
      • Fiddler Query Language
      • Model Versions
      • How to Effectively Use the Monitoring Chart UI
      • Performance Tracking
      • Model Segments: Analyze Cohorts for Performance Insights and Bias Detection
      • Statistics
      • Monitoring ML Model and LLM Traffic
      • Vector Monitoring
    • Enhance Model Insights with Fiddler's Slice and Explain
      • Events Table in RCA
      • Feature Analytics Creation
      • Metric Card Creation
      • Performance Charts Creation
      • Performance Charts Visualization
    • Master AI Monitoring: Create, Customize, and Compare Dashboards
      • Creating Dashboards
      • Dashboard Interactions
      • Dashboard Utilities
    • Adding and Editing Models in the UI
      • Model Editor UI
      • Model Schema Editing Guide
    • Fairness
    • Explainability
      • Model: Artifacts, Package, Surrogate
      • Global Explainability: Visualize Feature Impact and Importance in Fiddler
      • Point Explainability
      • Flexible Model Deployment
        • On Prem Manual Flexible Model Deployment XAI
  • Technical Reference
    • Python Client API Reference
    • Python Client Guides
      • Installation and Setup
      • Model Onboarding
        • Create a Project and Onboard a Model for Observation
        • Model Task Types
        • Customizing your Model Schema
        • Specifying Custom Missing Value Representations
      • Publishing Inference Data
        • Creating a Baseline Dataset
        • Publishing Batches Of Events
        • Publishing Ranking Events
        • Streaming Live Events
        • Updating Already Published Events
        • Deleting Events From Fiddler
      • Creating and Managing Alerts
      • Explainability Examples
        • Adding a Surrogate Model
        • Uploading Model Artifacts
        • Updating Model Artifacts
        • ML Framework Examples
          • Scikit Learn
          • Tensorflow HDF5
          • Tensorflow Savedmodel
          • Xgboost
        • Model Task Examples
          • Binary Classification
          • Multiclass Classification
          • Regression
          • Uploading A Ranking Model Artifact
    • Integrations
      • Data Pipeline Integrations
        • Airflow Integration
        • BigQuery Integration
        • Integration With S3
        • Kafka Integration
        • Sagemaker Integration
        • Snowflake Integration
      • ML Platform Integrations
        • Integrate Fiddler with Databricks for Model Monitoring and Explainability
        • Datadog Integration
        • ML Flow Integration
      • Alerting Integrations
        • PagerDuty Integration
    • Comprehensive REST API Reference
      • Projects REST API Guide
      • Model REST API Guide
      • File Upload REST API Guide
      • Custom Metrics REST API Guide
      • Segments REST API Guide
      • Baselines REST API Guide
      • Jobs REST API Guide
      • Alert Rules REST API Guide
      • Environments REST API Guide
      • Explainability REST API Guide
      • Server Info REST API Guide
      • Events REST API Guide
      • Fiddler Trust Service REST API Guide
    • Fiddler Free Guardrails Documentation
  • Configuration Guide
    • Authentication & Authorization
      • Adding Users
      • Overview of Role-Based Access Control
      • Email Authentication
      • Okta OIDC SSO Integration
      • Azure AD OIDC SSO Integration
      • Ping Identity SAML SSO Integration
      • Mapping LDAP Groups & Users to Fiddler Teams
    • Application Settings
    • Supported Browsers
  • History
    • Release Notes
    • Python Client History
    • Compatibility Matrix
    • Product Maturity Definitions
Powered by GitBook

© 2024 Fiddler Labs, Inc.

On this page
  • How Fiddler Uses Guardrails
  • Why Fiddler Guardrails Is Important
  • Types of Guardrails Protection
  • Challenges
  • Fiddler Guardrails Implementation Guide
  • Frequently Asked Questions
  • Related Terms
  • Related Resources

Was this helpful?

  1. Glossary
  2. Product Concepts

Fiddler Guardrails

PreviousEmbedding VisualizationNextFiddler Trust Service

Last updated 20 days ago

Was this helpful?

is a real-time content safety solution that evaluates and filters potentially harmful outputs from large language models (LLMs) before they reach end users. It serves as a protective layer between LLM systems and their users, ensuring that generated content adheres to organizational policies and safety standards.

Built on Fiddler's infrastructure, Fiddler Guardrails leverages purpose-built models optimized for efficient content evaluation. The system processes LLM outputs in real-time, detecting problematic content across multiple safety dimensions including harmfulness, toxicity, illegal activity, bias, and more. When violations are detected, Guardrails can either filter out the content entirely or provide detailed explanations of the specific policies that were violated.

Fiddler Guardrails is available both as a standalone service for organizations seeking content safety without full monitoring capabilities and as an integrated component within Fiddler's comprehensive LLM observability platform.

How Fiddler Uses Guardrails

Fiddler Guardrails integrates with the broader Fiddler platform to provide comprehensive protection and visibility for LLM and GenAI applications. As part of Fiddler's end-to-end LLM governance solution, Guardrails helps organizations maintain control over their generative AI deployments by ensuring outputs meet safety standards before reaching users.

When integrated with Fiddler's monitoring capabilities, Guardrails contributes to a complete approach where potentially harmful content is both blocked in real-time and analyzed for patterns that might indicate systemic issues. This dual approach enables organizations to maintain strong protections while continuously improving their LLM systems.

Fiddler offers Guardrails through a simple API that can be integrated into existing LLM application workflows, allowing organizations to implement protection without significant architectural changes.

Why Fiddler Guardrails Is Important

As organizations increasingly deploy generative AI applications across their operations, ensuring the safety and appropriateness of LLM outputs becomes a critical governance concern. Fiddler Guardrails addresses this need by providing real-time protection against potentially harmful, toxic, or inappropriate content generation.

Without robust guardrails in place, organizations face significant risks from LLM deployments, including reputational damage, compliance violations, and potential harm to users. By implementing Fiddler Guardrails, organizations can confidently deploy generative AI with protections that minimize these risks.

The importance of Guardrails extends beyond individual content filtering to enable trustworthy, responsible AI deployment at scale across an organization.

  • Risk Mitigation: Guardrails prevents harmful, toxic, or inappropriate content from reaching end users, protecting both users and the organization from potential negative impacts of unsafe LLM outputs.

  • Compliance Support: By enforcing content policies consistently, Guardrails helps organizations meet regulatory requirements and internal governance standards for responsible AI use.

  • Deployment Confidence: With protective measures in place, organizations can more confidently deploy LLM applications across a broader range of use cases and user groups.

  • Brand Protection: By filtering inappropriate content before it reaches customers, Guardrails helps protect brand reputation and maintain trust in AI-powered services.

  • Operational Efficiency: Real-time protection reduces the need for human review of all LLM outputs, allowing teams to focus on improving models rather than constantly monitoring for problematic content.

  • Educational Feedback: When content is filtered, Guardrails provides detailed explanations about policy violations, helping developers understand and address recurring issues in their prompting strategies or model configurations.

  • Customizable Protection: Organizations can tailor protection levels to their specific needs and risk tolerance, ensuring appropriate safeguards without unnecessarily restrictive filtering.

Types of Guardrails Protection

  • Safety Guardrails: Core protections against harmful, toxic, illegal, or objectionable content across multiple categories including violence, hate speech, explicit content, and illegal activities.

  • Custom Policy Guardrails: Organization-specific policy enforcement that can be configured to reflect particular industry requirements, company values, or audience sensitivities.

  • Standalone Guardrails Service: The Guardrails API offered as an independent service for organizations seeking content safety without full monitoring capabilities.

  • Integrated Platform Guardrails: Guardrails functionality within the comprehensive Fiddler observability platform, working alongside monitoring features for complete LLM governance.

  • Multi-level Protection: Tiered protection options allowing organizations to set different filtering thresholds for different applications or user contexts.

Challenges

Implementing effective content safety for LLM applications presents several challenges that Fiddler Guardrails is designed to address.

  • Latency Management: Adding protective layers can potentially slow response times, a challenge Guardrails addresses through efficient purpose-built models optimized for evaluation speed.

  • Nuanced Content Evaluation: Distinguishing between genuinely harmful content and benign discussions of sensitive topics requires sophisticated evaluation capabilities beyond simple keyword filtering.

  • False Positive Balance: Setting appropriate thresholds that protect against harmful content without excessive blocking of legitimate outputs requires careful calibration.

  • Context Awareness: Properly evaluating content requires understanding the broader context of the conversation, not just analyzing isolated responses.

  • Policy Customization: Different organizations have varying standards for acceptable content, requiring flexible guardrail systems that can be tailored to specific needs.

  • Multilingual Support: Ensuring consistent protection across content in different languages presents challenges in evaluation consistency.

  • Evolving Threat Landscape: As LLM capabilities advance and new exploitation techniques emerge, guardrail systems must continuously update to address novel risks.

Fiddler Guardrails Implementation Guide

  1. Define Your Content Safety Requirements

    • Identify which safety dimensions are most important for your LLM use cases.

    • Determine appropriate filtering thresholds based on your user base and risk tolerance.

  2. Choose Deployment Approach

    • Decide between standalone Guardrails API or integrated platform approach based on your needs.

    • Select between cloud-hosted or on-premises deployment options.

  3. Integrate Guardrails API

    • Implement API calls in your application flow to route LLM outputs through Guardrails before display.

    • Set up appropriate error handling for cases where content is filtered.

  4. Configure Safety Policies

    • Set appropriate thresholds for different safety dimensions.

    • Configure custom policies if needed for organization-specific requirements.

  5. Test and Refine

    • Verify protection effectiveness with representative test cases.

    • Refine thresholds based on observed false positive/negative rates.

  6. Monitor and Improve

    • Track guardrail filtering patterns to identify recurring issues.

    • Use insights to improve prompt engineering and LLM configuration.

Frequently Asked Questions

Q: How does Fiddler Guardrails differ from prompt engineering for safety?

While prompt engineering attempts to elicit safe behavior through careful input design, Guardrails provides a dedicated protection layer that evaluates outputs regardless of prompt quality. This approach is more robust as it doesn't rely on the effectiveness of prompts alone and can catch unsafe responses even when prompt constraints fail.

Q: What types of content can Fiddler Guardrails detect and filter?

Fiddler Guardrails can detect and filter content across multiple safety dimensions including but not limited to harmful content, toxicity, hate speech, violence, sexual content, illegal activities, bias, profanity, and content faithfulness.

Q: How does Guardrails impact response latency?

Fiddler Guardrails is designed to minimize latency impact through highly optimized models specifically built for efficient content evaluation. While there is some processing overhead, it's typically measured in milliseconds rather than seconds, and significantly faster than using general-purpose LLMs for evaluation.

Q: Can I customize the safety policies to fit my organization's needs?

Yes, Fiddler Guardrails allows for customization of safety policies and thresholds to align with specific organizational requirements, industry regulations, and risk tolerance levels.

Q: Is Guardrails available as a standalone service or only as part of the full Fiddler platform?

Fiddler Guardrails is available both as a standalone service for organizations that only need content safety capabilities and as an integrated component within the comprehensive Fiddler LLM observability platform.

Related Terms

Related Resources

Fiddler Guardrails
Trust Service
Fiddler Trust Service
Trust Score
Getting Started with Fiddler Guardrails
Guardrails in the Fiddler Platform
Guardrails FAQ
Fiddler Trust Service Overview
Fast Safety Enrichments