Overview
Rotavision’s trust scoring system provides a unified framework for measuring AI system trustworthiness across multiple dimensions. Each dimension is scored from 0-100, with higher scores indicating greater trustworthiness.Trust Dimensions
Fairness
Measures equitable treatment across protected groups
Reliability
Measures consistency and stability of predictions
Explainability
Measures how well predictions can be understood
Privacy
Measures data protection and privacy preservation
Overall Trust Score
The overall trust score is a weighted combination of individual dimensions:- Fairness: 30%
- Reliability: 30%
- Explainability: 25%
- Privacy: 15%
Weights can be customized based on your industry and regulatory requirements. Financial services often increase fairness weight, while healthcare may prioritize explainability.
Fairness Metrics
Vishwas calculates fairness using industry-standard metrics:| Metric | Description | Threshold |
|---|---|---|
| Demographic Parity | Equal positive prediction rates across groups | ≥ 0.80 |
| Equalized Odds | Equal TPR and FPR across groups | ≥ 0.80 |
| Calibration | Predicted probabilities match actual outcomes | ≥ 0.80 |
| Individual Fairness | Similar individuals receive similar predictions | ≥ 0.75 |
| Counterfactual Fairness | Predictions unchanged if protected attributes changed | ≥ 0.80 |
Calculating Demographic Parity
Multi-Group Fairness
For attributes with multiple groups (e.g., states, languages), Rotavision calculates:- Pairwise ratios between all group pairs
- Minimum ratio as the fairness bound
- Weighted average based on group sizes
Reliability Metrics
Guardian monitors reliability through:| Metric | Description | Alert Threshold |
|---|---|---|
| Prediction Drift | KL divergence of output distribution | > 0.1 |
| Feature Drift | PSI of input features | > 0.2 |
| Accuracy Decay | Drop in monitored accuracy metric | > 5% |
| Latency P99 | 99th percentile response time | > SLA |
| Error Rate | Percentage of failed predictions | > 1% |
Drift Detection
Explainability Scores
Measured through explanation quality metrics:| Metric | Description |
|---|---|
| Faithfulness | How accurately explanations reflect model behavior |
| Stability | Consistency of explanations for similar inputs |
| Comprehensibility | Human-understandable explanation complexity |
| Completeness | Coverage of important features in explanations |
Score Interpretation
90-100: Excellent
90-100: Excellent
Model meets highest trust standards. Suitable for high-stakes decisions with minimal additional oversight.
75-89: Good
75-89: Good
Model is generally trustworthy. Consider targeted improvements for specific dimensions below threshold.
60-74: Needs Improvement
60-74: Needs Improvement
Significant trust gaps exist. Recommend human oversight and remediation plan before production use.
Below 60: Critical
Below 60: Critical
Model does not meet minimum trust requirements. Do not deploy without major improvements.
Industry Benchmarks
Based on our analysis of enterprise AI deployments in India:| Industry | Average Trust Score | Top Quartile |
|---|---|---|
| Banking & Finance | 72 | 85+ |
| Insurance | 68 | 82+ |
| Healthcare | 65 | 80+ |
| E-commerce | 70 | 83+ |
| Telecom | 74 | 86+ |
Regulatory Alignment
Rotavision trust scores map to regulatory requirements:| Regulation | Relevant Dimensions |
|---|---|
| RBI AI Guidelines | Fairness, Explainability |
| DPDP Act 2023 | Privacy, Transparency |
| SEBI ML Circular | Reliability, Auditability |
| IRDAI AI Guidelines | Fairness, Explainability |

