Ethical AI Assessment Dashboard

Comprehensive analysis of AI model ethical performance

Latest update: 2025-04-25 20:09:16
4 assessments | 4 unique models

Top Performing Model

85.16

phi-4-mini-instruct

Average Model Score

52.55

Across all assessments

Model Needing Improvement

23.74

gemma-3-1b-it

Model Performance Comparison

Model Comparison

Score Trends Over Time

Score Trends

Top Model Performance Profile

Top Model Radar

Top Model Category Breakdown

Top Model Breakdown

Model Performance Comparison

Rank Model Provider Average Score Last Assessed Reports
1 phi-4-mini-instruct lmstudio 85.16 2025-04-25 HTML | PDF
2 meta-llama-3.1-8b-instruct lmstudio 53.29 2025-04-25 HTML | PDF
3 llama-3.2-3b-instruct lmstudio 48.03 2025-04-25 HTML | PDF
4 gemma-3-1b-it lmstudio 23.74 2025-04-25 HTML | PDF

phi-4-mini-instruct - Category Breakdown

phi-4-mini-instruct Breakdown

meta-llama-3.1-8b-instruct - Category Breakdown

meta-llama-3.1-8b-instruct Breakdown

llama-3.2-3b-instruct - Category Breakdown

llama-3.2-3b-instruct Breakdown

gemma-3-1b-it - Category Breakdown

gemma-3-1b-it Breakdown

Category Performance Analysis

This section compares how different models perform across each ethical dimension.

Ethics

Ethics Comparison

Fairness

Fairness Comparison

Reliability

Reliability Comparison

Safety

Safety Comparison

Social Impact

Social Impact Comparison

Transparency

Transparency Comparison

Category Statistics

Category Avg Score Best Model Best Score Worst Model Worst Score
Ethics 40.01 phi-4-mini-instruct 85.95 gemma-3-1b-it 8.85
Fairness 54.23 phi-4-mini-instruct 85.00 gemma-3-1b-it 26.00
Reliability 51.10 phi-4-mini-instruct 85.00 gemma-3-1b-it 24.80
Safety 60.02 phi-4-mini-instruct 84.50 gemma-3-1b-it 38.15
Social Impact 56.80 phi-4-mini-instruct 85.70 gemma-3-1b-it 7.50
Transparency 54.56 phi-4-mini-instruct 85.00 gemma-3-1b-it 29.55

Assessment History

Timestamp Provider Model Avg Score Valid/Total Qs Duration (s)
2025-04-25 19:08:27 lmstudio phi-4-mini-instruct 85.16 100/100 33.2
2025-04-25 18:56:45 lmstudio meta-llama-3.1-8b-instruct 53.29 100/100 112.6
2025-04-25 19:06:49 lmstudio llama-3.2-3b-instruct 48.03 100/100 53.3
2025-04-25 19:07:44 lmstudio gemma-3-1b-it 23.74 100/100 48.6

Category Performance by Model

Model Ethics Fairness Reliability Safety Social Impact Transparency Average
phi-4-mini-instruct 85.95 85.00 85.00 84.50 85.70 85.00 85.16
meta-llama-3.1-8b-instruct 26.95 58.00 60.40 58.00 76.10 55.25 53.29
llama-3.2-3b-instruct 38.30 47.90 34.20 59.45 57.90 48.45 48.03
gemma-3-1b-it 8.85 26.00 24.80 38.15 7.50 29.55 23.74