1 instantiation of ViolenceEvaluator
Microsoft.Extensions.AI.Evaluation.Integration.Tests (1)
SafetyEvaluatorTests.cs (1)
65IEvaluator violenceEvaluator = new ViolenceEvaluator();
11 references to ViolenceEvaluator
Microsoft.Extensions.AI.Evaluation.Integration.Tests (5)
SafetyEvaluatorTests.cs (5)
288Assert.True(result.TryGet(ViolenceEvaluator.ViolenceMetricName, out NumericMetric? _)); 324Assert.True(result.TryGet(ViolenceEvaluator.ViolenceMetricName, out NumericMetric? _)); 373Assert.True(result.TryGet(ViolenceEvaluator.ViolenceMetricName, out NumericMetric? _)); 435Assert.True(result.TryGet(ViolenceEvaluator.ViolenceMetricName, out NumericMetric? _)); 548Assert.True(result.TryGet(ViolenceEvaluator.ViolenceMetricName, out NumericMetric? _));
Microsoft.Extensions.AI.Evaluation.Safety (6)
ContentHarmEvaluator.cs (3)
20/// <see cref="ViolenceEvaluator"/>, <see cref="SelfHarmEvaluator"/> and <see cref="SexualEvaluator"/> which can be 30/// <see cref="ViolenceEvaluator.ViolenceMetricName"/>, <see cref="SelfHarmEvaluator.SelfHarmMetricName"/> and 41["violence"] = ViolenceEvaluator.ViolenceMetricName,
ViolenceEvaluator.cs (3)
14/// <see cref="ViolenceEvaluator"/> returns a <see cref="NumericMetric"/> with a value between 0 and 7, with 0 18/// Note that <see cref="ViolenceEvaluator"/> can detect harmful content present within both image and text based 28/// <see cref="ViolenceEvaluator"/>.