dingo
dingo copied to clipboard
Dingo: A Comprehensive AI Data Quality Evaluation Tool
``` PYTHONPATH="${PWD}" python3 examples/core/score.py Traceback (most recent call last): File "/Users//workspace/dingo/examples/core/score.py", line 5, in from dingo.model.llm.llm_text_quality_model_base import LLMTextQualityModelBase File "/Users//workspace/dingo/dingo/model/llm/llm_text_quality_model_base.py", line 4, in from dingo.model.llm.base_openai import BaseOpenAI File "/Users//workspace/dingo/dingo/model/llm/base_openai.py", line...
没有web可视化界面操作支持吗? 有web可视化界面支持就好了
接口请求内容包括但不限于文件、文件格式、检测规则; 接口返回内容包括但不限于概要报告和详细报告的数据。 感谢实现
When using LLM metrics for evaluation, users should not need to explicitly configure the scenario metrics: https://github.com/DataEval/dingo/blob/dev/examples/llm_and_rule/llm_and_rule_mix.py#L17
The results were inconsistent between the two evaluations using the inference model classification
The results were inconsistent when the inference model of DeepSeek-R1 was used for classification evaluation, and the DeepSeek-Chat model did not have this problem. What is the reason for this?