fix: fix the Data Evaluation Detail page (#154)

* fix: the Data Evaluation Detail Page should show the model used

* fix: fix the time format displayed

* fix: fix the Data Evaluation Detail page
This commit is contained in:
hefanli
2025-12-10 18:35:29 +08:00
committed by GitHub
parent 4ed049ca11
commit 99fd46cb70
6 changed files with 69 additions and 34 deletions

View File

@@ -24,6 +24,7 @@ from app.module.evaluation.schema.prompt_template import PromptTemplateResponse
from app.module.evaluation.service.prompt_template_service import PromptTemplateService
from app.module.evaluation.service.evaluation import EvaluationTaskService
from app.module.shared.schema.common import StandardResponse, TaskStatus
from app.module.system.service.common_service import get_model_by_id
router = APIRouter(
prefix="",
@@ -79,6 +80,10 @@ async def create_evaluation_task(
if existing_task.scalar_one_or_none():
raise HTTPException(status_code=400, detail=f"Evaluation task with name '{request.name}' already exists")
model_config = await get_model_by_id(db, request.eval_config.model_id)
if not model_config:
raise HTTPException(status_code=400, detail=f"Model with id '{request.eval_config.model_id}' not found")
# 创建评估任务
task = EvaluationTask(
id=str(uuid.uuid4()),
@@ -90,7 +95,8 @@ async def create_evaluation_task(
source_name=request.source_name,
eval_prompt=request.eval_prompt,
eval_config=json.dumps({
"model_id": request.eval_config.model_id,
"modelId": request.eval_config.model_id,
"modelName": model_config.model_name,
"dimensions": request.eval_config.dimensions,
}),
status=TaskStatus.PENDING.value,

View File

@@ -43,7 +43,7 @@ class EvaluationExecutor:
async def execute(self):
eval_config = json.loads(self.task.eval_config)
model_config = await get_model_by_id(self.db, eval_config.get("model_id"))
model_config = await get_model_by_id(self.db, eval_config.get("modelId"))
semaphore = asyncio.Semaphore(10)
files = (await self.db.execute(
select(EvaluationFile).where(EvaluationFile.task_id == self.task.id)