数据处理基础

This commit is contained in:
lhx
2025-09-28 15:17:09 +08:00
parent 247bb71d55
commit f89f2eef37
16 changed files with 1314 additions and 0 deletions

109
app/services/checkpoint.py Normal file
View File

@@ -0,0 +1,109 @@
from sqlalchemy.orm import Session
from typing import List, Optional, Dict, Any
from ..models.checkpoint import Checkpoint
from .base import BaseService
class CheckpointService(BaseService[Checkpoint]):
def __init__(self):
super().__init__(Checkpoint)
def get_by_section_id(self, db: Session, section_id: str) -> List[Checkpoint]:
"""根据断面ID获取观测点"""
return self.get_by_field(db, "section_id", section_id)
def get_by_point_id(self, db: Session, point_id: str) -> Optional[Checkpoint]:
"""根据观测点ID获取观测点"""
checkpoints = self.get_by_field(db, "point_id", point_id)
return checkpoints[0] if checkpoints else None
def search_checkpoints(self, db: Session,
aname: Optional[str] = None,
section_id: Optional[str] = None,
point_id: Optional[str] = None) -> List[Checkpoint]:
"""根据多个条件搜索观测点"""
conditions = {}
if aname is not None:
conditions["aname"] = aname
if section_id is not None:
conditions["section_id"] = section_id
if point_id is not None:
conditions["point_id"] = point_id
return self.search_by_conditions(db, conditions)
def batch_import_checkpoints(self, db: Session, data: List) -> Dict[str, Any]:
"""
批量导入观测点数据根据观测点ID判断是否重复重复数据改为更新操作
支持事务回滚,失败时重试一次
"""
import logging
logger = logging.getLogger(__name__)
total_count = len(data)
success_count = 0
failed_count = 0
failed_items = []
for attempt in range(2): # 最多重试1次
try:
db.begin()
success_count = 0
failed_count = 0
failed_items = []
for item_data in data:
try:
checkpoint = self.get_by_point_id(db, item_data.get('point_id'))
if checkpoint:
# 更新操作
checkpoint.aname = item_data.get('aname')
checkpoint.section_id = item_data.get('section_id')
checkpoint.burial_date = item_data.get('burial_date')
logger.info(f"Updated checkpoint: {item_data.get('point_id')}")
else:
# 新增操作
checkpoint = Checkpoint(
point_id=item_data.get('point_id'),
aname=item_data.get('aname'),
section_id=item_data.get('section_id'),
burial_date=item_data.get('burial_date')
)
db.add(checkpoint)
logger.info(f"Created checkpoint: {item_data.get('point_id')}")
success_count += 1
except Exception as e:
failed_count += 1
failed_items.append({
'data': item_data,
'error': str(e)
})
logger.error(f"Failed to process checkpoint {item_data.get('point_id')}: {str(e)}")
raise e
db.commit()
logger.info(f"Batch import checkpoints completed. Success: {success_count}, Failed: {failed_count}")
break
except Exception as e:
db.rollback()
logger.warning(f"Batch import attempt {attempt + 1} failed: {str(e)}")
if attempt == 1: # 最后一次重试失败
logger.error("Batch import checkpoints failed after retries")
return {
'success': False,
'message': f'批量导入失败: {str(e)}',
'total_count': total_count,
'success_count': 0,
'failed_count': total_count,
'failed_items': failed_items
}
return {
'success': failed_count == 0,
'message': '批量导入完成' if failed_count == 0 else f'部分导入失败',
'total_count': total_count,
'success_count': success_count,
'failed_count': failed_count,
'failed_items': failed_items
}