水准导入、沉降导入优化

This commit is contained in:
lhx
2025-11-16 10:39:30 +08:00
parent 2bd03f8ac8
commit cef473bda7
3 changed files with 251 additions and 117 deletions

View File

@@ -59,7 +59,9 @@ class LevelDataService(BaseService[LevelData]):
def batch_import_level_data(self, db: Session, data: List) -> Dict[str, Any]:
"""
批量导入水准数据根据期数ID和线路编码判断是否重复重复数据改为更新操作
批量导入水准数据 - 性能优化版
使用批量查询和批量操作,大幅提升导入速度
根据期数ID和线路编码判断是否重复重复数据改为更新操作
支持事务回滚,失败时重试一次
"""
import logging
@@ -70,6 +72,16 @@ class LevelDataService(BaseService[LevelData]):
failed_count = 0
failed_items = []
if total_count == 0:
return {
'success': False,
'message': '导入数据不能为空',
'total_count': 0,
'success_count': 0,
'failed_count': 0,
'failed_items': []
}
for attempt in range(2): # 最多重试1次
try:
db.begin()
@@ -77,49 +89,120 @@ class LevelDataService(BaseService[LevelData]):
failed_count = 0
failed_items = []
# ===== 性能优化1批量查询沉降数据IN查询 =====
nyid_list = list(set(str(item.get('NYID')) for item in data if item.get('NYID')))
logger.info(f"Checking {len(nyid_list)} unique NYIDs in settlement data")
settlements = db.query(SettlementData).filter(SettlementData.NYID.in_(nyid_list)).all()
settlement_map = {s.NYID: s for s in settlements}
missing_nyids = set(nyid_list) - set(settlement_map.keys())
if missing_nyids:
db.rollback()
return {
'success': False,
'message': f'以下期数在沉降表中不存在: {list(missing_nyids)}',
'total_count': total_count,
'success_count': 0,
'failed_count': total_count,
'failed_items': []
}
# ===== 性能优化2批量查询现有水准数据IN查询 =====
# 构建 (NYID, linecode) 组合键来查找重复数据
existing_data = db.query(LevelData).filter(
LevelData.NYID.in_(nyid_list)
).all()
# 使用组合键创建查找表key = f"{NYID}_{linecode}"
existing_map = {
f"{item.NYID}_{item.linecode}": item
for item in existing_data
}
logger.info(f"Found {len(existing_data)} existing level records")
# ===== 性能优化3批量处理插入和更新 =====
to_update = []
to_insert = []
for item_data in data:
try:
nyid = str(item_data.get('NYID'))
linecode = item_data.get('linecode')
# 判断期数id沉降数据是否存在
settlement = self._check_settlement_exists(db, item_data.get('NYID'))
if not settlement:
logger.error(f"Settlement {item_data.get('NYID')} not found")
raise Exception(f"Settlement {item_data.get('NYID')} not found")
# 构建组合键
key = f"{nyid}_{linecode}"
level_data = self.get_by_nyid_and_linecode(
db,
# item_data.get('linecode'),
nyid=item_data.get('NYID')
)
if level_data:
# 更新操作
level_data.benchmarkids = item_data.get('benchmarkids')
level_data.wsphigh = item_data.get('wsphigh')
level_data.mtype = item_data.get('mtype')
level_data.createDate = item_data.get('createDate')
logger.info(f"Updated level data: {item_data.get('linecode')}-{item_data.get('NYID')}")
else:
# 新增操作
level_data = LevelData(
linecode=item_data.get('linecode'),
benchmarkids=item_data.get('benchmarkids'),
wsphigh=item_data.get('wsphigh'),
mtype=item_data.get('mtype'),
NYID=item_data.get('NYID'),
createDate=item_data.get('createDate')
)
db.add(level_data)
logger.info(f"Created level data: {item_data.get('linecode')}-{item_data.get('NYID')}")
if key in existing_map:
# 记录需要更新的数据
existing_item = existing_map[key]
to_update.append((existing_item, item_data))
else:
# 记录需要插入的数据
to_insert.append(item_data)
success_count += 1
except Exception as e:
failed_count += 1
failed_items.append({
'data': item_data,
'error': str(e)
})
logger.error(f"Failed to process level data {item_data.get('linecode')}-{item_data.get('NYID')}: {str(e)}")
raise e
# ===== 执行批量更新 =====
if to_update:
logger.info(f"Updating {len(to_update)} existing records")
for existing_item, item_data in to_update:
try:
existing_item.benchmarkids = item_data.get('benchmarkids')
existing_item.wsphigh = item_data.get('wsphigh')
existing_item.mtype = item_data.get('mtype')
existing_item.createDate = item_data.get('createDate')
success_count += 1
except Exception as e:
failed_count += 1
failed_items.append({
'data': item_data,
'error': f'更新失败: {str(e)}'
})
logger.error(f"Failed to update level data: {str(e)}")
raise e
# ===== 执行批量插入 =====
if to_insert:
logger.info(f"Inserting {len(to_insert)} new records")
# 分批插入每批500条避免SQL过长
batch_size = 500
for i in range(0, len(to_insert), batch_size):
batch = to_insert[i:i + batch_size]
try:
level_data_list = [
LevelData(
linecode=item.get('linecode'),
benchmarkids=item.get('benchmarkids'),
wsphigh=item.get('wsphigh'),
mtype=item.get('mtype'),
NYID=str(item.get('NYID')),
createDate=item.get('createDate')
)
for item in batch
]
db.add_all(level_data_list)
success_count += len(batch)
logger.info(f"Inserted batch {i//batch_size + 1}: {len(batch)} records")
except Exception as e:
failed_count += len(batch)
failed_items.extend([
{
'data': item,
'error': f'插入失败: {str(e)}'
}
for item in batch
])
logger.error(f"Failed to insert batch: {str(e)}")
raise e
# 如果有失败记录,不提交事务
if failed_items:
db.rollback()
return {
'success': False,
'message': f'批量导入失败: {len(failed_items)}条记录处理失败',
'total_count': total_count,
'success_count': success_count,
'failed_count': failed_count,
'failed_items': failed_items
}
db.commit()
logger.info(f"Batch import level data completed. Success: {success_count}, Failed: {failed_count}")