spotless 적용

This commit is contained in:
2026-01-16 12:31:24 +09:00
parent fd8807b1d6
commit 520d1f975d

View File

@@ -19,15 +19,19 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
private final EntityManager em; private final EntityManager em;
private final QMapSheetAnalInferenceEntity inferenceEntity = private final QMapSheetAnalInferenceEntity inferenceEntity =
QMapSheetAnalInferenceEntity.mapSheetAnalInferenceEntity; QMapSheetAnalInferenceEntity.mapSheetAnalInferenceEntity;
/** tb_map_sheet_anal_data_inference */ /**
* tb_map_sheet_anal_data_inference
*/
private final QMapSheetAnalDataInferenceEntity inferenceDataEntity = private final QMapSheetAnalDataInferenceEntity inferenceDataEntity =
QMapSheetAnalDataInferenceEntity.mapSheetAnalDataInferenceEntity; QMapSheetAnalDataInferenceEntity.mapSheetAnalDataInferenceEntity;
/** tb_map_sheet_anal_data_inference_geom */ /**
* tb_map_sheet_anal_data_inference_geom
*/
private final QMapSheetAnalDataInferenceGeomEntity inferenceGeomEntity = private final QMapSheetAnalDataInferenceGeomEntity inferenceGeomEntity =
QMapSheetAnalDataInferenceGeomEntity.mapSheetAnalDataInferenceGeomEntity; QMapSheetAnalDataInferenceGeomEntity.mapSheetAnalDataInferenceGeomEntity;
// =============================== // ===============================
// Upsert (Native only) // Upsert (Native only)
@@ -36,7 +40,7 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
@Override @Override
public Long upsertGroupsFromMapSheetAnal(Long id) { public Long upsertGroupsFromMapSheetAnal(Long id) {
String sql = String sql =
""" """
INSERT INTO tb_map_sheet_anal_inference ( INSERT INTO tb_map_sheet_anal_inference (
stage, stage,
compare_yyyy, compare_yyyy,
@@ -77,8 +81,7 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
} }
/** /**
* inference_results 테이블을 기준으로 분석 데이터 단위(stage, compare_yyyy, target_yyyy, map_sheet_num)를 * inference_results 테이블을 기준으로 분석 데이터 단위(stage, compare_yyyy, target_yyyy, map_sheet_num)를 생성/갱신한다.
* 생성/갱신한다.
* *
* <p>- 최초 생성 시 file_created_yn = false - detecting_cnt는 inference_results 건수 기준 * <p>- 최초 생성 시 file_created_yn = false - detecting_cnt는 inference_results 건수 기준
* *
@@ -88,7 +91,7 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
public void upsertGroupsFromInferenceResults(Long analId) { public void upsertGroupsFromInferenceResults(Long analId) {
String sql = String sql =
""" """
INSERT INTO tb_map_sheet_anal_data_inference ( INSERT INTO tb_map_sheet_anal_data_inference (
anal_uid, anal_uid,
compare_yyyy, compare_yyyy,
@@ -144,7 +147,7 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
public void upsertGeomsFromInferenceResults(Long analUid) { public void upsertGeomsFromInferenceResults(Long analUid) {
String sql = String sql =
""" """
INSERT INTO tb_map_sheet_anal_data_inference_geom ( INSERT INTO tb_map_sheet_anal_data_inference_geom (
result_uid, result_uid,
stage, stage,
@@ -184,9 +187,9 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
SELECT SELECT
r.uid AS result_uid, r.uid AS result_uid,
msadi.stage, msadi.stage,
r.cd_prob, r.after_p as cd_prob,
r.input1 AS compare_yyyy, msl.compare_yyyy,
r.input2 AS target_yyyy, msl.target_yyyy,
CASE CASE
WHEN r.map_id ~ '^[0-9]+$' THEN r.map_id::bigint WHEN r.map_id ~ '^[0-9]+$' THEN r.map_id::bigint
ELSE NULL ELSE NULL
@@ -239,7 +242,7 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
public void upsertSttcFromInferenceResults(Long analUid) { public void upsertSttcFromInferenceResults(Long analUid) {
String sql = String sql =
""" """
INSERT INTO tb_map_sheet_anal_sttc INSERT INTO tb_map_sheet_anal_sttc
( (
compare_yyyy compare_yyyy
@@ -324,22 +327,22 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
public List<Long> findPendingDataUids(int limit, Long learnId) { public List<Long> findPendingDataUids(int limit, Long learnId) {
return queryFactory return queryFactory
.select(inferenceDataEntity.id) .select(inferenceDataEntity.id)
.from(inferenceEntity) .from(inferenceEntity)
.innerJoin(inferenceDataEntity) .innerJoin(inferenceDataEntity)
.on(inferenceEntity.id.eq(inferenceDataEntity.analUid)) .on(inferenceEntity.id.eq(inferenceDataEntity.analUid))
.where( .where(
inferenceEntity inferenceEntity
.learnId .learnId
.eq(learnId) .eq(learnId)
.and( .and(
inferenceDataEntity inferenceDataEntity
.fileCreatedYn .fileCreatedYn
.isFalse() .isFalse()
.or(inferenceDataEntity.fileCreatedYn.isNull()))) .or(inferenceDataEntity.fileCreatedYn.isNull())))
.orderBy(inferenceDataEntity.id.asc()) .orderBy(inferenceDataEntity.id.asc())
.limit(limit) .limit(limit)
.fetch(); .fetch();
} }
// =============================== // ===============================
@@ -358,13 +361,13 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
ZonedDateTime now = ZonedDateTime.now(); ZonedDateTime now = ZonedDateTime.now();
return (int) return (int)
queryFactory queryFactory
.update(inferenceDataEntity) .update(inferenceDataEntity)
.set(inferenceDataEntity.fileCreatedYn, false) .set(inferenceDataEntity.fileCreatedYn, false)
.set(inferenceDataEntity.fileCreatedDttm, (ZonedDateTime) null) .set(inferenceDataEntity.fileCreatedDttm, (ZonedDateTime) null)
.set(inferenceDataEntity.updatedDttm, now) .set(inferenceDataEntity.updatedDttm, now)
.where(inferenceDataEntity.id.eq(dataUid)) .where(inferenceDataEntity.id.eq(dataUid))
.execute(); .execute();
} }
/** /**
@@ -377,13 +380,13 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
ZonedDateTime now = ZonedDateTime.now(); ZonedDateTime now = ZonedDateTime.now();
return (int) return (int)
queryFactory queryFactory
.update(inferenceDataEntity) .update(inferenceDataEntity)
.set(inferenceDataEntity.fileCreatedYn, true) .set(inferenceDataEntity.fileCreatedYn, true)
.set(inferenceDataEntity.fileCreatedDttm, now) .set(inferenceDataEntity.fileCreatedDttm, now)
.set(inferenceDataEntity.updatedDttm, now) .set(inferenceDataEntity.updatedDttm, now)
.where(inferenceDataEntity.id.eq(dataUid)) .where(inferenceDataEntity.id.eq(dataUid))
.execute(); .execute();
} }
/** /**
@@ -396,13 +399,13 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
ZonedDateTime now = ZonedDateTime.now(); ZonedDateTime now = ZonedDateTime.now();
return (int) return (int)
queryFactory queryFactory
.update(inferenceGeomEntity) .update(inferenceGeomEntity)
.set(inferenceGeomEntity.fileCreatedYn, false) .set(inferenceGeomEntity.fileCreatedYn, false)
.set(inferenceGeomEntity.fileCreatedDttm, (ZonedDateTime) null) .set(inferenceGeomEntity.fileCreatedDttm, (ZonedDateTime) null)
.set(inferenceGeomEntity.updatedDttm, now) .set(inferenceGeomEntity.updatedDttm, now)
.where(inferenceGeomEntity.dataUid.eq(dataUid)) .where(inferenceGeomEntity.dataUid.eq(dataUid))
.execute(); .execute();
} }
/** /**
@@ -420,13 +423,13 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
ZonedDateTime now = ZonedDateTime.now(); ZonedDateTime now = ZonedDateTime.now();
return (int) return (int)
queryFactory queryFactory
.update(inferenceGeomEntity) .update(inferenceGeomEntity)
.set(inferenceGeomEntity.fileCreatedYn, true) .set(inferenceGeomEntity.fileCreatedYn, true)
.set(inferenceGeomEntity.fileCreatedDttm, now) .set(inferenceGeomEntity.fileCreatedDttm, now)
.set(inferenceGeomEntity.updatedDttm, now) .set(inferenceGeomEntity.updatedDttm, now)
.where(inferenceGeomEntity.geoUid.in(geoUids)) .where(inferenceGeomEntity.geoUid.in(geoUids))
.execute(); .execute();
} }
// =============================== // ===============================
@@ -440,18 +443,18 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
*/ */
@Override @Override
public List<MapSheetAnalDataInferenceGeomEntity> findGeomEntitiesByDataUid( public List<MapSheetAnalDataInferenceGeomEntity> findGeomEntitiesByDataUid(
Long dataUid, int limit) { Long dataUid, int limit) {
return queryFactory return queryFactory
.selectFrom(inferenceGeomEntity) .selectFrom(inferenceGeomEntity)
.where( .where(
inferenceGeomEntity.dataUid.eq(dataUid), inferenceGeomEntity.dataUid.eq(dataUid),
inferenceGeomEntity.geom.isNotNull(), inferenceGeomEntity.geom.isNotNull(),
inferenceGeomEntity inferenceGeomEntity
.fileCreatedYn .fileCreatedYn
.isFalse() .isFalse()
.or(inferenceGeomEntity.fileCreatedYn.isNull())) .or(inferenceGeomEntity.fileCreatedYn.isNull()))
.orderBy(inferenceGeomEntity.geoUid.asc()) .orderBy(inferenceGeomEntity.geoUid.asc())
.limit(limit) .limit(limit)
.fetch(); .fetch();
} }
} }