From 71aa3d5fae53144a59651ba7ffafda7d22d372d0 Mon Sep 17 00:00:00 2001 From: "gayoun.park" Date: Fri, 16 Jan 2026 15:07:28 +0900 Subject: [PATCH 1/3] =?UTF-8?q?=EC=B6=94=EB=A1=A0=EC=A2=85=EB=A3=8C=20?= =?UTF-8?q?=EC=8B=9C=20geom=20=EB=8D=B0=EC=9D=B4=ED=84=B0=20=EB=84=A3?= =?UTF-8?q?=EA=B8=B0,=20geom-list=20=EC=BF=BC=EB=A6=AC=20=EC=88=98?= =?UTF-8?q?=EC=A0=95?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .../inference/dto/InferenceDetailDto.java | 44 ++++++++--------- .../service/InferenceResultService.java | 4 ++ .../core/InferenceResultCoreService.java | 4 ++ .../InferenceResultRepositoryCustom.java | 3 ++ .../InferenceResultRepositoryImpl.java | 12 +++++ .../MapSheetLearnRepositoryImpl.java | 47 +++++++------------ 6 files changed, 61 insertions(+), 53 deletions(-) diff --git a/src/main/java/com/kamco/cd/kamcoback/inference/dto/InferenceDetailDto.java b/src/main/java/com/kamco/cd/kamcoback/inference/dto/InferenceDetailDto.java index 1bb42407..2a22fe61 100644 --- a/src/main/java/com/kamco/cd/kamcoback/inference/dto/InferenceDetailDto.java +++ b/src/main/java/com/kamco/cd/kamcoback/inference/dto/InferenceDetailDto.java @@ -2,9 +2,6 @@ package com.kamco.cd.kamcoback.inference.dto; import com.fasterxml.jackson.annotation.JsonIgnore; import com.fasterxml.jackson.annotation.JsonInclude; -import com.fasterxml.jackson.core.JsonProcessingException; -import com.fasterxml.jackson.databind.JsonNode; -import com.fasterxml.jackson.databind.ObjectMapper; import com.kamco.cd.kamcoback.common.enums.DetectionClassification; import com.kamco.cd.kamcoback.common.utils.interfaces.JsonFormatDttm; import com.kamco.cd.kamcoback.inference.dto.InferenceResultDto.DetectOption; @@ -307,10 +304,11 @@ public class InferenceDetailDto { Double classAfterProb; Long mapSheetNum; String mapSheetName; - @JsonIgnore String gemoStr; - @JsonIgnore String geomCenterStr; - JsonNode gemo; - JsonNode geomCenter; + + // @JsonIgnore String gemoStr; + // @JsonIgnore String geomCenterStr; + // JsonNode gemo; + // JsonNode geomCenter; public Geom( UUID uuid, @@ -323,9 +321,7 @@ public class InferenceDetailDto { String classAfterCd, Double classAfterProb, Long mapSheetNum, - String mapSheetName, - String gemoStr, - String geomCenterStr) { + String mapSheetName) { this.uuid = uuid; this.uid = uid; this.compareYyyy = compareYyyy; @@ -339,20 +335,20 @@ public class InferenceDetailDto { this.classAfterProb = classAfterProb; this.mapSheetNum = mapSheetNum; this.mapSheetName = mapSheetName; - this.gemoStr = gemoStr; - this.geomCenterStr = geomCenterStr; - - ObjectMapper mapper = new ObjectMapper(); - JsonNode geomJson; - JsonNode geomCenterJson; - try { - geomJson = mapper.readTree(gemoStr); - geomCenterJson = mapper.readTree(geomCenterStr); - } catch (JsonProcessingException e) { - throw new RuntimeException(e); - } - this.gemo = geomJson; - this.geomCenter = geomCenterJson; + // this.gemoStr = gemoStr; + // this.geomCenterStr = geomCenterStr; + // + // ObjectMapper mapper = new ObjectMapper(); + // JsonNode geomJson; + // JsonNode geomCenterJson; + // try { + // geomJson = mapper.readTree(gemoStr); + // geomCenterJson = mapper.readTree(geomCenterStr); + // } catch (JsonProcessingException e) { + // throw new RuntimeException(e); + // } + // this.gemo = geomJson; + // this.geomCenter = geomCenterJson; } } diff --git a/src/main/java/com/kamco/cd/kamcoback/inference/service/InferenceResultService.java b/src/main/java/com/kamco/cd/kamcoback/inference/service/InferenceResultService.java index b502c4cb..3586f22b 100644 --- a/src/main/java/com/kamco/cd/kamcoback/inference/service/InferenceResultService.java +++ b/src/main/java/com/kamco/cd/kamcoback/inference/service/InferenceResultService.java @@ -531,5 +531,9 @@ public class InferenceResultService { request.setUuid(dto.getUuid()); request.setUpdateUid(userUtil.getId()); inferenceResultCoreService.update(request); + + // upsertGeomData + Long learnId = inferenceResultCoreService.getInferenceLearnIdByUuid(dto.getUuid()); + inferenceResultCoreService.upsertGeomData(learnId); } } diff --git a/src/main/java/com/kamco/cd/kamcoback/postgres/core/InferenceResultCoreService.java b/src/main/java/com/kamco/cd/kamcoback/postgres/core/InferenceResultCoreService.java index aea7bcbf..16f5926d 100644 --- a/src/main/java/com/kamco/cd/kamcoback/postgres/core/InferenceResultCoreService.java +++ b/src/main/java/com/kamco/cd/kamcoback/postgres/core/InferenceResultCoreService.java @@ -437,4 +437,8 @@ public class InferenceResultCoreService { public void saveFail5k(UUID uuid, List failMapIds, String type) { mapSheetLearn5kRepository.saveFail5k(uuid, failMapIds, type); } + + public Long getInferenceLearnIdByUuid(UUID uuid) { + return inferenceResultRepository.getInferenceLearnIdByUuid(uuid); + } } diff --git a/src/main/java/com/kamco/cd/kamcoback/postgres/repository/Inference/InferenceResultRepositoryCustom.java b/src/main/java/com/kamco/cd/kamcoback/postgres/repository/Inference/InferenceResultRepositoryCustom.java index 022f6298..0d0887c7 100644 --- a/src/main/java/com/kamco/cd/kamcoback/postgres/repository/Inference/InferenceResultRepositoryCustom.java +++ b/src/main/java/com/kamco/cd/kamcoback/postgres/repository/Inference/InferenceResultRepositoryCustom.java @@ -2,6 +2,7 @@ package com.kamco.cd.kamcoback.postgres.repository.Inference; import com.kamco.cd.kamcoback.postgres.entity.MapSheetAnalDataInferenceGeomEntity; import java.util.List; +import java.util.UUID; public interface InferenceResultRepositoryCustom { @@ -24,4 +25,6 @@ public interface InferenceResultRepositoryCustom { int markGeomCreatedByGeoUids(List geoUids); List findGeomEntitiesByDataUid(Long dataUid, int limit); + + Long getInferenceLearnIdByUuid(UUID uuid); } diff --git a/src/main/java/com/kamco/cd/kamcoback/postgres/repository/Inference/InferenceResultRepositoryImpl.java b/src/main/java/com/kamco/cd/kamcoback/postgres/repository/Inference/InferenceResultRepositoryImpl.java index 755bf60e..bfce414c 100644 --- a/src/main/java/com/kamco/cd/kamcoback/postgres/repository/Inference/InferenceResultRepositoryImpl.java +++ b/src/main/java/com/kamco/cd/kamcoback/postgres/repository/Inference/InferenceResultRepositoryImpl.java @@ -1,5 +1,7 @@ package com.kamco.cd.kamcoback.postgres.repository.Inference; +import static com.kamco.cd.kamcoback.postgres.entity.QMapSheetLearnEntity.mapSheetLearnEntity; + import com.kamco.cd.kamcoback.postgres.entity.MapSheetAnalDataInferenceGeomEntity; import com.kamco.cd.kamcoback.postgres.entity.QMapSheetAnalDataInferenceEntity; import com.kamco.cd.kamcoback.postgres.entity.QMapSheetAnalDataInferenceGeomEntity; @@ -8,6 +10,7 @@ import com.querydsl.jpa.impl.JPAQueryFactory; import jakarta.persistence.EntityManager; import java.time.ZonedDateTime; import java.util.List; +import java.util.UUID; import lombok.RequiredArgsConstructor; import org.springframework.stereotype.Repository; @@ -454,4 +457,13 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC .limit(limit) .fetch(); } + + @Override + public Long getInferenceLearnIdByUuid(UUID uuid) { + return queryFactory + .select(mapSheetLearnEntity.id) + .from(mapSheetLearnEntity) + .where(mapSheetLearnEntity.uuid.eq(uuid)) + .fetchOne(); + } } diff --git a/src/main/java/com/kamco/cd/kamcoback/postgres/repository/Inference/MapSheetLearnRepositoryImpl.java b/src/main/java/com/kamco/cd/kamcoback/postgres/repository/Inference/MapSheetLearnRepositoryImpl.java index a0a5f1ab..296c7efa 100644 --- a/src/main/java/com/kamco/cd/kamcoback/postgres/repository/Inference/MapSheetLearnRepositoryImpl.java +++ b/src/main/java/com/kamco/cd/kamcoback/postgres/repository/Inference/MapSheetLearnRepositoryImpl.java @@ -378,19 +378,15 @@ public class MapSheetLearnRepositoryImpl implements MapSheetLearnRepositoryCusto // 기준년도 분류 if (searchGeoReq.getTargetClass() != null && !searchGeoReq.getTargetClass().equals("")) { builder.and( - mapSheetAnalDataInferenceGeomEntity - .classAfterCd - .toLowerCase() - .eq(searchGeoReq.getTargetClass().toLowerCase())); + mapSheetAnalDataInferenceGeomEntity.classAfterCd.eq( + searchGeoReq.getTargetClass().toLowerCase())); } // 비교년도 분류 if (searchGeoReq.getCompareClass() != null && !searchGeoReq.getCompareClass().equals("")) { builder.and( - mapSheetAnalDataInferenceGeomEntity - .classBeforeCd - .toLowerCase() - .eq(searchGeoReq.getCompareClass().toLowerCase())); + mapSheetAnalDataInferenceGeomEntity.classBeforeCd.eq( + searchGeoReq.getCompareClass().toLowerCase())); } // 분석도엽 @@ -414,11 +410,14 @@ public class MapSheetLearnRepositoryImpl implements MapSheetLearnRepositoryCusto mapSheetAnalDataInferenceGeomEntity.classAfterCd, mapSheetAnalDataInferenceGeomEntity.classAfterProb, mapSheetAnalDataInferenceGeomEntity.mapSheetNum, - mapInkx5kEntity.mapidNm, - Expressions.stringTemplate( - "ST_AsGeoJSON({0})", mapSheetAnalDataInferenceGeomEntity.geom), - Expressions.stringTemplate( - "ST_AsGeoJSON({0})", mapSheetAnalDataInferenceGeomEntity.geomCenter))) + mapInkx5kEntity.mapidNm + // Expressions.stringTemplate( + // "ST_AsGeoJSON({0})", + // mapSheetAnalDataInferenceGeomEntity.geom), + // Expressions.stringTemplate( + // "ST_AsGeoJSON({0})", + // mapSheetAnalDataInferenceGeomEntity.geomCenter) + )) .from(mapSheetAnalInferenceEntity) .join(mapSheetAnalDataInferenceGeomEntity) .on( @@ -429,10 +428,9 @@ public class MapSheetLearnRepositoryImpl implements MapSheetLearnRepositoryCusto mapSheetAnalDataInferenceGeomEntity.stage.eq(mapSheetAnalInferenceEntity.stage)) .join(mapInkx5kEntity) .on( - mapSheetAnalDataInferenceGeomEntity - .mapSheetNum - .stringValue() - .eq(mapInkx5kEntity.mapidcdNo)) + mapSheetAnalDataInferenceGeomEntity.mapSheetNum.eq( + Expressions.numberTemplate( + Long.class, "CAST({0} AS long)", mapInkx5kEntity.mapidcdNo))) .where(builder) .offset(pageable.getOffset()) .limit(pageable.getPageSize()) @@ -443,14 +441,6 @@ public class MapSheetLearnRepositoryImpl implements MapSheetLearnRepositoryCusto .select(mapSheetAnalDataInferenceGeomEntity.geoUid) .from(mapSheetAnalInferenceEntity) .join(mapSheetAnalDataInferenceGeomEntity) - .on( - mapSheetAnalDataInferenceGeomEntity.compareYyyy.eq( - mapSheetAnalInferenceEntity.compareYyyy), - mapSheetAnalDataInferenceGeomEntity.targetYyyy.eq( - mapSheetAnalInferenceEntity.targetYyyy), - mapSheetAnalDataInferenceGeomEntity.stage.eq(mapSheetAnalInferenceEntity.stage)) - .from(mapSheetAnalInferenceEntity) - .join(mapSheetAnalDataInferenceGeomEntity) .on( mapSheetAnalDataInferenceGeomEntity.compareYyyy.eq( mapSheetAnalInferenceEntity.compareYyyy), @@ -459,10 +449,9 @@ public class MapSheetLearnRepositoryImpl implements MapSheetLearnRepositoryCusto mapSheetAnalDataInferenceGeomEntity.stage.eq(mapSheetAnalInferenceEntity.stage)) .join(mapInkx5kEntity) .on( - mapSheetAnalDataInferenceGeomEntity - .mapSheetNum - .stringValue() - .eq(mapInkx5kEntity.mapidcdNo)) + mapSheetAnalDataInferenceGeomEntity.mapSheetNum.eq( + Expressions.numberTemplate( + Long.class, "CAST({0} AS long)", mapInkx5kEntity.mapidcdNo))) .where(builder) .fetchCount(); From 33672149299a750176da06798e6ae5621a5c4db4 Mon Sep 17 00:00:00 2001 From: teddy Date: Fri, 16 Jan 2026 15:10:11 +0900 Subject: [PATCH 2/3] =?UTF-8?q?spotless=20=EC=A0=81=EC=9A=A9?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .../core/InferenceResultCoreService.java | 122 +++++++++--------- 1 file changed, 61 insertions(+), 61 deletions(-) diff --git a/src/main/java/com/kamco/cd/kamcoback/postgres/core/InferenceResultCoreService.java b/src/main/java/com/kamco/cd/kamcoback/postgres/core/InferenceResultCoreService.java index aea7bcbf..396cf83a 100644 --- a/src/main/java/com/kamco/cd/kamcoback/postgres/core/InferenceResultCoreService.java +++ b/src/main/java/com/kamco/cd/kamcoback/postgres/core/InferenceResultCoreService.java @@ -108,7 +108,7 @@ public class InferenceResultCoreService { mapSheetLearnEntity.setMapSheetCnt(mapSheetName); mapSheetLearnEntity.setDetectingCnt((long) detectingCnt); mapSheetLearnEntity.setStage( - mapSheetLearnRepository.getLearnStage(req.getCompareYyyy(), req.getTargetYyyy())); + mapSheetLearnRepository.getLearnStage(req.getCompareYyyy(), req.getTargetYyyy())); // learn 테이블 저장 MapSheetLearnEntity savedLearn = mapSheetLearnRepository.save(mapSheetLearnEntity); @@ -170,9 +170,9 @@ public class InferenceResultCoreService { */ public InferenceDetailDto.AnalResSummary getInferenceResultSummary(Long id) { InferenceDetailDto.AnalResSummary summary = - mapSheetAnalDataRepository - .getInferenceResultSummary(id) - .orElseThrow(() -> new EntityNotFoundException("요약정보를 찾을 수 없습니다. " + id)); + mapSheetAnalDataRepository + .getInferenceResultSummary(id) + .orElseThrow(() -> new EntityNotFoundException("요약정보를 찾을 수 없습니다. " + id)); return summary; } @@ -193,7 +193,7 @@ public class InferenceResultCoreService { * @return */ public Page getInferenceResultGeomList( - Long id, InferenceDetailDto.SearchGeoReq searchGeoReq) { + Long id, InferenceDetailDto.SearchGeoReq searchGeoReq) { return mapSheetAnalDataRepository.getInferenceGeomList(id, searchGeoReq); } @@ -205,16 +205,16 @@ public class InferenceResultCoreService { */ @Transactional(readOnly = true) public Page listInferenceResultWithGeom( - @NotNull Long analyId, InferenceDetailDto.SearchGeoReq searchReq) { + @NotNull Long analyId, InferenceDetailDto.SearchGeoReq searchReq) { // 분석 ID 에 해당하는 dataids를 가져온다. List dataIds = - mapSheetAnalDataRepository.listAnalyGeom(analyId).stream() - .mapToLong(MapSheetAnalDataInferenceEntity::getId) - .boxed() - .toList(); + mapSheetAnalDataRepository.listAnalyGeom(analyId).stream() + .mapToLong(MapSheetAnalDataInferenceEntity::getId) + .boxed() + .toList(); // 해당데이터의 폴리곤데이터를 가져온다 Page mapSheetAnalDataGeomEntities = - mapSheetAnalDataRepository.listInferenceResultWithGeom(dataIds, searchReq); + mapSheetAnalDataRepository.listInferenceResultWithGeom(dataIds, searchReq); return mapSheetAnalDataGeomEntities.map(MapSheetAnalDataInferenceGeomEntity::toEntity); } @@ -231,14 +231,14 @@ public class InferenceResultCoreService { @Transactional(readOnly = true) public List listGetScenes5k(Long analyId) { List sceneCodes = - mapSheetAnalDataRepository.listAnalyGeom(analyId).stream() - .mapToLong(MapSheetAnalDataInferenceEntity::getMapSheetNum) - .mapToObj(String::valueOf) - .toList(); + mapSheetAnalDataRepository.listAnalyGeom(analyId).stream() + .mapToLong(MapSheetAnalDataInferenceEntity::getMapSheetNum) + .mapToObj(String::valueOf) + .toList(); return mapInkx5kRepository.listGetScenes5k(sceneCodes).stream() - .map(MapInkx5kEntity::toEntity) - .toList(); + .map(MapInkx5kEntity::toEntity) + .toList(); } /** @@ -246,11 +246,12 @@ public class InferenceResultCoreService { * * @param request 추론 실행 정보 */ + @Transactional public void update(SaveInferenceAiDto request) { MapSheetLearnEntity entity = - mapSheetLearnRepository - .getInferenceResultByUuid(request.getUuid()) - .orElseThrow(EntityNotFoundException::new); + mapSheetLearnRepository + .getInferenceResultByUuid(request.getUuid()) + .orElseThrow(EntityNotFoundException::new); // M1/M2/M3 영역 업데이트 if (request.getType() != null) { @@ -272,49 +273,46 @@ public class InferenceResultCoreService { private void applyModelUpdate(MapSheetLearnEntity entity, SaveInferenceAiDto request) { switch (request.getType()) { - case "M1" -> - applyModelFields( - request, - entity::setM1ModelBatchId, - entity::setM1ModelStartDttm, - entity::setM1ModelEndDttm, - entity::setM1PendingJobs, - entity::setM1RunningJobs, - entity::setM1CompletedJobs, - entity::setM1FailedJobs); - case "M2" -> - applyModelFields( - request, - entity::setM2ModelBatchId, - entity::setM2ModelStartDttm, - entity::setM2ModelEndDttm, - entity::setM2PendingJobs, - entity::setM2RunningJobs, - entity::setM2CompletedJobs, - entity::setM2FailedJobs); - case "M3" -> - applyModelFields( - request, - entity::setM3ModelBatchId, - entity::setM3ModelStartDttm, - entity::setM3ModelEndDttm, - entity::setM3PendingJobs, - entity::setM3RunningJobs, - entity::setM3CompletedJobs, - entity::setM3FailedJobs); + case "M1" -> applyModelFields( + request, + entity::setM1ModelBatchId, + entity::setM1ModelStartDttm, + entity::setM1ModelEndDttm, + entity::setM1PendingJobs, + entity::setM1RunningJobs, + entity::setM1CompletedJobs, + entity::setM1FailedJobs); + case "M2" -> applyModelFields( + request, + entity::setM2ModelBatchId, + entity::setM2ModelStartDttm, + entity::setM2ModelEndDttm, + entity::setM2PendingJobs, + entity::setM2RunningJobs, + entity::setM2CompletedJobs, + entity::setM2FailedJobs); + case "M3" -> applyModelFields( + request, + entity::setM3ModelBatchId, + entity::setM3ModelStartDttm, + entity::setM3ModelEndDttm, + entity::setM3PendingJobs, + entity::setM3RunningJobs, + entity::setM3CompletedJobs, + entity::setM3FailedJobs); default -> throw new IllegalArgumentException("Unknown type: " + request.getType()); } } private void applyModelFields( - SaveInferenceAiDto request, - Consumer setBatchId, - Consumer setStart, - Consumer setEnd, - Consumer setPending, - Consumer setRunning, - Consumer setCompleted, - Consumer setFailed) { + SaveInferenceAiDto request, + Consumer setBatchId, + Consumer setStart, + Consumer setEnd, + Consumer setPending, + Consumer setRunning, + Consumer setCompleted, + Consumer setFailed) { applyIfNotNull(request.getBatchId(), setBatchId); applyIfNotNull(request.getModelStartDttm(), setStart); applyIfNotNull(request.getModelEndDttm(), setEnd); @@ -341,7 +339,7 @@ public class InferenceResultCoreService { public InferenceBatchSheet getInferenceResultByStatus(String status) { MapSheetLearnEntity entity = - mapSheetLearnRepository.getInferenceResultByStatus(status).orElse(null); + mapSheetLearnRepository.getInferenceResultByStatus(status).orElse(null); if (entity == null) { return null; @@ -420,6 +418,7 @@ public class InferenceResultCoreService { * * @param id learn 테이블 id */ + @Transactional public void upsertGeomData(Long id) { Long analId = inferenceResultRepository.upsertGroupsFromMapSheetAnal(id); inferenceResultRepository.upsertGroupsFromInferenceResults(analId); @@ -430,10 +429,11 @@ public class InferenceResultCoreService { /** * 모델별 도엽별 실패여부 저장 * - * @param uuid learn 테이블 uuid + * @param uuid learn 테이블 uuid * @param failMapIds 실패한 도엽 목록 - * @param type 모델타입 + * @param type 모델타입 */ + @Transactional public void saveFail5k(UUID uuid, List failMapIds, String type) { mapSheetLearn5kRepository.saveFail5k(uuid, failMapIds, type); } From 691bf198368b1606de3c911174996e3502c9a480 Mon Sep 17 00:00:00 2001 From: "gayoun.park" Date: Fri, 16 Jan 2026 15:43:21 +0900 Subject: [PATCH 3/3] =?UTF-8?q?=EA=B0=9C=EB=B0=9C=EC=84=9C=EB=B2=84=20show?= =?UTF-8?q?-sql:false=EB=A1=9C=20=EC=88=98=EC=A0=95=20+=20spotless?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .../core/InferenceResultCoreService.java | 119 +++++++++--------- src/main/resources/application-dev.yml | 2 +- 2 files changed, 62 insertions(+), 59 deletions(-) diff --git a/src/main/java/com/kamco/cd/kamcoback/postgres/core/InferenceResultCoreService.java b/src/main/java/com/kamco/cd/kamcoback/postgres/core/InferenceResultCoreService.java index 69c1dd76..662e1897 100644 --- a/src/main/java/com/kamco/cd/kamcoback/postgres/core/InferenceResultCoreService.java +++ b/src/main/java/com/kamco/cd/kamcoback/postgres/core/InferenceResultCoreService.java @@ -108,7 +108,7 @@ public class InferenceResultCoreService { mapSheetLearnEntity.setMapSheetCnt(mapSheetName); mapSheetLearnEntity.setDetectingCnt((long) detectingCnt); mapSheetLearnEntity.setStage( - mapSheetLearnRepository.getLearnStage(req.getCompareYyyy(), req.getTargetYyyy())); + mapSheetLearnRepository.getLearnStage(req.getCompareYyyy(), req.getTargetYyyy())); // learn 테이블 저장 MapSheetLearnEntity savedLearn = mapSheetLearnRepository.save(mapSheetLearnEntity); @@ -170,9 +170,9 @@ public class InferenceResultCoreService { */ public InferenceDetailDto.AnalResSummary getInferenceResultSummary(Long id) { InferenceDetailDto.AnalResSummary summary = - mapSheetAnalDataRepository - .getInferenceResultSummary(id) - .orElseThrow(() -> new EntityNotFoundException("요약정보를 찾을 수 없습니다. " + id)); + mapSheetAnalDataRepository + .getInferenceResultSummary(id) + .orElseThrow(() -> new EntityNotFoundException("요약정보를 찾을 수 없습니다. " + id)); return summary; } @@ -193,7 +193,7 @@ public class InferenceResultCoreService { * @return */ public Page getInferenceResultGeomList( - Long id, InferenceDetailDto.SearchGeoReq searchGeoReq) { + Long id, InferenceDetailDto.SearchGeoReq searchGeoReq) { return mapSheetAnalDataRepository.getInferenceGeomList(id, searchGeoReq); } @@ -205,16 +205,16 @@ public class InferenceResultCoreService { */ @Transactional(readOnly = true) public Page listInferenceResultWithGeom( - @NotNull Long analyId, InferenceDetailDto.SearchGeoReq searchReq) { + @NotNull Long analyId, InferenceDetailDto.SearchGeoReq searchReq) { // 분석 ID 에 해당하는 dataids를 가져온다. List dataIds = - mapSheetAnalDataRepository.listAnalyGeom(analyId).stream() - .mapToLong(MapSheetAnalDataInferenceEntity::getId) - .boxed() - .toList(); + mapSheetAnalDataRepository.listAnalyGeom(analyId).stream() + .mapToLong(MapSheetAnalDataInferenceEntity::getId) + .boxed() + .toList(); // 해당데이터의 폴리곤데이터를 가져온다 Page mapSheetAnalDataGeomEntities = - mapSheetAnalDataRepository.listInferenceResultWithGeom(dataIds, searchReq); + mapSheetAnalDataRepository.listInferenceResultWithGeom(dataIds, searchReq); return mapSheetAnalDataGeomEntities.map(MapSheetAnalDataInferenceGeomEntity::toEntity); } @@ -231,14 +231,14 @@ public class InferenceResultCoreService { @Transactional(readOnly = true) public List listGetScenes5k(Long analyId) { List sceneCodes = - mapSheetAnalDataRepository.listAnalyGeom(analyId).stream() - .mapToLong(MapSheetAnalDataInferenceEntity::getMapSheetNum) - .mapToObj(String::valueOf) - .toList(); + mapSheetAnalDataRepository.listAnalyGeom(analyId).stream() + .mapToLong(MapSheetAnalDataInferenceEntity::getMapSheetNum) + .mapToObj(String::valueOf) + .toList(); return mapInkx5kRepository.listGetScenes5k(sceneCodes).stream() - .map(MapInkx5kEntity::toEntity) - .toList(); + .map(MapInkx5kEntity::toEntity) + .toList(); } /** @@ -249,9 +249,9 @@ public class InferenceResultCoreService { @Transactional public void update(SaveInferenceAiDto request) { MapSheetLearnEntity entity = - mapSheetLearnRepository - .getInferenceResultByUuid(request.getUuid()) - .orElseThrow(EntityNotFoundException::new); + mapSheetLearnRepository + .getInferenceResultByUuid(request.getUuid()) + .orElseThrow(EntityNotFoundException::new); // M1/M2/M3 영역 업데이트 if (request.getType() != null) { @@ -273,46 +273,49 @@ public class InferenceResultCoreService { private void applyModelUpdate(MapSheetLearnEntity entity, SaveInferenceAiDto request) { switch (request.getType()) { - case "M1" -> applyModelFields( - request, - entity::setM1ModelBatchId, - entity::setM1ModelStartDttm, - entity::setM1ModelEndDttm, - entity::setM1PendingJobs, - entity::setM1RunningJobs, - entity::setM1CompletedJobs, - entity::setM1FailedJobs); - case "M2" -> applyModelFields( - request, - entity::setM2ModelBatchId, - entity::setM2ModelStartDttm, - entity::setM2ModelEndDttm, - entity::setM2PendingJobs, - entity::setM2RunningJobs, - entity::setM2CompletedJobs, - entity::setM2FailedJobs); - case "M3" -> applyModelFields( - request, - entity::setM3ModelBatchId, - entity::setM3ModelStartDttm, - entity::setM3ModelEndDttm, - entity::setM3PendingJobs, - entity::setM3RunningJobs, - entity::setM3CompletedJobs, - entity::setM3FailedJobs); + case "M1" -> + applyModelFields( + request, + entity::setM1ModelBatchId, + entity::setM1ModelStartDttm, + entity::setM1ModelEndDttm, + entity::setM1PendingJobs, + entity::setM1RunningJobs, + entity::setM1CompletedJobs, + entity::setM1FailedJobs); + case "M2" -> + applyModelFields( + request, + entity::setM2ModelBatchId, + entity::setM2ModelStartDttm, + entity::setM2ModelEndDttm, + entity::setM2PendingJobs, + entity::setM2RunningJobs, + entity::setM2CompletedJobs, + entity::setM2FailedJobs); + case "M3" -> + applyModelFields( + request, + entity::setM3ModelBatchId, + entity::setM3ModelStartDttm, + entity::setM3ModelEndDttm, + entity::setM3PendingJobs, + entity::setM3RunningJobs, + entity::setM3CompletedJobs, + entity::setM3FailedJobs); default -> throw new IllegalArgumentException("Unknown type: " + request.getType()); } } private void applyModelFields( - SaveInferenceAiDto request, - Consumer setBatchId, - Consumer setStart, - Consumer setEnd, - Consumer setPending, - Consumer setRunning, - Consumer setCompleted, - Consumer setFailed) { + SaveInferenceAiDto request, + Consumer setBatchId, + Consumer setStart, + Consumer setEnd, + Consumer setPending, + Consumer setRunning, + Consumer setCompleted, + Consumer setFailed) { applyIfNotNull(request.getBatchId(), setBatchId); applyIfNotNull(request.getModelStartDttm(), setStart); applyIfNotNull(request.getModelEndDttm(), setEnd); @@ -339,7 +342,7 @@ public class InferenceResultCoreService { public InferenceBatchSheet getInferenceResultByStatus(String status) { MapSheetLearnEntity entity = - mapSheetLearnRepository.getInferenceResultByStatus(status).orElse(null); + mapSheetLearnRepository.getInferenceResultByStatus(status).orElse(null); if (entity == null) { return null; @@ -429,9 +432,9 @@ public class InferenceResultCoreService { /** * 모델별 도엽별 실패여부 저장 * - * @param uuid learn 테이블 uuid + * @param uuid learn 테이블 uuid * @param failMapIds 실패한 도엽 목록 - * @param type 모델타입 + * @param type 모델타입 */ @Transactional public void saveFail5k(UUID uuid, List failMapIds, String type) { diff --git a/src/main/resources/application-dev.yml b/src/main/resources/application-dev.yml index 0cb3c5ae..d7169713 100644 --- a/src/main/resources/application-dev.yml +++ b/src/main/resources/application-dev.yml @@ -4,7 +4,7 @@ spring: on-profile: dev jpa: - show-sql: true + show-sql: false hibernate: ddl-auto: validate properties: