diff --git a/src/main/java/com/kamco/cd/kamcoback/scheduler/service/MapSheetInferenceJobService.java b/src/main/java/com/kamco/cd/kamcoback/scheduler/service/MapSheetInferenceJobService.java index 59cdebec..91eee199 100644 --- a/src/main/java/com/kamco/cd/kamcoback/scheduler/service/MapSheetInferenceJobService.java +++ b/src/main/java/com/kamco/cd/kamcoback/scheduler/service/MapSheetInferenceJobService.java @@ -15,13 +15,15 @@ import com.kamco.cd.kamcoback.postgres.core.InferenceResultCoreService; import com.kamco.cd.kamcoback.scheduler.dto.BatchStatusDto; import com.kamco.cd.kamcoback.scheduler.dto.JobStatusDto; import java.time.ZonedDateTime; -import java.util.ArrayList; import java.util.HashSet; import java.util.List; import java.util.Map; +import java.util.Objects; import java.util.Optional; import java.util.Set; import java.util.UUID; +import java.util.stream.Collectors; +import java.util.stream.Stream; import lombok.RequiredArgsConstructor; import lombok.extern.log4j.Log4j2; import org.springframework.beans.factory.annotation.Value; @@ -244,15 +246,16 @@ public class MapSheetInferenceJobService { inferenceResultCoreService.upsertGeomData(sheet.getId()); // 추론 종료일때 shp 파일 생성 - List batchIds = new ArrayList<>(); - batchIds.add(sheet.getM1BatchId()); - batchIds.add(sheet.getM2BatchId()); - batchIds.add(sheet.getM3BatchId()); + List batchIds = + Stream.of(sheet.getM1BatchId(), sheet.getM2BatchId(), sheet.getM3BatchId()) + .filter(Objects::nonNull) + .distinct() + .toList(); - String batchId = sheet.getM1BatchId() + "," + sheet.getM2BatchId() + "," + sheet.getM3BatchId(); + String batchIdStr = batchIds.stream().map(String::valueOf).collect(Collectors.joining(",")); // shp 파일 비동기 생성 - shpPipelineService.runPipeline(jarPath, datasetDir, batchId, sheet.getUid()); + shpPipelineService.runPipeline(jarPath, datasetDir, batchIdStr, sheet.getUid()); } /**