Merge branch 'feat/infer_dev_260107' of https://kamco.gitea.gs.dabeeo.com/dabeeo/kamco-dabeeo-backoffice into feat/infer_dev_260107
This commit is contained in:
@@ -476,4 +476,14 @@ public class InferenceResultCoreService {
|
||||
dto.setUid(entity.getUid());
|
||||
return dto;
|
||||
}
|
||||
|
||||
/**
|
||||
* 실행중인 추론 도엽명 목록
|
||||
*
|
||||
* @param uuid 추론 실행중인 uuid
|
||||
* @return
|
||||
*/
|
||||
public List<String> getInferenceRunMapId(UUID uuid) {
|
||||
return mapSheetLearn5kRepository.getInferenceRunMapId(uuid);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -27,9 +27,8 @@ public class LabelAllocateCoreService {
|
||||
|
||||
private final LabelAllocateRepository labelAllocateRepository;
|
||||
|
||||
public List<AllocateInfoDto> fetchNextIds(
|
||||
Long lastId, Long batchSize, Integer compareYyyy, Integer targetYyyy, Integer stage) {
|
||||
return labelAllocateRepository.fetchNextIds(lastId, batchSize, compareYyyy, targetYyyy, stage);
|
||||
public List<AllocateInfoDto> fetchNextIds(Long lastId, Long batchSize, UUID uuid) {
|
||||
return labelAllocateRepository.fetchNextIds(lastId, batchSize, uuid);
|
||||
}
|
||||
|
||||
public void assignOwner(List<AllocateInfoDto> ids, String userId, Long analUid) {
|
||||
@@ -42,8 +41,8 @@ public class LabelAllocateCoreService {
|
||||
.toList();
|
||||
}
|
||||
|
||||
public Long findLabelUnAssignedCnt(Integer stage, Integer compareYyyy, Integer targetYyyy) {
|
||||
return labelAllocateRepository.findLabelUnAssignedCnt(stage, compareYyyy, targetYyyy);
|
||||
public Long findLabelUnAssignedCnt(UUID uuid) {
|
||||
return labelAllocateRepository.findLabelUnAssignedCnt(uuid);
|
||||
}
|
||||
|
||||
public void assignInspector(UUID assignmentUid, String inspectorUid) {
|
||||
@@ -144,8 +143,8 @@ public class LabelAllocateCoreService {
|
||||
return detail;
|
||||
}
|
||||
|
||||
public Long findMapSheetAnalInferenceUid(Integer compareYyyy, Integer targetYyyy, Integer stage) {
|
||||
return labelAllocateRepository.findMapSheetAnalInferenceUid(compareYyyy, targetYyyy, stage);
|
||||
public Long findMapSheetAnalInferenceUid(UUID uuid) {
|
||||
return labelAllocateRepository.findMapSheetAnalInferenceUid(uuid);
|
||||
}
|
||||
|
||||
public void insertInspector(Long analUid, String inspector) {
|
||||
@@ -227,4 +226,12 @@ public class LabelAllocateCoreService {
|
||||
public Page<WorkHistoryDto> workReviewerHistoryList(searchReq searchReq, String userId) {
|
||||
return labelAllocateRepository.workReviewerHistoryList(searchReq, userId);
|
||||
}
|
||||
|
||||
public void updateAnalInferenceMngState(UUID uuid, String status) {
|
||||
labelAllocateRepository.updateAnalInferenceMngState(uuid, status);
|
||||
}
|
||||
|
||||
public Long findLabelingIngProcessCnt() {
|
||||
return labelAllocateRepository.findLabelingIngProcessCnt();
|
||||
}
|
||||
}
|
||||
|
||||
@@ -4,6 +4,7 @@ import com.kamco.cd.kamcoback.postgres.repository.trainingdata.TrainingDataLabel
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.DefaultPaging;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.DetailRes;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.InferenceIdInfo;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.LabelFeatureRequest.LabelProperties;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.LabelingGeometryInfo;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.LabelingListDto;
|
||||
@@ -29,7 +30,7 @@ public class TrainingDataLabelCoreService {
|
||||
return trainingDataLabelRepository.findLabelingAssignedGeom(assignmentUid);
|
||||
}
|
||||
|
||||
public Long findLabelingAssignmentGeoUid(String assignmentUid) {
|
||||
public InferenceIdInfo findLabelingAssignmentGeoUid(String assignmentUid) {
|
||||
return trainingDataLabelRepository.findLabelingAssignmentGeoUid(assignmentUid);
|
||||
}
|
||||
|
||||
@@ -89,4 +90,8 @@ public class TrainingDataLabelCoreService {
|
||||
String mapSheetNum, Integer beforeYear, Integer afterYear) {
|
||||
return trainingDataLabelRepository.getCogImageUrl(mapSheetNum, beforeYear, afterYear);
|
||||
}
|
||||
|
||||
public void updateAnalInferenceMngState(Long analUid, String status) {
|
||||
trainingDataLabelRepository.updateAnalInferenceMngState(analUid, status);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,6 +1,7 @@
|
||||
package com.kamco.cd.kamcoback.postgres.core;
|
||||
|
||||
import com.kamco.cd.kamcoback.postgres.repository.trainingdata.TrainingDataReviewRepository;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.InferenceIdInfo;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataReviewDto;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataReviewDto.DefaultPaging;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataReviewDto.DetailRes;
|
||||
@@ -29,7 +30,7 @@ public class TrainingDataReviewCoreService {
|
||||
return trainingDataReviewRepository.findReviewAssignedGeom(operatorUid);
|
||||
}
|
||||
|
||||
public Long findReviewOperatorGeoUid(String operatorUid) {
|
||||
public InferenceIdInfo findReviewOperatorGeoUid(String operatorUid) {
|
||||
return trainingDataReviewRepository.findReviewOperatorGeoUid(operatorUid);
|
||||
}
|
||||
|
||||
@@ -89,4 +90,8 @@ public class TrainingDataReviewCoreService {
|
||||
String mapSheetNum, Integer beforeYear, Integer afterYear) {
|
||||
return trainingDataReviewRepository.getCogImageUrl(mapSheetNum, beforeYear, afterYear);
|
||||
}
|
||||
|
||||
public void updateAnalInferenceMngState(Long analUid, String status) {
|
||||
trainingDataReviewRepository.updateAnalInferenceMngState(analUid, status);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -37,4 +37,8 @@ public class TrainingDataReviewJobCoreService {
|
||||
public void lockInspectors(Long analUid, List<String> reviewerIds) {
|
||||
trainingDataReviewJobRepository.lockInspectors(analUid, reviewerIds);
|
||||
}
|
||||
|
||||
public void updateGeomUidTestState(List<Long> geomUids) {
|
||||
trainingDataReviewJobRepository.updateGeomUidTestState(geomUids);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -30,12 +30,6 @@ public class InferenceResultsTestingEntity {
|
||||
@Column(name = "map_id", length = Integer.MAX_VALUE)
|
||||
private String mapId;
|
||||
|
||||
@Column(name = "source_1", length = Integer.MAX_VALUE)
|
||||
private String source1;
|
||||
|
||||
@Column(name = "source_2", length = Integer.MAX_VALUE)
|
||||
private String source2;
|
||||
|
||||
@Column(name = "model_version", length = Integer.MAX_VALUE)
|
||||
private String modelVersion;
|
||||
|
||||
@@ -60,9 +54,6 @@ public class InferenceResultsTestingEntity {
|
||||
@Column(name = "area")
|
||||
private Double area;
|
||||
|
||||
@Column(name = "cd_prob")
|
||||
private Double cdProb;
|
||||
|
||||
@Column(name = "before_c", length = Integer.MAX_VALUE)
|
||||
private String beforeC;
|
||||
|
||||
@@ -75,12 +66,6 @@ public class InferenceResultsTestingEntity {
|
||||
@Column(name = "after_p")
|
||||
private Double afterP;
|
||||
|
||||
@Column(name = "input1")
|
||||
private Long input1;
|
||||
|
||||
@Column(name = "input2")
|
||||
private Long input2;
|
||||
|
||||
@Id
|
||||
@NotNull
|
||||
@ColumnDefault("nextval('inference_results_testing_seq_seq')")
|
||||
|
||||
@@ -187,7 +187,7 @@ public class MapSheetLearnEntity {
|
||||
private int m3FailedJobs = 0;
|
||||
|
||||
@Column(name = "uid", nullable = false)
|
||||
private String uid = UUID.randomUUID().toString().replace("-", "");
|
||||
private String uid = UUID.randomUUID().toString().replace("-", "").toUpperCase();
|
||||
|
||||
public InferenceResultDto.ResultList toDto() {
|
||||
return new InferenceResultDto.ResultList(
|
||||
|
||||
@@ -52,7 +52,8 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
|
||||
m1_model_batch_id,
|
||||
m2_model_batch_id,
|
||||
m3_model_batch_id,
|
||||
learn_id
|
||||
learn_id,
|
||||
anal_state
|
||||
)
|
||||
SELECT
|
||||
r.stage,
|
||||
@@ -64,7 +65,8 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
|
||||
r.m1_model_batch_id,
|
||||
r.m2_model_batch_id,
|
||||
r.m3_model_batch_id,
|
||||
r.id
|
||||
r.id,
|
||||
LabelMngState.PENDING.getId()
|
||||
FROM tb_map_sheet_learn r
|
||||
WHERE r.id = :id
|
||||
ON CONFLICT (stage, compare_yyyy, target_yyyy)
|
||||
@@ -217,11 +219,15 @@ public class InferenceResultRepositoryImpl implements InferenceResultRepositoryC
|
||||
msl.m2_model_batch_id,
|
||||
msl.m3_model_batch_id
|
||||
)
|
||||
inner join tb_map_sheet_anal_data_inference msadi
|
||||
on msadi.anal_uid = msl.anal_uid
|
||||
INNER JOIN tb_map_sheet_anal_data_inference msadi
|
||||
ON msadi.anal_uid = msl.anal_uid
|
||||
AND r.map_id ~ '^[0-9]+$'
|
||||
AND r.map_id::bigint = msadi.map_sheet_num
|
||||
where msl.anal_uid = :analUid
|
||||
AND r.map_id::bigint = msadi.map_sheet_num
|
||||
WHERE msl.anal_uid = :analUid
|
||||
AND r.before_c is not null
|
||||
AND r.before_p is not null
|
||||
AND r.after_c is not null
|
||||
AND r.after_p is not null
|
||||
ORDER BY r.uid, r.created_date DESC NULLS LAST
|
||||
) x
|
||||
ON CONFLICT (result_uid)
|
||||
|
||||
@@ -6,4 +6,6 @@ import java.util.UUID;
|
||||
public interface MapSheetLearn5kRepositoryCustom {
|
||||
|
||||
void saveFail5k(UUID uuid, List<Long> failMapIds, String type);
|
||||
|
||||
List<String> getInferenceRunMapId(UUID uuid);
|
||||
}
|
||||
|
||||
@@ -1,9 +1,11 @@
|
||||
package com.kamco.cd.kamcoback.postgres.repository.Inference;
|
||||
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QMapInkx5kEntity.mapInkx5kEntity;
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QMapSheetLearn5kEntity.mapSheetLearn5kEntity;
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QMapSheetLearnEntity.mapSheetLearnEntity;
|
||||
|
||||
import com.querydsl.core.types.dsl.BooleanPath;
|
||||
import com.querydsl.core.types.dsl.Expressions;
|
||||
import com.querydsl.jpa.impl.JPAQueryFactory;
|
||||
import java.util.List;
|
||||
import java.util.UUID;
|
||||
@@ -59,4 +61,25 @@ public class MapSheetLearn5kRepositoryImpl implements MapSheetLearn5kRepositoryC
|
||||
.and(mapSheetLearn5kEntity.mapSheetNum.in(failMapIds)))
|
||||
.execute();
|
||||
}
|
||||
|
||||
@Override
|
||||
public List<String> getInferenceRunMapId(UUID uuid) {
|
||||
return queryFactory
|
||||
.select(mapInkx5kEntity.mapidNm)
|
||||
.from(mapSheetLearnEntity)
|
||||
.innerJoin(mapSheetLearn5kEntity)
|
||||
.on(mapSheetLearn5kEntity.learn.id.eq(mapSheetLearnEntity.id))
|
||||
.innerJoin(mapInkx5kEntity)
|
||||
.on(
|
||||
Expressions.booleanTemplate(
|
||||
"function('regexp_match', {0}, '^[0-9]+$') is not null",
|
||||
mapInkx5kEntity.mapidcdNo)
|
||||
.and(
|
||||
mapSheetLearn5kEntity.mapSheetNum.eq(
|
||||
Expressions.numberTemplate(
|
||||
Long.class, "cast({0} as long)", mapInkx5kEntity.mapidcdNo))))
|
||||
.where(mapSheetLearnEntity.uuid.eq(uuid))
|
||||
.groupBy(mapInkx5kEntity.mapidNm)
|
||||
.fetch();
|
||||
}
|
||||
}
|
||||
|
||||
@@ -24,6 +24,7 @@ import com.kamco.cd.kamcoback.postgres.entity.MapSheetLearnEntity;
|
||||
import com.kamco.cd.kamcoback.postgres.entity.QModelMngEntity;
|
||||
import com.querydsl.core.BooleanBuilder;
|
||||
import com.querydsl.core.types.Projections;
|
||||
import com.querydsl.core.types.dsl.BooleanExpression;
|
||||
import com.querydsl.core.types.dsl.CaseBuilder;
|
||||
import com.querydsl.core.types.dsl.Expressions;
|
||||
import com.querydsl.core.types.dsl.NumberExpression;
|
||||
@@ -360,42 +361,56 @@ public class MapSheetLearnRepositoryImpl implements MapSheetLearnRepositoryCusto
|
||||
@Override
|
||||
public Page<Geom> getInferenceGeomList(UUID uuid, SearchGeoReq searchGeoReq) {
|
||||
Pageable pageable = searchGeoReq.toPageable();
|
||||
BooleanBuilder builder = new BooleanBuilder();
|
||||
BooleanBuilder where = new BooleanBuilder();
|
||||
|
||||
// analUid로 분석 정보 조회
|
||||
// 1) 분석 엔티티 조회
|
||||
MapSheetLearnEntity analEntity =
|
||||
queryFactory
|
||||
.selectFrom(mapSheetLearnEntity)
|
||||
.where(mapSheetLearnEntity.uuid.eq(uuid))
|
||||
.fetchOne();
|
||||
|
||||
if (Objects.isNull(analEntity)) {
|
||||
if (analEntity == null) {
|
||||
throw new CustomApiException("NOT_FOUND_DATA", HttpStatus.NOT_FOUND);
|
||||
}
|
||||
|
||||
// 추론결과 id
|
||||
builder.and(mapSheetAnalInferenceEntity.learnId.eq(analEntity.getId()));
|
||||
// 2) where 조건
|
||||
where.and(mapSheetAnalInferenceEntity.learnId.eq(analEntity.getId()));
|
||||
|
||||
// 기준년도 분류
|
||||
if (searchGeoReq.getTargetClass() != null && !searchGeoReq.getTargetClass().equals("")) {
|
||||
builder.and(
|
||||
if (searchGeoReq.getTargetClass() != null && !searchGeoReq.getTargetClass().isBlank()) {
|
||||
where.and(
|
||||
mapSheetAnalDataInferenceGeomEntity.classAfterCd.eq(
|
||||
searchGeoReq.getTargetClass().toLowerCase()));
|
||||
}
|
||||
|
||||
// 비교년도 분류
|
||||
if (searchGeoReq.getCompareClass() != null && !searchGeoReq.getCompareClass().equals("")) {
|
||||
builder.and(
|
||||
if (searchGeoReq.getCompareClass() != null && !searchGeoReq.getCompareClass().isBlank()) {
|
||||
where.and(
|
||||
mapSheetAnalDataInferenceGeomEntity.classBeforeCd.eq(
|
||||
searchGeoReq.getCompareClass().toLowerCase()));
|
||||
}
|
||||
|
||||
// 분석도엽
|
||||
if (searchGeoReq.getMapSheetNum() != null) {
|
||||
Long mapSheetNum = searchGeoReq.getMapSheetNum();
|
||||
builder.and(mapSheetAnalDataInferenceGeomEntity.mapSheetNum.like("%" + mapSheetNum + "%"));
|
||||
//
|
||||
// where.and(mapSheetAnalDataInferenceGeomEntity.mapSheetNum.eq(searchGeoReq.getMapSheetNum()));
|
||||
where.and(
|
||||
mapSheetAnalDataInferenceGeomEntity.mapSheetNum.like(
|
||||
"%" + searchGeoReq.getMapSheetNum() + "%"));
|
||||
}
|
||||
|
||||
where.and(mapSheetAnalDataInferenceGeomEntity.classAfterCd.isNotNull());
|
||||
where.and(mapSheetAnalDataInferenceGeomEntity.classAfterProb.isNotNull());
|
||||
where.and(mapSheetAnalDataInferenceGeomEntity.classBeforeCd.isNotNull());
|
||||
where.and(mapSheetAnalDataInferenceGeomEntity.classBeforeProb.isNotNull());
|
||||
|
||||
// 3) inkx 조인 조건: JPQL/HQL에서 '~' 불가 → function('regexp_match', ...) 사용
|
||||
BooleanExpression inkxIsNumeric =
|
||||
Expressions.booleanTemplate(
|
||||
"function('regexp_match', {0}, '^[0-9]+$') is not null", mapInkx5kEntity.mapidcdNo);
|
||||
|
||||
NumberExpression<Long> inkxNoAsLong =
|
||||
Expressions.numberTemplate(Long.class, "cast({0} as long)", mapInkx5kEntity.mapidcdNo);
|
||||
|
||||
// 4) content
|
||||
List<Geom> content =
|
||||
queryFactory
|
||||
.select(
|
||||
@@ -411,45 +426,32 @@ public class MapSheetLearnRepositoryImpl implements MapSheetLearnRepositoryCusto
|
||||
mapSheetAnalDataInferenceGeomEntity.classAfterCd,
|
||||
mapSheetAnalDataInferenceGeomEntity.classAfterProb,
|
||||
mapSheetAnalDataInferenceGeomEntity.mapSheetNum,
|
||||
mapInkx5kEntity.mapidNm
|
||||
// Expressions.stringTemplate(
|
||||
// "ST_AsGeoJSON({0})",
|
||||
// mapSheetAnalDataInferenceGeomEntity.geom),
|
||||
// Expressions.stringTemplate(
|
||||
// "ST_AsGeoJSON({0})",
|
||||
// mapSheetAnalDataInferenceGeomEntity.geomCenter)
|
||||
))
|
||||
mapInkx5kEntity.mapidNm))
|
||||
.from(mapSheetAnalInferenceEntity)
|
||||
.join(mapSheetAnalDataInferenceEntity)
|
||||
.on(mapSheetAnalDataInferenceEntity.analUid.eq(mapSheetAnalInferenceEntity.id))
|
||||
.join(mapSheetAnalDataInferenceGeomEntity)
|
||||
.on(mapSheetAnalDataInferenceGeomEntity.dataUid.eq(mapSheetAnalDataInferenceEntity.id))
|
||||
.join(mapInkx5kEntity)
|
||||
.on(
|
||||
mapSheetAnalDataInferenceGeomEntity.mapSheetNum.eq(
|
||||
Expressions.numberTemplate(
|
||||
Long.class, "CAST({0} AS long)", mapInkx5kEntity.mapidcdNo)))
|
||||
.where(builder)
|
||||
.on(inkxIsNumeric.and(mapSheetAnalDataInferenceGeomEntity.mapSheetNum.eq(inkxNoAsLong)))
|
||||
.where(where)
|
||||
.orderBy(mapSheetAnalDataInferenceGeomEntity.geoUid.desc())
|
||||
.offset(pageable.getOffset())
|
||||
.limit(pageable.getPageSize())
|
||||
.fetch();
|
||||
|
||||
long total =
|
||||
// 5) total (조인 최소화 유지)
|
||||
Long total =
|
||||
queryFactory
|
||||
.select(mapSheetAnalDataInferenceGeomEntity.geoUid)
|
||||
.select(mapSheetAnalDataInferenceGeomEntity.geoUid.count())
|
||||
.from(mapSheetAnalInferenceEntity)
|
||||
.join(mapSheetAnalDataInferenceEntity)
|
||||
.on(mapSheetAnalDataInferenceEntity.analUid.eq(mapSheetAnalInferenceEntity.id))
|
||||
.join(mapSheetAnalDataInferenceGeomEntity)
|
||||
.on(mapSheetAnalDataInferenceGeomEntity.dataUid.eq(mapSheetAnalDataInferenceEntity.id))
|
||||
.join(mapInkx5kEntity)
|
||||
.on(
|
||||
mapSheetAnalDataInferenceGeomEntity.mapSheetNum.eq(
|
||||
Expressions.numberTemplate(
|
||||
Long.class, "CAST({0} AS long)", mapInkx5kEntity.mapidcdNo)))
|
||||
.where(builder)
|
||||
.fetchCount();
|
||||
.where(where)
|
||||
.fetchOne();
|
||||
|
||||
return new PageImpl<>(content, pageable, total);
|
||||
return new PageImpl<>(content, pageable, total == null ? 0L : total);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -12,6 +12,7 @@ import com.fasterxml.jackson.core.JsonProcessingException;
|
||||
import com.fasterxml.jackson.databind.JsonNode;
|
||||
import com.fasterxml.jackson.databind.ObjectMapper;
|
||||
import com.kamco.cd.kamcoback.changedetection.dto.ChangeDetectionDto;
|
||||
import com.kamco.cd.kamcoback.changedetection.dto.ChangeDetectionDto.DetectSearchType;
|
||||
import com.kamco.cd.kamcoback.changedetection.dto.ChangeDetectionDto.MapScaleType;
|
||||
import com.kamco.cd.kamcoback.changedetection.dto.ChangeDetectionDto.MapSheetList;
|
||||
import com.kamco.cd.kamcoback.postgres.entity.MapSheetAnalDataInferenceGeomEntity;
|
||||
@@ -111,6 +112,26 @@ public class ChangeDetectionRepositoryImpl extends QuerydslRepositorySupport
|
||||
|
||||
@Override
|
||||
public ChangeDetectionDto.CogUrlDto getChangeDetectionCogUrl(ChangeDetectionDto.CogUrlReq req) {
|
||||
String mapSheetNum = req.getMapSheetNum();
|
||||
|
||||
if (req.getType().equals(DetectSearchType.MAPSHEET)
|
||||
&& req.getScale().equals(MapScaleType.SCALE_50K)) {
|
||||
mapSheetNum =
|
||||
queryFactory
|
||||
.select(mapSheetAnalDataInferenceEntity.mapSheetNum.stringValue())
|
||||
.from(mapSheetAnalInferenceEntity)
|
||||
.innerJoin(mapSheetAnalDataInferenceEntity)
|
||||
.on(mapSheetAnalInferenceEntity.id.eq(mapSheetAnalDataInferenceEntity.analUid))
|
||||
.where(
|
||||
mapSheetAnalInferenceEntity.uuid.eq(req.getUuid()),
|
||||
mapSheetAnalDataInferenceEntity
|
||||
.mapSheetNum
|
||||
.stringValue()
|
||||
.like("%" + req.getMapSheetNum() + "%"))
|
||||
.orderBy(mapSheetAnalDataInferenceEntity.mapSheetNum.asc())
|
||||
.fetchFirst();
|
||||
}
|
||||
|
||||
ChangeDetectionDto.CogUrlData data =
|
||||
queryFactory
|
||||
.select(
|
||||
@@ -128,7 +149,7 @@ public class ChangeDetectionRepositoryImpl extends QuerydslRepositorySupport
|
||||
.year
|
||||
.eq(req.getBeforeYear())
|
||||
.or(imageryEntity.year.eq(req.getAfterYear())),
|
||||
imageryEntity.scene5k.eq(req.getMapSheetNum()))
|
||||
imageryEntity.scene5k.eq(mapSheetNum))
|
||||
.groupBy(mapInkx5kEntity.geom)
|
||||
.fetchOne();
|
||||
|
||||
@@ -191,6 +212,7 @@ public class ChangeDetectionRepositoryImpl extends QuerydslRepositorySupport
|
||||
.where(
|
||||
mapSheetAnalInferenceEntity.uuid.eq(uuid),
|
||||
mapScaleTypeAnalDataSearchExpression(scale, mapSheetNum))
|
||||
.orderBy(mapSheetAnalDataInferenceGeomEntity.mapSheetNum.asc())
|
||||
.fetch();
|
||||
|
||||
ObjectMapper mapper = new ObjectMapper();
|
||||
@@ -313,7 +335,8 @@ public class ChangeDetectionRepositoryImpl extends QuerydslRepositorySupport
|
||||
.on(mapSheetAnalDataInferenceEntity.mapSheetNum.stringValue().eq(mapInkx5kEntity.mapidcdNo))
|
||||
.innerJoin(mapInkx5kEntity.mapInkx50k, mapInkx50kEntity)
|
||||
.where(mapSheetAnalInferenceEntity.uuid.eq(uuid))
|
||||
.orderBy(mapInkx5kEntity.mapidcdNo.asc())
|
||||
.groupBy(mapInkx50kEntity.mapidcdNo, mapInkx50kEntity.mapidNm)
|
||||
.orderBy(mapInkx50kEntity.mapidcdNo.asc())
|
||||
.fetch();
|
||||
}
|
||||
|
||||
|
||||
@@ -20,14 +20,13 @@ import org.springframework.data.domain.Page;
|
||||
|
||||
public interface LabelAllocateRepositoryCustom {
|
||||
|
||||
List<AllocateInfoDto> fetchNextIds(
|
||||
Long lastId, Long batchSize, Integer compareYyyy, Integer targetYyyy, Integer stage);
|
||||
List<AllocateInfoDto> fetchNextIds(Long lastId, Long batchSize, UUID uuid);
|
||||
|
||||
void assignOwner(List<AllocateInfoDto> ids, String userId, Long analUid);
|
||||
|
||||
List<LabelingAssignmentEntity> findAssignedLabelerList(Long analUid);
|
||||
|
||||
Long findLabelUnAssignedCnt(Integer stage, Integer compareYyyy, Integer targetYyyy);
|
||||
Long findLabelUnAssignedCnt(UUID uuid);
|
||||
|
||||
void assignInspector(UUID assignmentUid, String userId);
|
||||
|
||||
@@ -74,7 +73,7 @@ public interface LabelAllocateRepositoryCustom {
|
||||
/** UUID로 analUid 조회 */
|
||||
Long findAnalUidByUuid(String uuid);
|
||||
|
||||
Long findMapSheetAnalInferenceUid(Integer compareYyyy, Integer targetYyyy, Integer stage);
|
||||
Long findMapSheetAnalInferenceUid(UUID uuid);
|
||||
|
||||
void insertInspector(Long analUid, String inspector);
|
||||
|
||||
@@ -101,4 +100,8 @@ public interface LabelAllocateRepositoryCustom {
|
||||
Page<WorkHistoryDto> workLabelHistoryList(LabelAllocateDto.searchReq searchReq, String userId);
|
||||
|
||||
Page<WorkHistoryDto> workReviewerHistoryList(searchReq searchReq, String userId);
|
||||
|
||||
void updateAnalInferenceMngState(UUID uuid, String status);
|
||||
|
||||
Long findLabelingIngProcessCnt();
|
||||
}
|
||||
|
||||
@@ -13,6 +13,7 @@ import com.kamco.cd.kamcoback.label.dto.LabelAllocateDto;
|
||||
import com.kamco.cd.kamcoback.label.dto.LabelAllocateDto.AllocateInfoDto;
|
||||
import com.kamco.cd.kamcoback.label.dto.LabelAllocateDto.InferenceDetail;
|
||||
import com.kamco.cd.kamcoback.label.dto.LabelAllocateDto.InspectState;
|
||||
import com.kamco.cd.kamcoback.label.dto.LabelAllocateDto.LabelMngState;
|
||||
import com.kamco.cd.kamcoback.label.dto.LabelAllocateDto.LabelState;
|
||||
import com.kamco.cd.kamcoback.label.dto.LabelAllocateDto.LabelerDetail;
|
||||
import com.kamco.cd.kamcoback.label.dto.LabelAllocateDto.LabelingStatDto;
|
||||
@@ -66,8 +67,7 @@ public class LabelAllocateRepositoryImpl implements LabelAllocateRepositoryCusto
|
||||
@PersistenceContext private EntityManager em;
|
||||
|
||||
@Override
|
||||
public List<AllocateInfoDto> fetchNextIds(
|
||||
Long lastId, Long batchSize, Integer compareYyyy, Integer targetYyyy, Integer stage) {
|
||||
public List<AllocateInfoDto> fetchNextIds(Long lastId, Long batchSize, UUID uuid) {
|
||||
|
||||
return queryFactory
|
||||
.select(
|
||||
@@ -76,14 +76,17 @@ public class LabelAllocateRepositoryImpl implements LabelAllocateRepositoryCusto
|
||||
mapSheetAnalDataInferenceGeomEntity.geoUid,
|
||||
mapSheetAnalDataInferenceGeomEntity.mapSheetNum,
|
||||
mapSheetAnalDataInferenceGeomEntity.pnu))
|
||||
.from(mapSheetAnalDataInferenceGeomEntity)
|
||||
.where(
|
||||
lastId == null ? null : mapSheetAnalDataInferenceGeomEntity.geoUid.gt(lastId),
|
||||
.from(mapSheetAnalInferenceEntity)
|
||||
.innerJoin(mapSheetAnalDataInferenceEntity)
|
||||
.on(mapSheetAnalInferenceEntity.id.eq(mapSheetAnalDataInferenceEntity.analUid))
|
||||
.innerJoin(mapSheetAnalDataInferenceGeomEntity)
|
||||
.on(
|
||||
mapSheetAnalDataInferenceEntity.id.eq(mapSheetAnalDataInferenceGeomEntity.dataUid),
|
||||
mapSheetAnalDataInferenceGeomEntity.pnu.isNotNull(),
|
||||
mapSheetAnalDataInferenceGeomEntity.compareYyyy.eq(compareYyyy),
|
||||
mapSheetAnalDataInferenceGeomEntity.targetYyyy.eq(targetYyyy),
|
||||
mapSheetAnalDataInferenceGeomEntity.stage.eq(stage),
|
||||
mapSheetAnalDataInferenceGeomEntity.labelState.isNull())
|
||||
.where(
|
||||
mapSheetAnalInferenceEntity.uuid.eq(uuid),
|
||||
lastId == null ? null : mapSheetAnalDataInferenceGeomEntity.geoUid.gt(lastId))
|
||||
.orderBy(mapSheetAnalDataInferenceGeomEntity.mapSheetNum.asc())
|
||||
.limit(batchSize)
|
||||
.fetch();
|
||||
@@ -177,17 +180,19 @@ public class LabelAllocateRepositoryImpl implements LabelAllocateRepositoryCusto
|
||||
}
|
||||
|
||||
@Override
|
||||
public Long findLabelUnAssignedCnt(Integer stage, Integer compareYyyy, Integer targetYyyy) {
|
||||
public Long findLabelUnAssignedCnt(UUID uuid) {
|
||||
|
||||
return queryFactory
|
||||
.select(mapSheetAnalDataInferenceGeomEntity.geoUid.count())
|
||||
.from(mapSheetAnalDataInferenceGeomEntity)
|
||||
.where(
|
||||
.from(mapSheetAnalInferenceEntity)
|
||||
.innerJoin(mapSheetAnalDataInferenceEntity)
|
||||
.on(mapSheetAnalInferenceEntity.id.eq(mapSheetAnalDataInferenceEntity.analUid))
|
||||
.innerJoin(mapSheetAnalDataInferenceGeomEntity)
|
||||
.on(
|
||||
mapSheetAnalDataInferenceEntity.id.eq(mapSheetAnalDataInferenceGeomEntity.dataUid),
|
||||
mapSheetAnalDataInferenceGeomEntity.pnu.isNotNull(),
|
||||
mapSheetAnalDataInferenceGeomEntity.compareYyyy.eq(compareYyyy),
|
||||
mapSheetAnalDataInferenceGeomEntity.targetYyyy.eq(targetYyyy),
|
||||
mapSheetAnalDataInferenceGeomEntity.stage.eq(stage),
|
||||
mapSheetAnalDataInferenceGeomEntity.labelState.isNull())
|
||||
.where(mapSheetAnalInferenceEntity.uuid.eq(uuid))
|
||||
.fetchOne();
|
||||
}
|
||||
|
||||
@@ -912,14 +917,11 @@ public class LabelAllocateRepositoryImpl implements LabelAllocateRepositoryCusto
|
||||
}
|
||||
|
||||
@Override
|
||||
public Long findMapSheetAnalInferenceUid(Integer compareYyyy, Integer targetYyyy, Integer stage) {
|
||||
public Long findMapSheetAnalInferenceUid(UUID uuid) {
|
||||
return queryFactory
|
||||
.select(mapSheetAnalInferenceEntity.id)
|
||||
.from(mapSheetAnalInferenceEntity)
|
||||
.where(
|
||||
mapSheetAnalInferenceEntity.compareYyyy.eq(compareYyyy),
|
||||
mapSheetAnalInferenceEntity.targetYyyy.eq(targetYyyy),
|
||||
mapSheetAnalInferenceEntity.stage.eq(stage))
|
||||
.where(mapSheetAnalInferenceEntity.uuid.eq(uuid))
|
||||
.fetchOne();
|
||||
}
|
||||
|
||||
@@ -1784,4 +1786,23 @@ public class LabelAllocateRepositoryImpl implements LabelAllocateRepositoryCusto
|
||||
|
||||
return new PageImpl<>(list, pageable, countQuery != null ? countQuery : 0L);
|
||||
}
|
||||
|
||||
@Override
|
||||
public void updateAnalInferenceMngState(UUID uuid, String status) {
|
||||
queryFactory
|
||||
.update(mapSheetAnalInferenceEntity)
|
||||
.set(mapSheetAnalInferenceEntity.analState, status)
|
||||
.set(mapSheetAnalInferenceEntity.updatedDttm, ZonedDateTime.now())
|
||||
.where(mapSheetAnalInferenceEntity.uuid.eq(uuid))
|
||||
.execute();
|
||||
}
|
||||
|
||||
@Override
|
||||
public Long findLabelingIngProcessCnt() {
|
||||
return queryFactory
|
||||
.select(mapSheetAnalInferenceEntity.id.count())
|
||||
.from(mapSheetAnalInferenceEntity)
|
||||
.where(mapSheetAnalInferenceEntity.analState.eq(LabelMngState.ING.getId()))
|
||||
.fetchOne();
|
||||
}
|
||||
}
|
||||
|
||||
@@ -138,7 +138,9 @@ public class LabelWorkRepositoryImpl implements LabelWorkRepositoryCustom {
|
||||
mapSheetAnalDataInferenceGeomEntity
|
||||
.pnu
|
||||
.isNotNull()
|
||||
.and(mapSheetAnalDataInferenceGeomEntity.pnu.ne(0L))
|
||||
.and(
|
||||
mapSheetAnalDataInferenceGeomEntity.pnu
|
||||
.isNotNull()) // TODO: 이노팸 연동 후 0 이상이라고 해야할 듯
|
||||
.and(mapSheetAnalDataInferenceGeomEntity.passYn.eq(Boolean.FALSE)))
|
||||
.then(1L)
|
||||
.otherwise(0L)
|
||||
@@ -198,9 +200,8 @@ public class LabelWorkRepositoryImpl implements LabelWorkRepositoryCustom {
|
||||
.otherwise(0L)
|
||||
.sum(),
|
||||
mapSheetAnalDataInferenceGeomEntity.labelStateDttm.min(),
|
||||
// analState: tb_map_sheet_anal_inference.anal_state -> 우선은 미사용, java 단에서 로직화 해서
|
||||
// 내려줌
|
||||
// mapSheetAnalInferenceEntity.analState,
|
||||
// analState: tb_map_sheet_anal_inference.anal_state
|
||||
mapSheetAnalInferenceEntity.analState,
|
||||
// normalProgressCnt: stagnation_yn = 'N'인 건수 (서브쿼리)
|
||||
normalProgressCntSubQuery,
|
||||
// totalAssignmentCnt: 총 배정 건수 (서브쿼리)
|
||||
|
||||
@@ -16,4 +16,6 @@ public interface TrainingDataReviewJobRepositoryCustom {
|
||||
void assignReviewerBatch(List<UUID> assignmentUids, String reviewerId);
|
||||
|
||||
Tasks findAssignmentTask(String assignmentUid);
|
||||
|
||||
void updateGeomUidTestState(List<Long> geomUids);
|
||||
}
|
||||
|
||||
@@ -2,6 +2,7 @@ package com.kamco.cd.kamcoback.postgres.repository.scheduler;
|
||||
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QLabelingAssignmentEntity.labelingAssignmentEntity;
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QLabelingInspectorEntity.labelingInspectorEntity;
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QMapSheetAnalDataInferenceGeomEntity.mapSheetAnalDataInferenceGeomEntity;
|
||||
|
||||
import com.kamco.cd.kamcoback.label.dto.LabelAllocateDto.InspectState;
|
||||
import com.kamco.cd.kamcoback.label.dto.LabelAllocateDto.LabelState;
|
||||
@@ -99,6 +100,7 @@ public class TrainingDataReviewJobRepositoryImpl extends QuerydslRepositorySuppo
|
||||
.update(labelingAssignmentEntity)
|
||||
.set(labelingAssignmentEntity.inspectorUid, reviewerId)
|
||||
.set(labelingAssignmentEntity.inspectState, InspectState.UNCONFIRM.getId())
|
||||
.set(labelingAssignmentEntity.modifiedDate, ZonedDateTime.now())
|
||||
.where(labelingAssignmentEntity.assignmentUid.eq(assignmentUid))
|
||||
.execute();
|
||||
}
|
||||
@@ -115,6 +117,7 @@ public class TrainingDataReviewJobRepositoryImpl extends QuerydslRepositorySuppo
|
||||
.update(labelingAssignmentEntity)
|
||||
.set(labelingAssignmentEntity.inspectorUid, reviewerId)
|
||||
.set(labelingAssignmentEntity.inspectState, InspectState.UNCONFIRM.getId())
|
||||
.set(labelingAssignmentEntity.modifiedDate, ZonedDateTime.now())
|
||||
.where(labelingAssignmentEntity.assignmentUid.in(assignmentUids))
|
||||
.execute();
|
||||
}
|
||||
@@ -132,4 +135,14 @@ public class TrainingDataReviewJobRepositoryImpl extends QuerydslRepositorySuppo
|
||||
.where(labelingAssignmentEntity.assignmentUid.eq(UUID.fromString(assignmentUid)))
|
||||
.fetchOne();
|
||||
}
|
||||
|
||||
@Override
|
||||
public void updateGeomUidTestState(List<Long> geomUids) {
|
||||
queryFactory
|
||||
.update(mapSheetAnalDataInferenceGeomEntity)
|
||||
.set(mapSheetAnalDataInferenceGeomEntity.testState, InspectState.UNCONFIRM.getId())
|
||||
.set(mapSheetAnalDataInferenceGeomEntity.updatedDttm, ZonedDateTime.now())
|
||||
.where(mapSheetAnalDataInferenceGeomEntity.geoUid.in(geomUids))
|
||||
.execute();
|
||||
}
|
||||
}
|
||||
|
||||
@@ -3,6 +3,7 @@ package com.kamco.cd.kamcoback.postgres.repository.trainingdata;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.DefaultPaging;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.DetailRes;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.InferenceIdInfo;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.LabelFeatureRequest.LabelProperties;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.LabelingGeometryInfo;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.LabelingListDto;
|
||||
@@ -18,7 +19,7 @@ public interface TrainingDataLabelRepositoryCustom {
|
||||
|
||||
LabelingGeometryInfo findLabelingAssignedGeom(String assignmentUid);
|
||||
|
||||
Long findLabelingAssignmentGeoUid(String assignmentUid);
|
||||
InferenceIdInfo findLabelingAssignmentGeoUid(String assignmentUid);
|
||||
|
||||
void updateLabelingStateAssignment(String assignmentUid, String status);
|
||||
|
||||
@@ -37,4 +38,6 @@ public interface TrainingDataLabelRepositoryCustom {
|
||||
|
||||
TrainingDataLabelDto.CogImageResponse getCogImageUrl(
|
||||
String mapSheetNum, Integer beforeYear, Integer afterYear);
|
||||
|
||||
void updateAnalInferenceMngState(Long analUid, String status);
|
||||
}
|
||||
|
||||
@@ -4,6 +4,7 @@ import static com.kamco.cd.kamcoback.postgres.entity.QImageryEntity.imageryEntit
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QLabelingAssignmentEntity.labelingAssignmentEntity;
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QMapInkx5kEntity.mapInkx5kEntity;
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QMapSheetAnalDataInferenceGeomEntity.mapSheetAnalDataInferenceGeomEntity;
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QMapSheetAnalInferenceEntity.mapSheetAnalInferenceEntity;
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QMapSheetLearnDataGeomEntity.mapSheetLearnDataGeomEntity;
|
||||
|
||||
import com.fasterxml.jackson.databind.JsonNode;
|
||||
@@ -19,6 +20,7 @@ import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.DefaultPagin
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.DetailRes;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.InferenceDataGeometry;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.InferenceDataGeometry.InferenceProperties;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.InferenceIdInfo;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.InspectionResultInfo;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.LabelFeatureRequest.LabelProperties;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.LabelingGeometryInfo;
|
||||
@@ -174,9 +176,13 @@ public class TrainingDataLabelRepositoryImpl extends QuerydslRepositorySupport
|
||||
}
|
||||
|
||||
@Override
|
||||
public Long findLabelingAssignmentGeoUid(String assignmentUid) {
|
||||
public InferenceIdInfo findLabelingAssignmentGeoUid(String assignmentUid) {
|
||||
return queryFactory
|
||||
.select(labelingAssignmentEntity.inferenceGeomUid)
|
||||
.select(
|
||||
Projections.constructor(
|
||||
InferenceIdInfo.class,
|
||||
labelingAssignmentEntity.inferenceGeomUid,
|
||||
labelingAssignmentEntity.analUid))
|
||||
.from(labelingAssignmentEntity)
|
||||
.where(labelingAssignmentEntity.assignmentUid.eq(UUID.fromString(assignmentUid)))
|
||||
.fetchOne();
|
||||
@@ -838,6 +844,16 @@ public class TrainingDataLabelRepositoryImpl extends QuerydslRepositorySupport
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public void updateAnalInferenceMngState(Long analUid, String status) {
|
||||
queryFactory
|
||||
.update(mapSheetAnalInferenceEntity)
|
||||
.set(mapSheetAnalInferenceEntity.analState, status)
|
||||
.set(mapSheetAnalInferenceEntity.updatedDttm, ZonedDateTime.now())
|
||||
.where(mapSheetAnalInferenceEntity.id.eq(analUid))
|
||||
.execute();
|
||||
}
|
||||
|
||||
private StringExpression makeCogUrl(NumberPath<Integer> year) {
|
||||
return new CaseBuilder()
|
||||
.when(imageryEntity.year.eq(year))
|
||||
|
||||
@@ -1,5 +1,6 @@
|
||||
package com.kamco.cd.kamcoback.postgres.repository.trainingdata;
|
||||
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.InferenceIdInfo;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataReviewDto;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataReviewDto.DefaultPaging;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataReviewDto.DetailRes;
|
||||
@@ -18,7 +19,7 @@ public interface TrainingDataReviewRepositoryCustom {
|
||||
|
||||
ReviewGeometryInfo findReviewAssignedGeom(String operatorUid);
|
||||
|
||||
Long findReviewOperatorGeoUid(String operatorUid);
|
||||
InferenceIdInfo findReviewOperatorGeoUid(String operatorUid);
|
||||
|
||||
void updateReviewStateOperator(String operatorUid, String status);
|
||||
|
||||
@@ -37,4 +38,6 @@ public interface TrainingDataReviewRepositoryCustom {
|
||||
|
||||
TrainingDataReviewDto.CogImageResponse getCogImageUrl(
|
||||
String mapSheetNum, Integer beforeYear, Integer afterYear);
|
||||
|
||||
void updateAnalInferenceMngState(Long analUid, String status);
|
||||
}
|
||||
|
||||
@@ -4,6 +4,7 @@ import static com.kamco.cd.kamcoback.postgres.entity.QImageryEntity.imageryEntit
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QLabelingAssignmentEntity.labelingAssignmentEntity;
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QMapInkx5kEntity.mapInkx5kEntity;
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QMapSheetAnalDataInferenceGeomEntity.mapSheetAnalDataInferenceGeomEntity;
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QMapSheetAnalInferenceEntity.mapSheetAnalInferenceEntity;
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QMapSheetLearnDataGeomEntity.mapSheetLearnDataGeomEntity;
|
||||
import static com.kamco.cd.kamcoback.postgres.entity.QMemberEntity.memberEntity;
|
||||
|
||||
@@ -12,6 +13,7 @@ import com.fasterxml.jackson.databind.ObjectMapper;
|
||||
import com.kamco.cd.kamcoback.label.dto.LabelAllocateDto.InspectState;
|
||||
import com.kamco.cd.kamcoback.postgres.entity.LabelingAssignmentEntity;
|
||||
import com.kamco.cd.kamcoback.postgres.entity.MapSheetAnalDataInferenceGeomEntity;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataLabelDto.InferenceIdInfo;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataReviewDto;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataReviewDto.ChangeDetectionInfo;
|
||||
import com.kamco.cd.kamcoback.trainingdata.dto.TrainingDataReviewDto.ClassificationInfo;
|
||||
@@ -177,9 +179,13 @@ public class TrainingDataReviewRepositoryImpl extends QuerydslRepositorySupport
|
||||
}
|
||||
|
||||
@Override
|
||||
public Long findReviewOperatorGeoUid(String operatorUid) {
|
||||
public InferenceIdInfo findReviewOperatorGeoUid(String operatorUid) {
|
||||
return queryFactory
|
||||
.select(labelingAssignmentEntity.inferenceGeomUid)
|
||||
.select(
|
||||
Projections.constructor(
|
||||
InferenceIdInfo.class,
|
||||
labelingAssignmentEntity.inferenceGeomUid,
|
||||
labelingAssignmentEntity.analUid))
|
||||
.from(labelingAssignmentEntity)
|
||||
.where(labelingAssignmentEntity.assignmentUid.eq(UUID.fromString(operatorUid)))
|
||||
.fetchOne();
|
||||
@@ -209,6 +215,15 @@ public class TrainingDataReviewRepositoryImpl extends QuerydslRepositorySupport
|
||||
@Override
|
||||
public void updateReviewPolygonClass(
|
||||
Long inferenceGeomUid, Geometry geometry, Properties properties, String status) {
|
||||
|
||||
// inference_geom 테이블에 상태 업데이트
|
||||
queryFactory
|
||||
.update(mapSheetAnalDataInferenceGeomEntity)
|
||||
.set(mapSheetAnalDataInferenceGeomEntity.testStateDttm, ZonedDateTime.now())
|
||||
.set(mapSheetAnalDataInferenceGeomEntity.testState, status)
|
||||
.where(mapSheetAnalDataInferenceGeomEntity.geoUid.eq(inferenceGeomUid))
|
||||
.execute();
|
||||
|
||||
// inference_geom 테이블 정보 가져오기
|
||||
MapSheetAnalDataInferenceGeomEntity entity =
|
||||
queryFactory
|
||||
@@ -859,6 +874,16 @@ public class TrainingDataReviewRepositoryImpl extends QuerydslRepositorySupport
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public void updateAnalInferenceMngState(Long analUid, String status) {
|
||||
queryFactory
|
||||
.update(mapSheetAnalInferenceEntity)
|
||||
.set(mapSheetAnalInferenceEntity.analState, status)
|
||||
.set(mapSheetAnalInferenceEntity.updatedDttm, ZonedDateTime.now())
|
||||
.where(mapSheetAnalInferenceEntity.id.eq(analUid))
|
||||
.execute();
|
||||
}
|
||||
|
||||
private StringExpression makeCogUrl(NumberPath<Integer> year) {
|
||||
return new CaseBuilder()
|
||||
.when(imageryEntity.year.eq(year))
|
||||
|
||||
Reference in New Issue
Block a user