import { Inject, Injectable, Logger } from '@nestjs/common'; import { IAssetRepository, WithoutProperty } from '../asset'; import { MACHINE_LEARNING_ENABLED } from '../domain.constant'; import { usePagination } from '../domain.util'; import { IAssetJob, IBaseJob, IJobRepository, JobName, JOBS_ASSET_PAGINATION_SIZE } from '../job'; import { IMachineLearningRepository } from './machine-learning.interface'; import { ISmartInfoRepository } from './smart-info.repository'; @Injectable() export class SmartInfoService { private logger = new Logger(SmartInfoService.name); constructor( @Inject(IAssetRepository) private assetRepository: IAssetRepository, @Inject(IJobRepository) private jobRepository: IJobRepository, @Inject(ISmartInfoRepository) private repository: ISmartInfoRepository, @Inject(IMachineLearningRepository) private machineLearning: IMachineLearningRepository, ) {} async handleQueueObjectTagging({ force }: IBaseJob) { try { const assetPagination = usePagination(JOBS_ASSET_PAGINATION_SIZE, (pagination) => { return force ? this.assetRepository.getAll(pagination) : this.assetRepository.getWithout(pagination, WithoutProperty.OBJECT_TAGS); }); for await (const assets of assetPagination) { for (const asset of assets) { await this.jobRepository.queue({ name: JobName.CLASSIFY_IMAGE, data: { asset } }); await this.jobRepository.queue({ name: JobName.DETECT_OBJECTS, data: { asset } }); } } } catch (error: any) { this.logger.error(`Unable to queue object tagging`, error?.stack); } } async handleDetectObjects(data: IAssetJob) { const { asset } = data; if (!MACHINE_LEARNING_ENABLED || !asset.resizePath) { return; } try { const objects = await this.machineLearning.detectObjects({ thumbnailPath: asset.resizePath }); if (objects.length > 0) { await this.repository.upsert({ assetId: asset.id, objects }); await this.jobRepository.queue({ name: JobName.SEARCH_INDEX_ASSET, data: { ids: [asset.id] } }); } } catch (error: any) { this.logger.error(`Unable run object detection pipeline: ${asset.id}`, error?.stack); } } async handleClassifyImage(data: IAssetJob) { const { asset } = data; if (!MACHINE_LEARNING_ENABLED || !asset.resizePath) { return; } try { const tags = await this.machineLearning.classifyImage({ thumbnailPath: asset.resizePath }); if (tags.length > 0) { await this.repository.upsert({ assetId: asset.id, tags }); await this.jobRepository.queue({ name: JobName.SEARCH_INDEX_ASSET, data: { ids: [asset.id] } }); } } catch (error: any) { this.logger.error(`Unable to run image tagging pipeline: ${asset.id}`, error?.stack); } } async handleQueueEncodeClip({ force }: IBaseJob) { try { const assetPagination = usePagination(JOBS_ASSET_PAGINATION_SIZE, (pagination) => { return force ? this.assetRepository.getAll(pagination) : this.assetRepository.getWithout(pagination, WithoutProperty.CLIP_ENCODING); }); for await (const assets of assetPagination) { for (const asset of assets) { await this.jobRepository.queue({ name: JobName.ENCODE_CLIP, data: { asset } }); } } } catch (error: any) { this.logger.error(`Unable to queue clip encoding`, error?.stack); } } async handleEncodeClip(data: IAssetJob) { const { asset } = data; if (!MACHINE_LEARNING_ENABLED || !asset.resizePath) { return; } try { const clipEmbedding = await this.machineLearning.encodeImage({ thumbnailPath: asset.resizePath }); await this.repository.upsert({ assetId: asset.id, clipEmbedding: clipEmbedding }); await this.jobRepository.queue({ name: JobName.SEARCH_INDEX_ASSET, data: { ids: [asset.id] } }); } catch (error: any) { this.logger.error(`Unable run clip encoding pipeline: ${asset.id}`, error?.stack); } } }