added deduplication service

This commit is contained in:
Rushikesh Tote 2022-03-22 18:31:59 +05:30
parent 7a74b5e985
commit 497ea885e1

View file

@ -0,0 +1,126 @@
import { EnteFile } from 'types/file';
import { getEndpoint } from 'utils/common/apiUtil';
import { getToken } from 'utils/common/key';
import { logError } from 'utils/sentry';
import { getLocalFiles } from './fileService';
import HTTPService from './HTTPService';
const ENDPOINT = getEndpoint();
interface DuplicatesResponse {
duplicates: Array<{
fileIDs: number[];
size: number;
}>;
}
interface DuplicateFiles {
files: EnteFile[];
size: number;
}
class DeduplicationService {
public async getDuplicateFiles() {
try {
const dupes = await this.fetchDuplicateFileIDs();
let ids: number[] = [];
for (const dupe of dupes) {
ids = ids.concat(dupe.fileIDs);
}
const localFiles = await getLocalFiles();
const fileMap = new Map<number, EnteFile>();
for (const file of localFiles) {
fileMap.set(file.id, file);
}
const result: DuplicateFiles[] = [];
const missingFileIDs: number[] = [];
for (const dupe of dupes) {
const files: EnteFile[] = [];
for (const fileID of dupe.fileIDs) {
if (fileMap.has(fileID)) {
files.push(fileMap.get(fileID));
} else {
missingFileIDs.push(fileID);
}
}
if (files.length > 1) {
result.push({
files,
size: dupe.size,
});
}
}
return result;
} catch (e) {
logError(e, 'failed to get');
}
}
public async clubDuplicatesByTime(dupes: DuplicateFiles[]) {
const result: DuplicateFiles[] = [];
for (const dupe of dupes) {
let files: EnteFile[] = [];
const creationTimeCounter = new Map<number, number>();
let mostFreqCreationTime = 0;
let mostFreqCreationTimeCount = 0;
for (const file of dupe.files) {
const creationTime = file.metadata.creationTime;
if (creationTimeCounter.has(creationTime)) {
creationTimeCounter.set(
creationTime,
creationTimeCounter.get(creationTime) + 1
);
} else {
creationTimeCounter.set(creationTime, 1);
}
if (
creationTimeCounter.get(creationTime) >
mostFreqCreationTimeCount
) {
mostFreqCreationTime = creationTime;
mostFreqCreationTimeCount =
creationTimeCounter.get(creationTime);
}
files.push(file);
}
files = files.filter((file) => {
return file.metadata.creationTime === mostFreqCreationTime;
});
if (files.length > 1) {
result.push({
files,
size: dupe.size,
});
}
}
return result;
}
private async fetchDuplicateFileIDs() {
try {
const response = await HTTPService.get(
`${ENDPOINT}/files/duplicates`,
null,
{
'X-Auth-Token': getToken(),
}
);
return (response.data as DuplicatesResponse).duplicates;
} catch (e) {
logError(e, 'failed to get');
}
}
}
export default new DeduplicationService();