import 'dart:async'; import 'dart:core'; import 'dart:io'; import 'package:flutter/foundation.dart'; import 'package:logging/logging.dart'; import 'package:photos/core/configuration.dart'; import 'package:photos/core/constants.dart'; import 'package:photos/db/file_updation_db.dart'; import 'package:photos/db/files_db.dart'; import 'package:photos/extensions/stop_watch.dart'; import 'package:photos/models/file.dart' as ente; import 'package:photos/services/files_service.dart'; import 'package:photos/utils/file_uploader_util.dart'; import 'package:photos/utils/file_util.dart'; import 'package:shared_preferences/shared_preferences.dart'; // LocalFileUpdateService tracks all the potential local file IDs which have // changed/modified on the device and needed to be uploaded again. class LocalFileUpdateService { late FileUpdationDB _fileUpdationDB; late SharedPreferences _prefs; late Logger _logger; static const isLocationMigrationComplete = "fm_isLocationMigrationComplete"; static const isLocalImportDone = "fm_IsLocalImportDone"; static const isBadCreationTimeImportDone = 'fm_badCreationTime'; static const isBadCreationTimeMigrationComplete = 'fm_badCreationTimeCompleted'; Completer? _existingMigration; LocalFileUpdateService._privateConstructor() { _logger = Logger((LocalFileUpdateService).toString()); _fileUpdationDB = FileUpdationDB.instance; } void init(SharedPreferences preferences) { _prefs = preferences; } static LocalFileUpdateService instance = LocalFileUpdateService._privateConstructor(); bool isBadCreationMigrationCompleted() { return (_prefs.getBool(isBadCreationTimeMigrationComplete) ?? false); } Future markUpdatedFilesForReUpload() async { if (_existingMigration != null) { _logger.info("migration is already in progress, skipping"); return _existingMigration!.future; } _existingMigration = Completer(); try { await _markFilesWhichAreActuallyUpdated(); if (Platform.isAndroid) { await _migrationForFixingBadCreationTime(); } } catch (e, s) { _logger.severe('failed to perform migration', e, s); } finally { _existingMigration?.complete(); _existingMigration = null; } } // This method analyses all of local files for which the file // modification/update time was changed. It checks if the existing fileHash // is different from the hash of uploaded file. If fileHash are different, // then it marks the file for file update. Future _markFilesWhichAreActuallyUpdated() async { final sTime = DateTime.now().microsecondsSinceEpoch; // singleRunLimit indicates number of files to check during single // invocation of this method. The limit act as a crude way to limit the // resource consumed by the method const int singleRunLimit = 10; final localIDsToProcess = await _fileUpdationDB.getLocalIDsForPotentialReUpload( singleRunLimit, FileUpdationDB.modificationTimeUpdated, ); if (localIDsToProcess.isNotEmpty) { await _checkAndMarkFilesWithDifferentHashForFileUpdate( localIDsToProcess, ); final eTime = DateTime.now().microsecondsSinceEpoch; final d = Duration(microseconds: eTime - sTime); _logger.info( 'Performed hashCheck for ${localIDsToProcess.length} updated files ' 'completed in ${d.inSeconds.toString()} secs', ); } } Future _checkAndMarkFilesWithDifferentHashForFileUpdate( List localIDsToProcess, ) async { _logger.info("files to process ${localIDsToProcess.length} for reupload"); final List localFiles = await FilesDB.instance.getLocalFiles(localIDsToProcess); final Set processedIDs = {}; for (ente.File file in localFiles) { if (processedIDs.contains(file.localID)) { continue; } MediaUploadData uploadData; try { uploadData = await getUploadData(file); if (uploadData.hashData != null && file.hash != null && (file.hash == uploadData.hashData!.fileHash || file.hash == uploadData.hashData!.zipHash)) { _logger.info("Skip file update as hash matched ${file.tag}"); } else { _logger.info( "Marking for file update as hash did not match ${file.tag}", ); await clearCache(file); await FilesDB.instance.updateUploadedFile( file.localID!, file.title, file.location, file.creationTime!, file.modificationTime!, null, ); } processedIDs.add(file.localID!); } catch (e) { _logger.severe("Failed to get file uploadData", e); } finally {} } debugPrint("Deleting files ${processedIDs.length}"); await _fileUpdationDB.deleteByLocalIDs( processedIDs.toList(), FileUpdationDB.modificationTimeUpdated, ); } Future getUploadData(ente.File file) async { final mediaUploadData = await getUploadDataFromEnteFile(file); // delete the file from app's internal cache if it was copied to app // for upload. Shared Media should only be cleared when the upload // succeeds. if (Platform.isIOS && mediaUploadData.sourceFile != null) { await mediaUploadData.sourceFile?.delete(); } return mediaUploadData; } Future _migrationForFixingBadCreationTime() async { if (_prefs.containsKey(isBadCreationTimeMigrationComplete)) { return; } await _importFilesWithBadCreationTime(); const int singleRunLimit = 100; try { final generatedIDs = await _fileUpdationDB.getLocalIDsForPotentialReUpload( singleRunLimit, FileUpdationDB.badCreationTime, ); if (generatedIDs.isNotEmpty) { final List genIdIntList = []; for (String genIdString in generatedIDs) { final int? genIdInt = int.tryParse(genIdString); if (genIdInt != null) { genIdIntList.add(genIdInt); } } final filesWithBadTime = (await FilesDB.instance.getFilesFromGeneratedIDs(genIdIntList)) .values .toList(); filesWithBadTime.removeWhere( (e) => e.isUploaded && e.pubMagicMetadata?.editedTime != null, ); await FilesService.instance .bulkEditTime(filesWithBadTime, EditTimeSource.fileName); } else { // everything is done await _prefs.setBool(isBadCreationTimeMigrationComplete, true); } await _fileUpdationDB.deleteByLocalIDs( generatedIDs, FileUpdationDB.badCreationTime, ); } catch (e) { _logger.severe("Failed to fix bad creationTime", e); } } Future _importFilesWithBadCreationTime() async { if (_prefs.containsKey(isBadCreationTimeImportDone)) { return; } _logger.info('_importFilesWithBadCreationTime'); final EnteWatch watch = EnteWatch("_importFilesWithBadCreationTime"); final int ownerID = Configuration.instance.getUserID()!; final filesGeneratedID = await FilesDB.instance .getGeneratedIDForFilesOlderThan(jan011981Time, ownerID); await _fileUpdationDB.insertMultiple( filesGeneratedID, FileUpdationDB.badCreationTime, ); watch.log("imported ${filesGeneratedID.length} files"); _prefs.setBool(isBadCreationTimeImportDone, true); } }