Remove unused exports
This commit is contained in:
parent
87f60149e1
commit
58193c0d7f
|
@ -70,7 +70,7 @@ export const indexFaces = async (
|
||||||
return newMlFile;
|
return newMlFile;
|
||||||
};
|
};
|
||||||
|
|
||||||
export const syncFileAnalyzeFaces = async (fileContext: MLSyncFileContext) => {
|
const syncFileAnalyzeFaces = async (fileContext: MLSyncFileContext) => {
|
||||||
const { newMlFile } = fileContext;
|
const { newMlFile } = fileContext;
|
||||||
const startTime = Date.now();
|
const startTime = Date.now();
|
||||||
|
|
||||||
|
@ -118,7 +118,7 @@ const syncFileFaceDetections = async (fileContext: MLSyncFileContext) => {
|
||||||
*
|
*
|
||||||
* The model used is YOLO, running in an ONNX runtime.
|
* The model used is YOLO, running in an ONNX runtime.
|
||||||
*/
|
*/
|
||||||
export const detectFaces = async (
|
const detectFaces = async (
|
||||||
imageBitmap: ImageBitmap,
|
imageBitmap: ImageBitmap,
|
||||||
): Promise<Array<FaceDetection>> => {
|
): Promise<Array<FaceDetection>> => {
|
||||||
const maxFaceDistancePercent = Math.sqrt(2) / 100;
|
const maxFaceDistancePercent = Math.sqrt(2) / 100;
|
||||||
|
@ -284,7 +284,7 @@ const getFacesFromYOLOOutput = (
|
||||||
return faces;
|
return faces;
|
||||||
};
|
};
|
||||||
|
|
||||||
export const getRelativeDetection = (
|
const getRelativeDetection = (
|
||||||
faceDetection: FaceDetection,
|
faceDetection: FaceDetection,
|
||||||
dimensions: Dimensions,
|
dimensions: Dimensions,
|
||||||
): FaceDetection => {
|
): FaceDetection => {
|
||||||
|
@ -429,7 +429,7 @@ const ARC_FACE_5_LANDMARKS = [
|
||||||
*
|
*
|
||||||
* @param faceDetection A geometry indicating a face detected in an image.
|
* @param faceDetection A geometry indicating a face detected in an image.
|
||||||
*/
|
*/
|
||||||
export const faceAlignment = (faceDetection: FaceDetection): FaceAlignment => {
|
const faceAlignment = (faceDetection: FaceDetection): FaceAlignment => {
|
||||||
const landmarkCount = faceDetection.landmarks.length;
|
const landmarkCount = faceDetection.landmarks.length;
|
||||||
return getFaceAlignmentUsingSimilarityTransform(
|
return getFaceAlignmentUsingSimilarityTransform(
|
||||||
faceDetection,
|
faceDetection,
|
||||||
|
@ -494,10 +494,7 @@ function normalizeLandmarks(
|
||||||
/**
|
/**
|
||||||
* Laplacian blur detection.
|
* Laplacian blur detection.
|
||||||
*/
|
*/
|
||||||
export const detectBlur = (
|
const detectBlur = (alignedFaces: Float32Array, faces: Face[]): number[] => {
|
||||||
alignedFaces: Float32Array,
|
|
||||||
faces: Face[],
|
|
||||||
): number[] => {
|
|
||||||
const numFaces = Math.round(
|
const numFaces = Math.round(
|
||||||
alignedFaces.length /
|
alignedFaces.length /
|
||||||
(mobileFaceNetFaceSize * mobileFaceNetFaceSize * 3),
|
(mobileFaceNetFaceSize * mobileFaceNetFaceSize * 3),
|
||||||
|
@ -690,14 +687,14 @@ const syncFileFaceEmbeddings = async (
|
||||||
log.info("[MLService] facesWithEmbeddings: ", newMlFile.faces.length);
|
log.info("[MLService] facesWithEmbeddings: ", newMlFile.faces.length);
|
||||||
};
|
};
|
||||||
|
|
||||||
export const mobileFaceNetFaceSize = 112;
|
const mobileFaceNetFaceSize = 112;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Compute embeddings for the given {@link faceData}.
|
* Compute embeddings for the given {@link faceData}.
|
||||||
*
|
*
|
||||||
* The model used is MobileFaceNet, running in an ONNX runtime.
|
* The model used is MobileFaceNet, running in an ONNX runtime.
|
||||||
*/
|
*/
|
||||||
export const faceEmbeddings = async (
|
const faceEmbeddings = async (
|
||||||
faceData: Float32Array,
|
faceData: Float32Array,
|
||||||
): Promise<Array<FaceEmbedding>> => {
|
): Promise<Array<FaceEmbedding>> => {
|
||||||
const outputData = await workerBridge.faceEmbeddings(faceData);
|
const outputData = await workerBridge.faceEmbeddings(faceData);
|
||||||
|
@ -741,7 +738,7 @@ export const saveFaceCrop = async (imageBitmap: ImageBitmap, face: Face) => {
|
||||||
return blob;
|
return blob;
|
||||||
};
|
};
|
||||||
|
|
||||||
export const getFaceCrop = (
|
const getFaceCrop = (
|
||||||
imageBitmap: ImageBitmap,
|
imageBitmap: ImageBitmap,
|
||||||
faceDetection: FaceDetection,
|
faceDetection: FaceDetection,
|
||||||
): FaceCrop => {
|
): FaceCrop => {
|
||||||
|
|
Loading…
Reference in a new issue