diff --git a/bob/bio/face/config/baseline/helpers.py b/bob/bio/face/config/baseline/helpers.py index 0e0aefecd11d7f325fc390e1fb19bcebe1ce7340..0daeaf093a87f852bb5fdfa263a7852c3b847578 100644 --- a/bob/bio/face/config/baseline/helpers.py +++ b/bob/bio/face/config/baseline/helpers.py @@ -232,31 +232,3 @@ def embedding_transformer( ) return transformer - - -def embedding_transformer_224x224( - embedding, annotation_type, fixed_positions, color_channel="rgb" -): - """ - Creates a pipeline composed by and FaceCropper and an Embedding extractor. - This transformer is suited for Facenet based architectures - - .. warning:: - This will resize images to :math:`112 \times 112` - - """ - cropped_image_size = (224, 224) - if annotation_type == "eyes-center": - # Hard coding eye positions for backward consistency - cropped_positions = {"leye": (65, 150), "reye": (65, 77)} - else: - # Will use default - cropped_positions = dnn_default_cropping(cropped_image_size, annotation_type) - - return embedding_transformer( - cropped_image_size=cropped_image_size, - embedding=embedding, - cropped_positions=cropped_positions, - fixed_positions=fixed_positions, - color_channel=color_channel, - )