From 6b36db3557e3625a3e148cffadb9bdbf0739830c Mon Sep 17 00:00:00 2001 From: Tiago Freitas Pereira <tiagofrepereira@gmail.com> Date: Tue, 8 Dec 2020 09:16:24 +0100 Subject: [PATCH] Renamed CSVDatasetDevEval to CSVDataset --- bob/bio/base/database/__init__.py | 6 +++--- bob/bio/base/database/csv_dataset.py | 12 ++++++------ bob/bio/base/test/test_filelist.py | 26 +++++++++++++------------- doc/py_api.rst | 2 +- doc/vanilla_biometrics_features.rst | 12 ++++++------ 5 files changed, 29 insertions(+), 29 deletions(-) diff --git a/bob/bio/base/database/__init__.py b/bob/bio/base/database/__init__.py index c62f3bb7..62ca28e4 100644 --- a/bob/bio/base/database/__init__.py +++ b/bob/bio/base/database/__init__.py @@ -1,11 +1,11 @@ from .csv_dataset import ( - CSVDatasetDevEval, + CSVDataset, CSVToSampleLoader, CSVDatasetCrossValidation, CSVBaseSampleLoader, AnnotationsLoader, LSTToSampleLoader, - CSVDatasetDevEvalZTNorm, + CSVDatasetZTNorm, ) from .file import BioFile from .file import BioFileSet @@ -37,7 +37,7 @@ __appropriate__( BioDatabase, ZTBioDatabase, CSVBaseSampleLoader, - CSVDatasetDevEval, + CSVDataset, CSVToSampleLoader, CSVDatasetCrossValidation, ) diff --git a/bob/bio/base/database/csv_dataset.py b/bob/bio/base/database/csv_dataset.py index e1cae630..1ba5223d 100644 --- a/bob/bio/base/database/csv_dataset.py +++ b/bob/bio/base/database/csv_dataset.py @@ -260,7 +260,7 @@ def path_discovery(dataset_protocol_path, option1, option2): return op1 if op1 else find_element_in_tarball(dataset_protocol_path, option2) -class CSVDatasetDevEval(Database): +class CSVDataset(Database): """ Generic filelist dataset for :any:` bob.bio.base.pipelines.vanilla_biometrics.VanillaBiometricsPipeline` pipeline. Check :any:`vanilla_biometrics_features` for more details about the Vanilla Biometrics Dataset @@ -570,14 +570,14 @@ class CSVDatasetDevEval(Database): return groups -class CSVDatasetDevEvalZTNorm(Database): +class CSVDatasetZTNorm(Database): """ Generic filelist dataset for :any:`bob.bio.base.pipelines.vanilla_biometrics.ZTNormPipeline` pipelines. Check :any:`vanilla_biometrics_features` for more details about the Vanilla Biometrics Dataset interface. - This dataset interface takes as in put a :any:`CSVDatasetDevEval` as input and have two extra methods: - :any:`CSVDatasetDevEvalZTNorm.zprobes` and :any:`CSVDatasetDevEvalZTNorm.treferences`. + This dataset interface takes as in put a :any:`CSVDataset` as input and have two extra methods: + :any:`CSVDatasetZTNorm.zprobes` and :any:`CSVDatasetZTNorm.treferences`. To create a new dataset, you need to provide a directory structure similar to the one below: @@ -595,8 +595,8 @@ class CSVDatasetDevEvalZTNorm(Database): Parameters ---------- - database: :any:`CSVDatasetDevEval` - :any:`CSVDatasetDevEval` to be aggregated + database: :any:`CSVDataset` + :any:`CSVDataset` to be aggregated """ diff --git a/bob/bio/base/test/test_filelist.py b/bob/bio/base/test/test_filelist.py index 0658cd84..7a6ae781 100644 --- a/bob/bio/base/test/test_filelist.py +++ b/bob/bio/base/test/test_filelist.py @@ -8,12 +8,12 @@ import os import bob.io.base import bob.io.base.test_utils from bob.bio.base.database import ( - CSVDatasetDevEval, + CSVDataset, CSVToSampleLoader, CSVDatasetCrossValidation, AnnotationsLoader, LSTToSampleLoader, - CSVDatasetDevEvalZTNorm, + CSVDatasetZTNorm, ) import nose.tools from bob.pipelines import DelayedSample, SampleSet @@ -62,7 +62,7 @@ def check_all_true(list_of_something, something): def test_csv_file_list_dev_only(): - dataset = CSVDatasetDevEval(example_dir, "protocol_only_dev") + dataset = CSVDataset(example_dir, "protocol_only_dev") assert len(dataset.background_model_samples()) == 8 assert check_all_true(dataset.background_model_samples(), DelayedSample) @@ -75,7 +75,7 @@ def test_csv_file_list_dev_only(): def test_csv_file_list_dev_only_metadata(): - dataset = CSVDatasetDevEval(example_dir, "protocol_only_dev_metadata") + dataset = CSVDataset(example_dir, "protocol_only_dev_metadata") assert len(dataset.background_model_samples()) == 8 assert check_all_true(dataset.background_model_samples(), DelayedSample) @@ -107,7 +107,7 @@ def test_csv_file_list_dev_eval(): ) def run(filename): - dataset = CSVDatasetDevEval( + dataset = CSVDataset( filename, "protocol_dev_eval", csv_to_sample_loader=CSVToSampleLoader( @@ -161,7 +161,7 @@ def test_csv_file_list_dev_eval_score_norm(): ) def run(filename): - dataset = CSVDatasetDevEval( + dataset = CSVDataset( filename, "protocol_dev_eval", csv_to_sample_loader=CSVToSampleLoader( @@ -175,7 +175,7 @@ def test_csv_file_list_dev_eval_score_norm(): extension="", ), ) - znorm_dataset = CSVDatasetDevEvalZTNorm(dataset) + znorm_dataset = CSVDatasetZTNorm(dataset) assert len(znorm_dataset.background_model_samples()) == 8 assert check_all_true(znorm_dataset.background_model_samples(), DelayedSample) @@ -222,7 +222,7 @@ def test_csv_file_list_dev_eval_sparse(): ) ) - dataset = CSVDatasetDevEval( + dataset = CSVDataset( example_dir, "protocol_dev_eval_sparse", csv_to_sample_loader=CSVToSampleLoader( @@ -280,7 +280,7 @@ def test_csv_file_list_dev_eval_sparse(): def test_lst_file_list_dev_eval(): - dataset = CSVDatasetDevEval( + dataset = CSVDataset( legacy_example_dir, "", csv_to_sample_loader=LSTToSampleLoader( @@ -315,7 +315,7 @@ def test_lst_file_list_dev_eval(): def test_lst_file_list_dev_eval_sparse(): - dataset = CSVDatasetDevEval( + dataset = CSVDataset( legacy_example_dir, "", csv_to_sample_loader=LSTToSampleLoader( @@ -351,7 +351,7 @@ def test_lst_file_list_dev_eval_sparse(): def test_lst_file_list_dev_sparse_filelist2(): - dataset = CSVDatasetDevEval( + dataset = CSVDataset( legacy2_example_dir, "", csv_to_sample_loader=LSTToSampleLoader( @@ -369,7 +369,7 @@ def test_lst_file_list_dev_sparse_filelist2(): def test_csv_file_list_atnt(): - dataset = CSVDatasetDevEval(atnt_protocol_path, "idiap_protocol") + dataset = CSVDataset(atnt_protocol_path, "idiap_protocol") assert len(dataset.background_model_samples()) == 200 assert len(dataset.references()) == 20 assert len(dataset.probes()) == 100 @@ -419,7 +419,7 @@ def run_experiment(dataset): def test_atnt_experiment(): - dataset = CSVDatasetDevEval( + dataset = CSVDataset( dataset_protocol_path=atnt_protocol_path, protocol_name="idiap_protocol", csv_to_sample_loader=CSVToSampleLoader( diff --git a/doc/py_api.rst b/doc/py_api.rst index 927b4fc0..fe3a92bc 100644 --- a/doc/py_api.rst +++ b/doc/py_api.rst @@ -22,7 +22,7 @@ Database implementations .. autosummary:: -.. bob.bio.base.database.CSVDatasetDevEval +.. bob.bio.base.database.CSVDataset .. bob.bio.base.database.CSVDatasetCrossValidation Biometric Algorithm diff --git a/doc/vanilla_biometrics_features.rst b/doc/vanilla_biometrics_features.rst index 88e2eb94..f4528bf8 100644 --- a/doc/vanilla_biometrics_features.rst +++ b/doc/vanilla_biometrics_features.rst @@ -57,7 +57,7 @@ $ bob bio pipelines vanilla-biometrics -d my_database.py -p <pipeline_name> The ``database`` object defined in ``my_database.py`` is an instance of either: -- A :py:class:`~bob.bio.base.database.CSVDatasetDevEval` (see :ref:`here <bob.bio.base.database.csv_file_interface>`), +- A :py:class:`~bob.bio.base.database.CSVDataset` (see :ref:`here <bob.bio.base.database.csv_file_interface>`), - A :py:class:`~bob.bio.base.database.CSVDatasetCrossValidation` (see :ref:`here <bob.bio.base.database.csv_cross_validation>`), - Your own implementation of a :ref:`Database Interface <bob.bio.base.database.interface_class>`, - A :ref:`legacy Database connector <bob.bio.base.legacy.database_connector>`. @@ -73,7 +73,7 @@ This method is less complete and less flexible than implementing a :ref:`full in Protocol definition is possible and a set of csv files (at least ``dev_enroll.csv`` and ``dev_probe.csv``) in a folder must be created for each protocol. -The interface is created with :py:class:`~bob.bio.base.database.CSVDatasetDevEval`. +The interface is created with :py:class:`~bob.bio.base.database.CSVDataset`. This class takes as input the base directory, and the protocol sub-directory of the :ref:`csv file structure <bob.bio.base.database.csv_file_structure>`, and finally a ``csv_to_sample_loader`` that will load a sample data from a csv row read from the csv files. This csv_to_sample_loader needs to know the dataset base path and the extension of the dataset files. @@ -151,11 +151,11 @@ The following file structure and file naming must be followed, in order for the - The ``train.csv`` file (as shown in ``my_protocol_2``) is optional and contains the information of the *world* set. - The ``eval_enroll.csv`` and ``eval_probe.csv`` files (as shown in ``my_protocol_2``) are optional and contain the information of the *eval* set. -In this example, ``my_dataset`` would be the base path given to the ``dataset_protocol_path`` parameter of :py:class:`~bob.bio.base.database.CSVDatasetDevEval`, and ``my_protocol_1`` the ``protocol_name`` parameter: +In this example, ``my_dataset`` would be the base path given to the ``dataset_protocol_path`` parameter of :py:class:`~bob.bio.base.database.CSVDataset`, and ``my_protocol_1`` the ``protocol_name`` parameter: .. code-block:: python - from bob.bio.base.database import CSVDatasetDevEval, AnnotationsLoader + from bob.bio.base.database import CSVDataset, AnnotationsLoader # Define a loading function called for each sample with its path def my_load_function(full_path): @@ -172,7 +172,7 @@ In this example, ``my_dataset`` would be the base path given to the ``dataset_pr ) # Create the csv interface - database = CSVDatasetDevEval("my_dataset", "my_protocol_1", csv_to_sample_loader=my_sample_loader) + database = CSVDataset("my_dataset", "my_protocol_1", csv_to_sample_loader=my_sample_loader) This will create a database interface with: @@ -189,7 +189,7 @@ CSV file Cross-validation Database interface The :py:class:`~bob.bio.base.database.CSVDatasetCrossValidation` takes only one CSV file of identities and creates the necessary sets pseudo-randomly. -The format of the CSV file is the same as in :py:class:`~bob.bio.base.database.CSVDatasetDevEval`, comma separated with a header: +The format of the CSV file is the same as in :py:class:`~bob.bio.base.database.CSVDataset`, comma separated with a header: .. code-block:: text -- GitLab