diff --git a/bob/ip/binseg/configs/datasets/csv.py b/bob/ip/binseg/configs/datasets/csv.py
index b8fc6ccfe05fb9995c24fc8d38f250277c467b85..c33bc62ce1478ef067cd5242fc21db115b09a32a 100644
--- a/bob/ip/binseg/configs/datasets/csv.py
+++ b/bob/ip/binseg/configs/datasets/csv.py
@@ -42,10 +42,6 @@ configuration.  **Copy it locally to make changes**:
    $ bob binseg config copy csv-dataset-example mydataset.py
    # edit mydataset.py as explained here, follow the comments
 
-Keep in mind that specific models require that you feed images respecting
-certain restrictions (input dimensions, image centering, etc.).  Check the
-configuration that was used to train models and try to match it.
-
 Finally, the only object this file needs to provide is one named ``dataset``,
 and it should contain a dictionary mapping a name, such as ``train``, ``dev``,
 or ``test``, to objects of type :py:class:`torch.utils.data.Dataset`.  As you
@@ -101,12 +97,16 @@ from bob.ip.binseg.data.dataset import CSVDataset
 
 _raw_dataset = CSVDataset(
     # path to the CSV file(s) - you may add as many subsets as you want:
-    # * "train" is used for training a model (stock data augmentation is
+    # * "__train__" is used for training a model (stock data augmentation is
     #    applied via our "make_dataset()" connector)
     # * anything else can be used for prediction and/or evaluation (if labels
     #   are also provided in such a set).  Data augmentation is NOT applied
     #   using our "make_dataset()" connector.
-    subsets={"train": "<path/to/train.csv>", "test": "<path/to/test.csv>"},
+    subsets={
+        "__train__": "<path/to/train.csv>",  #applies data augmentation
+        "train": "<path/to/train.csv>",  #no data augmentation, evaluate it
+        "test": "<path/to/test.csv>",  #no data augmentation, evaluate it
+        },
     fieldnames=("data", "label"),  # these are the column names
     loader=_loader,
     keymaker=data_path_keymaker,