Skip to content
Snippets Groups Projects
Commit 74b26aaa authored by Daniel CARRON's avatar Daniel CARRON :b: Committed by André Anjos
Browse files

[tests] Update tests

parent c871c3a2
No related branches found
No related tags found
1 merge request!46Create common library
...@@ -12,7 +12,7 @@ from clapper.logging import setup ...@@ -12,7 +12,7 @@ from clapper.logging import setup
from mednet.libs.common.scripts.click import ConfigCommand from mednet.libs.common.scripts.click import ConfigCommand
from mednet.libs.segmentation.engine.evaluator import SUPPORTED_METRIC_TYPE from mednet.libs.segmentation.engine.evaluator import SUPPORTED_METRIC_TYPE
logger = setup(__name__.split(".")[0], format="%(levelname)s: %(message)s") logger = setup("mednet")
def _validate_threshold(threshold: float | str, splits: list[str]): def _validate_threshold(threshold: float | str, splits: list[str]):
......
...@@ -361,11 +361,15 @@ def test_evaluate_lwnet_drive(temporary_basedir): ...@@ -361,11 +361,15 @@ def test_evaluate_lwnet_drive(temporary_basedir):
assert (evaluation_path / "evaluation.rst").exists() assert (evaluation_path / "evaluation.rst").exists()
keywords = { keywords = {
r"^Analyzing split `train`...$": 1, r"^Writing run metadata at.*$": 1,
r"^Analyzing split `test`...$": 1, r"^Counting true/false positive/negatives at split.*$": 2,
r"^Creating and saving plot at .*$": 1, r"^Writing run metadata at.*$": 1,
r"^Evaluating threshold on.*$": 1,
r"^Tabulating performance summary...": 1, r"^Tabulating performance summary...": 1,
r"^Saving evaluation results at.*$": 1,
r"^Saving table at .*$": 1, r"^Saving table at .*$": 1,
r"^Plotting performance curves...": 1,
r"^Saving figures at .*$": 1,
} }
buf.seek(0) buf.seek(0)
logging_output = buf.read() logging_output = buf.read()
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment