Skip to content

Commit

Permalink
🩹🧪 Fix tests as far as possible
Browse files Browse the repository at this point in the history
  • Loading branch information
s-weigand committed Jun 30, 2024
1 parent b10882d commit abccfdf
Show file tree
Hide file tree
Showing 3 changed files with 12 additions and 15 deletions.
10 changes: 3 additions & 7 deletions tests/conftest.py
Original file line number Diff line number Diff line change
Expand Up @@ -8,12 +8,10 @@
import netCDF4 # noqa: F401
# isort: on

from dataclasses import replace

import pytest
from glotaran.optimization.optimize import optimize
from glotaran.testing.simulated_data.parallel_spectral_decay import SCHEME as SCHEME_PAR
from glotaran.testing.simulated_data.sequential_spectral_decay import SCHEME as SCHEME_SEQ
from glotaran.testing.simulated_data.shared_decay import PARAMETERS

from pyglotaran_extras.io.setup_case_study import get_script_dir

Expand All @@ -26,12 +24,10 @@ def wrapped_get_script_dir():
@pytest.fixture(scope="session")
def result_parallel_spectral_decay():
"""Test result from ``glotaran.testing.simulated_data.parallel_spectral_decay``."""
scheme = replace(SCHEME_PAR, maximum_number_function_evaluations=1)
return optimize(scheme)
return SCHEME_PAR.optimize(PARAMETERS, maximum_number_function_evaluations=1)


@pytest.fixture(scope="session")
def result_sequential_spectral_decay():
"""Test result from ``glotaran.testing.simulated_data.sequential_spectral_decay``."""
scheme = replace(SCHEME_SEQ, maximum_number_function_evaluations=1)
return optimize(scheme)
return SCHEME_SEQ.optimize(PARAMETERS, maximum_number_function_evaluations=1)
8 changes: 4 additions & 4 deletions tests/inspect/test_a_matrix.py
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@ def test_a_matrix_to_html_table(
encoding="utf8"
)
assert a_matrix_to_html_table(
result_parallel_spectral_decay.data["dataset_1"].a_matrix_megacomplex_parallel_decay,
result_parallel_spectral_decay.data["parallel-decay"].a_matrix_megacomplex_parallel_decay,
"megacomplex_parallel_decay",
**kwargs,
) == expected.rstrip("\n")
Expand Down Expand Up @@ -60,10 +60,10 @@ def test_show_a_matrixes(
)

result = result_parallel_spectral_decay
result.data["dataset_2"] = result_sequential_spectral_decay.data["dataset_1"]
result.data["dataset_2"] = result_sequential_spectral_decay.data["sequential-decay"]
# dummy data for filtering based on a-matrix size
single_entry_data = result_sequential_spectral_decay.data[
"dataset_1"
"sequential-decay"
].a_matrix_megacomplex_sequential_decay[:1, :1]
single_entry_data = single_entry_data.rename(
{
Expand All @@ -85,7 +85,7 @@ def test_show_a_matrixes_multiple_a_matrixes_in_dataset(
).read_text(encoding="utf8")

single_entry_data = result_sequential_spectral_decay.data[
"dataset_1"
"sequential-decay"
].a_matrix_megacomplex_sequential_decay[:1, :1]

a_matrix_one = single_entry_data.rename(
Expand Down
9 changes: 5 additions & 4 deletions tests/io/test_load_data.py
Original file line number Diff line number Diff line change
Expand Up @@ -43,7 +43,8 @@ def test_load_data(
result_sequential_spectral_decay: Result, tmp_path: Path, recwarn: WarningsRecorder
):
"""All input_type permutations result in a ``xr.Dataset``."""
compare_dataset = result_sequential_spectral_decay.data["dataset_1"]
compare_dataset = result_sequential_spectral_decay.data["sequential-decay"]
save_path = tmp_path / "result"

from_result = load_data(result_sequential_spectral_decay)

Expand All @@ -53,9 +54,9 @@ def test_load_data(

run_load_data_test(from_dataset, compare_dataset)

result_sequential_spectral_decay.save(tmp_path / "result.yml")
result_sequential_spectral_decay.save(save_path)

from_file = load_data(tmp_path / "dataset_1.nc")
from_file = load_data(save_path / "data/sequential-decay.nc")

run_load_data_test(from_file, compare_dataset)

Expand All @@ -69,7 +70,7 @@ def test_load_data(
assert len(filter_warnings(recwarn)) == 0

# Ensure not to mutate original fixture
result_multi_dataset = load_result(tmp_path / "result.yml")
result_multi_dataset = load_result(save_path)
result_multi_dataset.data["dataset_2"] = xr.Dataset({"foo": [1]})

from_result_multi_dataset = load_data(result_multi_dataset)
Expand Down

0 comments on commit abccfdf

Please sign in to comment.