Bug#1022255: python-xarray breaks satpy autopkgtest

Paul Gevers elbrus at debian.org
Sat Oct 22 20:36:41 BST 2022


Source: python-xarray, satpy
Control: found -1 python-xarray/2022.10.0-1
Control: found -1 satpy/0.37.1-1
Severity: serious
Tags: sid bookworm
User: debian-ci at lists.debian.org
Usertags: breaks needs-update

Dear maintainer(s),

With a recent upload of python-xarray the autopkgtest of satpy fails in 
testing when that autopkgtest is run with the binary packages of 
python-xarray from unstable. It passes when run with only packages from 
testing. In tabular form:

                        pass            fail
python-xarray          from testing    2022.10.0-1
satpy                  from testing    0.37.1-1
all others             from testing    from testing

I copied some of the output at the bottom of this report.

Currently this regression is blocking the migration of python-xarray to 
testing [1]. Due to the nature of this issue, I filed this bug report 
against both packages. Can you please investigate the situation and 
reassign the bug to the right package?

More information about this bug and the reason for filing it can be found on
https://wiki.debian.org/ContinuousIntegration/RegressionEmailInformation

Paul

[1] https://qa.debian.org/excuses.php?package=python-xarray

https://ci.debian.net/data/autopkgtest/testing/amd64/s/satpy/27389134/log.gz

____________________ TestNIRReflectance.test_no_sunz_no_co2 
____________________

self = <satpy.tests.test_modifiers.TestNIRReflectance 
testMethod=test_no_sunz_no_co2>
calculator = <MagicMock name='Calculator' id='139765627652032'>
apply_modifier_info = <MagicMock name='apply_modifier_info' 
id='139765627664464'>
sza = <MagicMock name='sun_zenith_angle' id='139765627656784'>

     @mock.patch('satpy.modifiers.spectral.sun_zenith_angle')
     @mock.patch('satpy.modifiers.NIRReflectance.apply_modifier_info')
     @mock.patch('satpy.modifiers.spectral.Calculator')
     def test_no_sunz_no_co2(self, calculator, apply_modifier_info, sza):
         """Test NIR reflectance compositor with minimal parameters."""
         calculator.return_value = mock.MagicMock(
             reflectance_from_tbs=self.refl_from_tbs)
         sza.return_value = self.da_sunz
         from satpy.modifiers.spectral import NIRReflectance
             comp = NIRReflectance(name='test')
         info = {'modifiers': None}
         res = comp([self.nir, self.ir_], optional_datasets=[], **info)
     >       self.get_lonlats.assert_called()

/usr/lib/python3/dist-packages/satpy/tests/test_modifiers.py:244: _ _ _ 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <MagicMock name='mock.get_lonlats' id='139765625209184'>

     def assert_called(self):
         """assert that the mock was called at least once
         """
         if self.call_count == 0:
             msg = ("Expected '%s' to have been called." %
                    (self._mock_name or 'mock'))
>           raise AssertionError(msg)
E           AssertionError: Expected 'get_lonlats' to have been called.

/usr/lib/python3.10/unittest/mock.py:888: AssertionError
_______________________ TestSceneLoading.test_load_comp4 
_______________________

self = <satpy.tests.test_scene.TestSceneLoading object at 0x7f1e012eb610>

     def test_load_comp4(self):
         """Test loading a composite that depends on a composite."""
         scene = Scene(filenames=['fake1_1.txt'], reader='fake1')
         scene.load(['comp4'])
         loaded_ids = list(scene._datasets.keys())
>       assert len(loaded_ids) == 1
E       AssertionError: assert 2 == 1
E        +  where 2 = len([DataID(name='comp2'), DataID(name='ds3', 
modifiers=())])

/usr/lib/python3/dist-packages/satpy/tests/test_scene.py:1059: 
AssertionError
------------------------------ Captured log call 
-------------------------------
WARNING  satpy.scene:scene.py:1275 The following datasets were not 
created and may require resampling to be generated: DataID(name='comp4')
_______________ TestSceneLoading.test_load_multiple_resolutions 
________________

self = <satpy.tests.test_scene.TestSceneLoading object at 0x7f1e012eb8b0>

     def test_load_multiple_resolutions(self):
         """Test loading a dataset has multiple resolutions available 
with different resolutions."""
         scene = Scene(filenames=['fake1_1.txt'], reader='fake1')
         comp25 = make_cid(name='comp25', resolution=1000)
         scene[comp25] = xr.DataArray([], attrs={'name': 'comp25', 
'resolution': 1000})
         scene.load(['comp25'], resolution=500)
             loaded_ids = list(scene._datasets.keys())
>       assert len(loaded_ids) == 2
E       AssertionError: assert 3 == 2
E        +  where 3 = len([DataID(name='comp24', resolution=500), 
DataID(name='comp25', resolution=1000), DataID(name='ds5', 
resolution=500, modifiers=())])

/usr/lib/python3/dist-packages/satpy/tests/test_scene.py:1070: 
AssertionError
------------------------------ Captured log call 
-------------------------------
WARNING  satpy.scene:scene.py:1275 The following datasets were not 
created and may require resampling to be generated: DataID(name='comp25')
_________________ TestSceneLoading.test_load_same_subcomposite 
_________________

self = <satpy.tests.test_scene.TestSceneLoading object at 0x7f1e012eb190>

     def test_load_same_subcomposite(self):
         """Test loading a composite and one of it's subcomposites at 
the same time."""
         scene = Scene(filenames=['fake1_1.txt'], reader='fake1')
         scene.load(['comp24', 'comp25'], resolution=500)
         loaded_ids = list(scene._datasets.keys())
         assert len(loaded_ids) == 2
         assert loaded_ids[0]['name'] == 'comp24'
         assert loaded_ids[0]['resolution'] == 500
>       assert loaded_ids[1]['name'] == 'comp25'
E       AssertionError: assert 'ds5' == 'comp25'
E         - comp25
E         + ds5

/usr/lib/python3/dist-packages/satpy/tests/test_scene.py:1084: 
AssertionError
------------------------------ Captured log call 
-------------------------------
WARNING  satpy.scene:scene.py:1275 The following datasets were not 
created and may require resampling to be generated: DataID(name='comp25')
_______________________ TestSceneLoading.test_load_comp9 
_______________________

self = <satpy.tests.test_scene.TestSceneLoading object at 0x7f1e012e9000>

     def test_load_comp9(self):
         """Test loading a composite that has a non-existent optional 
prereq."""
         # it is fine that an optional prereq doesn't exist
         scene = Scene(filenames=['fake1_1.txt'], reader='fake1')
         scene.load(['comp9'])
         loaded_ids = list(scene._datasets.keys())
>       assert len(loaded_ids) == 1
E       AssertionError: assert 2 == 1
E        +  where 2 = len([DataID(name='comp2'), DataID(name='ds1', 
resolution=250, calibration=<calibration.reflectance>, modifiers=())])

/usr/lib/python3/dist-packages/satpy/tests/test_scene.py:1114: 
AssertionError
------------------------------ Captured log call 
-------------------------------
WARNING  satpy.scene:scene.py:1275 The following datasets were not 
created and may require resampling to be generated: DataID(name='comp9')
______________________ TestSceneLoading.test_load_comp10 
_______________________

self = <satpy.tests.test_scene.TestSceneLoading object at 0x7f1e012e97e0>

     def test_load_comp10(self):
         """Test loading a composite that depends on a modified dataset."""
         # it is fine that an optional prereq doesn't exist
         scene = Scene(filenames=['fake1_1.txt'], reader='fake1')
         scene.load(['comp10'])
         loaded_ids = list(scene._datasets.keys())
>       assert len(loaded_ids) == 1
E       AssertionError: assert 2 == 1
E        +  where 2 = len([DataID(name='comp2'), DataID(name='ds1', 
resolution=250, calibration=<calibration.reflectance>, 
modifiers=('mod1',))])

/usr/lib/python3/dist-packages/satpy/tests/test_scene.py:1123: 
AssertionError
------------------------------ Captured log call 
-------------------------------
WARNING  satpy.scene:scene.py:1275 The following datasets were not 
created and may require resampling to be generated: DataID(name='comp10')
______________________ TestSceneLoading.test_load_comp19 
_______________________

self = <satpy.tests.test_scene.TestSceneLoading object at 0x7f1e011abfa0>

     def test_load_comp19(self):
         """Test loading a composite that shares a dep with a dependency.
             More importantly test that loading a dependency that depends on
         the same dependency as this composite (a sibling dependency) and
         that sibling dependency includes a modifier. This test makes sure
         that the Node in the dependency tree is the exact same node.
             """
         # Check dependency tree nodes
         # initialize the dep tree without loading the data
         scene = Scene(filenames=['fake1_1.txt'], reader='fake1')
         scene._update_dependency_tree({'comp19'}, None)
             this_node = scene._dependency_tree['comp19']
         shared_dep_id = make_dataid(name='ds5', modifiers=('res_change',))
         shared_dep_expected_node = scene._dependency_tree[shared_dep_id]
         # get the node for the first dep in the prereqs list of the
         # comp13 node
         shared_dep_node = scene._dependency_tree['comp13'].data[1][0]
         shared_dep_node2 = this_node.data[1][0]
         assert shared_dep_expected_node is shared_dep_node
         assert shared_dep_expected_node is shared_dep_node2
             scene.load(['comp19'])
             loaded_ids = list(scene._datasets.keys())
>       assert len(loaded_ids) == 1
E       AssertionError: assert 3 == 1
E        +  where 3 = len([DataID(name='comp13'), DataID(name='ds2', 
resolution=250, calibration=<calibration.reflectance>, modifiers=()), 
DataID(name='ds5', resolution=250, modifiers=('res_change',))])

/usr/lib/python3/dist-packages/satpy/tests/test_scene.py:1266: 
AssertionError
------------------------------ Captured log call 
-------------------------------
WARNING  satpy.scene:scene.py:1275 The following datasets were not 
created and may require resampling to be generated: DataID(name='comp19')
_____________ TestSceneLoading.test_load_dataset_after_composite2 
______________

self = <satpy.tests.test_scene.TestSceneLoading object at 0x7f1e011ab280>

     def test_load_dataset_after_composite2(self):
         """Test load complex composite followed by other datasets."""
         from satpy.readers.yaml_reader import FileYAMLReader
         from satpy.tests.utils import FakeCompositor, FakeModifier
         load_mock = spy_decorator(FileYAMLReader.load)
         comp_mock = spy_decorator(FakeCompositor.__call__)
         mod_mock = spy_decorator(FakeModifier.__call__)
         with mock.patch.object(FileYAMLReader, 'load', load_mock), \
              mock.patch.object(FakeCompositor, '__call__', comp_mock), \
              mock.patch.object(FakeModifier, '__call__', mod_mock):
             lmock = load_mock.mock
             scene = Scene(filenames=['fake1_1.txt'], reader='fake1')
             scene.load(['comp10'])
             assert lmock.call_count == 1
             loaded_ids = list(scene._datasets.keys())
>           assert len(loaded_ids) == 1
E           AssertionError: assert 2 == 1
E            +  where 2 = len([DataID(name='comp2'), DataID(name='ds1', 
resolution=250, calibration=<calibration.reflectance>, 
modifiers=('mod1',))])

/usr/lib/python3/dist-packages/satpy/tests/test_scene.py:1353: 
AssertionError
------------------------------ Captured log call 
-------------------------------
WARNING  satpy.scene:scene.py:1275 The following datasets were not 
created and may require resampling to be generated: DataID(name='comp10')
___________________ TestSceneLoading.test_no_generate_comp10 
___________________

self = <satpy.tests.test_scene.TestSceneLoading object at 0x7f1e01467520>

     def test_no_generate_comp10(self):
         """Test generating a composite after loading."""
         # it is fine that an optional prereq doesn't exist
         scene = Scene(filenames=['fake1_1.txt'], reader='fake1')
         scene.load(['comp10'], generate=False)
         assert any(ds_id['name'] == 'comp10' for ds_id in scene._wishlist)
         assert 'comp10' not in scene._datasets
         # two dependencies should have been loaded
         assert len(scene._datasets) == 2
         assert len(scene.missing_datasets) == 1
             scene._generate_composites_from_loaded_datasets()
         assert any(ds_id['name'] == 'comp10' for ds_id in scene._wishlist)
>       assert 'comp10' in scene._datasets
E       AssertionError: assert 'comp10' in {DataID(name='ds1', 
resolution=250, calibration=<calibration.reflectance>, modifiers=()): 
<xarray.DataArray 'zeros_like-1d67fcaf0431507eb1a9b35f8d43c7dc' (y: 20, 
x: 20)>\ndask.array<zeros_like, shape=(20, 20), dtype=float64, 
chunksize=(20, 20), chunktype=numpy.ndarray>\nCoordinates:\n    crs 
object +proj=longlat +ellps=WGS84 +type=crs\nDimensions without 
coordinates: y, x\nAttributes: (12/14)\n    start_time: 
2020-01-01 00:00:00\n    end_time:             2020-01-01 01:00:00\n 
name:                 ds1\n    resolution:           250\n 
calibration:          reflectance\n    modifiers:            ()\n    ... 
                   ...\n    sensor:               fake_sensor\n 
platform_name:        fake_platform\n    reader:               fake1\n 
  area:                 Shape: (20, 20)\nLons: <xarray.DataArray 
'zeros_lik...\n    _satpy_id:            DataID(name='ds1', 
resolution=250, calibration=<cal...\n    ancillary_variables:  [], 
DataID(name='ds2', resolution=250, 
calibration=<calibration.reflectance>, modifiers=()): <xarray.DataArray 
'zeros_like-1d67fcaf0431507eb1a9b35f8d43c7dc' (y: 20, x: 
20)>\ndask.array<zeros_like, shape=(20, 20), dtype=float64, 
chunksize=(20, 20), chunktyp...:           2020-01-01 00:00:00\n 
end_time:             2020-01-01 01:00:00\n    name: 
ds1\n    resolution:           250\n    calibration: 
reflectance\n    modifiers:            ('mod1',)\n    ... 
    ...\n    sensor:               fake_sensor\n    platform_name: 
  fake_platform\n    reader:               fake1\n    area: 
    Shape: (20, 20)\nLons: <xarray.DataArray 'zeros_lik...\n 
_satpy_id:            DataID(name='ds1', resolution=250, 
calibration=<cal...\n    ancillary_variables:  [], DataID(name='comp2'): 
<xarray.DataArray 'zeros_like-92483e10e77d6b5cc4ab18a829286920' (y: 20, 
x: 20,\n 
bands: 3)>\ndask.array<zeros_like, shape=(20, 20, 3), dtype=float64, 
chunksize=(20, 20, 3), chunktype=numpy.ndarray>\nCoordinates:\n  * bands 
    (bands) <U1 'R' 'G' 'B'\nDimensions without coordinates: y, 
x\nAttributes:\n    _satpy_id:               DataID(name='comp2')\n 
name:                    comp2\n    prerequisites:           ['ds1', 
'ds2']\n    optional_prerequisites:  []\n    optional_datasets: 
[]\n    area:                    Shape: (20, 20)\nLons: 
<xarray.DataArray 'zeros_...}
E        +  where {DataID(name='ds1', resolution=250, 
calibration=<calibration.reflectance>, modifiers=()): <xarray.DataArray 
'zeros_like-1d67fcaf0431507eb1a9b35f8d43c7dc' (y: 20, x: 
20)>\ndask.array<zeros_like, shape=(20, 20), dtype=float64, 
chunksize=(20, 20), chunktype=numpy.ndarray>\nCoordinates:\n    crs 
object +proj=longlat +ellps=WGS84 +type=crs\nDimensions without 
coordinates: y, x\nAttributes: (12/14)\n    start_time: 
2020-01-01 00:00:00\n    end_time:             2020-01-01 01:00:00\n 
name:                 ds1\n    resolution:           250\n 
calibration:          reflectance\n    modifiers:            ()\n    ... 
                   ...\n    sensor:               fake_sensor\n 
platform_name:        fake_platform\n    reader:               fake1\n 
  area:                 Shape: (20, 20)\nLons: <xarray.DataArray 
'zeros_lik...\n    _satpy_id:            DataID(name='ds1', 
resolution=250, calibration=<cal...\n    ancillary_variables:  [], 
DataID(name='ds2', resolution=250, 
calibration=<calibration.reflectance>, modifiers=()): <xarray.DataArray 
'zeros_like-1d67fcaf0431507eb1a9b35f8d43c7dc' (y: 20, x: 
20)>\ndask.array<zeros_like, shape=(20, 20), dtype=float64, 
chunksize=(20, 20), chunktyp...:           2020-01-01 00:00:00\n 
end_time:             2020-01-01 01:00:00\n    name: 
ds1\n    resolution:           250\n    calibration: 
reflectance\n    modifiers:            ('mod1',)\n    ... 
    ...\n    sensor:               fake_sensor\n    platform_name: 
  fake_platform\n    reader:               fake1\n    area: 
    Shape: (20, 20)\nLons: <xarray.DataArray 'zeros_lik...\n 
_satpy_id:            DataID(name='ds1', resolution=250, 
calibration=<cal...\n    ancillary_variables:  [], DataID(name='comp2'): 
<xarray.DataArray 'zeros_like-92483e10e77d6b5cc4ab18a829286920' (y: 20, 
x: 20,\n 
bands: 3)>\ndask.array<zeros_like, shape=(20, 20, 3), dtype=float64, 
chunksize=(20, 20, 3), chunktype=numpy.ndarray>\nCoordinates:\n  * bands 
    (bands) <U1 'R' 'G' 'B'\nDimensions without coordinates: y, 
x\nAttributes:\n    _satpy_id:               DataID(name='comp2')\n 
name:                    comp2\n    prerequisites:           ['ds1', 
'ds2']\n    optional_prerequisites:  []\n    optional_datasets: 
[]\n    area:                    Shape: (20, 20)\nLons: 
<xarray.DataArray 'zeros_...} = <satpy.scene.Scene object at 
0x7f1dd68203a0>._datasets

/usr/lib/python3/dist-packages/satpy/tests/test_scene.py:1433: 
AssertionError
_____________ TestSceneResampling.test_resample_reduce_data_toggle 
_____________

self = {DataID(name='ds2', resolution=250, 
calibration=<calibration.reflectance>, modifiers=()): <xarray.DataArray 
'zeros_lik...:  []
     optional_datasets:       []
     area:                    Shape: (20, 20)\nLons: <xarray.DataArray 
'zeros_...}
item = 'comp19'

     def __getitem__(self, item):
         """Get item from container."""
         try:
             # short circuit - try to get the object without more work
>           return super(DatasetDict, self).__getitem__(item)
E           KeyError: 'comp19'

/usr/lib/python3/dist-packages/satpy/dataset/data_dict.py:169: KeyError

During handling of the above exception, another exception occurred:

self = <satpy.tests.test_scene.TestSceneResampling object at 0x7f1e014664a0>
rs = <MagicMock name='resample_dataset' id='139766129303824'>

     @mock.patch('satpy.scene.resample_dataset')
     def test_resample_reduce_data_toggle(self, rs):
         """Test that the Scene can be reduced or not reduced during 
resampling."""
         from pyresample.geometry import AreaDefinition
             rs.side_effect = self._fake_resample_dataset_force_20x20
         proj_str = ('+proj=lcc +datum=WGS84 +ellps=WGS84 '
                     '+lon_0=-95. +lat_0=25 +lat_1=25 +units=m +no_defs')
         target_area = AreaDefinition('test', 'test', 'test', proj_str, 
4, 4, (-1000., -1500., 1000., 1500.))
         area_def = AreaDefinition('test', 'test', 'test', proj_str, 5, 
5, (-1000., -1500., 1000., 1500.))
         area_def.get_area_slices = mock.MagicMock()
         get_area_slices = area_def.get_area_slices
         get_area_slices.return_value = (slice(0, 3, None), slice(0, 3, 
None))
         area_def_big = AreaDefinition('test', 'test', 'test', proj_str, 
10, 10, (-1000., -1500., 1000., 1500.))
         area_def_big.get_area_slices = mock.MagicMock()
         get_area_slices_big = area_def_big.get_area_slices
         get_area_slices_big.return_value = (slice(0, 6, None), slice(0, 
6, None))
             # Test that data reduction can be disabled
         scene = Scene(filenames=['fake1_1.txt'], reader='fake1')
         scene.load(['comp19'])
>       scene['comp19'].attrs['area'] = area_def

/usr/lib/python3/dist-packages/satpy/tests/test_scene.py:1665: _ _ _ _ _ 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
/usr/lib/python3/dist-packages/satpy/scene.py:803: in __getitem__
     return self._datasets[key]
/usr/lib/python3/dist-packages/satpy/dataset/data_dict.py:171: in 
__getitem__
     key = self.get_key(item)
/usr/lib/python3/dist-packages/satpy/dataset/data_dict.py:158: in get_key
     return get_key(match_key, self.keys(), num_results=num_results,
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
_ _ _ _
key = DataQuery(name='comp19')
key_container = [DataID(name='comp13'), DataID(name='ds2', 
resolution=250, calibration=<calibration.reflectance>, modifiers=()), 
DataID(name='ds5', resolution=250, modifiers=('res_change',))]
num_results = 1, best = True, query = None, kwargs = {}, res = []

     def get_key(key, key_container, num_results=1, best=True, query=None,
                 **kwargs):
         """Get the fully-specified key best matching the provided key.
             Only the best match is returned if `best` is `True` 
(default). See
         `get_best_dataset_key` for more information on how this is 
determined.
             `query` is provided as a convenience to filter by multiple 
parameters
         at once without having to filter by multiple `key` inputs.
             Args:
             key (DataID): DataID of query parameters to use for
                              searching. Any parameter that is `None`
                              is considered a wild card and any match is
                              accepted.
             key_container (dict or set): Container of DataID objects that
                                          uses hashing to quickly access 
items.
             num_results (int): Number of results to return. Use `0` for all
                                matching results. If `1` then the single 
matching
                                key is returned instead of a list of 
length 1.
                                (default: 1)
             best (bool): Sort results to get "best" result first
                          (default: True). See `get_best_dataset_key` 
for details.
             query (DataQuery): filter for the key which can contain for 
example:
                     resolution (float, int, or list): Resolution of the 
dataset in
                                                 dataset units (typically
                                                 meters). This can also be a
                                                 list of these numbers.
                 calibration (str or list): Dataset calibration
                                         (ex.'reflectance'). This can 
also be a
                                         list of these strings.
                 polarization (str or list): Dataset polarization
                                             (ex.'V'). This can also be a
                                             list of these strings.
                 level (number or list): Dataset level (ex. 100). This 
can also be a
                                         list of these numbers.
                 modifiers (list): Modifiers applied to the dataset. Unlike
                                 resolution and calibration this is the 
exact
                                 desired list of modifiers for one 
dataset, not
                                 a list of possible modifiers.
             Returns:
             list or DataID: Matching key(s)
             Raises: KeyError if no matching results or if more than one 
result is
                 found when `num_results` is `1`.
             """
         key = create_filtered_query(key, query)
             res = key.filter_dataids(key_container)
         if not res:
>           raise KeyError("No dataset matching '{}' found".format(str(key)))
E           KeyError: "No dataset matching 'DataQuery(name='comp19')' found"

/usr/lib/python3/dist-packages/satpy/dataset/data_dict.py:107: KeyError
------------------------------ Captured log call 
-------------------------------
WARNING  satpy.scene:scene.py:1275 The following datasets were not 
created and may require resampling to be generated: DataID(name='comp19')
_________________ TestSceneResampling.test_resample_ancillary 
__________________

self = {DataID(name='ds2', resolution=250, 
calibration=<calibration.reflectance>, modifiers=()): <xarray.DataArray 
'zeros_lik...:  []
     optional_datasets:       []
     area:                    Shape: (10, 10)\nLons: <xarray.DataArray 
'zeros_...}
item = 'comp19'

     def __getitem__(self, item):
         """Get item from container."""
         try:
             # short circuit - try to get the object without more work
>           return super(DatasetDict, self).__getitem__(item)
E           KeyError: 'comp19'

/usr/lib/python3/dist-packages/satpy/dataset/data_dict.py:169: KeyError

During handling of the above exception, another exception occurred:

self = <satpy.tests.test_scene.TestSceneResampling object at 0x7f1e01464700>

     def test_resample_ancillary(self):
         """Test that the Scene reducing data does not affect final 
output."""
         from pyresample.geometry import AreaDefinition
         from pyresample.utils import proj4_str_to_dict
         proj_dict = proj4_str_to_dict('+proj=lcc +datum=WGS84 
+ellps=WGS84 '
                                       '+lon_0=-95. +lat_0=25 +lat_1=25 '
                                       '+units=m +no_defs')
         area_def = AreaDefinition('test', 'test', 'test', proj_dict, 5, 
5, (-1000., -1500., 1000., 1500.))
         scene = Scene(filenames=['fake1_1.txt'], reader='fake1')
             scene.load(['comp19', 'comp20'])
>       scene['comp19'].attrs['area'] = area_def

/usr/lib/python3/dist-packages/satpy/tests/test_scene.py:1707: _ _ _ _ _ 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
/usr/lib/python3/dist-packages/satpy/scene.py:803: in __getitem__
     return self._datasets[key]
/usr/lib/python3/dist-packages/satpy/dataset/data_dict.py:171: in 
__getitem__
     key = self.get_key(item)
/usr/lib/python3/dist-packages/satpy/dataset/data_dict.py:158: in get_key
     return get_key(match_key, self.keys(), num_results=num_results,
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
_ _ _ _
key = DataQuery(name='comp19')
key_container = [DataID(name='comp13'), DataID(name='comp20'), 
DataID(name='ds2', resolution=250, 
calibration=<calibration.reflectance>, modifiers=()), DataID(name='ds5', 
resolution=250, modifiers=('res_change',))]
num_results = 1, best = True, query = None, kwargs = {}, res = []

     def get_key(key, key_container, num_results=1, best=True, query=None,
                 **kwargs):
         """Get the fully-specified key best matching the provided key.
             Only the best match is returned if `best` is `True` 
(default). See
         `get_best_dataset_key` for more information on how this is 
determined.
             `query` is provided as a convenience to filter by multiple 
parameters
         at once without having to filter by multiple `key` inputs.
             Args:
             key (DataID): DataID of query parameters to use for
                              searching. Any parameter that is `None`
                              is considered a wild card and any match is
                              accepted.
             key_container (dict or set): Container of DataID objects that
                                          uses hashing to quickly access 
items.
             num_results (int): Number of results to return. Use `0` for all
                                matching results. If `1` then the single 
matching
                                key is returned instead of a list of 
length 1.
                                (default: 1)
             best (bool): Sort results to get "best" result first
                          (default: True). See `get_best_dataset_key` 
for details.
             query (DataQuery): filter for the key which can contain for 
example:
                     resolution (float, int, or list): Resolution of the 
dataset in
                                                 dataset units (typically
                                                 meters). This can also be a
                                                 list of these numbers.
                 calibration (str or list): Dataset calibration
                                         (ex.'reflectance'). This can 
also be a
                                         list of these strings.
                 polarization (str or list): Dataset polarization
                                             (ex.'V'). This can also be a
                                             list of these strings.
                 level (number or list): Dataset level (ex. 100). This 
can also be a
                                         list of these numbers.
                 modifiers (list): Modifiers applied to the dataset. Unlike
                                 resolution and calibration this is the 
exact
                                 desired list of modifiers for one 
dataset, not
                                 a list of possible modifiers.
             Returns:
             list or DataID: Matching key(s)
             Raises: KeyError if no matching results or if more than one 
result is
                 found when `num_results` is `1`.
             """
         key = create_filtered_query(key, query)
             res = key.filter_dataids(key_container)
         if not res:
>           raise KeyError("No dataset matching '{}' found".format(str(key)))
E           KeyError: "No dataset matching 'DataQuery(name='comp19')' found"

/usr/lib/python3/dist-packages/satpy/dataset/data_dict.py:107: KeyError
------------------------------ Captured log call 
-------------------------------
WARNING  satpy.scene:scene.py:1275 The following datasets were not 
created and may require resampling to be generated: DataID(name='comp19')
________________ TestSceneResampling.test_resample_reduce_data 
_________________

self = {DataID(name='ds2', resolution=250, 
calibration=<calibration.reflectance>, modifiers=()): <xarray.DataArray 
'zeros_lik...:  []
     optional_datasets:       []
     area:                    Shape: (20, 20)\nLons: <xarray.DataArray 
'zeros_...}
item = 'comp19'

     def __getitem__(self, item):
         """Get item from container."""
         try:
             # short circuit - try to get the object without more work
>           return super(DatasetDict, self).__getitem__(item)
E           KeyError: 'comp19'

/usr/lib/python3/dist-packages/satpy/dataset/data_dict.py:169: KeyError

During handling of the above exception, another exception occurred:

self = <satpy.tests.test_scene.TestSceneResampling object at 0x7f1e01467a90>

     def test_resample_reduce_data(self):
         """Test that the Scene reducing data does not affect final 
output."""
         from pyresample.geometry import AreaDefinition
         proj_str = ('+proj=lcc +datum=WGS84 +ellps=WGS84 '
                     '+lon_0=-95. +lat_0=25 +lat_1=25 +units=m +no_defs')
         area_def = AreaDefinition('test', 'test', 'test', proj_str, 20, 
20, (-1000., -1500., 1000., 1500.))
         scene = Scene(filenames=['fake1_1.txt'], reader='fake1')
             scene.load(['comp19'])
>       scene['comp19'].attrs['area'] = area_def

/usr/lib/python3/dist-packages/satpy/tests/test_scene.py:1728: _ _ _ _ _ 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
/usr/lib/python3/dist-packages/satpy/scene.py:803: in __getitem__
     return self._datasets[key]
/usr/lib/python3/dist-packages/satpy/dataset/data_dict.py:171: in 
__getitem__
     key = self.get_key(item)
/usr/lib/python3/dist-packages/satpy/dataset/data_dict.py:158: in get_key
     return get_key(match_key, self.keys(), num_results=num_results,
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
_ _ _ _
key = DataQuery(name='comp19')
key_container = [DataID(name='comp13'), DataID(name='ds2', 
resolution=250, calibration=<calibration.reflectance>, modifiers=()), 
DataID(name='ds5', resolution=250, modifiers=('res_change',))]
num_results = 1, best = True, query = None, kwargs = {}, res = []

     def get_key(key, key_container, num_results=1, best=True, query=None,
                 **kwargs):
         """Get the fully-specified key best matching the provided key.
             Only the best match is returned if `best` is `True` 
(default). See
         `get_best_dataset_key` for more information on how this is 
determined.
             `query` is provided as a convenience to filter by multiple 
parameters
         at once without having to filter by multiple `key` inputs.
             Args:
             key (DataID): DataID of query parameters to use for
                              searching. Any parameter that is `None`
                              is considered a wild card and any match is
                              accepted.
             key_container (dict or set): Container of DataID objects that
                                          uses hashing to quickly access 
items.
             num_results (int): Number of results to return. Use `0` for all
                                matching results. If `1` then the single 
matching
                                key is returned instead of a list of 
length 1.
                                (default: 1)
             best (bool): Sort results to get "best" result first
                          (default: True). See `get_best_dataset_key` 
for details.
             query (DataQuery): filter for the key which can contain for 
example:
                     resolution (float, int, or list): Resolution of the 
dataset in
                                                 dataset units (typically
                                                 meters). This can also be a
                                                 list of these numbers.
                 calibration (str or list): Dataset calibration
                                         (ex.'reflectance'). This can 
also be a
                                         list of these strings.
                 polarization (str or list): Dataset polarization
                                             (ex.'V'). This can also be a
                                             list of these strings.
                 level (number or list): Dataset level (ex. 100). This 
can also be a
                                         list of these numbers.
                 modifiers (list): Modifiers applied to the dataset. Unlike
                                 resolution and calibration this is the 
exact
                                 desired list of modifiers for one 
dataset, not
                                 a list of possible modifiers.
             Returns:
             list or DataID: Matching key(s)
             Raises: KeyError if no matching results or if more than one 
result is
                 found when `num_results` is `1`.
             """
         key = create_filtered_query(key, query)
             res = key.filter_dataids(key_container)
         if not res:
>           raise KeyError("No dataset matching '{}' found".format(str(key)))
E           KeyError: "No dataset matching 'DataQuery(name='comp19')' found"

/usr/lib/python3/dist-packages/satpy/dataset/data_dict.py:107: KeyError
------------------------------ Captured log call 
-------------------------------
WARNING  satpy.scene:scene.py:1275 The following datasets were not 
created and may require resampling to be generated: DataID(name='comp19')
_________________ TestSceneResampling.test_no_generate_comp10 
__________________

self = <satpy.tests.test_scene.TestSceneResampling object at 0x7f1e014665f0>
rs = <MagicMock name='resample_dataset' id='139767046526560'>

     @mock.patch('satpy.scene.resample_dataset')
     def test_no_generate_comp10(self, rs):
         """Test generating a composite after loading."""
         from pyresample.geometry import AreaDefinition
         from pyresample.utils import proj4_str_to_dict
             rs.side_effect = self._fake_resample_dataset
         proj_dict = proj4_str_to_dict('+proj=lcc +datum=WGS84 
+ellps=WGS84 '
                                       '+lon_0=-95. +lat_0=25 +lat_1=25 '
                                       '+units=m +no_defs')
         area_def = AreaDefinition(
             'test',
             'test',
             'test',
             proj_dict,
             200,
             400,
             (-1000., -1500., 1000., 1500.),
         )
             # it is fine that an optional prereq doesn't exist
         scene = Scene(filenames=['fake1_1.txt'], reader='fake1')
         scene.load(['comp10'], generate=False)
         assert any(ds_id['name'] == 'comp10' for ds_id in scene._wishlist)
         assert 'comp10' not in scene
         # two dependencies should have been loaded
         assert len(scene._datasets) == 2
         assert len(scene.missing_datasets) == 1
             new_scn = scene.resample(area_def, generate=False)
         assert 'comp10' not in scene
         # two dependencies should have been loaded
         assert len(scene._datasets) == 2
         assert len(scene.missing_datasets) == 1
             new_scn._generate_composites_from_loaded_datasets()
         assert any(ds_id['name'] == 'comp10' for ds_id in 
new_scn._wishlist)
>       assert 'comp10' in new_scn
E       AssertionError: assert 'comp10' in <satpy.scene.Scene object at 
0x7f1db8963550>

/usr/lib/python3/dist-packages/satpy/tests/test_scene.py:1778: 
AssertionError
______ TestSceneResampling.test_comp_loading_after_resampling_new_sensor 
_______

self = <satpy.tests.test_scene.TestSceneResampling object at 0x7f1e01466380>

     def test_comp_loading_after_resampling_new_sensor(self):
         """Test requesting a composite after resampling when the sensor 
composites weren't loaded before."""
         # this is our base Scene with sensor "fake_sensor2"
         scene1 = Scene(filenames=['fake2_3ds_1.txt'], reader='fake2_3ds')
         scene1.load(["ds2"])
         new_scn = scene1.resample(resampler='native')
             # Can't load from readers after resampling
         with pytest.raises(KeyError):
             new_scn.load(["ds3"])
             # Can't load the composite from fake_sensor composites yet
         # 'ds1' is missing
         with pytest.raises(KeyError):
             new_scn.load(["comp2"])
             # artificial DataArray "created by the user"
         # mimics a user adding their own data with the same sensor
         user_da = scene1["ds2"].copy()
         user_da.attrs["name"] = "ds1"
         user_da.attrs["sensor"] = {"fake_sensor2"}
         # Add 'ds1' that doesn't provide the 'fake_sensor' sensor
         new_scn["ds1"] = user_da
         with pytest.raises(KeyError):
             new_scn.load(["comp2"])
         assert "comp2" not in new_scn
             # artificial DataArray "created by the user"
         # mimics a user adding their own data with its own sensor to 
the Scene
         user_da = scene1["ds2"].copy()
         user_da.attrs["name"] = "ds1"
         user_da.attrs["sensor"] = {"fake_sensor"}
         # Now 'fake_sensor' composites have been loaded
         new_scn["ds1"] = user_da
         new_scn.load(["comp2"])
>       assert "comp2" in new_scn
E       AssertionError: assert 'comp2' in <satpy.scene.Scene object at 
0x7f1db8d45a50>

/usr/lib/python3/dist-packages/satpy/tests/test_scene.py:1837: 
AssertionError
------------------------------ Captured log call 
-------------------------------
WARNING  satpy.scene:scene.py:1275 The following datasets were not 
created and may require resampling to be generated: DataID(name='comp2')
=============================== warnings summary 
===============================
../../../../usr/lib/python3/dist-packages/satpy/readers/seviri_base.py:453
   /usr/lib/python3/dist-packages/satpy/readers/seviri_base.py:453: 
DeprecationWarning: `np.bool` is a deprecated alias for the builtin 
`bool`. To silence this warning, use `bool` by itself. Doing this will 
not modify any behavior and is safe. If you specifically wanted the 
numpy scalar type, use `np.bool_` here.
   Deprecated in NumPy 1.20; for more details and guidance: 
https://numpy.org/devdocs/release/1.20.0-notes.html#deprecations
     ('GsicsCalMode', np.bool),

../../../../usr/lib/python3/dist-packages/satpy/readers/seviri_base.py:454
   /usr/lib/python3/dist-packages/satpy/readers/seviri_base.py:454: 
DeprecationWarning: `np.bool` is a deprecated alias for the builtin 
`bool`. To silence this warning, use `bool` by itself. Doing this will 
not modify any behavior and is safe. If you specifically wanted the 
numpy scalar type, use `np.bool_` here.
   Deprecated in NumPy 1.20; for more details and guidance: 
https://numpy.org/devdocs/release/1.20.0-notes.html#deprecations
     ('GsicsCalValidity', np.bool),

../../../../usr/lib/python3/dist-packages/satpy/tests/reader_tests/test_mviri_l1b_fiduceo_nc.py:541
 
/usr/lib/python3/dist-packages/satpy/tests/reader_tests/test_mviri_l1b_fiduceo_nc.py:541: 
PytestUnknownMarkWarning: Unknown pytest.mark.file_handler_data - is 
this a typo?  You can register custom marks to avoid this warning - for 
details, see https://docs.pytest.org/en/stable/how-to/mark.html
     @pytest.mark.file_handler_data(mask_bad_quality=False)

tests/test_composites.py::TestMatchDataArrays::test_nondimensional_coords
tests/test_composites.py::TestMatchDataArrays::test_nondimensional_coords
tests/reader_tests/test_goes_imager_nc.py::GOESNCEUMFileHandlerRadianceTest::test_get_dataset_radiance
tests/reader_tests/test_goes_imager_nc.py::GOESNCEUMFileHandlerRadianceTest::test_get_dataset_radiance
tests/reader_tests/test_goes_imager_nc.py::GOESNCEUMFileHandlerRadianceTest::test_get_dataset_radiance
tests/reader_tests/test_goes_imager_nc.py::GOESNCEUMFileHandlerRadianceTest::test_get_dataset_radiance
tests/reader_tests/test_goes_imager_nc.py::GOESNCEUMFileHandlerReflectanceTest::test_get_dataset_reflectance
   /usr/lib/python3/dist-packages/xarray/core/dataarray.py:2837: 
PendingDeprecationWarning: dropping variables using `drop` will be 
deprecated; using drop_vars is encouraged.
     ds = self._to_temp_dataset().drop(labels, dim, errors=errors, 
**labels_kwargs)

tests/test_config.py: 137 warnings
tests/test_multiscene.py: 8 warnings
tests/test_resample.py: 14 warnings
tests/test_scene.py: 4 warnings
tests/test_writers.py: 4 warnings
tests/test_yaml_reader.py: 3 warnings
tests/modifier_tests/test_parallax.py: 39 warnings
tests/reader_tests/test_ahi_hsd.py: 2 warnings
tests/reader_tests/test_cmsaf_claas.py: 12 warnings
tests/reader_tests/test_fci_l1c_nc.py: 16 warnings
tests/reader_tests/test_fci_l2_nc.py: 1 warning
tests/reader_tests/test_geocat.py: 6 warnings
tests/reader_tests/test_geos_area.py: 1 warning
tests/reader_tests/test_goes_imager_hrit.py: 1 warning
tests/reader_tests/test_gpm_imerg.py: 1 warning
tests/reader_tests/test_hrit_base.py: 1 warning
tests/reader_tests/test_mviri_l1b_fiduceo_nc.py: 12 warnings
tests/reader_tests/test_nwcsaf_msg.py: 1 warning
tests/reader_tests/test_nwcsaf_nc.py: 3 warnings
tests/reader_tests/test_oceancolorcci_l3_nc.py: 1 warning
tests/reader_tests/test_seviri_l1b_hrit.py: 3 warnings
tests/reader_tests/test_seviri_l1b_native.py: 2 warnings
tests/writer_tests/test_mitiff.py: 24 warnings
   /usr/lib/python3/dist-packages/pyproj/crs/crs.py:1286: UserWarning: 
You will likely lose important projection information when converting to 
a PROJ string from another format. See: 
https://proj.org/faq.html#what-is-the-best-format-for-describing-coordinate-reference-systems
     proj = self._crs.to_proj4(version=version)

tests/test_config.py::TestPluginsConfigs::test_plugin_reader_available_readers
   /usr/lib/python3/dist-packages/pygac/reader.py:45: 
DeprecationWarning: The distutils package is deprecated and slated for 
removal in Python 3.12. Use setuptools or check PEP 632 for potential 
alternatives
     from distutils.version import LooseVersion

tests/test_config.py::TestPluginsConfigs::test_plugin_writer_available_writers
   /usr/lib/python3/dist-packages/pyninjotiff/tifffile.py:154: 
UserWarning: failed to import the optional _tifffile C extension module.
   Loading of some compressed images will be slow.
   Tifffile.c can be obtained at http://www.lfd.uci.edu/~gohlke/
     warnings.warn(

tests/test_dataset.py::test_combine_dicts_close
   /usr/lib/python3/dist-packages/satpy/tests/test_dataset.py:329: 
DeprecationWarning: `np.str` is a deprecated alias for the builtin 
`str`. To silence this warning, use `str` by itself. Doing this will not 
modify any behavior and is safe. If you specifically wanted the numpy 
scalar type, use `np.str_` here.
   Deprecated in NumPy 1.20; for more details and guidance: 
https://numpy.org/devdocs/release/1.20.0-notes.html#deprecations
     'e': np.str('bar'),

tests/test_dataset.py::test_combine_dicts_close
   /usr/lib/python3/dist-packages/satpy/tests/test_dataset.py:342: 
DeprecationWarning: `np.str` is a deprecated alias for the builtin 
`str`. To silence this warning, use `str` by itself. Doing this will not 
modify any behavior and is safe. If you specifically wanted the numpy 
scalar type, use `np.str_` here.
   Deprecated in NumPy 1.20; for more details and guidance: 
https://numpy.org/devdocs/release/1.20.0-notes.html#deprecations
     'e': np.str('bar'),

tests/test_dataset.py::test_combine_dicts_different[test_mda5]
   /usr/lib/python3/dist-packages/satpy/dataset/metadata.py:199: 
FutureWarning: elementwise comparison failed; returning scalar instead, 
but in the future will perform elementwise comparison
     res = comp_func(a, b)

tests/test_dataset.py::TestIDQueryInteractions::test_seviri_hrv_has_priority_over_vis008
   /usr/lib/python3/dist-packages/satpy/tests/test_dataset.py:688: 
UserWarning: Attribute access to DataIDs is deprecated, use key access 
instead.
     assert res[0].name == "HRV"

tests/test_dependency_tree.py::TestMultipleSensors::test_compositor_loaded_sensor_order
 
/usr/lib/python3/dist-packages/satpy/tests/test_dependency_tree.py:222: 
UserWarning: Attribute access to DataIDs is deprecated, use key access 
instead.
     self.assertEqual(comp_nodes[0].name.resolution, 500)

tests/test_readers.py::TestReaderLoader::test_missing_requirements
   /usr/lib/python3/dist-packages/satpy/readers/yaml_reader.py:509: 
UserWarning: No handler for reading requirement 'HRIT_EPI' for 
H-000-MSG4__-MSG4________-IR_108___-000006___-201809050900-__
     warnings.warn(msg)

tests/test_readers.py::TestReaderLoader::test_missing_requirements
   /usr/lib/python3/dist-packages/satpy/readers/yaml_reader.py:509: 
UserWarning: No handler for reading requirement 'HRIT_PRO' for 
H-000-MSG4__-MSG4________-IR_108___-000006___-201809050900-__
     warnings.warn(msg)

tests/test_readers.py::TestReaderLoader::test_missing_requirements
   /usr/lib/python3/dist-packages/satpy/readers/yaml_reader.py:512: 
UserWarning: No matching requirement file of type HRIT_PRO for 
H-000-MSG4__-MSG4________-IR_108___-000006___-201809051000-__
     warnings.warn(str(err) + ' for {}'.format(filename))

tests/test_regressions.py::test_1088
   /usr/lib/python3/dist-packages/numpy/lib/function_base.py:1292: 
RuntimeWarning: invalid value encountered in subtract
     a = op(a[slice1], a[slice2])

tests/test_resample.py::TestHLResample::test_type_preserve
tests/test_resample.py::TestHLResample::test_type_preserve
   /usr/lib/python3/dist-packages/pyresample/geometry.py:658: 
DeprecationWarning: This function is deprecated. See: 
https://pyproj4.github.io/pyproj/stable/gotchas.html#upgrading-to-pyproj-2-from-pyproj-1
     xyz = np.stack(transform(src, dst, lons, lats, alt), axis=1)

tests/test_resample.py::TestKDTreeResampler::test_check_numpy_cache
   /usr/lib/python3/dist-packages/satpy/resample.py:554: UserWarning: 
Using Numpy files as resampling cache is deprecated.
     warnings.warn("Using Numpy files as resampling cache is "

tests/test_resample.py::TestBucketAvg::test_compute_and_not_use_skipna_handling
tests/test_resample.py::TestBucketAvg::test_compute_and_not_use_skipna_handling
tests/test_resample.py::TestBucketSum::test_compute_and_not_use_skipna_handling
tests/test_resample.py::TestBucketSum::test_compute_and_not_use_skipna_handling
   /usr/lib/python3/dist-packages/satpy/resample.py:1100: 
DeprecationWarning: Argument mask_all_nan is deprecated.Please update 
Pyresample and use skipna for missing values handling.
     warnings.warn('Argument mask_all_nan is deprecated.'

tests/test_resample.py::TestBucketAvg::test_compute_and_use_skipna_handling
tests/test_resample.py::TestBucketSum::test_compute_and_use_skipna_handling
   /usr/lib/python3/dist-packages/satpy/resample.py:1095: 
DeprecationWarning: Argument mask_all_nan is deprecated. Please use 
skipna for missing values handling. Continuing with default skipna=True, 
if not provided differently.
     warnings.warn('Argument mask_all_nan is deprecated. Please use 
skipna for missing values handling. '

tests/test_scene.py: 2 warnings
tests/test_writers.py: 14 warnings
tests/writer_tests/test_geotiff.py: 5 warnings
   /usr/lib/python3/dist-packages/rasterio/__init__.py:287: 
NotGeoreferencedWarning: Dataset has no geotransform, gcps, or rpcs. The 
identity matrix will be returned.
     dataset = writer(

tests/test_scene.py: 3 warnings
tests/test_writers.py: 10 warnings
tests/reader_tests/test_aapp_l1b.py: 3 warnings
tests/writer_tests/test_geotiff.py: 3 warnings
tests/writer_tests/test_mitiff.py: 5 warnings
tests/writer_tests/test_simple_image.py: 2 warnings
   /usr/lib/python3/dist-packages/dask/core.py:119: RuntimeWarning: 
divide by zero encountered in true_divide
     return func(*(_execute_task(a, cache) for a in args))

tests/test_scene.py: 3 warnings
tests/test_writers.py: 10 warnings
tests/writer_tests/test_geotiff.py: 3 warnings
tests/writer_tests/test_mitiff.py: 5 warnings
tests/writer_tests/test_simple_image.py: 2 warnings
   /usr/lib/python3/dist-packages/dask/core.py:119: RuntimeWarning: 
invalid value encountered in multiply
     return func(*(_execute_task(a, cache) for a in args))

tests/enhancement_tests/test_enhancements.py::TestEnhancementStretch::test_crefl_scaling
   /usr/lib/python3/dist-packages/satpy/enhancements/__init__.py:129: 
DeprecationWarning: 'crefl_scaling' is deprecated, use 
'piecewise_linear_stretch' instead.
     warnings.warn("'crefl_scaling' is deprecated, use 
'piecewise_linear_stretch' instead.", DeprecationWarning)

tests/enhancement_tests/test_enhancements.py::TestColormapLoading::test_cmap_list
   /usr/lib/python3/dist-packages/trollimage/colormap.py:229: 
UserWarning: Colormap 'colors' should be flotaing point numbers between 
0 and 1.
     warnings.warn("Colormap 'colors' should be flotaing point numbers 
between 0 and 1.")

tests/modifier_tests/test_parallax.py: 46 warnings
   /usr/lib/python3/dist-packages/pyresample/geometry.py:1545: 
PendingDeprecationWarning: 'name' is deprecated, use 'description' instead.
     warnings.warn("'name' is deprecated, use 'description' instead.", 
PendingDeprecationWarning)

tests/modifier_tests/test_parallax.py: 37 warnings
   /usr/lib/python3/dist-packages/satpy/modifiers/parallax.py:390: 
UserWarning: Overlap checking not impelemented. Waiting for fix for 
https://github.com/pytroll/pyresample/issues/329
     warnings.warn(

tests/modifier_tests/test_parallax.py::TestParallaxCorrectionClass::test_correct_area_clearsky_different_resolutions[0.08-0.3]
tests/modifier_tests/test_parallax.py::TestParallaxCorrectionClass::test_correct_area_clearsky_different_resolutions[0.3-0.08]
   /usr/lib/python3/dist-packages/_pytest/python.py:192: 
PytestRemovedIn8Warning: Passing None has been deprecated.
   See 
https://docs.pytest.org/en/latest/how-to/capture-warnings.html#additional-use-cases-of-warnings-in-tests 
for alternatives in common use cases.
     result = testfunction(**testargs)

tests/modifier_tests/test_parallax.py::TestParallaxCorrectionModifier::test_parallax_modifier_interface_with_cloud
   /usr/lib/python3/dist-packages/dask/core.py:119: RuntimeWarning: 
invalid value encountered in cos
     return func(*(_execute_task(a, cache) for a in args))

tests/modifier_tests/test_parallax.py::TestParallaxCorrectionModifier::test_parallax_modifier_interface_with_cloud
   /usr/lib/python3/dist-packages/dask/core.py:119: RuntimeWarning: 
invalid value encountered in sin
     return func(*(_execute_task(a, cache) for a in args))

tests/modifier_tests/test_parallax.py::TestParallaxCorrectionModifier::test_parallax_modifier_interface_with_cloud
   /usr/lib/python3/dist-packages/dask/core.py:119: RuntimeWarning: 
invalid value encountered in remainder
     return func(*(_execute_task(a, cache) for a in args))

tests/reader_tests/test_aapp_l1b.py::TestAAPPL1BAllChannelsPresent::test_navigation
tests/reader_tests/test_aapp_l1b.py::TestAAPPL1BAllChannelsPresent::test_navigation
   /usr/lib/python3/dist-packages/geotiepoints/geointerpolator.py:81: 
RuntimeWarning: invalid value encountered in arcsin
     lats = np.sign(z__) * (90 - np.rad2deg(np.arcsin(np.sqrt(x__ ** 2 + 
y__ ** 2) / radius)))

tests/reader_tests/test_aapp_l1b.py::TestAAPPL1BAllChannelsPresent::test_read
tests/reader_tests/test_aapp_l1b.py::TestAAPPL1BAllChannelsPresent::test_read
tests/reader_tests/test_aapp_l1b.py::TestAAPPL1BAllChannelsPresent::test_read
tests/reader_tests/test_aapp_l1b.py::TestAAPPL1BAllChannelsPresent::test_read
tests/reader_tests/test_aapp_l1b.py::TestAAPPL1BAllChannelsPresent::test_read
tests/reader_tests/test_aapp_l1b.py::TestAAPPL1BAllChannelsPresent::test_read
tests/reader_tests/test_aapp_l1b.py::TestAAPPL1BAllChannelsPresent::test_read
tests/reader_tests/test_ahi_hsd.py::TestAHICalibration::test_default_calibrate
   /usr/lib/python3/dist-packages/dask/core.py:119: RuntimeWarning: 
invalid value encountered in log
     return func(*(_execute_task(a, cache) for a in args))

tests/reader_tests/test_aapp_mhs_amsub_l1c.py::TestMHS_AMSUB_AAPPL1CReadData::test_read
tests/reader_tests/test_aapp_mhs_amsub_l1c.py::TestMHS_AMSUB_AAPPL1CReadData::test_read
tests/reader_tests/test_aapp_mhs_amsub_l1c.py::TestMHS_AMSUB_AAPPL1CReadData::test_read
tests/reader_tests/test_aapp_mhs_amsub_l1c.py::TestMHS_AMSUB_AAPPL1CReadData::test_read
tests/reader_tests/test_aapp_mhs_amsub_l1c.py::TestMHS_AMSUB_AAPPL1CReadData::test_read
 
/usr/lib/python3/dist-packages/satpy/readers/aapp_mhs_amsub_l1c.py:402: 
DeprecationWarning: `np.float` is a deprecated alias for the builtin 
`float`. To silence this warning, use `float` by itself. Doing this will 
not modify any behavior and is safe. If you specifically wanted the 
numpy scalar type, use `np.float64` here.
   Deprecated in NumPy 1.20; for more details and guidance: 
https://numpy.org/devdocs/release/1.20.0-notes.html#deprecations
     channel = channel.astype(np.float)

tests/reader_tests/test_abi_l2_nc.py::TestMCMIPReading::test_mcmip_get_dataset
   /usr/lib/python3/dist-packages/satpy/readers/abi_l2_nc.py:41: 
UserWarning: Attribute access to DataIDs is deprecated, use key access 
instead.
     var += "_" + key.name

tests/reader_tests/test_ahi_hsd.py: 5 warnings
tests/reader_tests/test_utils.py: 6 warnings
   /usr/lib/python3/dist-packages/satpy/readers/utils.py:352: 
DeprecationWarning: This function is deprecated. See: 
https://pyproj4.github.io/pyproj/stable/gotchas.html#upgrading-to-pyproj-2-from-pyproj-1
     x, y, z = pyproj.transform(latlong, geocent, lon, lat, 0.)

tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_scene_loading
   /usr/lib/python3/dist-packages/satpy/readers/ahi_hsd.py:515: 
UserWarning: Actual block1 header size does not match expected
     warnings.warn(f"Actual {block} header size does not match expected")

tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_scene_loading
   /usr/lib/python3/dist-packages/satpy/readers/ahi_hsd.py:515: 
UserWarning: Actual block2 header size does not match expected
     warnings.warn(f"Actual {block} header size does not match expected")

tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_scene_loading
   /usr/lib/python3/dist-packages/satpy/readers/ahi_hsd.py:515: 
UserWarning: Actual block3 header size does not match expected
     warnings.warn(f"Actual {block} header size does not match expected")

tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_scene_loading
   /usr/lib/python3/dist-packages/satpy/readers/ahi_hsd.py:515: 
UserWarning: Actual block4 header size does not match expected
     warnings.warn(f"Actual {block} header size does not match expected")

tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_scene_loading
   /usr/lib/python3/dist-packages/satpy/readers/ahi_hsd.py:515: 
UserWarning: Actual block5 header size does not match expected
     warnings.warn(f"Actual {block} header size does not match expected")

tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_scene_loading
   /usr/lib/python3/dist-packages/satpy/readers/ahi_hsd.py:515: 
UserWarning: Actual block6 header size does not match expected
     warnings.warn(f"Actual {block} header size does not match expected")

tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_scene_loading
   /usr/lib/python3/dist-packages/satpy/readers/ahi_hsd.py:515: 
UserWarning: Actual block7 header size does not match expected
     warnings.warn(f"Actual {block} header size does not match expected")

tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_scene_loading
   /usr/lib/python3/dist-packages/satpy/readers/ahi_hsd.py:515: 
UserWarning: Actual block8 header size does not match expected
     warnings.warn(f"Actual {block} header size does not match expected")

tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_scene_loading
   /usr/lib/python3/dist-packages/satpy/readers/ahi_hsd.py:515: 
UserWarning: Actual block9 header size does not match expected
     warnings.warn(f"Actual {block} header size does not match expected")

tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_scene_loading
   /usr/lib/python3/dist-packages/satpy/readers/ahi_hsd.py:515: 
UserWarning: Actual block10 header size does not match expected
     warnings.warn(f"Actual {block} header size does not match expected")

tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_read_band
tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_scene_loading
   /usr/lib/python3/dist-packages/satpy/readers/ahi_hsd.py:515: 
UserWarning: Actual block11 header size does not match expected
     warnings.warn(f"Actual {block} header size does not match expected")

tests/reader_tests/test_ahi_hsd.py::TestAHIHSDFileHandler::test_time_rounding
   /usr/lib/python3/dist-packages/satpy/readers/ahi_hsd.py:461: 
UserWarning: Observation timeline is fill value, not rounding 
observation time.
     warnings.warn("Observation timeline is fill value, not rounding 
observation time.")

tests/reader_tests/test_ami_l1b.py::TestAMIL1bNetCDF::test_get_dataset
tests/reader_tests/test_ami_l1b.py::TestAMIL1bNetCDF::test_get_dataset_counts
tests/reader_tests/test_ami_l1b.py::TestAMIL1bNetCDF::test_get_dataset_vis
tests/reader_tests/test_ami_l1b.py::TestAMIL1bNetCDFIRCal::test_default_calibrate
tests/reader_tests/test_ami_l1b.py::TestAMIL1bNetCDFIRCal::test_gsics_radiance_corr
tests/reader_tests/test_ami_l1b.py::TestAMIL1bNetCDFIRCal::test_infile_calibrate
tests/reader_tests/test_ami_l1b.py::TestAMIL1bNetCDFIRCal::test_user_radiance_corr
   /usr/lib/python3/dist-packages/satpy/readers/ami_l1b.py:165: 
DeprecationWarning: This function is deprecated. See: 
https://pyproj4.github.io/pyproj/stable/gotchas.html#upgrading-to-pyproj-2-from-pyproj-1
     sc_position = pyproj.transform(

tests/reader_tests/test_avhrr_l0_hrpt.py::TestHRPTGetCalibratedReflectances::test_calibrated_reflectances_values
tests/reader_tests/test_avhrr_l0_hrpt.py::TestHRPTGetCalibratedBT::test_calibrated_bt_values
tests/reader_tests/test_avhrr_l0_hrpt.py::TestHRPTChannel3::test_channel_3a_masking
tests/reader_tests/test_avhrr_l0_hrpt.py::TestHRPTChannel3::test_channel_3b_masking
tests/reader_tests/test_avhrr_l0_hrpt.py::TestHRPTNavigation::test_latitudes_are_returned
tests/reader_tests/test_avhrr_l0_hrpt.py::TestHRPTNavigation::test_longitudes_are_returned
   /usr/lib/python3/dist-packages/satpy/readers/hrpt.py:80: 
DeprecationWarning: parsing timezone aware datetimes is deprecated; this 
will raise an error in the future
     return (np.datetime64(

tests/reader_tests/test_avhrr_l0_hrpt.py::TestHRPTGetCalibratedReflectances::test_calibrated_reflectances_values
tests/reader_tests/test_avhrr_l0_hrpt.py::TestHRPTChannel3::test_channel_3a_masking
   /usr/lib/python3/dist-packages/satpy/readers/hrpt.py:227: 
DeprecationWarning: parsing timezone aware datetimes is deprecated; this 
will raise an error in the future
     - np.datetime64(str(self.year) + '-01-01T00:00:00Z'))

tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCFileHandler::test_area_definition
tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCFileHandler::test_dataset
tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCFileHandler::test_dataset_with_layer
tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCFileHandler::test_dataset_with_total_cot
tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCReadingByteData::test_byte_extraction
tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCReadingByteData::test_byte_extraction
   /usr/lib/python3/dist-packages/satpy/readers/fci_l2_nc.py:255: 
UserWarning: Attribute access to DataIDs is deprecated, use key access 
instead.
     res = dataset_id.resolution

tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCSegmentFileHandler::test_dataset_with_adef
tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCSegmentFileHandler::test_dataset_with_adef_and_wrongs_dims
   /usr/lib/python3/dist-packages/satpy/readers/fci_l2_nc.py:363: 
UserWarning: Attribute access to DataIDs is deprecated, use key access 
instead.
     res = dataset_id.resolution

tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCSegmentFileHandler::test_dataset_with_adef
tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCSegmentFileHandler::test_dataset_with_adef
tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCSegmentFileHandler::test_dataset_with_adef_and_wrongs_dims
   /usr/lib/python3/dist-packages/pyresample/geometry.py:1291: 
PendingDeprecationWarning: 'x_size' is deprecated, use 'width' instead.
     warnings.warn("'x_size' is deprecated, use 'width' instead.", 
PendingDeprecationWarning)

tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCSegmentFileHandler::test_dataset_with_adef
tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCSegmentFileHandler::test_dataset_with_adef
tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCSegmentFileHandler::test_dataset_with_adef_and_wrongs_dims
   /usr/lib/python3/dist-packages/pyresample/geometry.py:1297: 
PendingDeprecationWarning: 'y_size' is deprecated, use 'height' instead.
     warnings.warn("'y_size' is deprecated, use 'height' instead.", 
PendingDeprecationWarning)

tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCReadingByteData::test_byte_extraction
tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCReadingByteData::test_byte_extraction
   /usr/lib/python3/dist-packages/satpy/readers/fci_l2_nc.py:239: 
RuntimeWarning: Mean of empty slice.
     scale_factor = (x[1:]-x[0:-1]).values.mean()

tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCReadingByteData::test_byte_extraction
tests/reader_tests/test_fci_l2_nc.py::TestFciL2NCReadingByteData::test_byte_extraction
   /usr/lib/python3/dist-packages/numpy/core/_methods.py:189: 
RuntimeWarning: invalid value encountered in true_divide
     ret = ret.dtype.type(ret / rcount)

tests/reader_tests/test_generic_image.py::TestGenericImage::test_GenericImageFileHandler
tests/reader_tests/test_generic_image.py::TestGenericImage::test_GenericImageFileHandler_datasetid
tests/reader_tests/test_generic_image.py::TestGenericImage::test_GenericImageFileHandler_nodata
tests/reader_tests/test_generic_image.py::TestGenericImage::test_geotiff_scene
tests/reader_tests/test_generic_image.py::TestGenericImage::test_geotiff_scene
tests/reader_tests/test_generic_image.py::TestGenericImage::test_geotiff_scene_nan
tests/reader_tests/test_generic_image.py::TestGenericImage::test_geotiff_scene_nan
tests/reader_tests/test_generic_image.py::TestGenericImage::test_png_scene
tests/reader_tests/test_generic_image.py::TestGenericImage::test_png_scene
   /usr/lib/python3/dist-packages/satpy/readers/generic_image.py:78: 
DeprecationWarning: open_rasterio is Deprecated in favor of rioxarray. 
For information about transitioning, see: 
https://corteva.github.io/rioxarray/stable/getting_started/getting_started.html
     data = xr.open_rasterio(dataset, chunks=(1, CHUNK_SIZE, CHUNK_SIZE))

tests/reader_tests/test_generic_image.py::TestGenericImage::test_png_scene
tests/reader_tests/test_generic_image.py::TestGenericImage::test_png_scene
   /usr/lib/python3/dist-packages/rasterio/__init__.py:277: 
NotGeoreferencedWarning: Dataset has no geotransform, gcps, or rpcs. The 
identity matrix will be returned.
     dataset = DatasetReader(path, driver=driver, sharing=sharing, **kwargs)

tests/reader_tests/test_modis_l1b.py::TestModisL1b::test_load_longitude_latitude[modis_l1b_nasa_mod021km_file-True-False-False-1000]
 
/usr/lib/python3/dist-packages/geotiepoints/simple_modis_interpolator.py:38: 
DeprecationWarning: Please use `map_coordinates` from the 
`scipy.ndimage` namespace, the `scipy.ndimage.interpolation` namespace 
is deprecated.
     from scipy.ndimage.interpolation import map_coordinates

tests/reader_tests/test_msi_safe.py::TestMTDXML::test_satellite_zenith_array
   /usr/lib/python3/dist-packages/satpy/readers/msi_safe.py:320: 
RuntimeWarning: Mean of empty slice
     angles = np.nanmean(np.dstack(arrays), -1)

tests/reader_tests/test_olci_nc.py::TestOLCIReader::test_olci_angles
tests/reader_tests/test_olci_nc.py::TestOLCIReader::test_olci_angles
tests/reader_tests/test_olci_nc.py::TestOLCIReader::test_olci_angles
tests/reader_tests/test_olci_nc.py::TestOLCIReader::test_olci_angles
tests/reader_tests/test_olci_nc.py::TestOLCIReader::test_olci_meteo
tests/reader_tests/test_olci_nc.py::TestOLCIReader::test_olci_meteo
tests/reader_tests/test_olci_nc.py::TestOLCIReader::test_olci_meteo
tests/reader_tests/test_olci_nc.py::TestOLCIReader::test_olci_meteo
   /usr/lib/python3/dist-packages/geotiepoints/interpolator.py:239: 
DeprecationWarning: elementwise comparison failed; this will raise an 
error in the future.
     if np.all(self.hrow_indices == self.row_indices):

tests/reader_tests/test_satpy_cf_nc.py: 8 warnings
tests/writer_tests/test_cf.py: 20 warnings
   /usr/lib/python3/dist-packages/satpy/writers/cf_writer.py:571: 
FutureWarning: The default behaviour of the CF writer will soon change 
to not compress data by default.
     warnings.warn("The default behaviour of the CF writer will soon 
change to not compress data by default.",

tests/reader_tests/test_satpy_cf_nc.py: 8 warnings
tests/writer_tests/test_cf.py: 24 warnings
   /usr/lib/python3/dist-packages/satpy/writers/cf_writer.py:746: 
UserWarning: Dtype int64 not compatible with CF-1.7.
     warnings.warn('Dtype {} not compatible with 
{}.'.format(str(ds.dtype), CF_VERSION))

tests/reader_tests/test_satpy_cf_nc.py: 18 warnings
   /usr/lib/python3/dist-packages/satpy/readers/satpy_cf_nc.py:240: 
DeprecationWarning: The truth value of an empty array is ambiguous. 
Returning False, but in future this will result in an error. Use 
`array.size > 0` to check that an array is not empty.
     if 'modifiers' in ds_info and not ds_info['modifiers']:

tests/reader_tests/test_satpy_cf_nc.py::TestCFReader::test_read_prefixed_channels_by_user_no_prefix
tests/writer_tests/test_cf.py::TestCFWriter::test_save_dataset_a_digit_no_prefix_include_attr
   /usr/lib/python3/dist-packages/satpy/writers/cf_writer.py:566: 
UserWarning: Invalid NetCDF dataset name: 1 starts with a digit.
     warnings.warn('Invalid NetCDF dataset name: {} starts with a 
digit.'.format(name))

tests/reader_tests/test_seviri_base.py::TestOrbitPolynomialFinder::test_get_orbit_polynomial[orbit_polynomials1-time1-orbit_polynomial_exp1]
tests/reader_tests/test_seviri_base.py::TestOrbitPolynomialFinder::test_get_orbit_polynomial_exceptions[orbit_polynomials1-time1]
tests/reader_tests/test_seviri_l1b_native.py::TestNativeMSGDataset::test_satpos_no_valid_orbit_polynomial
   /usr/lib/python3/dist-packages/satpy/readers/seviri_base.py:774: 
UserWarning: No orbit polynomial valid for 2006-01-01T12:15:00.000000. 
Using closest match.
     warnings.warn(

tests/reader_tests/test_seviri_base.py::TestOrbitPolynomialFinder::test_get_orbit_polynomial_exceptions[orbit_polynomials0-time0]
   /usr/lib/python3/dist-packages/satpy/readers/seviri_base.py:774: 
UserWarning: No orbit polynomial valid for 2006-01-02T12:15:00.000000. 
Using closest match.
     warnings.warn(

tests/reader_tests/test_seviri_l1b_hrit.py::TestHRITMSGFileHandler::test_satpos_no_valid_orbit_polynomial
   /usr/lib/python3/dist-packages/satpy/readers/seviri_base.py:774: 
UserWarning: No orbit polynomial valid for 2006-01-01T12:15:09.304888. 
Using closest match.
     warnings.warn(

tests/reader_tests/test_seviri_l1b_nc.py::TestNCSEVIRIFileHandler::test_satpos_no_valid_orbit_polynomial
   /usr/lib/python3/dist-packages/satpy/readers/seviri_base.py:774: 
UserWarning: No orbit polynomial valid for 2020-01-01T00:00:00.000000. 
Using closest match.
     warnings.warn(

tests/reader_tests/test_seviri_l2_grib.py::Test_SeviriL2GribFileHandler::test_data_reading
   /usr/lib/python3/dist-packages/satpy/readers/seviri_l2_grib.py:114: 
UserWarning: Attribute access to DataIDs is deprecated, use key access 
instead.
     self._res = dataset_id.resolution

tests/reader_tests/test_slstr_l1b.py::TestSLSTRReader::test_instantiate
   /usr/lib/python3/dist-packages/satpy/readers/slstr_l1b.py:173: 
UserWarning: Warning: No radiance adjustment supplied for channel foo_nadir
     warnings.warn("Warning: No radiance adjustment supplied " +

tests/reader_tests/test_smos_l2_wind.py::TestSMOSL2WINDReader::test_load_lat
tests/reader_tests/test_smos_l2_wind.py::TestSMOSL2WINDReader::test_load_lon
tests/reader_tests/test_smos_l2_wind.py::TestSMOSL2WINDReader::test_load_wind_speed
tests/writer_tests/test_mitiff.py::TestMITIFFWriter::test_convert_proj4_string
tests/writer_tests/test_mitiff.py::TestMITIFFWriter::test_convert_proj4_string
tests/writer_tests/test_mitiff.py::TestMITIFFWriter::test_convert_proj4_string
tests/writer_tests/test_mitiff.py::TestMITIFFWriter::test_convert_proj4_string
tests/writer_tests/test_mitiff.py::TestMITIFFWriter::test_convert_proj4_string
   /usr/lib/python3/dist-packages/pyproj/crs/crs.py:141: FutureWarning: 
'+init=<authority>:<code>' syntax is deprecated. '<authority>:<code>' is 
the preferred initialization method. When making the change, be mindful 
of axis order changes: 
https://pyproj4.github.io/pyproj/stable/gotchas.html#axis-order-changes-in-proj-6
     in_crs_string = _prepare_from_proj_string(in_crs_string)

tests/reader_tests/test_viirs_compact.py::TestCompact::test_distributed
tests/reader_tests/test_viirs_compact.py::TestCompact::test_distributed
   /usr/lib/python3/dist-packages/tornado/ioloop.py:265: 
DeprecationWarning: There is no current event loop
     loop = asyncio.get_event_loop()

tests/reader_tests/test_viirs_compact.py::TestCompact::test_distributed
   /usr/lib/python3/dist-packages/tornado/ioloop.py:350: 
DeprecationWarning: make_current is deprecated; start the event loop first
     self.make_current()

tests/reader_tests/test_viirs_compact.py::TestCompact::test_distributed
   /usr/lib/python3/dist-packages/tornado/platform/asyncio.py:360: 
DeprecationWarning: There is no current event loop
     self.old_asyncio = asyncio.get_event_loop()

tests/reader_tests/test_viirs_compact.py::TestCompact::test_distributed
   /usr/lib/python3/dist-packages/_pytest/threadexception.py:73: 
PytestUnhandledThreadExceptionWarning: Exception in thread Profile
     Traceback (most recent call last):
     File "/usr/lib/python3/dist-packages/distributed/profile.py", line 
115, in process
       d = state["children"][ident]
   KeyError: 
'write;/usr/lib/python3/dist-packages/distributed/comm/tcp.py;245'
     During handling of the above exception, another exception occurred:
     Traceback (most recent call last):
     File "/usr/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
       self.run()
     File "/usr/lib/python3.10/threading.py", line 953, in run
       self._target(*self._args, **self._kwargs)
     File "/usr/lib/python3/dist-packages/distributed/profile.py", line 
274, in _watch
       process(frame, None, recent, omit=omit)
     File "/usr/lib/python3/dist-packages/distributed/profile.py", line 
119, in process
       "description": info_frame(frame),
     File "/usr/lib/python3/dist-packages/distributed/profile.py", line 
72, in info_frame
       line = linecache.getline(co.co_filename, frame.f_lineno, 
frame.f_globals).lstrip()
     File "/usr/lib/python3.10/linecache.py", line 31, in getline
       if 1 <= lineno <= len(lines):
   TypeError: '<=' not supported between instances of 'int' and 'NoneType'
       warnings.warn(pytest.PytestUnhandledThreadExceptionWarning(msg))

tests/writer_tests/test_awips_tiled.py::TestAWIPSTiledWriter::test_lettered_tiles_no_valid_data
   /usr/lib/python3/dist-packages/dask/array/reductions.py:569: 
RuntimeWarning: All-NaN slice encountered
     return np.nanmax(x_chunk, axis=axis, keepdims=keepdims)

tests/writer_tests/test_awips_tiled.py::TestAWIPSTiledWriter::test_lettered_tiles_no_valid_data
   /usr/lib/python3/dist-packages/dask/array/reductions.py:540: 
RuntimeWarning: All-NaN slice encountered
     return np.nanmin(x_chunk, axis=axis, keepdims=keepdims)

tests/writer_tests/test_awips_tiled.py::TestAWIPSTiledWriter::test_basic_numbered_tiles_rgb
 
/usr/lib/python3/dist-packages/satpy/tests/writer_tests/test_awips_tiled.py:427: 
UserWarning: rename 'y' to 'y' does not create an index anymore. Try 
using swap_dims instead or use set_index after rename to create an 
indexed coordinate.
     ds = ds.rename(dict((old, new) for old, new in zip(ds.dims, 
['bands', 'y', 'x'])))

tests/writer_tests/test_awips_tiled.py::TestAWIPSTiledWriter::test_basic_numbered_tiles_rgb
 
/usr/lib/python3/dist-packages/satpy/tests/writer_tests/test_awips_tiled.py:427: 
UserWarning: rename 'x' to 'x' does not create an index anymore. Try 
using swap_dims instead or use set_index after rename to create an 
indexed coordinate.
     ds = ds.rename(dict((old, new) for old, new in zip(ds.dims, 
['bands', 'y', 'x'])))

tests/writer_tests/test_awips_tiled.py: 54 warnings
   /usr/lib/python3/dist-packages/satpy/writers/awips_tiled.py:942: 
UserWarning: Production location attribute is longer than 31 characters 
(AWIPS limit). Set it to a smaller value with the 'ORGANIZATION' 
environment variable. Defaults to hostname and is currently set to 
'11111111111111111111111111111111111111111111111111'.
     warnings.warn("Production location attribute is longer than 31 "

tests/writer_tests/test_cf.py::TestCFWriter::test_groups
   /usr/lib/python3/dist-packages/satpy/writers/cf_writer.py:361: 
UserWarning: Cannot pretty-format "acq_time" coordinates because they 
are not unique among the given datasets
     warnings.warn('Cannot pretty-format "{}" coordinates because they 
are not unique among the '

tests/writer_tests/test_cf.py::TestCFWriter::test_link_coords
   /usr/lib/python3/dist-packages/satpy/writers/cf_writer.py:305: 
UserWarning: Coordinate "not_exist" referenced by dataarray var4 does 
not exist, dropping reference.
     warnings.warn('Coordinate "{}" referenced by dataarray {} does not 
exist, dropping reference.'

tests/writer_tests/test_cf.py::TestCFWriter::test_save_with_compression
   /usr/lib/python3/dist-packages/satpy/writers/cf_writer.py:576: 
FutureWarning: The `compression` keyword will soon be deprecated. Please 
use the `encoding` of the DataArrays to tune compression from now on.
     warnings.warn("The `compression` keyword will soon be deprecated. 
Please use the `encoding` of the "

tests/writer_tests/test_mitiff.py: 23 warnings
   /usr/lib/python3/dist-packages/libtiff/libtiff_ctypes.py:644: 
DeprecationWarning: `np.bool` is a deprecated alias for the builtin 
`bool`. To silence this warning, use `bool` by itself. Doing this will 
not modify any behavior and is safe. If you specifically wanted the 
numpy scalar type, use `np.bool_` here.
   Deprecated in NumPy 1.20; for more details and guidance: 
https://numpy.org/devdocs/release/1.20.0-notes.html#deprecations
     elif arr.dtype in np.sctypes['uint'] + [np.bool]:

-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
=========================== short test summary info 
============================
FAILED tests/test_modifiers.py::TestNIRReflectance::test_no_sunz_no_co2 
- Ass...
FAILED tests/test_scene.py::TestSceneLoading::test_load_comp4 - 
AssertionErro...
FAILED tests/test_scene.py::TestSceneLoading::test_load_multiple_resolutions
FAILED 
tests/test_scene.py::TestSceneLoading::test_load_same_subcomposite - A...
FAILED tests/test_scene.py::TestSceneLoading::test_load_comp9 - 
AssertionErro...
FAILED tests/test_scene.py::TestSceneLoading::test_load_comp10 - 
AssertionErr...
FAILED tests/test_scene.py::TestSceneLoading::test_load_comp19 - 
AssertionErr...
FAILED 
tests/test_scene.py::TestSceneLoading::test_load_dataset_after_composite2
FAILED tests/test_scene.py::TestSceneLoading::test_no_generate_comp10 - 
Asser...
FAILED 
tests/test_scene.py::TestSceneResampling::test_resample_reduce_data_toggle
FAILED tests/test_scene.py::TestSceneResampling::test_resample_ancillary 
- Ke...
FAILED 
tests/test_scene.py::TestSceneResampling::test_resample_reduce_data - ...
FAILED tests/test_scene.py::TestSceneResampling::test_no_generate_comp10 
- As...
FAILED 
tests/test_scene.py::TestSceneResampling::test_comp_loading_after_resampling_new_sensor
= 14 failed, 1681 passed, 8 skipped, 64 deselected, 7 xfailed, 786 
warnings in 243.92s (0:04:03) =
autopkgtest [21:51:10]: test python3

-------------- next part --------------
A non-text attachment was scrubbed...
Name: OpenPGP_signature
Type: application/pgp-signature
Size: 495 bytes
Desc: OpenPGP digital signature
URL: <http://alioth-lists.debian.net/pipermail/debian-science-maintainers/attachments/20221022/b7349a86/attachment-0001.sig>


More information about the debian-science-maintainers mailing list