Skip to content

Memory error using dask-image ndmeasure.label #391

@maxbeegee

Description

@maxbeegee

I am obtaining an OOM error when using ndmeasure.label with a large array.

Minimal Complete Verifiable Example:

nx = 5120 # things are OK < 2500
arr = da.random.random(size=(nx,nx,nx))
darr_bin = arr > 0.8
# The next line will fail
label_image, num_labels = dask_image.ndmeasure.label(darr_bin)

Note that this problem occurs already in the last line, and not when executing the computation via, e.g., num_labels.compute().
This also means that I have the same problem when using a (large) cluster as the OOM always occurs on node 1.

Environment:
[I could reproduce this problem on several machines, below is one particular environment]

  • Dask version: 2024.9.1
  • Python version: Python 3.12.6
  • Operating System: Mac OS 12.2
  • Install method (conda, pip, source): conda / mamba

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions