Index index by Group index by Distribution index by Vendor index by creation date index by Name Mirrors Help Search

python311-distributed-2024.2.1-1.2 RPM for noarch

From OpenSuSE Tumbleweed for noarch

Name: python311-distributed Distribution: openSUSE Tumbleweed
Version: 2024.2.1 Vendor: openSUSE
Release: 1.2 Build date: Tue Mar 5 22:29:43 2024
Group: Unspecified Build host: reproducible
Size: 9100458 Source RPM: python-distributed-2024.2.1-1.2.src.rpm
Packager: https://bugs.opensuse.org
Url: https://distributed.dask.org
Summary: Library for distributed computing with Python
Dask.distributed is a library for distributed computing in Python. It
extends both the concurrent.futures and dask APIs to moderate sized
clusters.

Provides

Requires

License

BSD-3-Clause

Changelog

* Tue Mar 05 2024 Ben Greiner <code@bnavigator.de>
  - Update to 2024.2.1
    * Allow silencing dask.DataFrame deprecation warning
    * More robust distributed scheduler for rare key collisions
    * More robust adaptive scaling on large clusters
  - Drop distributed-ignore-daskdepr.patch
* Wed Feb 14 2024 Ben Greiner <code@bnavigator.de>
  - Update to 2024.2.0
    * Deprecate Dask DataFrame implementation
    * Improved tokenization
    * https://docs.dask.org/en/stable/changelog.html#v2024-2-0
  - Add distributed-ignore-daskdepr.patch gh#dask/distributed#8504
  - Drop python39 test flavor
  - Fix python312 tests
* Sun Feb 04 2024 Ben Greiner <code@bnavigator.de>
  - Add python312 test flavor
* Tue Jan 30 2024 Dirk Müller <dmueller@suse.com>
  - update to 2024.1.1:
    * see corresponding dask update:
    https://docs.dask.org/en/stable/changelog.html#v2024-1-1
    * compatibility updates for latest pandas and scipy
* Mon Jan 22 2024 Dirk Müller <dmueller@suse.com>
  - skip one more test for 32bit
* Sat Jan 20 2024 Dirk Müller <dmueller@suse.com>
  - update to 2024.1.0:
    * see corresponding dask update:
    https://docs.dask.org/en/stable/changelog.html#v2024-1-0
* Mon Dec 18 2023 Dirk Müller <dmueller@suse.com>
  - update to 2023.12.1:
    * see corresponding dask update:
    https://docs.dask.org/en/stable/changelog.html#v2023-12-1
* Sat Dec 02 2023 Dirk Müller <dmueller@suse.com>
  - update to 2023.12.0:
    * see corresponding dask update:
    * https://docs.dask.org/en/stable/changelog.html#v2023-12-0
    * Bokeh 3.3.0 compatibility
* Thu Nov 16 2023 Ondřej Súkup <mimi.vx@gmail.com>
  - Update to 2023.11.0
  - add distributed-ignore-rerun.patch to ignore pytest runners during
    test code introspection
    * Zero-copy P2P Array Rechunking
    * Deprecating PyArrow <14.0.1
    * Improved PyArrow filesystem for Parquet
    * Improve Type Reconciliation in P2P Shuffling
    * official support for Python 3.12
    * Reduced memory pressure for multi array reductions
    * improved P2P shuffling robustness
    * Reduced scheduler CPU load for large graphs
* Sun Sep 10 2023 Ben Greiner <code@bnavigator.de>
  - Update to 2023.9.1
    [#]# Enhancements
    Raise in P2P if column dtype is wrong (GH#8167) Hendrik Makait
    * Auto-fail tasks with deps larger than the worker memory (GH#8135) crusaderky
    * Make workers table sortable (GH#8153) Jacob Tomlinson
    * Support for unsetting environment variables (GH#8144) crusaderky
    [#]# Deprecations
    * Deprecate asynchronous Listener.stop() (GH#8151) Hendrik Makait
  - Release 2023.9.0
    [#]# Enhancements
    Don’t capture functools frames in code (GH#8146) crusaderky
    * Ignore asyncio when scraping code (GH#8122) Matthew Rocklin
    * Remove stringification (GH#8083) Florian Jetter
    [#]# Bug Fixes
    * Ensure NannyPlugin are always installed (GH#8107) Florian Jetter
    * Don’t use exception hooks to shorten tracebacks (GH#8127) crusaderky
    * Fix P2P shuffle with LocalCluster(..., processes=False) (GH#8125) Hendrik Makait
    * Merge with P2P shuffle fails if left_index or right_index is True (GH#8121) Patrick Hoefler
  - Release 2023.8.1
    [#]# New Features
    * Add memray integration (GH#8044) Florian Jetter
    [#]# Enhancements
    * Await async listener.stop in Worker.close (GH#8118) Hendrik Makait
    * Minor fixes in memray (GH#8113) Florian Jetter
    * Enable basic p2p shuffle for dask-cudf (GH#7743) Richard (Rick) Zamora
    * Don’t shut down unresponsive workers on gather() (GH#8101) crusaderky
    * Propagate CancelledError in gather_from_workers (GH#8089) crusaderky
    * Better logging for anomalous task termination (GH#8082) crusaderky
    [#]# Bug Fixes
    * Handle null partitions in P2P shuffling (GH#8116) Hendrik Makait
    * Handle CancelledError properly in ConnectionPool (GH#8110) Florian Jetter
    * Fix additional race condition that can cause P2P restart to deadlock (GH#8094) Hendrik Makait
    * Ensure x-axis is uniform when plotting (GH#8093) Florian Jetter
    * Fix deadlock in P2P restarts (GH#8091) Hendrik Makait
  - Release 2023.8.0
    [#]# Enhancements
    * Offload CPU intensive sections of update graph to unblock event loop (GH#8049) Florian Jetter
    * Log worker close reason in events (GH#8042) Florian Jetter
    * Exclude comm handshake from connect timeout (GH#7698) Florian Jetter
    * Automatically restart P2P shuffles when output worker leaves (GH#7970) Hendrik Makait
    * Add Client.unregister_scheduler_plugin method (GH#7968) Brian Phillips
    * Fix log message (GH#8029) Hendrik Makait
    * Send shards grouped by input chunk in P2P rechunking (GH#8010) Hendrik Makait
    [#]# Bug Fixes
    * Close state machine and add-ins first in Worker.close (GH#8066) Hendrik Makait
    * Fix decide_worker picking a closing worker (GH#8032) crusaderky
    * Raise CommClosedError in get_stream_address (GH#8020) jochenott
    * Respect average nthreads in adaptive (GH#8041) Matthew Rocklin
    * Use queued tasks in adaptive target (GH#8037) Matthew Rocklin
    * Restore support for yield unsafe Client context managers and deprecate that support (GH#7987) Thomas Grainger
  - Release 2023.7.1
    [#]# Enhancements
    * gather_dep should handle CancelledError (GH#8013) crusaderky
    * Pass stimulus_id to SchedulerPlugin.remove_worker and SchedulerPlugin.transition (GH#7974) Hendrik Makait
    * Log stimulus_id in retire_worker (GH#8003) crusaderky
    * Use BufferOutputStream in P2P (GH#7991) Florian Jetter
    * Add Coiled to ignored modules for code sniffing (GH#7986) Matthew Rocklin
    * Progress bar can group tasks by span (GH#7952) Irina Truong
    * Improved error messages for P2P shuffling (GH#7979) Hendrik Makait
    * Reduce removing comms log to debug level (GH#7972) Florian Jetter
    [#]# Bug Fixes
    * Fix for TypeError: '<' not supported in graph dashboard (GH#8017) Irina Truong
    * Fix shuffle code to work with pyarrow 13 (GH#8009) Joris Van den Bossche
  - Release 2023.7.0
    [#]# Enhancements
    * Propagate spans to tasks (GH#7898) crusaderky
    * Make Fine Performance Metrics bar graph horizontal (GH#7966) crusaderky
    * Don’t pile up context_meter callbacks (GH#7961) crusaderky
    * Polish Fine Performance Metrics plot (GH#7963) crusaderky
    * Sign task-erred with run_id and reject outdated responses (GH#7933) Hendrik Makait
    * Set Client.as_current when entering ctx (GH#6527) Florian Jetter
    * Re-run erred task on ComputeTaskEvent (GH#7967) Hendrik Makait
    [#]# Bug Fixes
    * Fix crash in spans when time() is not monotonic (GH#7960) crusaderky
  - Release 2023.6.1
    [#]# Enhancements
    * Add idle time to fine performance metrics (GH#7938) crusaderky
    * Spans: capture code snippets (GH#7930) crusaderky
    * Improve memory footprint of P2P rechunking (GH#7897) Hendrik Makait
    * Improve error message on invalid state in _handle_remove_replicas (GH#7920) Hendrik Makait
    * Make ShuffleSchedulerExtension.remove_worker more robust (GH#7921) Hendrik Makait
    * Provide more information if occupancy drops below zero (GH#7924) Hendrik Makait
    * Improved conversion between pyarrow and pandas in P2P shuffling (GH#7896) Hendrik Makait
    [#]# Bug Fixes
    Add Cluster.called_from_running_loop and fix Cluster.asynchronous (GH#7941) Jacob Tomlinson
    * Fix annotations and spans leaking between threads (GH#7935) Irina Truong
    * Handle null partitions in P2P shuffling (GH#7922) Jonathan De Troye
    * Fix race condition in Fine Performance Metrics sync (GH#7927) crusaderky
    * Avoid (GH#7923) by starting run_id at 1 (GH#7925) Hendrik Makait
    * Fix glitches in Fine Performance Metrics stacked graph (GH#7919) crusaderky
  - Release 2023.6.0
    [#]# Enhancements
    * Post fine performance metrics to spans (GH#7885) crusaderky
    * Unique Spans (GH#7882) crusaderky
    * Add a timeout to client.as_completed that mirrors concurrent.futures.as_completed timeout (GH#7811) Thomas Grainger
    * Enforce dtypes in P2P shuffle (GH#7879) Hendrik Makait
    * Support load= keyword for Client.upload_file (GH#7873) James Bourbeau
    * Support get_worker() and worker_client() in async tasks (GH#7844) Thomas Grainger
    * Capture line number for code frames (GH#7786) Miles
    [#]# Bug Fixes
    * Avoid meta roundtrip in P2P shuffle (GH#7895) Hendrik Makait
    * Fix Fine Performance Metrics mis-aligned ColumnData lengths (GH#7893) Miles
    * Fix Fine Performance Metrics spilling crash (GH#7878) Miles
    * Fix spans bug when scatter or client_desires_new_key creates a task (GH#7886) crusaderky
    * Fix Fine Performance Metrics w/ Bokeh 3 (GH#7874) Miles
    * TaskGroup.start can move backwards (GH#7867) crusaderky
    * Use properly imported MatDescriptor for cupy dispatch registration (GH#7868) Charles Blackmon-Luca
    * Ensure retire_workers works if AMM extension hasn’t been loaded (GH#7863) crusaderky
* Mon Jun 05 2023 Steve Kowalik <steven.kowalik@suse.com>
  - Ratchet up bokeh to >= 3.1, since upstream supports it.
  - Also BuildRequire sparse for Python 3.11.
* Fri May 26 2023 Ben Greiner <code@bnavigator.de>
  - Update to 2023.5.1
    * This release drops support for Python 3.8
    [#]# Enhancements
    * Exclude IPython code from computations (GH#7788) Miles
    * Drop Python 3.8 support (GH#7840) Thomas Grainger
    * Add storage_options to performance_report (GH#7636) ypogorelova
    * Don’t warn about mismatched msgpack (GH#7839) Irina Truong
    * Clean up sys.path on Server shutdown (GH#7838) James Bourbeau
    * Dashboard: Fine Performance Metrics (GH#7725) Miles
    [#]# Bug Fixes
    * Properly handle unknown chunk sizes in P2P rechunking (GH#7856)
      Hendrik Makait
    * Minimal change to work around (GH#7726) / support for UCX
      (GH#7851) Benjamin Zaitlen
    * Don’t end computations until cluster is truly idle (GH#7790)
      crusaderky
  - Release 2023.5.0
    [#]# Enhancements
    * Client.upload_file send to both Workers and Scheduler and
      rename scratch directory (GH#7802) Miles
    * Allow dashboard to be used with bokeh prereleases (GH#7814)
      James Bourbeau
    [#]# Bug Fixes
    * Ensure log_event of non-msgpack serializable object do not kill
      servers (GH#7472) Florian Jetter
    * Fix test_nanny.py duplicated pytestmark definitions (GH#7819)
      Thomas Grainger
    * Fix flaky
      test_dask_worker.py::test_single_executable_deprecated
      (GH#7817) Thomas Grainger
  - Release 2023.4.2
    [#]# Enhancements
    * Enable GIL monitoring when gilknocker installed (GH#7730) Miles
    * By default only set logging handler if no other handler has
      been set to avoid double logging (GH#7750) Thomas Grainger
    * Cluster wait (GH#6700) Iain Dorrington
    * Add Prometheus counter for SystemMonitor.last_time (GH#7785)
      Miles
    [#]# Bug Fixes
    * Partial revert defaultclient config setting (GH#7803) Florian
      Jetter
    * Delay awaiting async SchedulerPlugin.{add|remove}_worker hooks
      in order to immediately execute all sync ones (GH#7799) Hendrik
      Makait
    * Fix check_idle not returning the correct value if no change to
      idleness (GH#7781) Jacob Tomlinson
  - Release 2023.4.0
    * With this release we are making a change which will require the
      Dask scheduler to have consistent software and hardware
      capabilities as the client and workers.
    * It’s always been recommended that your client and workers have
      a consistent software and hardware environment so that data
      structures and dependencies can be pickled and passed between
      them. However recent changes to the Dask scheduler mean that we
      now also require your scheduler to have the same consistent
      environment as everything else
    [#]# Enhancements
    * Meter queue time to the offload executor (GH#7758) crusaderky
    * Add GIL contention metric to Prometheus (GH#7651) Miles
    * Add methods Client.forward_logging() and
      Client.unforward_logging(). (GH#7276) Max Bane
    * Optionally capture more frames in computations (GH#7656) Gabe
      Joseph
    * Consider Jupyter activity in idle timeout (GH#7687) Gabe Joseph
    * Add a dashboard component that displays RMM memory (GH#7718)
      Peter Andreas Entschev
    * Improve error message if shuffle/rechunk lost annotations
      (GH#7707) Hendrik Makait
    * Exception chaining in P2P shuffling (GH#7706) Hendrik Makait
    * Use pickle for graph submissions from client to scheduler
      (GH#7564) Florian Jetter
    [#]# Bug Fixes
    * Fix crash on missing env var in dashboard link formatting
      (GH#7729) Miles
    * Fix randbytes() on Python 3.8 (GH#7771) crusaderky
    * Run scheduler of SubprocessCluster in subprocess (GH#7727)
      Hendrik Makait
    * Drop id from RMM dashboard component (GH#7739) James Bourbeau
* Tue Apr 04 2023 Ben Greiner <code@bnavigator.de>
  - Drop python38 test flavor
* Mon Mar 27 2023 Ben Greiner <code@bnavigator.de>
  - Update to 2023.3.2
    [#]# Enhancements
    * Enhanced thread-safety in zict.File (GH#7691) crusaderky
    * Future deserialization without available client (GH#7580)
      Florian Jetter
    * Support adjusting GIL monitoring interval (GH#7650) Miles
    * Gracefully stop GIL monitoring if running (GH#7652) Miles
    * Fine performance metrics for execute, gather_dep, etc.
      (GH#7586) crusaderky
    * Add GIL metric to dashboard (GH#7646) Miles
    * Expose scheduler idle via RPC and HTTP API (GH#7642) Jacob
      Tomlinson
    * Add full dashboard link in scheduler logs (GH#7631) Miles
    [#]# Bug Fixes
    * Tell workers when their peers have left (so they don’t hang
      fetching data from them) (GH#7574) Thomas Grainger
    * Fix regression in dashboard after (GH#7586) (GH#7683)
      crusaderky
    * Fix OverflowError in Cluster._sync_cluster_info() (GH#7648)
      Hendrik Makait
    * Ensure that serialized data is measured correctly (GH#7593)
      Florian Jetter
    [#]# Maintenance
    * Avoid bool object has no attribute close in @gen_cluster
      (GH#7657) Thomas Grainger
    * Fix failing test_server_close_stops_gil_monitoring (GH#7659)
      James Bourbeau
    * Remove weakref finalizer for Offload Executor (GH#7644) Florian
      Jetter
  - Enable python311 build and test: numba not a strict requirement
    of dask
* Tue Mar 14 2023 Ben Greiner <code@bnavigator.de>
  - Skip more flaky tests
* Sat Mar 11 2023 Ben Greiner <code@bnavigator.de>
  - Update to 2023.3.1
    [#]# Enhancements
    * Add Jupyter link to dashboard menu if --jupyter flag is set
      (GH#7638) Jacob Tomlinson
    * Bump minimum click version from 7.0 to 8.0 (GH#7637) Miles
    * Extend dask metapackage dependencies (GH#7630) James Bourbeau
    * Further improvements to Client.restart_workers (GH#7620) Miles
    * P2P offload get_output_partition (GH#7587) Florian Jetter
    * Initial integration of GIL contention metric (GH#7624) Miles
    * Add dashboard documentation links (GH#7610) Miles
    * Rename shuffle/rechunk config option/kwarg to method (GH#7623)
      Hendrik Makait
    * Return results in restart_workers (GH#7606) Miles
    * Ensure client key cancellation uses ordered messages (GH#7583)
      Florian Jetter
    [#]# Bug Fixes
    * Fix undefined async_wait_for -> async_poll_for (GH#7627) Miles
    * Don’t send client heartbeat without a scheduler_comm (GH#7612)
      James Bourbeau
    * Do not unspill on free-keys (GH#7607) crusaderky
    [#]# Documentation
    * Add notes to Client.submit, Client.map, and Client.scatter with
      the description of the current task graph resolution algorithm
      limitations (GH#7588) Eugene Druzhynin
    [#]# Maintenance
    * Use range with pickle protocol versions (GH#7635) jakirkham
    * Share thread pool among P2P shuffle runs (GH#7621) Hendrik
      Makait
    * Replace psutil suspend with BlockedGatherDep in
      test_failing_worker_with_additional_replicas_on_cluster
      (GH#7633) Thomas Grainger
    * Ignore pkg_resources DeprecationWarning for mindeps (GH#7626)
      Miles
    * Implement wait_for using asyncio.timeout() on 3.11 (GH#7571)
      Thomas Grainger
    * Use tmp_path fixture instead of outdated tmpdir fixture
      (GH#7582) ypogorelova
    * Only one crick callback (GH#7614) crusaderky
    * Add mindeps + numpy job to tests CI (GH#7609) Miles
    * Do not xfail whole tests due to (GH#6705) (GH#7611) crusaderky
* Wed Mar 08 2023 Dirk Müller <dmueller@suse.com>
  - update to 2023.3.0:
    * Remove pyarrow dependency for rechunking
    * Update rechunk_transfer and rechunk_unpack errors
    * Remove dead code and document arguments to ShardBuffer
      constructors
    * Fix tests for p2p by default
    * Remove obsolete cast
* Sun Feb 26 2023 Ben Greiner <code@bnavigator.de>
  - Prepare test multiflavors for python311, but skip python311
    * Numba is not ready for python 3.11 yet gh#numba/numba#8304
* Fri Feb 17 2023 Ben Greiner <code@bnavigator.de>
  - Update to 2023.2.0
    [#]# Enhancements
    * Rate limit the worker memory logs (GH#7529) Florian Jetter
    * Move P2P barrier logic to scheduler extension (GH#7519) Hendrik
      Makait
* Sun Feb 05 2023 Ben Greiner <code@bnavigator.de>
  - Update to 2023.1.1
    [#]# Enhancements
    * P2P shuffle deduplicates data and can be run several times
      (GH#7486) Hendrik Makait
    * Reverse order of get_logs() and get_worker_logs() (GH#7475)
      Nicholas R. Knezek
    * Add prometheus metric for time and memory used per task prefix
      (GH#7406) Thomas Grainger
    * Additive worker counts in Prometheus (GH#7468) crusaderky
    * Add help tool for taskstream (GH#7478) Florian Jetter
    [#]# Bug Fixes
    * Do not allow for a worker to reject a drop replica request
      (GH#7490) Hendrik Makait
    * Fix un/packing for namedtuples with custom constructors
      (GH#7465) antonymayi
    * Documentation
    * Remove timeout= from docstring example for worker_client
      (GH#7497) Florian Jetter
    [#]# Maintenance
    * Ignore get_default_shuffle_algorithm linting issue (GH#7506)
      Florian Jetter
    * Remove set_config when using default client (GH#7482) Florian
      Jetter
    * Update gpuCI RAPIDS_VER to 23.04 (GH#7501)
    * Fix test_balance_expensive_tasks and improve helper functions
      in test_steal.py (GH#7253) Hendrik Makait
    * Sign every compute task with run ID to correlate response
      (GH#7463) Hendrik Makait
  - Release 2023.1.0
    [#]# New Features
    * Add local SubprocessCluster that runs workers in separate
      processes (GH#7431) Hendrik Makait
    [#]# Enhancements
    * Ensure client session is quiet after cluster.close() or
      client.shutdown() (GH#7429) James Bourbeau
    * Set lifetime-stagger default value to None (GH#7445)
      bstadlbauer
    * Memory thresholds should never be exactly 0.0 (GH#7458) Stuart
      Berg
    * Remove the Incorrect-Sizeof-Warning (GH#7450) Mads R. B.
      Kristensen
    * Log exceptions in P2P shuffle tasks (GH#7442) Hendrik Makait
    * Add support for packing namedtuple and add test for future
      resolution in submit (GH#7292) Andrew
    * Avoid deep copy on lz4 decompression (GH#7437) crusaderky
    * Avoid deep copy of numpy buffers on unspill (GH#7435)
      crusaderky
    * Don’t error when clicking on empty task stream plot (GH#7432)
      James Bourbeau
    * Do not count spilled memory when comparing vs. process memory
      (GH#7430) crusaderky
    * Stop Client periodic callbacks during shutdown() (GH#7428)
      James Bourbeau
    * Add dask spec CLI (GH#7427) Matthew Rocklin
    * Create new zstd (de)compressor for each compression call
      (GH#7404) Dylan Wragge
    * Rename managed_in_memory etc. to match GUI (GH#7418) crusaderky
    * Warn users when sizeof() returns inflated output (GH#7419)
      crusaderky
    [#]# Bug Fixes
    * Ensure dicts are properly recognized as msgpack serializable
      (GH#7473) Florian Jetter
    * Reset state of ShuffleSchedulerExtension on restart (GH#7446)
      Hendrik Makait
    * Reject non-string column names in P2P shuffle (GH#7447) Hendrik
      Makait
    * Avoid int32 in dashboard (GH#7443) Matthew Rocklin
    * Fix P2PShuffle serialization for categorical data (GH#7410)
      Hendrik Makait
    * WorkerPorcess blocks on kill if still starting (GH#7424)
      Matthew Rocklin
    [#]# Documentation
    * Move Prometheus docs from dask/dask (GH#7405) crusaderky
    [#]# Maintenance
    * Various cleanups in semaphore (GH#5885) Florian Jetter
    * test_rlimit fails on MacOSX (GH#7457) crusaderky
    * Bump actions/checkout from 3.2.0 to 3.3.0 (GH#7464)
    * Remove conditional imports of psutil (GH#7462) crusaderky
    * Drop support for zict < 2.1.0 (GH#7456) crusaderky
    * Fix flaky test_digests (GH#7454) crusaderky
    * Add minimum dependency testing to CI (GH#7285) Charles
      Blackmon-Luca
    * Avoid overflow in statitics.mean (GH#7426) Matthew Rocklin
    * Ignore numpy bool8 deprecation (GH#7423) Matthew Rocklin
    * Add missing skips for pyarrow (GH#7416) Elliott Sales de
      Andrade
    * Be more permissive about expected ciphers in tests (GH#7417)
      Elliott Sales de Andrade
    * Revert “TST: Fetch executables from build root (GH#2551)”
      (GH#7415) Elliott Sales de Andrade
* Mon Jan 02 2023 Ben Greiner <code@bnavigator.de>
  - Update to 2022.12.1
    [#]# Enhancements
    * SpillBuffer metrics (GH#7368) crusaderky
    * Prometheus: measure how much spilling blocks the event loop (GH#7370) crusaderky
    * Add transfer_outgoing_bytes_total metric (GH#7388) Gabe Joseph
    * Fail P2PShuffle gracefully upon worker failure (GH#7326) Hendrik Makait
    [#]# Bug Fixes
    * Select queued tasks in stimuli, not transitions (GH#7402) Gabe Joseph
    * Check ContextVar in default_client (GH#7369) Matthew Rocklin
    * Fix sending event messages to non-subscribers (GH#7014) Laurence Watts
    * Set sizing mode on Tabs to avoid layout collapse (GH#7365) Mateusz Paprocki
    [#]# Maintenance
    * Restructure P2PShuffle extensions (GH#7390) Hendrik Makait
    * Add Python 3.11 classifier (GH#7408) James Bourbeau
    * Add support for Python 3.11 (GH#7249) Thomas Grainger
    * Add test for using annotations with client.submit and client.map (GH#7399) James Bourbeau
    * Bump actions/checkout from 3.1.0 to 3.2.0 (GH#7393)
    * Remove superfluous ShuffleSchedulerExtension.barriers (GH#7389) Hendrik Makait
    * Remove ignore annotation-unchecked (GH#7379) crusaderky
    * Remove tornado max version from nightly recipe (GH#7376) Charles Blackmon-Luca
    * Remove the experimental feature warning for Semaphore (GH#7373) Florian Jetter
  - Release 2022.12.0
    [#]# Enhancements
    * Expose event loop health metrics in Prometheus (GH#7360) Hendrik Makait
    * Allow log propagation by default (GH#5669) Florian Jetter
    * Clean up of unpack_remotedata() (GH#7322) Mads R. B. Kristensen
    * Upgrade to tornado 6.2 (GH#7286) Thomas Grainger
    * Introduce Server level comm counters (GH#7332) Florian Jetter
    * Prometheus debug log (GH#7302) Florian Jetter
    [#]# Bug Fixes
    * Catch BaseException s from user tasks (GH#5997) Gabe Joseph
    * Impossible use case of erred deps in transition to waiting (GH#7354) crusaderky
    * Fix a deadlock when queued tasks are resubmitted quickly in succession (GH#7348) Florian Jetter
    [#]# Maintenance
    * Fetch all artifacts (GH#7355) Enrico Minack
    * Delay fsspec and urllib3 import time (GH#6659) Florian Jetter
    * Bump mypy (GH#7349) crusaderky
    * Bump mypy and remove win specific run (GH#7344) Florian Jetter
    * Finish overhaul of SchedulerState annotations (GH#7333) crusaderky
    * Fix flaky test_pause_while_spilling (GH#7334) Gabe Joseph
    * Update gpuCI RAPIDS_VER to 23.02 (GH#7337)
  - Drop distributed-pr7286-tornado-6-2.patch
  - Move to PEP517 build
* Mon Nov 21 2022 Ben Greiner <code@bnavigator.de>
  - Remove allow-bokeh3.patch
    * gh#dask/distributed#7329, gh#dask/dask#9659
    * we provide a legacy bokeh2 instead
* Sun Nov 20 2022 Ben Greiner <code@bnavigator.de>
  - Update to version 2022.11.1
    [#]# Enhancements
    * Restrict bokeh=3 support (GH#7329) Gabe Joseph (ignored in rpm
      fixed by bokek 3.0.2, see gh#dask/dask#9659)
    * Respect death timeout when waiting for scheduler file (GH#7296)
      Florian Jetter
    * Always raise exception if P2PShuffle s send fails (GH#7317)
      Hendrik Makait
    [#]# Maintenance
    * Remove exception handling from transitions (GH#7316) crusaderky
    * Turn private functions into private SchedulerState methods
      (GH#7260) Hendrik Makait
    * Bump toolz minimum version to 0.10.0 (GH#7309) Sam Grayson
  - Release 2022.11.0
    [#]# Note
    * This release changes the default scheduling mode to use
      queuing. This will significantly reduce cluster memory use in
      most cases, and generally improve stability and performance.
      Learn more here and please provide any feedback on this
      discussion.
    * In rare cases, this could make some workloads slower. See the
      documentation for more information, and how to switch back to
      the old mode.
    [#]# New Features
    * Add ForwardOutput worker plugin to forward stdout and stderr to
      client. (GH#7297) Hendrik Makait
    * Duration counters on prefix level (GH#7288) Florian Jetter
    * Include button for launching JupyterLab layout in repr
      (GH#7218) Ian Rose
    [#]# Enhancements
    * Support MIG parsing during CUDA context creation in UCX
      initialization (GH#6720) Peter Andreas Entschev
    * Handle /metrics endpoint without prometheus-client installed
      (GH#7234) Hendrik Makait
    * Add support for unpacking namedtuples in remote data (GH#7282)
      Andrew
    * Enable queuing by default (GH#7279) Florian Jetter
    * Fix exists->``exist`` typo in scheduler error messages
      (GH#7281) Matthew Plough
    * If there’s an exception in the Client async context manager
      body then close fast (GH#6920) Thomas Grainger
    [#]# Bug Fixes
    * Ensure category is optional when logging "warn" events
      (GH#7169) James Bourbeau
    * Edge and impossible transitions to memory (GH#7205) crusaderky
    [#]# Maintenance
    * Allow bokeh=3 (GH#5648) James Bourbeau
    * Fix typos in P2P shuffle code (GH#7304) Hendrik Makait
    * Reenable test_bad_disk (GH#7300) Florian Jetter
    * Reduce max-runs in test reports (GH#7299) Florian Jetter
    * Revert idle classification when worker-saturation is set
      (GH#7278) Florian Jetter
    * Fix flaky deadline_expiration (GH#7287) Florian Jetter
    * Rewrite of P2P control flow (GH#7268) Florian Jetter
    * Bump minimum bokeh version to 2.4.2 (GH#7271) James Bourbeau
    * Remove deprecated code calls to IOLoop.make_current() (GH#7240)
      Thomas Grainger
    * Improved test for balancing expensive tasks (GH#7272) Hendrik
      Makait
    * Refactor semaphore._Watch into general-purpose Deadline utility
      (GH#7238) Hendrik Makait
    * Require Click 7.0+ (GH#7226) jakirkham
    * Drop tests (GH#7269) Hendrik Makait
    * Replace test_(do_not_)steal_communication_heavy_tasks tests
      with more robust versions (GH#7243) Hendrik Makait
    * xfail test_bad_disk (GH#7265) crusaderky
    * Move transition_log from Scheduler to SchedulerState (GH#7254)
      crusaderky
    * Remove Scheduler.log (GH#7258) crusaderky
    * Use latest pickle (GH#5826) jakirkham
    * Polish parsing of worker-saturation from config (GH#7255)
      crusaderky
    * Avoid expensive occupancy calculation when unused (GH#7257)
      Gabe Joseph
    * Un-skip test_nested_compute (GH#7247) Gabe Joseph
    * Review test_do_not_steal_communication_heavy_tasks (GH#7250)
      crusaderky
    * Fix test_stress_creation_and_deletion (GH#7215) crusaderky
    * Raise exceptions in Server.handle_stream instead of
      swallowing/logging (GH#7162) Hendrik Makait
    * Fix _update_scheduler_info hanging failed tests (GH#7225) Gabe
      Joseph
    * Bump xarray-contrib/ci-trigger from 1.1 to 1.2 (GH#7232)
  - Release 2022.10.2
    * Reverted a bug where Bokeh was accidentally made non-optional
      (GH#7230) Oliver Holworthy
    * Schedule a queued task when a task secedes (GH#7224) Gabe
      Joseph
    * This was a hotfix release
  - Release 2022.10.1
    [#]# New Features
    * Add Client.restart_workers method (GH#7154) James Bourbeau
    * Implement PackageInstall plugin for pip and conda (GH#7126)
      Hendrik Makait
    [#]# Enhancements
    * Add prometheus collector for work-stealing (GH#7206) Hendrik
      Makait
    * Track reason of workers closing and restarting (GH#7166)
      Hendrik Makait
    * Show no-worker on task progress bar (GH#7171) Florian Jetter
    * Set OPENBLAS_NUM_THREADS by default (GH#7177) James Bourbeau
    * Optionally provide local directory to data constructor
      (GH#7153) Lawrence Mitchell
    * Introduce distributed.comm.ucx.environment config slot
      (GH#7164) Lawrence Mitchell
    * Log information about memory limit (GH#7160) Florian Jetter
    * Improve log messages on scheduler for restart (GH#7150) Florian
      Jetter
    * More comprehensive WorkerState task counters (GH#7167)
      crusaderky
    [#]# Bug Fixes
    * Update minimum bokeh version message (GH#7172) James Bourbeau
    * Revamped implementations of remote print() and warn(), fixing
      [#7095] (GH#7129) Max Bane
    [#]# Maintenance
    * Temporarily restrict bokeh<3 (GH#7219) James Bourbeau
    * Make Scheduler.reschedule private (GH#7216) crusaderky
    * Fix decide_worker_rootish_queuing_disabled assert (GH#7065)
      Gabe Joseph
    * Fix flaky test_include_communication_in_occupancy (GH#7212)
      Gabe Joseph
    * Do not raise on leaked websockets (GH#7199) Florian Jetter
    * Update nightly recipes with CLI tests, dependency changes
      (GH#7201) Charles Blackmon-Luca
    * Make p2p shuffle submodules private (GH#7186) Florian Jetter
    * Backport tornado PeriodicCallback (GH#7165) Florian Jetter
    * Fix mypy failure on CI (GH#7198) Florian Jetter
    * User a layer for p2p shuffle (GH#7180) Florian Jetter
    * Type annotations for shuffle (GH#7185) Florian Jetter
    * Do not close worker on comm error in heartbeat (GH#7163)
      Hendrik Makait
    * Errors when setting TCP timeouts log as error (GH#7161) Florian
      Jetter
    * Remove incorrect advice from pre-commit config (GH#7159)
      crusaderky
    * Use functionalities network for codecov uploader (GH#7148)
      Florian Jetter
    * Use counter metric type where appropriate, incoming_count was
      reporting bytes (GH#7125) Nat Tabris
  - Add allow-bokeh3.patch -- raise upper bokeh version, reverts part
    of gh#dask/distributed#7329, see also gh#dask/dask#9659
  - Replace support-tornado-6-2.patch with
    distributed-pr7286-tornado-6-2.patch -- gh#dask/distributed#7286
* Fri Oct 21 2022 Ben Greiner <code@bnavigator.de>
  - Update to version 2022.10.0
    * Use of new dask CLI (GH#6735) Doug Davis
    * Refactor occupancy (GH#7075) Hendrik Makait
    * Expose managed/unmanaged/spilled memory to Prometheus (GH#7112)
      crusaderky
    * Round up saturation-factor (GH#7116) Gabe Joseph
    * Return default on KeyError at any level in get_metadata
      (GH#7109) Hendrik Makait
    * Count task states per task prefix and expose to Prometheus
      (GH#7088) Nat Tabris
    * Add scheduler-sni option for dask workers (GH#6290) Burt
      Holzman
    * Improve exception catching in UCX communication (GH#7132) Peter
      Andreas Entschev
    * Improve robustness of PipInstall plugin (GH#7111) Hendrik
      Makait
    * Fix dependencies that should point to dask/dask (GH#7138) James
      Bourbeau
    * Hold on to z.sum() until test completes (GH#7136) Lawrence
      Mitchell
    * Update typing for system_monitor after python/typeshed#8829
      (GH#7131) Lawrence Mitchell
    * Fix two potentially flaky queuing tests (GH#7124) Gabe Joseph
    * Revamp SystemMonitor (GH#7097) crusaderky
    * Adjust hardware benchmarks bokeh test (GH#7096) Florian Jetter
    * Multi-platform mypy checks (GH#7094) crusaderky
* Fri Sep 30 2022 Arun Persaud <arun@gmx.de>
  - update to version 2022.9.2:
    * Enhancements
      + Smarter stealing with dependencies (GH#7024) Hendrik Makait
      + Enable Active Memory Manager by default (GH#7042) crusaderky
      + Allow timeout strings in distributed.wait (GH#7081) James
      Bourbeau
      + Make AMM memory measure configurable (GH#7062) crusaderky
      + AMM support for actors (GH#7072) crusaderky
      + Expose message-bytes-limit in config (GH#7074) Hendrik Makait
      + Detect mismatching Python version in scheduler (GH#7018) Hendrik
      Makait
      + Improve KilledWorker message users see (GH#7043) James Bourbeau
      + Support for cgroups v2 and respect soft limits (GH#7051)
      Samantha Hughes
    * Bug Fixes
      + Catch BaseException on UCX read error (GH#6996) Peter Andreas
      Entschev
      + Fix transfer limiting in _select_keys_for_gather (GH#7071)
      Hendrik Makait
      + Parse worker-saturation if a string (GH#7064) Gabe Joseph
      + Nanny(config=...) parameter overlays global dask config
      (GH#7069) crusaderky
      + Ensure default clients don’t propagate to subprocesses (GH#7028)
      Florian Jetter
    * Documentation
      + Improve documentation of message-bytes-limit (GH#7077) Hendrik
      Makait
      + Minor tweaks to Sphinx documentation (GH#7041) crusaderky
      + Improve upload_file API documentation (GH#7040) Florian Jetter
    * Maintenance
      + test_serialize_numba: Workaround issue with np.empty_like in NP
      1.23 (GH#7089) Graham Markall
      + Type platform constants for mypy (GH#7091) jakirkham
      + dask-worker-space (GH#7054) crusaderky
      + Remove failing test case (GH#7087) Hendrik Makait
      + test_default_client (GH#7058) crusaderky
      + Fix pre-commit fails with recent versions of mypy and pandas
      (GH#7068) crusaderky
      + Add factorization utility (GH#7048) James Bourbeau
* Fri Sep 16 2022 Arun Persaud <arun@gmx.de>
  - update to version 2022.9.1:
    * Enhancements
      + Add dashboard component for size of open data transfers
      (GH#6982) Hendrik Makait
      + Allow very fast keys and very expensive transfers as stealing
      candidates (GH#7022) Florian Jetter
    * Bug Fixes
      + No longer double count transfer cost in stealing (GH#7036)
      Hendrik Makait
    * Maintenance
      + Make test_wait_first_completed robust (GH#7039) Florian Jetter
      + Partial annotations for SchedulerState (GH#7023) crusaderky
      + Add more type annotations to stealing.py (GH#7009) Florian
      Jetter
      + Update codecov settings (GH#7015) Florian Jetter
      + Speed up test_balance (GH#7008) Florian Jetter
      + Fix test report after queuing job added (GH#7012) Gabe Joseph
      + Clean up env variables in Gihub Actions (GH#7001) crusaderky
      + Make test_steal_reschedule_reset_in_flight_occupancy non timing
      dependent (GH#7010) Florian Jetter
      + Replaced distributed.utils.key_split with dask.utils.key_split
      (GH#7005) Luke Conibear
      + Revert “Revert “Limit incoming data transfers by amount of data”
      (GH#6994)” (:pr:`7007) Florian Jetter
      + CI job running tests with queuing on (GH#6989) Gabe Joseph
      + Fix distributed/tests/test_client_executor.py::test_wait
      (GH#6990) Florian Jetter
* Sat Sep 10 2022 Arun Persaud <arun@gmx.de>
  - update to version 2022.9.0:
    * Enhancements
      + Limit incoming data transfers by amount of data (GH#6975)
      Hendrik Makait
      + Expose transfer-related metrics in Worker.get_metrics and
      WorkerMetricCollector (GH#6936) Hendrik Makait
      + Withhold root tasks (no co assignment) (GH#6614) Gabe Joseph
      + Improve differentiation between incoming/outgoing connections
      and transfers (GH#6933) Hendrik Makait
      + Change memory bars color on spilling/paused status (GH#6959)
      crusaderky
      + Ensure restart clears taskgroups et al (GH#6944) Florian Jetter
      + Optimise scheduler.get_comm_cost set difference (GH#6931)
      Lawrence Mitchell
      + Expose setting multiple protocols and ports via the
      dask-scheduler CLI (GH#6898) Jacob Tomlinson
      + Make TextProgressBar clear the line when finished (GH#5968)
      Vincenzo Eduardo Padulano
    * Bug Fixes
      + Revert getaddrinfo fast path for Python 3.8 (GH#6978) Florian
      Jetter
      + cancelled/resumed->long-running transitions (GH#6916) crusaderky
    * Deprecations
      + Deprecate default value for Client.wait_for_workers (GH#6942)
      Florian Jetter
    * Documentation
      + Document Scheduler and Worker state machine (GH#6948) crusaderky
      + Insert memory_limit parameter into LocalCluster docstring
      (GH#6839) Crislana Rafael
    * Maintenance
      + Revert “Limit incoming data transfers by amount of data”
      (GH#6994) Florian Jetter
      + Cache conda environment between CI test runs (GH#6855) Charles
      Blackmon-Luca
      + Revert “Fix co-assignment for binary operations” (GH#6985) Gabe
      Joseph
      + Cache test_report shelves in CI (GH#6937) Florian Jetter
      + Cleanup ipywidgets mocking (GH#6918) Thomas Grainger
      + Improve testing of {Scheduler|Worker}MetricCollector (GH#6945)
      Hendrik Makait
      + Clean up nanny WorkerProcess.kill (GH#6972) Gabe Joseph
      + Rewrite test_reconnect to use subprocess to kill scheduler
      reliably (GH#6967) Florian Jetter
      + Task state domain on the scheduler side (GH#6929) crusaderky
      + Remove @avoid_ci from test_steal (GH#6872) crusaderky
      + Use async with Worker in tests (GH#6958) crusaderky
      + Ignore spurious warnings in test_quiet_close_process (GH#6955)
      crusaderky
      + Fix tests on Windows (GH#6954) Hendrik Makait
      + Prevent duplicates in HeapSet.sorted() (GH#6952) crusaderky
      + Propagate worker address and improve _remove_from_processing
      behavior (GH#6946) Hendrik Makait
      + Add HeapSet._sorted internal flag (GH#6949) Gabe Joseph
      + Add HeapSet.peekn (GH#6947) Gabe Joseph
      + Fix pyright error when importing from distributed (GH#6904) Ian
      Liu Rodrigues
      + Always return ws.address from _remove_from_processing (GH#6884)
      Hendrik Makait
      + Use async with Client: in tests (GH#6921) Thomas Grainger
      + Ensure relative memory limits work as percentage of system
      memory (GH#6923) Florian Jetter
* Thu Sep 01 2022 Steve Kowalik <steven.kowalik@suse.com>
  - Update to 2022.8.1:
    * Display unexpected state in Worker.execute validation (:pr:`6856`)
      James Bourbeau
    * cancelled / resumed -> rescheduled transition (:pr:`6913`) crusaderky
    * Fix resource deallocation for resumed tasks (:pr:`6914`) crusaderky
    * Only close scheduler in SpecCluster if it exists (:pr:`6888`)
      Matthew Rocklin
    * Always close BatchedSend write coroutines (:pr:`6865`) Gabe Joseph
    * Harden preamble of Worker.execute against race conditions (:pr:`6878`)
      crusaderky
    * Duplicated code: in_flight_tasks validation (:pr:`6917`) crusaderky
    * Overhaul transitions for the resumed state (:pr:`6699`) crusaderky
    * Clean up cluster process reaping (:pr:`6840`) Gabe Joseph
    * Don't use bokeh Figure in tests (:pr:`6721`) Bryan Van de Ven
    * Work around incompatibility of crick with setuptools 65 (:pr:`6887`)
      crusaderky
    * Replace port = random.randint( with port = d.utils.open_port (:pr:`6883`)
      Thomas Grainger
    * Fix flaky test_wall_clock (:pr:`6879`) crusaderky
    * Add descriptive error message to assert (:pr:`6871`) Hendrik Makait
    * Increase timeout in test_quiet_process (:pr:`6857`) Florian Jetter
    * Remove redundant use of with clean(): (:pr:`6852`) Thomas Grainger
    * Update pre-commit dependencies (:pr:`6851`) Hendrik Makait
    * Call exit callback even if AsyncProcess is reaped elsewhere (:pr:`6684`)
      Thomas Grainger
    * Avoid function calls in argument defaults (:pr:`6812`) Hendrik Makait
    * Add Cluster.get_client() method (:pr:`6745`) Julia Signell
    * Start bokeh app to activate bokeh's clean session callbacks (:pr:`6728`)
      Martí Zamora
    * Don't connect to cluster subprocesses at shutdown (:pr:`6829`) Gabe Joseph
    * Fix restart wait for workers edge case (:pr:`6823`) Gabe Joseph
    * Fix spilled size calculation in Slow (:pr:`6789`) Hendrik Makait
    * Deprecate passing stopped loops to LoopRunner (and therefore
      Client/Cluster) (:pr:`6680`) Thomas Grainger
    * Only set 5s connect timeout in gen_cluster tests (:pr:`6822`) Gabe Joseph
    * Fix flaky test_worker_who_has_clears_after_failed_connection (:pr:`6832`)
      Gabe Joseph
    * Assert otherwise pointless comparisons (B015) (:pr:`6811`) Hendrik Makait
    * Fix Jupyter security note (:pr:`6818`) Jacob Tomlinson
    * Improve check_thread_leak output (:pr:`6797`) Gabe Joseph
    * Use contextmanager to ensure clients are closed and do not leak
      (:pr:`6817`) Hendrik Makait
    * Robust thread termination in test_watch and
      test_watch_requires_lock_to_run (:pr:`6788`) Hendrik Makait
    * Avoid unused loop control variable or name them _ (:pr:`6813`)
      Hendrik Makait
    * Avoid mutable argument defaults in tests (:pr:`6810`) Hendrik Makait
    * Avoid mutable argument defaults outside of tests (:pr:`6665`)
      Hendrik Makait
    * Wait for workers to return in Client.restart (:pr:`6714`) Gabe Joseph
    * Remove global mutable Cluster._cluster_info (:pr:`6487`) Thomas Grainger
    * Fix: nvml no early init (:pr:`6678`) Lawrence Mitchell
    * Fix bug when restarting client (:pr:`6654`) Iain Dorrington
    * Failure to spill breaks available resources (:pr:`6703`) crusaderky
    * Fix resource allocation for tasks with dependencies (:pr:`6676`)
      Hendrik Makait
    * Remove test_restart_fast_sync, test_fast_kill (:pr:`6750`) Gabe Joseph
    * Fix flaky test_async_task_group_call_later_executes_delayed_task_in_background
      (:pr:`6744`) Hendrik Makait
    * Store ready and constrained tasks in heapsets (:pr:`6711`) crusaderky
    * Future-proof Bokeh value import (:pr:`6707`) Bryan Van de Ven
    * Validate constrained tasks (:pr:`6698`) crusaderky
    * Minor quality-of-life tweaks to cancelled state (:pr:`6701`) crusaderky
    * Pickle worker state machine exceptions (:pr:`6702`) crusaderky
    * Partial matches for worker state machine instructions (:pr:`6704`)
      crusaderky
    * Automatically mark all WorkerState tests (:pr:`6706`) crusaderky
  - Add patch support-tornado-6-2.patch:
    * Ignore two new deprecation warnings from Tornado 6.2.
  - Stop pining tornado to <6.2
  - Refreshed patches.
* Tue Jul 12 2022 Ben Greiner <code@bnavigator.de>
  - Fix runtime requirement with tornado <6.2
  - Move one more test to notparallel
  - Skip failing test on Python 3.10 i586 gh#dask/distributed#6718
* Mon Jul 11 2022 Arun Persaud <arun@gmx.de>
  - update to version 2022.7.0:
    * Enhancements
      + Use a tempdir path by default instead of cwd for the worker
      scratch dir (GH#6658) Florian Jetter
      + Add WorkerState.all_running_tasks (GH#6690) Hendrik Makait
      + Scheduler.reschedule() works only by accident (GH#6339)
      crusaderky
      + Remove spurious UnpauseEvent at worker start (GH#6652)
      crusaderky
      + Log if closing an executor is not possible in thread (GH#6644)
      Florian Jetter
      + Cloudpickle register by value (GH#6466) Ian Rose
      + Adding replicas to a task in fetch now sends it to flight
      immediately (GH#6594) crusaderky
    * Bug Fixes
      + Fix dump output of parameter-less events (GH#6695) crusaderky
      + Set MALLOC_TRIM_THRESHOLD_ before interpreter start (GH#6681)
      crusaderky
      + Fix deadlocks around rescheduled and resumed states (GH#6673)
      crusaderky
      + has_arg returns True for keyword-only arguments (GH#6648)
      Lawrence Mitchell
      + Transitions caused by worker death use old ‘worker-connect’
      stimulus_id (GH#6657) crusaderky
      + A key is forgotten while acquire-replicas is running (GH#6638)
      crusaderky
    * Maintenance
      + Revisit WorkerState.long_running set (GH#6697) crusaderky
      + WorkerState unit tests for resumed state (GH#6688) crusaderky
      + Bump version of pandas-stubs (GH#6691) crusaderky
      + Add dummy factory methods for ExecuteSuccessEvent and
      ExecuteFailureEvent (GH#6687) Hendrik Makait
      + Pin tornado<6.2 in nightly conda recipes (GH#6675) Peter Andreas
      Entschev
      + Refactor resource restriction handling in WorkerState (GH#6672)
      Hendrik Makait
      + test_signal and test_terminate occasionally return SIGKILL on
      MacOS (GH#6671) crusaderky
      + Use the loop fixture in even more tests (GH#6674) Thomas
      Grainger
      + Inconditionally import ssl (GH#6670) Thomas Grainger
      + Use the loop fixture in more tests (GH#6642) Thomas Grainger
      + Pin tornado to <6.2 (GH#6668) Florian Jetter
      + Handle AsyncTaskGroupClosedError (GH#6664) Hendrik Makait
      + Replace occurences of large delay slowinc with locks (GH#6656)
      Florian Jetter
      + Merge extend-ignore and ignore values for flake8 (GH#6660)
      Hendrik Makait
      + Remove server close background task grace period (GH#6633)
      Thomas Grainger
      + Do not use tempfile in utils_test (GH#6651) Florian Jetter
      + close_worker cleanup (GH#6650) crusaderky
      + Rewrite test_cancelled_resumed_after_flight_with_dependencies
      using WorkerState (GH#6645) crusaderky
      + Log the actual duration to create a directory (GH#6647) Florian
      Jetter
      + pandas type stubs (GH#6635) crusaderky
      + Remove unused __started Event in Server (GH#6615) Florian Jetter
      + Use safe temp directory in gen_cluster (GH#6628) Florian Jetter
      + Print CI host info (GH#6629) crusaderky
      + Deduplicate data_needed (GH#6587) crusaderky
      + Remove EnsureCommunicatingAfterTransitions (GH#6462) crusaderky
      + Pickle WorkerState (GH#6623) crusaderky
      + Harden vs. TaskState collisions (GH#6593) crusaderky
      + Do not interact with the event loop when the cluster is garbage
      collected (GH#6627) Thomas Grainger
* Fri Jun 24 2022 Ben Greiner <code@bnavigator.de>
  - Update to 2022.6.1
    * Highlights
    - This release includes the Worker State Machine refactor. The
      expectation should be that the worker state is its own
      synchronous subclass. Pulling all the state out into its own
      class allows us to write targeted unit tests without invoking
      any concurrent or asynchronous code.
    * Enhancements
    - Make worker state machine methods private (GH#6564)
      crusaderky
    - Yank state machine out of Worker class (GH#6566) crusaderky
    - Track worker_state_machine.TaskState instances (GH#6525)
      Hendrik Makait
    - Trivial tweaks to the Worker State Machine (GH#6586)
      crusaderky
    - Replace loop.call_later and loop.add_callback with background
      tasks added to Server. (GH#6603) Thomas Grainger
    - Support for neater WorkerState tests (GH#6609) crusaderky
    - Limit TCP writes with Tornado to 2GB (GH#6557) hhuuggoo
    - Enable no_implicit_optional for scheduler (GH#6622) Thomas
      Grainger
    * Bug Fixes
    - Partial revert of compute-task message format (GH#6626)
      Florian Jetter
    - Restore log message about received signals in CLI (GH#6618)
      Florian Jetter
    - Handle empty memoryviews of bytearrays when (de)serializing
      (GH#6576) Benjamin Zaitlen
    - Ensure steal requests from same-IP but distinct workers are
      rejected (GH#6585) Florian Jetter
    - Fix tls_(min|max)_ version having no effect on openssl 1.1.0g
      or lower (GH#6562) Thomas Grainger
    - Fix idle_timeout and unxfail test (GH#6563) Matthew Rocklin
    - Fix crashing debug statement in _purge_state (GH#6589)
      crusaderky
    - Abort connections on CancelledError (GH#6574) Thomas Grainger
    - Fix Active Memory Manager ignores nbytes thresholds (GH#6583)
      crusaderky
    * Deprecations
    - Deprecate WorkerState accessors (GH#6579)
      crusaderky
  - Release 2022.6.0
    * Enhancements
    - Make disk access in system monitor configurable (GH#6537)
      Matthew Rocklin
    - Log and except errors on preload start (GH#6553) Matthew
      Rocklin
    - Fix Scheduler.restart logic (GH#6504) Gabe Joseph
    - Don’t heartbeat while Worker is closing (GH#6543) Gabe Joseph
    - No longer retry LocalCluster on errno.EADDRINUSE (GH#6369)
      Thomas Grainger
    - Don’t invoke log_event from state machine (GH#6512)
      crusaderky
    - Add config option to disable profiling and disable it in many
      tests per default (GH#6490) Hendrik Makait
    - Encapsulate Worker.batched_stream.send() (GH#6475) crusaderky
    * Bug Fixes
    - refresh-who-has can break the worker state machine (GH#6529)
      crusaderky
    - Restart worker if it’s unrecognized by scheduler (GH#6505)
      Gabe Joseph
    - Fix import error when distributed.rmm.pool-size is set
      (GH#6482) KoyamaSohei
    * Deprecations
    - Restore signature compatibility for dask-gateway (GH#6561)
      Tom Augspurger
    - Deprecate the io_loop and loop kwarg to Server, Worker, and
      Nanny (GH#6473) Thomas Grainger
    - Deprecate the loop kwarg to Scheduler (GH#6443) Thomas
      Grainger
  - Release 2022.05.2
    * Enhancements
    - Add a lock to distributed.profile for better concurrency
      control (GH#6421) Hendrik Makait
    - Send SIGKILL after SIGTERM when passing 95% memory (GH#6419)
      crusaderky
    * Bug Fixes
    - Log rather than raise exceptions in preload.teardown()
      (GH#6458) Matthew Rocklin
    - Handle failing plugin.close() calls during scheduler shutdown
      (GH#6450) Matthew Rocklin
    - Fix slicing bug in ensure_memoryview (GH#6449) jakirkham
    - Generalize UCX errors on connect() and correct pytest
      fixtures (GH#6434) Peter Andreas Entschev
    - Run cluster widget periodic callbacks on the correct event
      loop (GH#6444) Thomas Grainger
  - Release 2022.05.1
    * New Features
    - Add HTTP API to scheduler (GH#6270) Matthew Murray
    - Shuffle Service with Scheduler Logic (GH#6007) Matthew
      Rocklin
    * Enhancements
    - Follow-up on removing report and safe from Worker.close
      (GH#6423) Gabe Joseph
    - Server close faster (GH#6415) Florian Jetter
    - Disable HTTP API by default (GH#6420) Jacob Tomlinson
    - Remove report and safe from Worker.close (GH#6363) Florian
      Jetter
    - Allow deserialized plugins in register_scheduler_plugin
      (GH#6401) Matthew Rocklin
    - WorkerState are different for different addresses (GH#6398)
      Florian Jetter
    - Do not filter tasks before gathering data (GH#6371)
      crusaderky
    - Remove worker reconnect (GH#6361) Gabe Joseph
    - Add SchedulerPlugin.log_event handler (GH#6381) Matthew
      Rocklin
    - Ensure occupancy tracking works as expected for long running
      tasks (GH#6351) Florian Jetter
    - stimulus_id for all Instructions (GH#6347) crusaderky
    - Refactor missing-data command (GH#6332) crusaderky
    - Add idempotent to register_scheduler_plugin client (GH#6328)
      Alex Ford
    - Add option to specify a scheduler address for workers to use
      (GH#5944) Enric Tejedor
    * Bug Fixes
    - Remove stray breakpoint (GH#6417) Thomas Grainger
    - Fix API JSON MIME type (GH#6397) Jacob Tomlinson
    - Remove wrong assert in handle compute (GH#6370) Florian
      Jetter
    - Ensure multiple clients can cancel their key without
      interference (GH#6016) Florian Jetter
    - Fix Nanny shutdown assertion (GH#6357) Gabe Joseph
    - Fix fail_hard for sync functions (GH#6269) Gabe Joseph
    - Prevent infinite transition loops; more aggressive
      validate_state() (GH#6318) crusaderky
    - Ensure cleanup of many GBs of spilled data on terminate
      (GH#6280) crusaderky
    - Fix WORKER_ANY_RUNNING regression (GH#6297) Florian Jetter
    - Race conditions from fetch to compute while AMM requests
      replica (GH#6248) Florian Jetter
    - Ensure resumed tasks are not accidentally forgotten (GH#6217)
      Florian Jetter
    - Do not allow closing workers to be awaited again (GH#5910)
      Florian Jetter
    * Deprecations
    - Move wait_for_signals to private module and deprecate
      distributed.cli.utils (GH#6367) Hendrik Makait
  - Release 2022.05.0
    * Highlights
    - This is a bugfix release for this issue.
    * Enhancements
    - Handle writeable in buffer_callback (GH#6238) jakirkham
    - Use .data with NumPy array allocation (GH#6242) jakirkham
    * Bug Fixes
    - Close executor in event loop if interpreter is closing
      (GH#6256) Matthew Rocklin
  - Release 2022.04.2
    * Enhancements
    - Unblock event loop while waiting for ThreadpoolExecutor to
      shut down (GH#6091) Florian Jetter
    - RetireWorker policy is done if removed (GH#6234) Gabe Joseph
    - Pause to disable dependency gathering (GH#6195) crusaderky
    - Add EOFError to nanny multiprocessing.queue except list
      (GH#6213) Matthew Rocklin
    - Re-interpret error in lost worker scenario (GH#6193) Matthew
      Rocklin
    - Add Stimulus IDs to Scheduler (GH#6161) Florian Jetter
    - Set a five minute TTL for Dask workers (GH#6200) Matthew
      Rocklin
    - Add distributed.metrics.monotonic (GH#6181) crusaderky
    - Send worker validation errors to scheduler and err on test
      completion (GH#6192) Matthew Rocklin
    - Redesign worker exponential backoff on busy-gather (GH#6173)
      crusaderky
    - Log all invalid worker transitions to scheduler (GH#6134)
      Matthew Rocklin
    - Make Graph dashboard plot have invisible axes (GH#6149)
      Matthew Rocklin
    - Remove Nanny auto_restart state (GH#6138) Matthew Rocklin
    * Bug Fixes
    - Ensure scheduler events do not hold on to TaskState objects
      (GH#6226) Florian Jetter
    - Allow pausing and choke event loop while spilling (GH#6189)
      crusaderky
    - Do not use UUID in stealing (GH#6179) Florian Jetter
    - Handle int worker names in info page (GH#6158) Brett Naul
    - Fix psutil dio counters none (GH#6093) ungarj
    - Join Nanny watch thread (GH#6146) Matthew Rocklin
    - Improve logging when closing workers (GH#6129) Matthew
      Rocklin
    - Avoid stack overflow in profiling (GH#6141) Matthew Rocklin
    - Clean up SSHCluster if failure to start (GH#6130) Matthew
      Rocklin
    * Deprecations
    - Deprecate rpc synchronous context manager (GH#6171) Thomas
      Grainger
  - Release 2022.04.1
    * New Features
    - Add KillWorker Plugin (GH#6126) Matthew Rocklin
    * Enhancements
    - Sort worker list in info pages (GH#6135) Matthew Rocklin
    - Add back Worker.transition_fetch_missing (GH#6112) Matthew
      Rocklin
    - Log state machine events (GH#6092) crusaderky
    - Migrate ensure_executing transitions to new WorkerState event
      mechanism - part 1 (GH#6003) crusaderky
    - Migrate ensure_executing transitions to new WorkerState event
      mechanism - part 2 (GH#6062) crusaderky
    - Annotate worker transitions to error (GH#6012) crusaderky
    * Bug Fixes
    - Avoid transitioning from memory/released to missing in worker
      (GH#6123) Matthew Rocklin
    - Don’t try to reconnect client on interpreter shutdown
      (GH#6120) Matthew Rocklin
    - Wrap UCX init warnings in importable functions (GH#6121)
      Charles Blackmon-Luca
    - Cancel asyncio tasks on worker close (GH#6098) crusaderky
    - Avoid port collisions when defining port ranges (GH#6054)
      crusaderky
  - Release 2022.04.0
    * This is the first release with support for Python 3.10
    * New Features
    - Add Python 3.10 support (GH#5952) Thomas Grainger
    - New cluster dump utilities (GH#5920) Simon Perkins
    - New ClusterDump SchedulerPlugin for dumping cluster state on
      close (GH#5983) Simon Perkins
    - Track Event Loop intervals in dashboard plot (GH#5964)
      Matthew Rocklin
    - ToPickle - Unpickle on the Scheduler (GH#5728) Mads R. B.
      Kristensen
    * Enhancements
    - Retry on transient error codes in preload (GH#5982) Matthew
      Rocklin
    - More idiomatic mypy configuration (GH#6022) crusaderky
    - Name extensions and enable extension heartbeats (GH#5957)
      Matthew Rocklin
    - Better error message on misspelled executor annotation
      (GH#6009) crusaderky
    - Clarify that SchedulerPlugin must be subclassed (GH#6008)
      crusaderky
    - Remove duplication from stealing (GH#5787) Duncan McGregor
    - Remove cache in iscoroutinefunction to avoid holding on to
      refs (GH#5985) Florian Jetter
    - Add title to individual plots (GH#5967) Matthew Rocklin
    - Specify average in timeseries titles (GH#5974) Matthew
      Rocklin
    * Bug Fixes
    - Do not catch CancelledError in CommPool (GH#6005) Florian
      Jetter
    * Deprecations
    - Remove distributed._ipython_utils and dependents (GH#6036)
      Thomas Grainger
    - Remove support for PyPy (GH#6029) James Bourbeau
    - Drop runtime dependency to setuptools (GH#6017) crusaderky
    - Remove heartbeats from events (GH#5989) Matthew Rocklin
  - Cythonization has been abandoned upstream: Back to noarch.
  - Add distributed-ignore-offline.patch
  - Drop distributed-pr5952-py310.patch
* Fri Mar 25 2022 Ben Greiner <code@bnavigator.de>
  - Update to 2022.03.0
    * Support dumping cluster state to URL (GH#5863) Gabe Joseph
    * Prevent data duplication on unspill (GH#5936) crusaderky
    * Encapsulate spill buffer and memory_monitor (GH#5904)
      crusaderky
    * Drop pkg_resources in favour of importlib.metadata (GH#5923)
      Thomas Grainger
    * Worker State Machine refactor: redesign TaskState and scheduler
      messages (GH#5922) crusaderky
    * Tidying of OpenSSL 1.0.2/Python 3.9 (and earlier) handling
      (GH#5854) jakirkham
    * zict type annotations (GH#5905) crusaderky
    * Add key to compute failed message (GH#5928) Florian Jetter
    * Change default log format to include timestamp (GH#5897)
      Florian Jetter
    * Improve type annotations in worker.py (GH#5814) crusaderky
  - Add distributed-pr5952-py310.patch -- gh#dask/distributed#5952
  - Add distributed-ignore-thread-leaks.patch
  - Make the distributed/dask update sync requirement even more
    obvious.
* Tue Mar 08 2022 Matej Cepl <mcepl@suse.com>
  - Update to 2022.02.1:
      Add the ability for Client to run preload code
      Optionally use NumPy to allocate buffers
      Add git hash to distributed-impl version
      Immediately raise exception when trying to connect to a closed cluster
      Lazily get dask version information
      Remove the requirements to add comm to every handler
      Raise on unclosed comms in check_instances
      Constrained spill
      Remove redundant str() conversions
      Cluster dump now excludes run_spec by default
      Dump more objects with dump_cluster_state
      Do not connect to any sockets on import
      Avoid deadlock when two tasks are concurrently waiting for an unresolved ActorFuture
      Drop Python 3.7
      Remove support for UCX < 1.11.1
      Document and test spill->target hysteresis cycle
      Fix flaky test_remove_replicas_while_computing
      Fix time based test_assert_worker_story_malformed_story parameterize
      Remove xfail from test_no_unnecessary_imports_on_worker
      Start building pre-releases with cythonized scheduler
      Do not mark tests xfail if they don't come up in time
      Use gen_cluster where possible in test_dask_worker.py
      Generate junit report when pytest-timeout kills pytest
      Decrease timeout-minutes for GHA jobs
      Bump pre-release version to be greater than stable releases
      Do not run schedule jobs on forks
      Remove pillow<9 pin in CI
      Show scheduled test runs in report
      Add obvious exclusions with pragma statement
      Add coverage exclusions for cli files
      Add pragma statements
      Remove pragma: no cover from distributed.cli.dask_ssh
      Add pragma - worker.py, client.py, stealing.py
      Relax distributed / dask-core dependencies for pre-releases
      Remove test_ucx_config_w_env_var flaky condition
  - Update to 2022.02.0:
      Update client.scheduler_info in wait_for_workers
      Increase robustness to TimeoutError during connect
      Respect KeyboardInterrupt in sync
      Add workflow / recipe to generate Dask/distributed pre-releases
      Review Scheduler / Worker display repr
      AMM: Graceful Worker Retirement
      AMM: tentatively stabilize flaky tests around worker pause
      AMM: speed up and stabilize test_memory
      Defer pandas import on worker in P2P shuffle
      Fix for distributed.worker.memory.target=False and spill=0.7
      Transition flight to missing if no who_has
      Remove deprecated ncores
      Deprecate registering plugins by class
      Deprecate --nprocs option for dask-worker CLI
      Fix imbalanced backticks
      xfail test_worker_reconnects_mid_compute
      Fix linting CI build
      Update pre-commit versions
      Reactivate pytest_resourceleaks
      Set test assumption for test_client_timeout
      Remove client timeout from test_ucx_config_w_env_var
      Remove test_failed_worker_without_warning
      Fix longitudinal report
      Fix flaky test_robust_to_bad_sizeof_estimates
      Revert "Pin coverage to 6.2
      Trigger test runs periodically to increases failure statistics
      More fault tolerant test report
      Pin pillow<9 to work around torch incompatability
      Overhaul check_process_leak
      Fix flaky test_exit_callback test
      Generate tests summary
      Upload different architectured pre-releases separately
      Ignore non-test directories
      Bump gpuCI PYTHON_VER to 3.9
      Regression: threads noted down before they start
  - Remove upstreamed patches:
    - 5709-avoid-deadlock-ActorFuture.patch
* Mon Feb 14 2022 Matej Cepl <mcepl@suse.com>
  - Add 5709-avoid-deadlock-ActorFuture.patch to avoid deadlock in
    ActorFuture (gh#dask/distributed#5709).
* Sat Jan 29 2022 Ben Greiner <code@bnavigator.de>
  - Update to version 2022.1.1
    * P2P shuffle skeleton
    * Fix <Task pending name='...' coro=<Client._handle_report()>
    * Add distributed.client.security-loader config
    * Avoid Client._handle_report cancelling itself on Client._close
    * Paused workers shouldn't steal tasks
    * Add option for timestamps from output of Node.get_logs
    * Don't use time.time() or IOLoop.time()
    * Raise plugin exceptions on Worker.start()
    * Fixing docstrings
    * Fix typo in Client.run docstring
    * Update client.py docstrings
    * Skip shuffle tests if pandas / dask.dataframe not installed
    * Improve test coverage
    * Test report improvements
    * P2P shuffle: ignore row order in tests
    * Fix flaky test_no_reconnect[--no-nanny]
    * Pin coverage to 6.2
    * Check for new name of timeouts artifact and be more fault
      tolerant
    * Revisit rebalance unit tests
    * Update comment in rearrange_by_column_p2p
    * Update gpuCI RAPIDS_VER to 22.04
    * Fix groupby test after meta requirements got stricter in Dask
      PR#8563
    * Fix flaky test_close_gracefully and test_lifetime
    * Fix flaky test_workspace_concurrency
    * Fix flaky test_shuffle_extension.py::test_get_partition
    * Fix flaky test_dump_cluster_unresponsive_remote_worker
    * Dump cluster state on all test failures
    * Update license format
    * Fix flaky test_drop_with_paused_workers_with_running_tasks_3_4
    * Do not raise an exception if the GitHub token cannot be found
* Tue Jan 25 2022 Ben Greiner <code@bnavigator.de>
  - Revert distributed-fix-python310.patch addition
  - The patch is gh#dask/distributed#5353
    * Is not merged upstream
    * Does not fix the issues
    * Breaks other flavors
* Tue Jan 25 2022 Dirk Müller <dmueller@suse.com>
  - add distributed-fix-python310.patch to fix build with python 3.10
* Thu Jan 20 2022 Ben Greiner <code@bnavigator.de>
  - Update to version 2022.1.0
    * full changelog at
      https://distributed.readthedocs.io/en/latest/changelog.html
    * New Features:
    - Task group stacked area chart (GH#5320) Ian Rose
    - Support configuring TLS min/max version (GH#5594) Jim
      Crist-Harif
    - Use asyncio for TCP/TLS comms (GH#5450) Jim Crist-Harif
  - Enhancement release notes for 2021.12.0
    * Support pytest fixures and parametrize with gen_test (GH#5532)
      Fábio Rosado
    * Allow idempotent scheduler plugins to be registered via the RPC
      (GH#5545) Jacob Tomlinson
    * AMM logging (GH#5530) crusaderky
    * Raise error if asyncssh isn’t installed when using SSHCluster
      (GH#5535) Fábio Rosado
    * Allow None in UCX configuration schema (GH#5534) Fábio Rosado
    * Add distributed.comm.ucx.create-cuda-context config (GH#5526)
      Peter Andreas Entschev
  - Security release note for 2021.10.0:
    * This release fixed a potential security vulnerability relating
      to single-machine Dask clusters. Clusters started with
      dask.distributed.LocalCluster or dask.distributed.Client()
      (which defaults to using LocalCluster) would mistakenly
      configure their respective Dask workers to listen on external
      interfaces (typically with a randomly selected high port)
      rather than only on localhost. A Dask cluster created using
      this method AND running on a machine that has these ports
      exposed could be used by a sophisticated attacker to enable
      remote code execution. Users running on machines with standard
      firewalls in place should not be affected. This vulnerability
      is documented in CVE-2021-42343, and is fixed in this release
      (GH#5427). Thanks to Jean-Pierre van Riel for discovering and
      reporting the issue.
  - Skip python310: Not supported yet -- gh#dask/distributed#5350
* Wed Sep 22 2021 Ben Greiner <code@bnavigator.de>
  - Update to version 2021.9.1
    * Revert AMM ReduceReplicas and parallel AMMs updates
    * Run multiple AMMs in parallel
    * AMM ReduceReplicas to iterate only on replicated tasks
    * Add type annotations to various functions within
      distributed.worker
    * Mark test_ucx_config_w_env_var flaky on UCX < 1.11
    * Warn if CUDA context is created on incorrect device in UCX
    * Remove redundant timeouts from test_client
    * Allow Client to subscribe to events // Remote printing and
      warning
    * Test pickle protocols 4 & 5
    * Fix-up test_pickle_empty
    * Increase timeout for
      test_worker_reconnects_mid_compute_multiple_states_on_scheduler
    * Add synced dict between cluster and scheduler to store cluster
      info
    * Update test_sub_submit_priority
    * Revert "Add test setup fixture
    * Fix flaky test_worker_reconnects_mid_compute
    * Use gen_test in test_adaptive
    * Increase worker.suspicious_counter threshold
    * Active Memory Manager framework + discard excess replicas
    * Add test setup fixture
  - Release version 2021.9.0
    * Fix add_plugin warnings
    * Add list around iterator in handle_missing_dep
    * Jupyter-client 7 compatiblity
    * Replace atop with blockwise
    * Add pytest color to CI
    * Fix test_map and others
    * Use name argument with Scheduler.remove_plugin calls
    * Downgrade to jupyter_client 6
    * Migrate Security HTML repr to Jinja2
    * Migrate ProcessInterface HTML repr to Jinja2
    * Add support for diskless machines to system monitor
    * Avoid during-iteration scheduler plugin changes
    * Remove GroupProgress scheduler plugin
    * Properly check for ipv6 availability
    * Improved IPv6 dask-worker support
    * Overwrite worker plugins
    * Refactor scheduler plugins; store in a dictionary
    * Fix "then" -> "than" typo in docs
    * Fix typo (remove extra verb "creates") in docs
    * Fix "fractiom" -> "fraction" typo in docstring
    * Fix "schedulers" -> "scheduler" typo in docs
    * Use non-histogram plots up to 100 workers
  - Release 2021.08.1
    * Rename plots to fit in the labextension
    * Log messages for CommClosedError now includes information
      about remote address
    * Add target='_blank' for redirects of dashboard link
    * Update computation code retrieval logic
    * Minor polish on cfexecutor
    * Use development version of dask in gpuCI build
    * Use upstream dask.widgets
    * Fix flaky test_worker_reconnects_mid_compute
    * Update WorkerPlugin docstring about usage of TaskState objects
    * Worker Network Timeseries
    * Add HTML Repr for ProcessInterface class and all its subclasses
    * Fix an issue where a reconnecting worker could cause an
      invalid transition
    * Minor fixes for cfexecutor
    * Add HTML Repr for Security class
    * Fix performance report sizing issue
    * Drop RMM compatibility code from RAPIDS < 0.11
  - Release 2021.08.0
    * Include addresses in closed comm repr
    * Test nanny.environ precedence
    * Migrating HTML reprs to jinja2
    * Fix test_process_executor_kills_process flakyness
    * Remove urllib3 as a dependency downloading preloads
    * Download preload urls in the Preload constructor
    * Avoid recursion error in profile.merge
    * Add text exceptions to the Scheduler
    * Use kwarg for Theme filename
    * Add a .git-ignore-revs file
    * Replace not not with bool()
    * Resolve deadlock cause by transition error after fetching
      dependency
    * Set z-index of data-table lower
    * Add no-worker - memory transition to scheduler
    * Deprecate worker plugin overwrite policy
    * Fix flaky tests in CI
    * Instructions for jemalloc with brew on macOS
    * Bump RAPIDS_VER to 21.10
    * Tweak verbiage around async functions
    * Use Python 3 super() calls
    * Support asynchronous tasks
    * Rename total comm bytes and provide doc string
    * Add GPU executor if GPU is present
    * Fix RMM and UCX tests
    * Remove excessive timeout of
      test_steal_during_task_deserialization
    * Add gpuCI build script
    * Demote Worker.ensure_computing to function
* Sun Aug 08 2021 Ben Greiner <code@bnavigator.de>
  - Update to version 2021.7.2
    * Fix a deadlock connected to task stealing and task
      deserialization
    * Include maximum shard size in second to_frames method
    * Minor dashboard style updates
    * Cap maximum shard size at the size of an integer
    * Document automatic MALLOC_TRIM_THRESHOLD_ environment variable
    * Mark ucx-py tests for GPU
    * Update individual profile plot sizing
    * Handle NVMLError_Unknown in NVML diagnostics
    * Unit tests to use a random port for the dashboard
    * Ensure worker reconnect registers existing tasks properly
    * Halve CI runtime!
    * Add NannyPlugins
    * Add WorkerNetworkBandwidth chart to dashboard
    * Set nanny environment variables in config
    * Read smaller frames to workaround OpenSSL bug
    * Move UCX/RMM config variables to Distributed namespace
    * Allow ws(s) messages greater than 10Mb
    * Short-circuit root-ish check for many deps
  -Release 2021.07.1
    * Remove experimental feature warning from actors docs
    * Keep dependents in worker dependency if TS is still known
    * Add Scheduler.set_restrictions
    * Make Actor futures awaitable and work with as_completed
    * Simplify test_secede_balances
    * Computation class
    * Some light dashboard cleanup
    * Don't package tests
    * Add pytest marker for GPU tests
    * Actor: don't hold key references on workers
    * Collapse nav to hamburger sooner
    * Verify that actors survive pickling
    * Reenable UCX-Py tests that used to segfault
    * Better support ProcessPoolExecutors
    * Simplify test_worker_heartbeat_after_cancel
    * Avoid property validation in Bokeh
    * Reduce default websocket frame size and make configurable
    * Disable pytest-timeout SIGALARM on MacOS
    * rebalance() resilience to computations
    * Improve CI stability
    * Ensure heartbeats after cancelation do not raise KeyError s
    * Add more useful exception message on TLS cert mismatch
    * Add bokeh mode parameter to performance reports
  - Use the GitHub tarball because the PyPI sdist does to provide the
    tests anymore
    * Remove extra conftest.py source
  - Drop distributed-pr5022-improve_ci.patch merged upstream
* Fri Jul 16 2021 Ben Greiner <code@bnavigator.de>
  - Update to version 2021.7.0
    * Fix Nbytes jitter - less expensive
    * Use native GH actions cancel feature
    * Don't require workers to report to scheduler if scheduler
      shuts down
    * Add pandas to the list of checked packages for client.
      get_versions()
    * Move worker preload before scheduler address is set
    * Fix flaky test_oversubscribing_leases
    * Update scheduling policy docs for #4967
    * Add echo handler to Server class
    * Also include pngs when bundling package
    * Remove duplicated dashboard panes
    * Fix worker memory dashboard flickering
    * Tabs on bottom left corner on dashboard
    * Rename nbytes widgets
    * Co-assign root-ish tasks
    * OSError tweaks
    * Update imports to cudf.testing._utils
    * Ensure shuffle split default durations uses proper prefix
    * Follow up pyupgrade formatting
    * Rename plot dropdown
    * Pyupgrade
    * Misc Sphinx tweaks
    * No longer hold dependencies of erred tasks in memory
    * Add maximum shard size to config
    * Ensure shuffle split operations are blacklisted from work
      stealing
    * Add dropdown menu to access individual plots
    * Edited the path to scheduler.py
    * Task Group Graph Visualization
    * Remove more internal references to deprecated utilities
    * Restructure nbytes hover
    * Except more errors in pynvml.nvmlInit()
    * Add occupancy as individual plot
    * Deprecate utilities which have moved to dask
    * Ensure connectionpool does not leave comms if closed mid
      connect
    * Add support for registering scheduler plugins from Client
    * Stealing dashboard fixes
    * Allow requirements verification to be ignored when loading
      backends from entrypoints
    * Add Log and Logs to API docs
    * Support fixtures and pytest.mark.parametrize with gen_cluster
  - Release 2021.06.2
    * Revert refactor to utils.Log[s] and Cluster.get_logs
    * Use deprecation utility from Dask
    * Add transition counter to Scheduler
    * Remove nbytes_in_memory
  - Release 2021.06.1
    * Fix deadlock in handle_missing_dep if additional replicas are
      available
    * Add configuration to enable/disable NVML diagnostics
    * Add scheduler log tab to performance reports
    * Add HTML repr to scheduler_info and incorporate into client
      and cluster reprs
    * Fix error state typo
    * Allow actor exceptions to propagate
    * Remove importing apply from dask.compatibility
    * Use more informative default name for WorkerPlugin s
    * Removed unused utility functions
    * Locally rerun successfully completed futures
    * Forget erred tasks and fix deadlocks on worker
    * Handle HTTPClientError in websocket connector
    * Update dask_cuda usage in SSHCluster docstring
    * Remove tests for process_time and thread_time
    * Flake8 config cleanup
    * Don't strip scheduler protocol when determining host
    * Add more documentation on memory management
    * Add range_query tests to NVML test suite
    * No longer cancel result future in async process when using
      timeouts
  - Release 2021.06.0
    * Multiple worker executors
    * Ensure PyNVML works correctly when installed with no GPUs
    * Show more in test summary
    * Move SystemMonitor s GPU initialization back to constructor
    * Mark test_server_comms_mark_active_handlers with pytest.mark.asyncio
    * Who has has what html reprs v2
    * O(1) rebalance
    * Ensure repr and eq for cluster always works
  - Release 2021.05.1
    * Drop usage of WhoHas & WhatHas from Client
    * Ensure adaptive scaling is properly awaited and closed
    * Fix WhoHas/ HasWhat async usage
    * Add HTML reprs for Client.who_has and Client.has_what
    * Prevent accidentally starting multiple Worker s in the same
      process
    * Add system tab to performance reports
    * Let servers close faster if there are no active handlers
    * Fix UCX scrub config logging
    * Ensure worker clients are closed
    * Fix warning for attribute error when deleting a client
    * Ensure exceptions are raised if workers are incorrectly started
    * Update handling of UCX exceptions on endpoint closing
    * Ensure busy workloads properly look up who_has
    * Check distributed.scheduler.pickle in Scheduler.run_function
    * Add performance_report to API docs
    * Use dict _workers_dv in unordered use cases
    * Bump pre-commit hook versions
    * Do not mindlessly spawn workers when no memory limit is set
    * test_memory to use gen_cluster
    * Increase timeout of gen_test to 30s
  - Work on the very flaky testsuite:
    * Add missing conftest.py not packaged on PyPI
    * Add distributed-pr5022-improve_ci.patch in the hope for better
      stability -- gh#dask/distributed#5022
    * Do not use pytest-xdist
  - Add Cython as runtime dep because the scheduler checks the
    presence
* Tue May 18 2021 Ben Greiner <code@bnavigator.de>
  - update to version 2021.5.0
    * Merge global annotations on the client (GH#4691) Mads R. B.
      Kristensen
    * Add support for click 8 (GH#4810) James Bourbeau
    * Add HTML reprs to some scheduler classes (GH#4795) James
      Bourbeau
    * Use JupyterLab theme variables (GH#4796) Ian Rose
    * Allow the dashboard to run on multiple ports (GH#4786) Jacob
      Tomlinson
    * Remove release_dep from WorkerPlugin API (GH#4791) James
      Bourbeau
    * Support for UCX 1.10+ (GH#4787) Peter Andreas Entschev
    * Reduce complexity of test_gather_allow_worker_reconnect
      (GH#4739) Florian Jetter
    * Fix doctests in utils.py (GH#4785) Jacob Tomlinson
    * Ensure deps are actually logged in worker (GH#4753) Florian
      Jetter
    * Add stacklevel keyword into performance_report() to allow for
      selecting calling code to be displayed (GH#4777) Nathan
      Danielsen
    * Unregister worker plugin (GH#4748) Naty Clementi
    * Fixes some pickling issues in the Cythonized Scheduler
      (GH#4768) jakirkham
    * Improve graceful shutdown if nanny is involved (GH#4725)
      Florian Jetter
    * Update cythonization in CI (GH#4764) James Bourbeau
    * Use contextlib.nullcontext (GH#4763) James Bourbeau
    * Cython fixes for MemoryState (GH#4761) jakirkham
    * Fix errors in check_thread_leak (GH#4747) James Bourbeau
    * Handle missing key case in report_on_key (GH#4755) jakirkham
    * Drop temporary set variables s (GH#4758) jakirkham
  - skip even more flaky tests
  - stress that upstream wants to use dask with a matching version
    number
* Mon May 03 2021 Arun Persaud <arun@gmx.de>
  - update to version 2021.4.1:
    * Avoid active_threads changing size during iteration (GH#4729)
      James Bourbeau
    * Fix UnboundLocalError in AdaptiveCore.adapt() (GH#4731) Anderson
      Banihirwe
    * Minor formatting updates for HTTP endpoints doc (GH#4736) James
      Bourbeau
    * Unit test for metrics["memory"]=None (GH#4727) crusaderky
    * Enable configuration of prometheus metrics namespace (GH#4722)
      Jacob Tomlinson
    * Reintroduce weight function (GH#4723) James Bourbeau
    * Add ready->memory to transitions in worker (GH#4728) Gil Forsyth
    * Fix regressions in GH#4651 (GH#4719) crusaderky
    * Add descriptions for UCX config options (GH#4683) Charles
      Blackmon-Luca
    * Split RAM measure into dask keys/other old/other new (GH#4651)
      crusaderky
    * Fix DeprecationWarning on Python 3.9 (GH#4717) George Sakkis
    * ipython causes test_profile_nested_sizeof crash on windows
      (GH#4713) crusaderky
    * Add iterate_collection argument to serialize (GH#4641) Richard J
      Zamora
    * When closing Server, close all listeners (GH#4704) Florian Jetter
    * Fix timeout in client.restart (GH#4690) Matteo De Wint
    * Avoid repeatedly using the same worker on first task with quiet
      cluster (GH#4638) Doug Davis
    * Grab func for finish case only if used (GH#4702) jakirkham
    * Remove hostname check in test_dashboard (GH#4706) James Bourbeau
    * Faster tests_semaphore::test_worker_dies (GH#4703) Florian Jetter
    * Clean up test_dashboard (GH#4700) crusaderky
    * Add timing information to TaskGroup (GH#4671) Matthew Rocklin
    * Remove WSSConnector TLS presence check (GH#4695) Marcos Moyano
    * Fix typo and remove unused time.time import (GH#4689) Hristo
      Georgiev
    * Don’t initialize CUDA context in monitor (GH#4688) Charles
      Blackmon-Luca
    * Add support for extra conn args for HTTP protocols (GH#4682)
      Marcos Moyano
    * Adjust timings in test_threadpoolworkers (GH#4681) Florian Jetter
    * Add GPU metrics to SystemMonitor (GH#4661) Charles Blackmon-Luca
    * Removing dumps_msgpack() and loads_msgpack() (GH#4677) Mads
      R. B. Kristensen
    * Expose worker SystemMonitor s to scheduler via RPC (GH#4657)
      Charles Blackmon-Luca
* Sun Apr 04 2021 Arun Persaud <arun@gmx.de>
  - update to version 2021.4.0:
    * Fix un-merged frames (GH#4666) Matthew Rocklin
    * Add informative error message to install uvloop (GH#4664) Matthew
      Rocklin
    * Remove incorrect comment regarding default LocalCluster creation
      (GH#4660) cameron16
    * Treat empty/missing writeable as a no-op (GH#4659) jakirkham
    * Avoid list mutation in pickle_loads (GH#4653) Matthew Rocklin
    * Ignore OSError exception when scaling down (GH#4633) Gerald
    * Add isort to pre-commit hooks, package resorting (GH#4647) Charles
      Blackmon-Luca
    * Use powers-of-two when displaying RAM (GH#4649) crusaderky
    * Support Websocket communication protocols (GH#4396) Marcos Moyano
    * scheduler.py / worker.py code cleanup (GH#4626) crusaderky
    * Update out-of-date references to config.yaml (GH#4643) Hristo
      Georgiev
    * Suppress OSError on SpecCluster shutdown (GH#4567) Jacob Tomlinson
    * Replace conda with mamba (GH#4585) crusaderky
    * Expand documentation on pure functions (GH#4644) James Lamb
* Fri Apr 02 2021 Ben Greiner <code@bnavigator.de>
  - Disable Cythonization for 32-bit gh#dask/dask#7489
* Tue Mar 30 2021 Ben Greiner <code@bnavigator.de>
  - Update to 2021.3.1
    * Add standalone dashboard page for GPU usage (GH#4556) Jacob
      Tomlinson
    * Handle stream is None case in TCP comm finalizer (GH#4631)
      James Bourbeau
    * Include LIST_PICKLE in NumPy array serialization (GH#4632)
      James Bourbeau
    * Rename annotation plugin in test_highlevelgraph.py (GH#4618)
      James Bourbeau
    * UCX use nbytes instead of len (GH#4621) Mads R. B. Kristensen
    * Skip NumPy and pandas tests if not importable (GH#4563) Ben
      Greiner
    * Remove utils.shutting_down in favor of sys.is_finalizing
      (GH#4624) James Bourbeau
    * Handle async clients when closing (GH#4623) Matthew Rocklin
    * Drop log from remove_key_from_stealable (GH#4609) jakirkham
    * Introduce events log length config option (GH#4615) Fabian
      Gebhart
    * Upstream config serialization and inheritance (GH#4372) Jacob
      Tomlinson
    * Add check to scheduler creation in SpecCluster (GH#4605) Jacob
      Tomlinson
    * Make length of events deque configurable (GH#4604) Fabian
      Gebhart
    * Add explicit fetch state to worker TaskState (GH#4470) Gil
      Forsyth
    * Update develop.rst (GH#4603) Florian Jetter
    * pickle_loads(): Handle empty memoryview (GH#4595) Mads R. B.
      Kristensen
    * Switch documentation builds for PRs to readthedocs (GH#4599)
      James Bourbeau
    * Track frame sizes along with frames (GH#4593) jakirkham
    * Add support for a list of keys when using batch_size in client.
      map (GH#4592) Sultan Orazbayev
    * If SpecCluster fails to start attempt to gracefully close out
      again (GH#4590) Jacob Tomlinson
    * Multi-lock extension (GH#4503) Mads R. B. Kristensen
    * Update PipInstall plugin command (GH#4584) James Bourbeau
    * IPython magics: remove deprecated ioloop workarounds (GH#4530)
      Min RK
    * Add GitHub actions workflow to cancel duplicate builds
      (GH#4581) James Bourbeau
    * Remove outdated macOS build badge from README (GH#4576) James
      Bourbeau
    * Dask master -> main (GH#4569) Julia Signell
    * Drop support for Python 3.6 (GH#4390) James Bourbeau
    * Add docstring for dashboard_link property (GH#4572) Doug Davis
    * Change default branch from master to main (GH#4495) Julia
      Signell
    * Msgpack handles extract serialize (GH#4531) Mads R. B.
      Kristensen
  - Drop importorskip-numpy-tests.patch merged upstream
  - Skip python36 build. Upstream dropped support for Python < 3.7
* Sun Mar 07 2021 Ben Greiner <code@bnavigator.de>
  - Update to 2021.3.0
    * This is the first release with support for Python 3.9 and the
      last release with support for Python 3.6
    * tcp.write(): cast memoryview to byte itemsize (#4555)
      Mads R. B. Kristensen
    * Refcount the thread_state.asynchronous flag (#4557) Mads
      R. B. Kristensen
    * Python 3.9 (#4460) crusaderky
    * Better bokeh defaults for dashboard (#4554) Benjamin
      Zaitlen
    * Expose system monitor dashboard as individual plot for lab
      extension (#4540) Jacob Tomlinson
    * Pass on original temp dir from nanny to worker (#4549)
      Martin Durant
    * Serialize and split (#4541) Mads R. B. Kristensen
    * Use the new HLG pack/unpack API in Dask (#4489) Mads R.
      B. Kristensen
    * Handle annotations for culled tasks (#4544) Tom
      Augspurger
    * Make sphinx autosummary and autoclass consistent (#4367)
      Casey Clements
    * Move _transition* to SchedulerState (#4545) jakirkham
    * Migrate from travis to GitHub actions (#4504) crusaderky
    * Move new_task to SchedulerState (#4527) jakirkham
    * Batch more Scheduler sends (#4526) jakirkham
    * transition_memory_released and get_nbytes() optimizations
      (#4516) jakirkham
    * Pin black pre-commit (#4533) James Bourbeau
    * Read & write all frames in one pass (#4506) jakirkham
    * Skip stream.write call for empty frames (#4507) jakirkham
    * Prepend frame metadata header (#4505) jakirkham
    * transition_processing_memory optimizations, etc. (#4487)
      jakirkham
    * Attempt to get client from worker in Queue and Variable
      (#4490) James Bourbeau
    * Use main branch for zict (#4499) jakirkham
    * Use a callback to close TCP Comms, rather than check every
      time (#4453) Matthew Rocklin
* Sun Feb 21 2021 Ben Greiner <code@bnavigator.de>
  - Another package not available in python36: bokeh
  - Now that the (previously still lingering) python36-numpy package
    really has been purged from the dependency tree, it became
    apparent that it is assumed to be present in some tests.
    * gh#dask/distributed#4529
    * Add importorskip-numpy-tests.patch
* Mon Feb 15 2021 Ben Greiner <code@bnavigator.de>
  - Unskip python36 build. Just skip installing optional
    python36-sparse for the tests.
* Fri Feb 12 2021 Dirk Müller <dmueller@suse.com>
  - skip python 36 build
* Mon Feb 08 2021 Ben Greiner <code@bnavigator.de>
  - Update to version 2021.2.0
    * Bump minimum Dask to 2021.02.0 (GH#4486) James Bourbeau
    * Update TaskState documentation about dependents attribute
      (GH#4440) Florian Jetter
    * DOC: Autoreformat all functions docstrings (GH#4475) Matthias
      Bussonnier
    * Use cached version of is_coroutine_function in stream handling
      to (GH#4481) Ian Rose
    * Optimize transitions (GH#4451) jakirkham
    * Create PULL_REQUEST_TEMPLATE.md (GH#4476) Ray Bell
    * DOC: typo, directives ends with 2 colons :: (GH#4472) Matthias
      Bussonnier
    * DOC: Proper numpydoc syntax for distributed/protocol/*.py
      (GH#4473) Matthias Bussonnier
    * Update pytest.skip usage in test_server_listen (GH#4467) James
      Bourbeau
    * Unify annotations (GH#4406) Ian Rose
    * Added worker resources from config (GH#4456) Tom Augspurger
    * Fix var name in worker validation func (GH#4457) Gil Forsyth
    * Refactor task_groups & task_prefixes (GH#4452) jakirkham
    * Use parent._tasks in heartbeat (GH#4450) jakirkham
    * Refactor SchedulerState from Scheduler (GH#4365) jakirkham
  - Do not pin to tornado < 6.1. The referenced pull request reverted
    that change and we have asyncio test problems anyway.
  - Use new python_module boolean deps capability
* Thu Jan 28 2021 Ben Greiner <code@bnavigator.de>
  - Update to version 2021.1.1
    * Make system monitor interval configurable (GH#4447) Matthew
      Rocklin
    * Add uvloop config value (GH#4448) Matthew Rocklin
    * Additional optimizations to stealing (GH#4445) jakirkham
    * Give clusters names (GH#4426) Jacob Tomlinson
    * Use worker comm pool in Semaphore (GH#4195) Florian Jetter
    * Set runspec on all new tasks to avoid deadlocks (GH#4432)
      Florian Jetter
    * Support TaskState objects in story methods (GH#4434) Matthew
      Rocklin
    * Support missing event loop in Client.asynchronous (GH#4436)
      Matthew Rocklin
    * Don’t require network to inspect tests (GH#4433) Matthew
      Rocklin
  - Changelog for 2021.1.0
    * Add time started to scheduler info (GH#4425) Jacob Tomlinson
    * Log adaptive error (GH#4422) Jacob Tomlinson
    * Xfail normalization tests (GH#4411) Jacob Tomlinson
    * Use dumps_msgpack and loads_msgpack when packing high level
      graphs (GH#4409) Mads R. B. Kristensen
    * Add nprocs auto option to dask-worker CLI (GH#4377) Jacob
      Tomlinson
    * Type annotation of _reevaluate_occupancy_worker (GH#4398)
      jakirkham
    * Type TaskGroup in active_states (GH#4408) jakirkham
    * Fix test_as_current_is_thread_local (GH#4402) jakirkham
    * Use list comprehensions to bind TaskGroup type (GH#4401)
      jakirkham
    * Make tests pass after 2028 (GH#4403) Bernhard M. Wiedemann
    * Fix compilation warnings, decide_worker now a C func, stealing
      improvements (GH#4375) jakirkham
    * Drop custom __eq__ from Status (GH#4270) jakirkham
    * test_performance_report: skip without bokeh (GH#4388) Bruno
      Pagani
    * Nanny now respects dask settings from ctx mgr (GH#4378)
      Florian Jetter
    * Better task duration estimates for outliers (GH#4213) selshowk
    * Dask internal inherit config (GH#4364) Jacob Tomlinson
    * Provide setup.py option to profile Cython code (GH#4362)
      jakirkham
    * Optimizations of *State and Task* objects and stealing
      (GH#4358) jakirkham
    * Cast SortedDict s to dict s in a few key places & other minor
      changes (GH#4355) jakirkham
    * Use task annotation priorities for user-level priorities
      (GH#4354) James Bourbeau
    * Added docs to highlevelgraph pack/unpack (GH#4352) Mads R. B.
      Kristensen
    * Optimizations in notable functions used by transitions
      (GH#4351) jakirkham
    * Silence exception when releasing futures on process shutdown
      (GH#4309) Benjamin Zaitlen
  - Deboard blind passenger requirement: scikit-build is only
    available for Python > 3.6 (because of SciPy) and not needed here
* Sat Jan 09 2021 Benjamin Greiner <code@bnavigator.de>
  - update to version 2020.12.0
    * Switched to CalVer for versioning scheme.
    * The scheduler can now receives Dask HighLevelGraph s instead of
      raw dictionary task graphs. This allows for a much more
      efficient communication of task graphs from the client to the
      scheduler.
    * Added support for using custom Layer-level annotations like
      priority, retries, etc. with the dask.annotations context
      manager.
    * Updated minimum supported version of Dask to 2020.12.0.
    * Added many type annotations and updates to allow for gradually
      Cythonizing the scheduler.
  - Full changelog on
    https://distributed.readthedocs.io/en/latest/changelog.html
  - Cythonize scheduler: is platform dependent package now
  - Iterative deselection of failing tests: kept adding tests to
    pytest -k parameter until the test suite did not fail anymore.
    * pytest-dev/pytest-asyncio#168, pytest-dev/pytest-asyncio#188
      still not resolved
    * upstream just pinned pytest-asyncio gh#dask/distributed#4212

Files

/etc/alternatives/dask-scheduler
/etc/alternatives/dask-ssh
/etc/alternatives/dask-worker
/usr/bin/dask-scheduler
/usr/bin/dask-scheduler-3.11
/usr/bin/dask-ssh
/usr/bin/dask-ssh-3.11
/usr/bin/dask-worker
/usr/bin/dask-worker-3.11
/usr/lib/python3.11/site-packages/distributed
/usr/lib/python3.11/site-packages/distributed-2024.2.1.dist-info
/usr/lib/python3.11/site-packages/distributed-2024.2.1.dist-info/AUTHORS.md
/usr/lib/python3.11/site-packages/distributed-2024.2.1.dist-info/INSTALLER
/usr/lib/python3.11/site-packages/distributed-2024.2.1.dist-info/LICENSE.txt
/usr/lib/python3.11/site-packages/distributed-2024.2.1.dist-info/METADATA
/usr/lib/python3.11/site-packages/distributed-2024.2.1.dist-info/RECORD
/usr/lib/python3.11/site-packages/distributed-2024.2.1.dist-info/REQUESTED
/usr/lib/python3.11/site-packages/distributed-2024.2.1.dist-info/WHEEL
/usr/lib/python3.11/site-packages/distributed-2024.2.1.dist-info/entry_points.txt
/usr/lib/python3.11/site-packages/distributed-2024.2.1.dist-info/top_level.txt
/usr/lib/python3.11/site-packages/distributed/__init__.py
/usr/lib/python3.11/site-packages/distributed/__pycache__
/usr/lib/python3.11/site-packages/distributed/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/_asyncio.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/_asyncio.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/_concurrent_futures_thread.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/_concurrent_futures_thread.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/_signals.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/_signals.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/_stories.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/_stories.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/_version.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/_version.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/active_memory_manager.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/active_memory_manager.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/actor.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/actor.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/batched.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/batched.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/bokeh.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/bokeh.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/cfexecutor.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/cfexecutor.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/chaos.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/chaos.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/client.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/client.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/cluster_dump.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/cluster_dump.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/collections.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/collections.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/compatibility.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/compatibility.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/config.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/config.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/core.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/core.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/counter.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/counter.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/diskutils.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/diskutils.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/event.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/event.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/exceptions.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/exceptions.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/itertools.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/itertools.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/lock.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/lock.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/metrics.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/metrics.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/multi_lock.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/multi_lock.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/nanny.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/nanny.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/node.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/node.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/objects.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/objects.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/preloading.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/preloading.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/process.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/process.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/proctitle.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/proctitle.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/profile.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/profile.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/publish.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/publish.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/pubsub.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/pubsub.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/pytest_resourceleaks.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/pytest_resourceleaks.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/queues.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/queues.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/recreate_tasks.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/recreate_tasks.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/scheduler.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/scheduler.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/security.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/security.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/semaphore.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/semaphore.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/sizeof.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/sizeof.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/spans.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/spans.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/spill.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/spill.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/stealing.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/stealing.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/system.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/system.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/system_monitor.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/system_monitor.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/threadpoolexecutor.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/threadpoolexecutor.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/utils.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/utils.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/utils_comm.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/utils_comm.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/utils_perf.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/utils_perf.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/utils_test.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/utils_test.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/variable.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/variable.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/versions.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/versions.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/worker.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/worker.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/worker_client.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/worker_client.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/worker_memory.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/worker_memory.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/worker_state_machine.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/__pycache__/worker_state_machine.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/_asyncio.py
/usr/lib/python3.11/site-packages/distributed/_concurrent_futures_thread.py
/usr/lib/python3.11/site-packages/distributed/_signals.py
/usr/lib/python3.11/site-packages/distributed/_stories.py
/usr/lib/python3.11/site-packages/distributed/_version.py
/usr/lib/python3.11/site-packages/distributed/active_memory_manager.py
/usr/lib/python3.11/site-packages/distributed/actor.py
/usr/lib/python3.11/site-packages/distributed/batched.py
/usr/lib/python3.11/site-packages/distributed/bokeh.py
/usr/lib/python3.11/site-packages/distributed/cfexecutor.py
/usr/lib/python3.11/site-packages/distributed/chaos.py
/usr/lib/python3.11/site-packages/distributed/cli
/usr/lib/python3.11/site-packages/distributed/cli/__init__.py
/usr/lib/python3.11/site-packages/distributed/cli/__pycache__
/usr/lib/python3.11/site-packages/distributed/cli/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/cli/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/cli/__pycache__/dask_scheduler.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/cli/__pycache__/dask_scheduler.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/cli/__pycache__/dask_spec.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/cli/__pycache__/dask_spec.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/cli/__pycache__/dask_ssh.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/cli/__pycache__/dask_ssh.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/cli/__pycache__/dask_worker.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/cli/__pycache__/dask_worker.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/cli/__pycache__/utils.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/cli/__pycache__/utils.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/cli/dask_scheduler.py
/usr/lib/python3.11/site-packages/distributed/cli/dask_spec.py
/usr/lib/python3.11/site-packages/distributed/cli/dask_ssh.py
/usr/lib/python3.11/site-packages/distributed/cli/dask_worker.py
/usr/lib/python3.11/site-packages/distributed/cli/utils.py
/usr/lib/python3.11/site-packages/distributed/client.py
/usr/lib/python3.11/site-packages/distributed/cluster_dump.py
/usr/lib/python3.11/site-packages/distributed/collections.py
/usr/lib/python3.11/site-packages/distributed/comm
/usr/lib/python3.11/site-packages/distributed/comm/__init__.py
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/addressing.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/addressing.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/core.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/core.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/inproc.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/inproc.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/registry.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/registry.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/tcp.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/tcp.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/ucx.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/ucx.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/utils.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/utils.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/ws.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/comm/__pycache__/ws.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/comm/addressing.py
/usr/lib/python3.11/site-packages/distributed/comm/core.py
/usr/lib/python3.11/site-packages/distributed/comm/inproc.py
/usr/lib/python3.11/site-packages/distributed/comm/registry.py
/usr/lib/python3.11/site-packages/distributed/comm/tcp.py
/usr/lib/python3.11/site-packages/distributed/comm/ucx.py
/usr/lib/python3.11/site-packages/distributed/comm/utils.py
/usr/lib/python3.11/site-packages/distributed/comm/ws.py
/usr/lib/python3.11/site-packages/distributed/compatibility.py
/usr/lib/python3.11/site-packages/distributed/config.py
/usr/lib/python3.11/site-packages/distributed/core.py
/usr/lib/python3.11/site-packages/distributed/counter.py
/usr/lib/python3.11/site-packages/distributed/dashboard
/usr/lib/python3.11/site-packages/distributed/dashboard/__init__.py
/usr/lib/python3.11/site-packages/distributed/dashboard/__pycache__
/usr/lib/python3.11/site-packages/distributed/dashboard/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/__pycache__/core.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/__pycache__/core.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/__pycache__/export_tool.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/__pycache__/export_tool.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/__pycache__/scheduler.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/__pycache__/scheduler.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/__pycache__/utils.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/__pycache__/utils.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/__pycache__/worker.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/__pycache__/worker.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/components
/usr/lib/python3.11/site-packages/distributed/dashboard/components/__init__.py
/usr/lib/python3.11/site-packages/distributed/dashboard/components/__pycache__
/usr/lib/python3.11/site-packages/distributed/dashboard/components/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/components/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/components/__pycache__/nvml.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/components/__pycache__/nvml.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/components/__pycache__/rmm.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/components/__pycache__/rmm.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/components/__pycache__/scheduler.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/components/__pycache__/scheduler.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/components/__pycache__/shared.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/components/__pycache__/shared.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/components/__pycache__/worker.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/components/__pycache__/worker.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/components/nvml.py
/usr/lib/python3.11/site-packages/distributed/dashboard/components/rmm.py
/usr/lib/python3.11/site-packages/distributed/dashboard/components/scheduler.py
/usr/lib/python3.11/site-packages/distributed/dashboard/components/shared.py
/usr/lib/python3.11/site-packages/distributed/dashboard/components/worker.py
/usr/lib/python3.11/site-packages/distributed/dashboard/core.py
/usr/lib/python3.11/site-packages/distributed/dashboard/export_tool.js
/usr/lib/python3.11/site-packages/distributed/dashboard/export_tool.py
/usr/lib/python3.11/site-packages/distributed/dashboard/scheduler.py
/usr/lib/python3.11/site-packages/distributed/dashboard/templates
/usr/lib/python3.11/site-packages/distributed/dashboard/templates/__init__.py
/usr/lib/python3.11/site-packages/distributed/dashboard/templates/__pycache__
/usr/lib/python3.11/site-packages/distributed/dashboard/templates/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/templates/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/dashboard/templates/performance_report.html
/usr/lib/python3.11/site-packages/distributed/dashboard/theme.yaml
/usr/lib/python3.11/site-packages/distributed/dashboard/utils.py
/usr/lib/python3.11/site-packages/distributed/dashboard/worker.py
/usr/lib/python3.11/site-packages/distributed/deploy
/usr/lib/python3.11/site-packages/distributed/deploy/__init__.py
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/adaptive.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/adaptive.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/adaptive_core.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/adaptive_core.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/cluster.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/cluster.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/local.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/local.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/old_ssh.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/old_ssh.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/spec.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/spec.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/ssh.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/ssh.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/subprocess.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/subprocess.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/utils.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/__pycache__/utils.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/deploy/adaptive.py
/usr/lib/python3.11/site-packages/distributed/deploy/adaptive_core.py
/usr/lib/python3.11/site-packages/distributed/deploy/cluster.py
/usr/lib/python3.11/site-packages/distributed/deploy/local.py
/usr/lib/python3.11/site-packages/distributed/deploy/old_ssh.py
/usr/lib/python3.11/site-packages/distributed/deploy/spec.py
/usr/lib/python3.11/site-packages/distributed/deploy/ssh.py
/usr/lib/python3.11/site-packages/distributed/deploy/subprocess.py
/usr/lib/python3.11/site-packages/distributed/deploy/utils.py
/usr/lib/python3.11/site-packages/distributed/diagnostics
/usr/lib/python3.11/site-packages/distributed/diagnostics/__init__.py
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/cluster_dump.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/cluster_dump.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/cudf.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/cudf.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/eventstream.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/eventstream.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/graph_layout.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/graph_layout.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/memory_sampler.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/memory_sampler.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/memray.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/memray.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/nvml.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/nvml.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/plugin.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/plugin.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/progress.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/progress.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/progress_stream.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/progress_stream.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/progressbar.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/progressbar.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/rmm.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/rmm.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/task_stream.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/task_stream.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/websocket.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/__pycache__/websocket.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/diagnostics/cluster_dump.py
/usr/lib/python3.11/site-packages/distributed/diagnostics/cudf.py
/usr/lib/python3.11/site-packages/distributed/diagnostics/eventstream.py
/usr/lib/python3.11/site-packages/distributed/diagnostics/graph_layout.py
/usr/lib/python3.11/site-packages/distributed/diagnostics/memory_sampler.py
/usr/lib/python3.11/site-packages/distributed/diagnostics/memray.py
/usr/lib/python3.11/site-packages/distributed/diagnostics/nvml.py
/usr/lib/python3.11/site-packages/distributed/diagnostics/plugin.py
/usr/lib/python3.11/site-packages/distributed/diagnostics/progress.py
/usr/lib/python3.11/site-packages/distributed/diagnostics/progress_stream.py
/usr/lib/python3.11/site-packages/distributed/diagnostics/progressbar.py
/usr/lib/python3.11/site-packages/distributed/diagnostics/rmm.py
/usr/lib/python3.11/site-packages/distributed/diagnostics/task_stream.py
/usr/lib/python3.11/site-packages/distributed/diagnostics/websocket.py
/usr/lib/python3.11/site-packages/distributed/diskutils.py
/usr/lib/python3.11/site-packages/distributed/distributed-schema.yaml
/usr/lib/python3.11/site-packages/distributed/distributed.yaml
/usr/lib/python3.11/site-packages/distributed/event.py
/usr/lib/python3.11/site-packages/distributed/exceptions.py
/usr/lib/python3.11/site-packages/distributed/http
/usr/lib/python3.11/site-packages/distributed/http/__init__.py
/usr/lib/python3.11/site-packages/distributed/http/__pycache__
/usr/lib/python3.11/site-packages/distributed/http/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/__pycache__/health.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/__pycache__/health.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/__pycache__/prometheus.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/__pycache__/prometheus.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/__pycache__/proxy.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/__pycache__/proxy.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/__pycache__/routing.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/__pycache__/routing.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/__pycache__/statics.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/__pycache__/statics.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/__pycache__/utils.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/__pycache__/utils.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/health.py
/usr/lib/python3.11/site-packages/distributed/http/prometheus.py
/usr/lib/python3.11/site-packages/distributed/http/proxy.py
/usr/lib/python3.11/site-packages/distributed/http/routing.py
/usr/lib/python3.11/site-packages/distributed/http/scheduler
/usr/lib/python3.11/site-packages/distributed/http/scheduler/__init__.py
/usr/lib/python3.11/site-packages/distributed/http/scheduler/__pycache__
/usr/lib/python3.11/site-packages/distributed/http/scheduler/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/__pycache__/api.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/__pycache__/api.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/__pycache__/info.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/__pycache__/info.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/__pycache__/json.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/__pycache__/json.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/__pycache__/missing_bokeh.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/__pycache__/missing_bokeh.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/api.py
/usr/lib/python3.11/site-packages/distributed/http/scheduler/info.py
/usr/lib/python3.11/site-packages/distributed/http/scheduler/json.py
/usr/lib/python3.11/site-packages/distributed/http/scheduler/missing_bokeh.py
/usr/lib/python3.11/site-packages/distributed/http/scheduler/prometheus
/usr/lib/python3.11/site-packages/distributed/http/scheduler/prometheus/__init__.py
/usr/lib/python3.11/site-packages/distributed/http/scheduler/prometheus/__pycache__
/usr/lib/python3.11/site-packages/distributed/http/scheduler/prometheus/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/prometheus/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/prometheus/__pycache__/core.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/prometheus/__pycache__/core.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/prometheus/__pycache__/semaphore.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/prometheus/__pycache__/semaphore.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/prometheus/__pycache__/stealing.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/prometheus/__pycache__/stealing.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/scheduler/prometheus/core.py
/usr/lib/python3.11/site-packages/distributed/http/scheduler/prometheus/semaphore.py
/usr/lib/python3.11/site-packages/distributed/http/scheduler/prometheus/stealing.py
/usr/lib/python3.11/site-packages/distributed/http/static
/usr/lib/python3.11/site-packages/distributed/http/static/__init__.py
/usr/lib/python3.11/site-packages/distributed/http/static/__pycache__
/usr/lib/python3.11/site-packages/distributed/http/static/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/static/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/static/css
/usr/lib/python3.11/site-packages/distributed/http/static/css/__init__.py
/usr/lib/python3.11/site-packages/distributed/http/static/css/__pycache__
/usr/lib/python3.11/site-packages/distributed/http/static/css/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/static/css/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/static/css/base.css
/usr/lib/python3.11/site-packages/distributed/http/static/css/gpu.css
/usr/lib/python3.11/site-packages/distributed/http/static/css/individual-cluster-map.css
/usr/lib/python3.11/site-packages/distributed/http/static/css/sortable.min.css
/usr/lib/python3.11/site-packages/distributed/http/static/css/status.css
/usr/lib/python3.11/site-packages/distributed/http/static/images
/usr/lib/python3.11/site-packages/distributed/http/static/images/__init__.py
/usr/lib/python3.11/site-packages/distributed/http/static/images/__pycache__
/usr/lib/python3.11/site-packages/distributed/http/static/images/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/static/images/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/static/images/dask-logo.svg
/usr/lib/python3.11/site-packages/distributed/http/static/images/fa-bars.svg
/usr/lib/python3.11/site-packages/distributed/http/static/images/favicon.ico
/usr/lib/python3.11/site-packages/distributed/http/static/images/jupyter.svg
/usr/lib/python3.11/site-packages/distributed/http/static/images/numpy.png
/usr/lib/python3.11/site-packages/distributed/http/static/images/pandas.png
/usr/lib/python3.11/site-packages/distributed/http/static/images/python.png
/usr/lib/python3.11/site-packages/distributed/http/static/individual-cluster-map.html
/usr/lib/python3.11/site-packages/distributed/http/static/js
/usr/lib/python3.11/site-packages/distributed/http/static/js/__init__.py
/usr/lib/python3.11/site-packages/distributed/http/static/js/__pycache__
/usr/lib/python3.11/site-packages/distributed/http/static/js/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/static/js/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/static/js/anime.min.js
/usr/lib/python3.11/site-packages/distributed/http/static/js/individual-cluster-map.js
/usr/lib/python3.11/site-packages/distributed/http/static/js/reconnecting-websocket.min.js
/usr/lib/python3.11/site-packages/distributed/http/static/js/sortable.min.js
/usr/lib/python3.11/site-packages/distributed/http/statics.py
/usr/lib/python3.11/site-packages/distributed/http/templates
/usr/lib/python3.11/site-packages/distributed/http/templates/__init__.py
/usr/lib/python3.11/site-packages/distributed/http/templates/__pycache__
/usr/lib/python3.11/site-packages/distributed/http/templates/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/templates/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/templates/base.html
/usr/lib/python3.11/site-packages/distributed/http/templates/call-stack.html
/usr/lib/python3.11/site-packages/distributed/http/templates/exceptions.html
/usr/lib/python3.11/site-packages/distributed/http/templates/gpu.html
/usr/lib/python3.11/site-packages/distributed/http/templates/json-index.html
/usr/lib/python3.11/site-packages/distributed/http/templates/logs.html
/usr/lib/python3.11/site-packages/distributed/http/templates/main.html
/usr/lib/python3.11/site-packages/distributed/http/templates/simple.html
/usr/lib/python3.11/site-packages/distributed/http/templates/status.html
/usr/lib/python3.11/site-packages/distributed/http/templates/task.html
/usr/lib/python3.11/site-packages/distributed/http/templates/worker-table.html
/usr/lib/python3.11/site-packages/distributed/http/templates/worker.html
/usr/lib/python3.11/site-packages/distributed/http/templates/workers.html
/usr/lib/python3.11/site-packages/distributed/http/utils.py
/usr/lib/python3.11/site-packages/distributed/http/worker
/usr/lib/python3.11/site-packages/distributed/http/worker/__init__.py
/usr/lib/python3.11/site-packages/distributed/http/worker/__pycache__
/usr/lib/python3.11/site-packages/distributed/http/worker/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/worker/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/worker/prometheus
/usr/lib/python3.11/site-packages/distributed/http/worker/prometheus/__init__.py
/usr/lib/python3.11/site-packages/distributed/http/worker/prometheus/__pycache__
/usr/lib/python3.11/site-packages/distributed/http/worker/prometheus/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/worker/prometheus/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/worker/prometheus/__pycache__/core.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/http/worker/prometheus/__pycache__/core.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/http/worker/prometheus/core.py
/usr/lib/python3.11/site-packages/distributed/itertools.py
/usr/lib/python3.11/site-packages/distributed/lock.py
/usr/lib/python3.11/site-packages/distributed/metrics.py
/usr/lib/python3.11/site-packages/distributed/multi_lock.py
/usr/lib/python3.11/site-packages/distributed/nanny.py
/usr/lib/python3.11/site-packages/distributed/node.py
/usr/lib/python3.11/site-packages/distributed/objects.py
/usr/lib/python3.11/site-packages/distributed/preloading.py
/usr/lib/python3.11/site-packages/distributed/process.py
/usr/lib/python3.11/site-packages/distributed/proctitle.py
/usr/lib/python3.11/site-packages/distributed/profile.py
/usr/lib/python3.11/site-packages/distributed/protocol
/usr/lib/python3.11/site-packages/distributed/protocol/__init__.py
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/arrow.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/arrow.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/compression.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/compression.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/core.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/core.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/cuda.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/cuda.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/cupy.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/cupy.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/h5py.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/h5py.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/keras.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/keras.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/netcdf4.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/netcdf4.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/numba.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/numba.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/numpy.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/numpy.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/pickle.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/pickle.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/rmm.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/rmm.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/scipy.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/scipy.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/serialize.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/serialize.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/sparse.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/sparse.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/torch.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/torch.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/utils.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/utils.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/utils_test.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/__pycache__/utils_test.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/protocol/arrow.py
/usr/lib/python3.11/site-packages/distributed/protocol/compression.py
/usr/lib/python3.11/site-packages/distributed/protocol/core.py
/usr/lib/python3.11/site-packages/distributed/protocol/cuda.py
/usr/lib/python3.11/site-packages/distributed/protocol/cupy.py
/usr/lib/python3.11/site-packages/distributed/protocol/h5py.py
/usr/lib/python3.11/site-packages/distributed/protocol/keras.py
/usr/lib/python3.11/site-packages/distributed/protocol/netcdf4.py
/usr/lib/python3.11/site-packages/distributed/protocol/numba.py
/usr/lib/python3.11/site-packages/distributed/protocol/numpy.py
/usr/lib/python3.11/site-packages/distributed/protocol/pickle.py
/usr/lib/python3.11/site-packages/distributed/protocol/rmm.py
/usr/lib/python3.11/site-packages/distributed/protocol/scipy.py
/usr/lib/python3.11/site-packages/distributed/protocol/serialize.py
/usr/lib/python3.11/site-packages/distributed/protocol/sparse.py
/usr/lib/python3.11/site-packages/distributed/protocol/torch.py
/usr/lib/python3.11/site-packages/distributed/protocol/utils.py
/usr/lib/python3.11/site-packages/distributed/protocol/utils_test.py
/usr/lib/python3.11/site-packages/distributed/publish.py
/usr/lib/python3.11/site-packages/distributed/pubsub.py
/usr/lib/python3.11/site-packages/distributed/py.typed
/usr/lib/python3.11/site-packages/distributed/pytest_resourceleaks.py
/usr/lib/python3.11/site-packages/distributed/queues.py
/usr/lib/python3.11/site-packages/distributed/recreate_tasks.py
/usr/lib/python3.11/site-packages/distributed/scheduler.py
/usr/lib/python3.11/site-packages/distributed/security.py
/usr/lib/python3.11/site-packages/distributed/semaphore.py
/usr/lib/python3.11/site-packages/distributed/shuffle
/usr/lib/python3.11/site-packages/distributed/shuffle/__init__.py
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_arrow.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_arrow.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_buffer.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_buffer.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_comms.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_comms.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_core.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_core.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_disk.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_disk.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_exceptions.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_exceptions.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_limiter.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_limiter.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_memory.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_memory.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_merge.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_merge.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_pickle.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_pickle.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_rechunk.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_rechunk.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_scheduler_plugin.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_scheduler_plugin.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_shuffle.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_shuffle.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_worker_plugin.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/__pycache__/_worker_plugin.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/shuffle/_arrow.py
/usr/lib/python3.11/site-packages/distributed/shuffle/_buffer.py
/usr/lib/python3.11/site-packages/distributed/shuffle/_comms.py
/usr/lib/python3.11/site-packages/distributed/shuffle/_core.py
/usr/lib/python3.11/site-packages/distributed/shuffle/_disk.py
/usr/lib/python3.11/site-packages/distributed/shuffle/_exceptions.py
/usr/lib/python3.11/site-packages/distributed/shuffle/_limiter.py
/usr/lib/python3.11/site-packages/distributed/shuffle/_memory.py
/usr/lib/python3.11/site-packages/distributed/shuffle/_merge.py
/usr/lib/python3.11/site-packages/distributed/shuffle/_pickle.py
/usr/lib/python3.11/site-packages/distributed/shuffle/_rechunk.py
/usr/lib/python3.11/site-packages/distributed/shuffle/_scheduler_plugin.py
/usr/lib/python3.11/site-packages/distributed/shuffle/_shuffle.py
/usr/lib/python3.11/site-packages/distributed/shuffle/_worker_plugin.py
/usr/lib/python3.11/site-packages/distributed/sizeof.py
/usr/lib/python3.11/site-packages/distributed/spans.py
/usr/lib/python3.11/site-packages/distributed/spill.py
/usr/lib/python3.11/site-packages/distributed/stealing.py
/usr/lib/python3.11/site-packages/distributed/system.py
/usr/lib/python3.11/site-packages/distributed/system_monitor.py
/usr/lib/python3.11/site-packages/distributed/threadpoolexecutor.py
/usr/lib/python3.11/site-packages/distributed/utils.py
/usr/lib/python3.11/site-packages/distributed/utils_comm.py
/usr/lib/python3.11/site-packages/distributed/utils_perf.py
/usr/lib/python3.11/site-packages/distributed/utils_test.py
/usr/lib/python3.11/site-packages/distributed/variable.py
/usr/lib/python3.11/site-packages/distributed/versions.py
/usr/lib/python3.11/site-packages/distributed/widgets
/usr/lib/python3.11/site-packages/distributed/widgets/__init__.py
/usr/lib/python3.11/site-packages/distributed/widgets/__pycache__
/usr/lib/python3.11/site-packages/distributed/widgets/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/widgets/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/widgets/templates
/usr/lib/python3.11/site-packages/distributed/widgets/templates/__init__.py
/usr/lib/python3.11/site-packages/distributed/widgets/templates/__pycache__
/usr/lib/python3.11/site-packages/distributed/widgets/templates/__pycache__/__init__.cpython-311.opt-1.pyc
/usr/lib/python3.11/site-packages/distributed/widgets/templates/__pycache__/__init__.cpython-311.pyc
/usr/lib/python3.11/site-packages/distributed/widgets/templates/client.html.j2
/usr/lib/python3.11/site-packages/distributed/widgets/templates/cluster.html.j2
/usr/lib/python3.11/site-packages/distributed/widgets/templates/computation.html.j2
/usr/lib/python3.11/site-packages/distributed/widgets/templates/future.html.j2
/usr/lib/python3.11/site-packages/distributed/widgets/templates/has_what.html.j2
/usr/lib/python3.11/site-packages/distributed/widgets/templates/local_cluster.html.j2
/usr/lib/python3.11/site-packages/distributed/widgets/templates/log.html.j2
/usr/lib/python3.11/site-packages/distributed/widgets/templates/logs.html.j2
/usr/lib/python3.11/site-packages/distributed/widgets/templates/process_interface.html.j2
/usr/lib/python3.11/site-packages/distributed/widgets/templates/scheduler.html.j2
/usr/lib/python3.11/site-packages/distributed/widgets/templates/scheduler_info.html.j2
/usr/lib/python3.11/site-packages/distributed/widgets/templates/security.html.j2
/usr/lib/python3.11/site-packages/distributed/widgets/templates/task_state.html.j2
/usr/lib/python3.11/site-packages/distributed/widgets/templates/who_has.html.j2
/usr/lib/python3.11/site-packages/distributed/widgets/templates/worker_state.html.j2
/usr/lib/python3.11/site-packages/distributed/worker.py
/usr/lib/python3.11/site-packages/distributed/worker_client.py
/usr/lib/python3.11/site-packages/distributed/worker_memory.py
/usr/lib/python3.11/site-packages/distributed/worker_state_machine.py
/usr/share/doc/packages/python311-distributed
/usr/share/doc/packages/python311-distributed/README.rst
/usr/share/licenses/python311-distributed
/usr/share/licenses/python311-distributed/LICENSE.txt


Generated by rpm2html 1.8.1

Fabrice Bellet, Wed May 1 23:23:10 2024