• 设为首页
  • 点击收藏
  • 手机版
    手机扫一扫访问
    迪恩网络手机版
  • 关注官方公众号
    微信扫一扫关注
    迪恩网络公众号

dask/dask-tutorial: Dask tutorial

原作者: [db:作者] 来自: 网络 收藏 邀请

开源软件名称(OpenSource Name):

dask/dask-tutorial

开源软件地址(OpenSource Url):

https://github.com/dask/dask-tutorial

开源编程语言(OpenSource Language):

Jupyter Notebook 92.0%

开源软件介绍(OpenSource Introduction):

Dask Tutorial

NOTE: This tutorial will be presented again at SciPy 2022! We are in the progress of making some modifications (see PR #236 for more info)

This tutorial was last given at SciPy 2020 which was a virtual conference. A video of the SciPy 2020 tutorial is available online.

Binder Build Status

Dask provides multi-core execution on larger-than-memory datasets.

We can think of dask at a high and a low level

  • High-level collections: Dask provides high-level Array, Bag, and DataFrame collections that mimic NumPy, lists, and Pandas but can operate in parallel on datasets that don't fit into main memory. Dask's high-level collections are alternatives to NumPy and Pandas for large datasets.
  • Low-level schedulers: Dask provides dynamic task schedulers that execute task graphs in parallel. These execution engines power the high-level collections mentioned above but can also power custom, user-defined workloads. These schedulers are low-latency (around 1ms) and work hard to run computations in a small memory footprint. Dask's schedulers are an alternative to direct use of threading or multiprocessing libraries in complex cases or other task scheduling systems like Luigi or IPython parallel.

Different users operate at different levels but it is useful to understand both. This tutorial will interleave between high-level use of dask.array and dask.dataframe (even sections) and low-level use of dask graphs and schedulers (odd sections.)

Prepare

1. You should clone this repository

git clone http://github.com/dask/dask-tutorial

and then install necessary packages. There are three different ways to achieve this, pick the one that best suits you, and only pick one option. They are, in order of preference:

2a) Create a conda environment (preferred)

In the main repo directory

conda env create -f binder/environment.yml
conda activate dask-tutorial
jupyter labextension install dask-labextension
jupyter labextension install @jupyter-widgets/jupyterlab-manager
jupyter labextension install @bokeh/jupyter_bokeh

2b) Install into an existing environment

You will need the following core libraries

conda install numpy pandas h5py pillow matplotlib scipy toolz pytables snakeviz scikit-image dask distributed -c conda-forge

You may find the following libraries helpful for some exercises

conda install python-graphviz -c conda-forge

Note that these options will alter your existing environment, potentially changing the versions of packages you already have installed.

2c) Use Dockerfile

You can build a docker image from the provided Dockerfile.

$ docker build . # This will build using the same env as in a)

Run a container, replacing the ID with the output of the previous command

$ docker run -it -p 8888:8888 -p 8787:8787 <container_id_or_tag>

The above command will give an URL (Like http://(container_id or 127.0.0.1):8888/?token=<sometoken>) which can be used to access the notebook from browser. You may need to replace the given hostname with "localhost" or "127.0.0.1".

You should follow only one of the options above!

Launch notebook

From the repo directory

jupyter notebook

Or

jupyter lab

This was already done for method c) and does not need repeating.

Links

Outline

  1. Overview - dask's place in the universe.

  2. Delayed - the single-function way to parallelize general python code.

1x. Lazy - some of the principles behind lazy execution, for the interested.

  1. Bag - the first high-level collection: a generalized iterator for use with a functional programming style and to clean messy data.

  2. Array - blocked numpy-like functionality with a collection of numpy arrays spread across your cluster.

  3. Dataframe - parallelized operations on many pandas dataframes spread across your cluster.

  4. Distributed - Dask's scheduler for clusters, with details of how to view the UI.

  5. Advanced Distributed - further details on distributed computing, including how to debug.

  6. Dataframe Storage - efficient ways to read and write dataframes to disc.

  7. Machine Learning - applying dask to machine-learning problems.




鲜花

握手

雷人

路过

鸡蛋
该文章已有0人参与评论

请发表评论

全部评论

专题导读
热门推荐
阅读排行榜

扫描微信二维码

查看手机版网站

随时了解更新最新资讯

139-2527-9053

在线客服(服务时间 9:00~18:00)

在线QQ客服
地址:深圳市南山区西丽大学城创智工业园
电邮:jeky_zhao#qq.com
移动电话:139-2527-9053

Powered by 互联科技 X3.4© 2001-2213 极客世界.|Sitemap