[Python-modules-commits] [dask] 01/03: Import dask_0.12.0.orig.tar.gz
Diane Trout
diane at moszumanska.debian.org
Wed Jan 11 06:23:32 UTC 2017
This is an automated email from the git hooks/post-receive script.
diane pushed a commit to branch master
in repository dask.
commit 5a9d26d757827b79ee01489d2da240fd2266a232
Author: Diane Trout <diane at ghic.org>
Date: Tue Jan 10 19:50:45 2017 -0800
Import dask_0.12.0.orig.tar.gz
---
.coveragerc | 6 +
.github/CONTRIBUTING.md | 2 +
.gitignore | 13 +
.travis.yml | 57 +
AUTHORS.md | 121 +
LICENSE.txt | 28 +
MANIFEST.in | 11 +
README.rst | 28 +
appveyor.yml | 39 +
conftest.py | 15 +
continuous_integration/build.cmd | 6 +
continuous_integration/run_tests.cmd | 11 +
continuous_integration/run_with_env.cmd | 90 +
continuous_integration/setup_conda_environment.cmd | 37 +
dask/__init__.py | 17 +
dask/_version.py | 484 +++
dask/array/NUMPY_LICENSE.txt | 30 +
dask/array/__init__.py | 34 +
dask/array/chunk.py | 188 +
dask/array/conftest.py | 3 +
dask/array/core.py | 3699 ++++++++++++++++++++
dask/array/creation.py | 139 +
dask/array/fft.py | 104 +
dask/array/ghost.py | 422 +++
dask/array/image.py | 71 +
dask/array/learn.py | 102 +
dask/array/linalg.py | 773 ++++
dask/array/numpy_compat.py | 356 ++
dask/array/optimization.py | 204 ++
dask/array/percentile.py | 170 +
dask/array/random.py | 406 +++
dask/array/rechunk.py | 238 ++
dask/array/reductions.py | 646 ++++
dask/array/slicing.py | 717 ++++
dask/array/tests/__init__.py | 0
dask/array/tests/test_array_core.py | 2450 +++++++++++++
dask/array/tests/test_chunk.py | 114 +
dask/array/tests/test_creation.py | 94 +
dask/array/tests/test_fft.py | 110 +
dask/array/tests/test_ghost.py | 345 ++
dask/array/tests/test_image.py | 49 +
dask/array/tests/test_learn.py | 39 +
dask/array/tests/test_linalg.py | 430 +++
dask/array/tests/test_linearoperator.py | 34 +
dask/array/tests/test_optimization.py | 133 +
dask/array/tests/test_percentiles.py | 62 +
dask/array/tests/test_random.py | 218 ++
dask/array/tests/test_rechunk.py | 190 +
dask/array/tests/test_reductions.py | 347 ++
dask/array/tests/test_slicing.py | 552 +++
dask/array/tests/test_ufunc.py | 154 +
dask/array/tests/test_wrap.py | 54 +
dask/array/ufunc.py | 164 +
dask/array/utils.py | 70 +
dask/array/wrap.py | 79 +
dask/async.py | 626 ++++
dask/bag/__init__.py | 8 +
dask/bag/core.py | 1679 +++++++++
dask/bag/tests/__init__.py | 0
dask/bag/tests/test_bag.py | 1099 ++++++
dask/bag/tests/test_text.py | 47 +
dask/bag/text.py | 109 +
dask/base.py | 368 ++
dask/bytes/__init__.py | 9 +
dask/bytes/compression.py | 57 +
dask/bytes/core.py | 387 ++
dask/bytes/local.py | 136 +
dask/bytes/s3.py | 202 ++
dask/bytes/tests/__init__.py | 0
dask/bytes/tests/test_bytes_utils.py | 68 +
dask/bytes/tests/test_compression.py | 40 +
dask/bytes/tests/test_local.py | 281 ++
dask/bytes/tests/test_s3.py | 280 ++
dask/bytes/utils.py | 124 +
dask/cache.py | 66 +
dask/callbacks.py | 121 +
dask/compatibility.py | 258 ++
dask/context.py | 42 +
dask/core.py | 437 +++
dask/dataframe/__init__.py | 16 +
dask/dataframe/accessor.py | 176 +
dask/dataframe/categorical.py | 40 +
dask/dataframe/core.py | 3364 ++++++++++++++++++
dask/dataframe/groupby.py | 871 +++++
dask/dataframe/indexing.py | 223 ++
dask/dataframe/io/__init__.py | 8 +
dask/dataframe/io/csv.py | 307 ++
dask/dataframe/io/demo.py | 90 +
dask/dataframe/io/hdf.py | 318 ++
dask/dataframe/io/io.py | 566 +++
dask/dataframe/io/tests/__init__.py | 0
dask/dataframe/io/tests/test_csv.py | 801 +++++
dask/dataframe/io/tests/test_demo.py | 34 +
dask/dataframe/io/tests/test_hdf.py | 558 +++
dask/dataframe/io/tests/test_io.py | 546 +++
dask/dataframe/methods.py | 221 ++
dask/dataframe/multi.py | 631 ++++
dask/dataframe/optimize.py | 30 +
dask/dataframe/partitionquantiles.py | 457 +++
dask/dataframe/reshape.py | 153 +
dask/dataframe/rolling.py | 236 ++
dask/dataframe/shuffle.py | 406 +++
dask/dataframe/tests/__init__.py | 0
dask/dataframe/tests/test_arithmetics_reduction.py | 942 +++++
dask/dataframe/tests/test_categorical.py | 142 +
dask/dataframe/tests/test_dataframe.py | 2504 +++++++++++++
dask/dataframe/tests/test_groupby.py | 724 ++++
dask/dataframe/tests/test_indexing.py | 323 ++
dask/dataframe/tests/test_multi.py | 818 +++++
dask/dataframe/tests/test_optimize_dataframe.py | 60 +
dask/dataframe/tests/test_reshape.py | 187 +
dask/dataframe/tests/test_rolling.py | 230 ++
dask/dataframe/tests/test_shuffle.py | 257 ++
dask/dataframe/tests/test_ufunc.py | 280 ++
dask/dataframe/tests/test_utils_dataframe.py | 218 ++
dask/dataframe/tseries/__init__.py | 0
dask/dataframe/tseries/resample.py | 161 +
dask/dataframe/tseries/tests/__init__.py | 0
dask/dataframe/tseries/tests/test_resample.py | 56 +
dask/dataframe/utils.py | 538 +++
dask/delayed.py | 462 +++
dask/diagnostics/__init__.py | 5 +
dask/diagnostics/profile.py | 345 ++
dask/diagnostics/profile_visualize.py | 410 +++
dask/diagnostics/progress.py | 134 +
dask/diagnostics/tests/__init__.py | 0
dask/diagnostics/tests/test_profiler.py | 303 ++
dask/diagnostics/tests/test_progress.py | 117 +
dask/distributed.py | 6 +
dask/dot.py | 246 ++
dask/multiprocessing.py | 88 +
dask/optimize.py | 612 ++++
dask/order.py | 195 ++
dask/rewrite.py | 438 +++
dask/store/__init__.py | 1 +
dask/store/core.py | 114 +
dask/store/tests/__init__.py | 0
dask/store/tests/test_store.py | 50 +
dask/tests/__init__.py | 0
dask/tests/test_async.py | 207 ++
dask/tests/test_base.py | 351 ++
dask/tests/test_cache.py | 62 +
dask/tests/test_callbacks.py | 30 +
dask/tests/test_context.py | 36 +
dask/tests/test_core.py | 228 ++
dask/tests/test_delayed.py | 367 ++
dask/tests/test_distributed.py | 6 +
dask/tests/test_dot.py | 169 +
dask/tests/test_multiprocessing.py | 103 +
dask/tests/test_optimize.py | 392 +++
dask/tests/test_order.py | 179 +
dask/tests/test_rewrite.py | 140 +
dask/tests/test_threaded.py | 102 +
dask/tests/test_utils.py | 215 ++
dask/threaded.py | 80 +
dask/utils.py | 899 +++++
dask/utils_test.py | 127 +
docs/Makefile | 153 +
docs/README.rst | 10 +
docs/make.bat | 190 +
docs/release-procedure.md | 25 +
docs/requirements-docs.txt | 7 +
docs/source/_static/dask-simple.png | Bin 0 -> 11021 bytes
docs/source/_static/profile.html | 31 +
docs/source/_static/stacked_profile.html | 31 +
docs/source/_static/style.css | 3 +
docs/source/_templates/layout.html | 2 +
docs/source/array-api.rst | 462 +++
docs/source/array-creation.rst | 245 ++
docs/source/array-design.rst | 148 +
docs/source/array-ghost.rst | 171 +
docs/source/array-linear-operator.rst | 28 +
docs/source/array-overview.rst | 77 +
docs/source/array-slicing.rst | 44 +
docs/source/array-stack.rst | 54 +
docs/source/array.rst | 23 +
docs/source/bag-api.rst | 77 +
docs/source/bag-creation.rst | 119 +
docs/source/bag-overview.rst | 96 +
docs/source/bag.rst | 13 +
docs/source/bytes.rst | 74 +
docs/source/caching.rst | 171 +
docs/source/changelog.rst | 426 +++
docs/source/cheatsheet.rst | 5 +
docs/source/cite.rst | 44 +
docs/source/conf.py | 300 ++
docs/source/custom-graphs.rst | 90 +
docs/source/daskcheatsheet.pdf | Bin 0 -> 262284 bytes
docs/source/dataframe-api.rst | 190 +
docs/source/dataframe-create.rst | 177 +
docs/source/dataframe-design.rst | 113 +
docs/source/dataframe-groupby.rst | 101 +
docs/source/dataframe-overview.rst | 158 +
docs/source/dataframe.rst | 31 +
docs/source/delayed-api.rst | 11 +
docs/source/delayed-collections.rst | 39 +
docs/source/delayed-overview.rst | 119 +
docs/source/delayed.rst | 15 +
docs/source/develop.rst | 239 ++
docs/source/diagnostics.rst | 295 ++
docs/source/distributed.rst | 8 +
docs/source/examples-tutorials.rst | 75 +
docs/source/examples/array-extend.rst | 102 +
docs/source/examples/array-hdf5.rst | 29 +
docs/source/examples/array-numpy.rst | 20 +
docs/source/examples/array-random.rst | 12 +
docs/source/examples/bag-json.rst | 17 +
docs/source/examples/bag-word-count-hdfs.rst | 210 ++
docs/source/examples/dataframe-csv.rst | 72 +
docs/source/examples/dataframe-hdf5.rst | 83 +
docs/source/examples/delayed-array.rst | 49 +
docs/source/examples/delayed-custom.rst | 40 +
docs/source/faq.rst | 126 +
docs/source/graphs.rst | 86 +
docs/source/images/array.png | Bin 0 -> 13488 bytes
docs/source/images/array.svg | 421 +++
docs/source/images/async-embarrassing.gif | Bin 0 -> 521955 bytes
docs/source/images/collections-schedulers.png | Bin 0 -> 24588 bytes
docs/source/images/collections-schedulers.svg | 520 +++
docs/source/images/crosstalk.png | Bin 0 -> 190750 bytes
docs/source/images/crosstalk.svg | 1868 ++++++++++
docs/source/images/dask-array-black-text.svg | 288 ++
docs/source/images/dask-dataframe.svg | 225 ++
docs/source/images/dask_horizontal.svg | 1 +
docs/source/images/dask_horizontal_white.svg | 1 +
docs/source/images/dask_icon.svg | 1 +
docs/source/images/dask_stacked.svg | 1 +
docs/source/images/dask_stacked_white.svg | 1 +
docs/source/images/delayed-inc-double-add.svg | 328 ++
docs/source/images/distributed-layout.png | Bin 0 -> 37459 bytes
docs/source/images/distributed-layout.svg | 437 +++
docs/source/images/distributed-network.png | Bin 0 -> 36839 bytes
docs/source/images/distributed-network.svg | 678 ++++
docs/source/images/frame-shuffle.png | Bin 0 -> 27033 bytes
docs/source/images/frame-shuffle.svg | 438 +++
docs/source/images/frame-sort.png | Bin 0 -> 9225 bytes
docs/source/images/frame-sort.svg | 516 +++
docs/source/images/frame.png | Bin 0 -> 22041 bytes
docs/source/images/frame.svg | 275 ++
docs/source/images/ghost.svg | 937 +++++
docs/source/images/ghosted-blocks.png | Bin 0 -> 4275 bytes
docs/source/images/ghosted-blocks.svg | 444 +++
docs/source/images/ghosted-neighbors.png | Bin 0 -> 1972 bytes
docs/source/images/ghosted-neighbors.svg | 129 +
docs/source/images/optimize_dask1.png | Bin 0 -> 89539 bytes
docs/source/images/optimize_dask1.svg | 278 ++
docs/source/images/optimize_dask2.png | Bin 0 -> 74766 bytes
docs/source/images/optimize_dask2.svg | 198 ++
docs/source/images/optimize_dask3.png | Bin 0 -> 48939 bytes
docs/source/images/optimize_dask3.svg | 148 +
docs/source/images/optimize_dask4.png | Bin 0 -> 44299 bytes
docs/source/images/optimize_dask4.svg | 123 +
docs/source/images/optimize_dask5.png | Bin 0 -> 19515 bytes
docs/source/images/optimize_dask5.svg | 43 +
docs/source/images/pipeline.png | Bin 0 -> 39275 bytes
docs/source/images/scaling-edges.png | Bin 0 -> 50209 bytes
docs/source/images/scaling-nodes.png | Bin 0 -> 68476 bytes
docs/source/images/simple-dask.png | Bin 0 -> 21294 bytes
docs/source/images/trivial.png | Bin 0 -> 122177 bytes
docs/source/images/trivial.svg | 1643 +++++++++
docs/source/images/unghosted-neighbors.png | Bin 0 -> 546 bytes
docs/source/images/unghosted-neighbors.svg | 76 +
docs/source/index.rst | 272 ++
docs/source/inspect.rst | 59 +
docs/source/install.rst | 68 +
docs/source/logos.rst | 15 +
docs/source/optimize.rst | 335 ++
docs/source/scheduler-choice.rst | 115 +
docs/source/scheduler-overview.rst | 175 +
docs/source/scheduling-policy.rst | 84 +
docs/source/scripts/scheduling.py | 110 +
docs/source/shared.rst | 150 +
docs/source/spark.rst | 184 +
docs/source/spec.rst | 153 +
docs/source/support.rst | 32 +
docs/source/use-cases.rst | 406 +++
setup.cfg | 29 +
setup.py | 37 +
versioneer.py | 1774 ++++++++++
279 files changed, 65662 insertions(+)
diff --git a/.coveragerc b/.coveragerc
new file mode 100644
index 0000000..be2863c
--- /dev/null
+++ b/.coveragerc
@@ -0,0 +1,6 @@
+[run]
+omit =
+ */test_*.py
+ dask/compatibility.py
+source =
+ dask
diff --git a/.github/CONTRIBUTING.md b/.github/CONTRIBUTING.md
new file mode 100644
index 0000000..2578981
--- /dev/null
+++ b/.github/CONTRIBUTING.md
@@ -0,0 +1,2 @@
+See [developer documentation](http://dask.pydata.org/en/latest/develop.html)
+for tips on how to get started.
diff --git a/.gitignore b/.gitignore
new file mode 100644
index 0000000..f7c7996
--- /dev/null
+++ b/.gitignore
@@ -0,0 +1,13 @@
+*.pyc
+*.egg-info
+docs/build
+build/
+dist/
+.idea/
+log.*
+log
+.coverage
+.DS_Store
+*.swp
+*.swo
+.cache/
diff --git a/.travis.yml b/.travis.yml
new file mode 100644
index 0000000..502a764
--- /dev/null
+++ b/.travis.yml
@@ -0,0 +1,57 @@
+language: python
+sudo: false
+
+env:
+ matrix:
+ - PYTHON=2.7 NUMPY=1.10.4 PANDAS=0.19.0 COVERAGE='true' PARALLEL='false' XTRATESTARGS=
+ - PYTHON=2.7 NUMPY=1.11.0 PANDAS=0.18.1 COVERAGE='false' PARALLEL='true' PYTHONOPTIMIZE=2 XTRATESTARGS=--ignore=dask/diagnostics
+ - PYTHON=3.3 NUMPY=1.9.2 PANDAS=0.18.1 COVERAGE='false' PARALLEL='true' XTRATESTARGS=
+ - PYTHON=3.4 NUMPY=1.10.4 PANDAS=0.18.0 COVERAGE='false' PARALLEL='true' PYTHONOPTIMIZE=2 XTRATESTARGS=--ignore=dask/diagnostics
+ - PYTHON=3.5 NUMPY=1.11.0 PANDAS=0.19.0 COVERAGE='false' PARALLEL='true' XTRATESTARGS=
+
+addons:
+ apt:
+ packages:
+ - graphviz
+ - liblzma-dev
+
+install:
+ # Install conda
+ - wget http://repo.continuum.io/miniconda/Miniconda-latest-Linux-x86_64.sh -O miniconda.sh
+ - bash miniconda.sh -b -p $HOME/miniconda
+ - export PATH="$HOME/miniconda/bin:$PATH"
+ - conda config --set always_yes yes --set changeps1 no
+ - conda update conda
+
+ # Install dependencies
+ - conda create -n test-environment python=$PYTHON
+ - source activate test-environment
+ - conda install -c conda-forge numpy=$NUMPY scipy pytables h5py bcolz pytest coverage toolz scikit-learn cytoolz chest blosc cython psutil ipython
+ - if [[ $PYTHON != '3.3' ]]; then conda install -c conda-forge pandas=$PANDAS distributed cloudpickle bokeh python-lmdb; fi
+ - if [[ $PYTHON != '3.3' ]]; then pip install git+https://github.com/dask/zict --upgrade --no-deps; fi
+ - if [[ $PYTHON != '3.3' ]]; then pip install git+https://github.com/dask/distributed --upgrade --no-deps; fi
+ - if [[ $PYTHON == '3.3' ]]; then pip install cloudpickle pandas==$PANDAS; fi
+ - if [[ $PYTHON == '2.7' ]]; then pip install backports.lzma mock; fi
+ - pip install git+https://github.com/mrocklin/partd --upgrade
+ - pip install git+https://github.com/mrocklin/cachey --upgrade
+ - pip install blosc --upgrade
+ - pip install graphviz moto flake8
+ - if [[ $PYTHON < '3' ]]; then pip install git+https://github.com/Blosc/castra; fi
+ # For parallel testing (`-n` argument in XTRATESTARGS)
+ - pip install pytest-xdist
+
+ # Install dask
+ - pip install --no-deps -e .[complete]
+
+script:
+ # Need to make test order deterministic when parallelizing tests, hence PYTHONHASHSEED
+ # (see https://github.com/pytest-dev/pytest-xdist/issues/63)
+ - if [[ $PARALLEL == 'true' ]]; then export XTRATESTARGS="-n3 $XTRATESTARGS"; export PYTHONHASHSEED=42; fi
+ - if [[ $COVERAGE == 'true' ]]; then coverage run $(which py.test) dask --runslow --doctest-modules --verbose $XTRATESTARGS; else py.test dask --runslow --verbose $XTRATESTARGS; fi
+ - flake8 dask
+
+after_success:
+ - if [[ $coverage == 'true' ]]; then coverage report --show-missing; pip install coveralls ; coveralls ; fi
+
+notifications:
+ email: false
diff --git a/AUTHORS.md b/AUTHORS.md
new file mode 100644
index 0000000..6fbce19
--- /dev/null
+++ b/AUTHORS.md
@@ -0,0 +1,121 @@
+[Arve](http://arve0.github.io/)
+
+[Bartosz Telenczuk](https://github.com/btel)
+
+[Benjamin Regan-Kelley](https://github.com/minrk)
+
+[Blake Griffith](http://github.com/cowlicks)
+
+[Bradley McElroy](https://github.com/limx0)
+
+[Carlos De la Guardia](https://github.com/dela3499)
+
+[Casey Clements](https://github.com/caseyclements)
+
+[Christine Doig](https://github.com/chdoig)
+
+[Clark Fitzgerald](https://github.com/clarkfitzg)
+
+[Daniel Davis](https://github.com/wabu)
+
+[Erik Welch](http://github.com/eriknw/)
+
+[Erkka Rinne](https://github.com/ererkka)
+
+[Eyad Sibai](https://github.com/eyadsibai)
+
+[Gabriele Lanaro](http://gabrielelanaro.github.io/)
+
+[Hussain Sultan](https://github.com/hussainsultan)
+
+[Isaac Slaavitt](https://github.com/isms)
+
+[Israel Saeta Perez](https://github.com/dukebody)
+
+[jgoutin](https://github.com/JGoutin)
+
+[Jim Crist](http://jcrist.github.io/)
+
+[John Kirkham](https://github.com/jakirkham)
+
+[Jon Renner](https://github.com/jrenner/)
+
+[Joshua Corbin](https://github.com/jcorbin)
+
+[jslmann](https://github.com/jslmann)
+
+[Kristopher Overholt](https://github.com/koverholt)
+
+[Kurt Smith](https://github.com/kwmsmith)
+
+[Linwood Creekmore](https://github.com/linwoodc3)
+
+[Mariano Tepper](http://www.marianotepper.com.ar/)
+
+[Matthew Rocklin](http://matthewrocklin.com/)
+
+[Max Hutchinson](https://github.com/maxhutch)
+
+[Michael Heilman](https://github.com/mheilman)
+
+[Mike Graham](https://github.com/mikegraham)
+
+[Nicholaus Jackson](https://github.com/NoonienSoong/)
+
+[Nir](https://github.com/nirizr)
+
+[Olivier Grisel](http://ogrisel.com/)
+
+[Pedro Duarte](https://github.com/PedroMDuarte)
+
+[Peter Steinberg](https://github.com/PeterDSteinberg/)
+
+[Peter Quackenbush](https://github.com/thequackdaddy/)
+
+[Phillip Cloud](https://github.com/cpcloud)
+
+[Phil Elson](https://github.com/pelson)
+
+[Phillip Wolfram](https://github.com/pwolfram)
+
+[Rami Chowdhury](https://github.com/necaris)
+
+[Rich Postelnik](https://github.com/postelrich)
+
+[Rolando Espinoza](https://github.com/rolando)
+
+[Ruggero Turra](https://github.com/wiso)
+
+[Scott Sanderson](https://github.com/ssanderson)
+
+[Sigurd Spieckermann](https://github.com/sisp)
+
+[sinhrks](https://github.com/sinhrks)
+
+[Skipper Seabold](https://github.com/jseabold)
+
+[Stefan Seefeld](https://github.com/stefanseefeld)
+
+[Stephan Hoyer](http://stephanhoyer.com)
+
+[Stuart Owen](http://stuartowen.com/)
+
+[Thomas Smith](https://github.com/tgs)
+
+[Tom Augspurger](https://github.com/TomAugspurger)
+
+[Tomas Ostasevicius](https://github.com/to266)
+
+[Travis Oliphant](https://github.com/teoliphant)
+
+[Valentin Haenel](http://haenel.co/)
+
+[Vikhyat Korrapati](http://vikhyat.net/)
+
+[Vlad Frolov](https://github.com/frol)
+
+[Wesley Emeneker](http://github.com/nevermindewe/)
+
+[wikiped](https://github.com/wikiped)
+
+[Will Warner](https://github.com/electronwill/)
diff --git a/LICENSE.txt b/LICENSE.txt
new file mode 100644
index 0000000..0c01700
--- /dev/null
+++ b/LICENSE.txt
@@ -0,0 +1,28 @@
+Copyright (c) 2014-2015, Continuum Analytics, Inc. and contributors
+All rights reserved.
+
+Redistribution and use in source and binary forms, with or without modification,
+are permitted provided that the following conditions are met:
+
+Redistributions of source code must retain the above copyright notice,
+this list of conditions and the following disclaimer.
+
+Redistributions in binary form must reproduce the above copyright notice,
+this list of conditions and the following disclaimer in the documentation
+and/or other materials provided with the distribution.
+
+Neither the name of Continuum Analytics nor the names of any contributors
+may be used to endorse or promote products derived from this software
+without specific prior written permission.
+
+THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
+AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
+LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
+CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
+SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
+INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
+CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
+ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
+THE POSSIBILITY OF SUCH DAMAGE.
diff --git a/MANIFEST.in b/MANIFEST.in
new file mode 100644
index 0000000..fbd1ed7
--- /dev/null
+++ b/MANIFEST.in
@@ -0,0 +1,11 @@
+recursive-include dask *.py
+recursive-include docs *.rst
+
+include setup.py
+include README.rst
+include LICENSE.txt
+include MANIFEST.in
+
+prune docs/_build
+include versioneer.py
+include dask/_version.py
diff --git a/README.rst b/README.rst
new file mode 100644
index 0000000..4dede02
--- /dev/null
+++ b/README.rst
@@ -0,0 +1,28 @@
+Dask
+====
+
+|Build Status| |Coverage| |Doc Status| |Gitter| |Version Status|
+
+Dask is a flexible parallel computing library for analytics. See
+documentation_ for more information.
+
+
+LICENSE
+-------
+
+New BSD. See `License File <https://github.com/dask/dask/blob/master/LICENSE.txt>`__.
+
+.. _documentation: http://dask.pydata.org/en/latest/
+.. |Build Status| image:: https://travis-ci.org/dask/dask.svg
+ :target: https://travis-ci.org/dask/dask
+.. |Coverage| image:: https://coveralls.io/repos/dask/dask/badge.svg
+ :target: https://coveralls.io/r/dask/dask
+ :alt: Coverage status
+.. |Doc Status| image:: http://readthedocs.org/projects/dask/badge/?version=latest
+ :target: https://readthedocs.io/projects/dask/?badge=latest
+ :alt: Documentation Status
+.. |Gitter| image:: https://badges.gitter.im/Join%20Chat.svg
+ :alt: Join the chat at https://gitter.im/dask/dask
+ :target: https://gitter.im/dask/dask?utm_source=badge&utm_medium=badge&utm_campaign=pr-badge&utm_content=badge
+.. |Version Status| image:: https://img.shields.io/pypi/v/dask.svg
+ :target: https://pypi.python.org/pypi/dask/
diff --git a/appveyor.yml b/appveyor.yml
new file mode 100644
index 0000000..22660b4
--- /dev/null
+++ b/appveyor.yml
@@ -0,0 +1,39 @@
+# Environment loosely based on https://github.com/conda/conda/blob/master/appveyor.yml
+
+environment:
+
+ global:
+ # SDK v7.0 MSVC Express 2008's SetEnv.cmd script will fail if the
+ # /E:ON and /V:ON options are not enabled in the batch script intepreter
+ # See: http://stackoverflow.com/a/13751649/163740
+ CMD_IN_ENV: "cmd /E:ON /V:ON /C .\\continuous_integration\\run_with_env.cmd"
+
+ matrix:
+ # Since appveyor is quite slow, we only use a single configuration
+ - PYTHON: "3.5"
+ ARCH: "64"
+ NUMPY: "1.11"
+ # Note Pandas 0.19 crashes in test_split_apply_combine_on_series,
+ # see https://ci.appveyor.com/project/pitrou/dask/build/1.0.40#L316
+ PANDAS: "0.18"
+ CONDA_ENV: testenv
+
+init:
+ # Use AppVeyor's provided Miniconda: https://www.appveyor.com/docs/installed-software#python
+ - if "%ARCH%" == "64" set MINICONDA=C:\Miniconda35-x64
+ - if "%ARCH%" == "32" set MINICONDA=C:\Miniconda35
+ - set PATH=%MINICONDA%;%MINICONDA%/Scripts;%MINICONDA%/Library/bin;%PATH%
+
+install:
+ # Update to a known good conda
+ # (to workaround http://help.appveyor.com/discussions/problems/4910)
+ - conda install -q -y conda=4.2.9
+ - continuous_integration\\setup_conda_environment.cmd
+
+build_script:
+ - continuous_integration\\build.cmd
+
+test_script:
+ # %CMD_IN_ENV% is needed for distutils/setuptools-based tests
+ # on certain build configurations.
+ - "%CMD_IN_ENV% continuous_integration\\run_tests.cmd"
diff --git a/conftest.py b/conftest.py
new file mode 100644
index 0000000..35e83a4
--- /dev/null
+++ b/conftest.py
@@ -0,0 +1,15 @@
+import pytest
+
+
+def pytest_addoption(parser):
+ parser.addoption("--runslow", action="store_true", help="run slow tests")
+
+
+def pytest_runtest_setup(item):
+ if 'slow' in item.keywords and not item.config.getoption("--runslow"):
+ pytest.skip("need --runslow option to run")
+
+
+def pytest_ignore_collect(path, config):
+ if 'run_test.py' in str(path):
+ return True
diff --git a/continuous_integration/build.cmd b/continuous_integration/build.cmd
new file mode 100644
index 0000000..1065404
--- /dev/null
+++ b/continuous_integration/build.cmd
@@ -0,0 +1,6 @@
+call activate %CONDA_ENV%
+
+ at echo on
+
+ at rem Install Dask
+%PIP_INSTALL% --no-deps -e .[complete]
diff --git a/continuous_integration/run_tests.cmd b/continuous_integration/run_tests.cmd
new file mode 100644
index 0000000..803bbd9
--- /dev/null
+++ b/continuous_integration/run_tests.cmd
@@ -0,0 +1,11 @@
+call activate %CONDA_ENV%
+
+ at echo on
+
+set PYTHONFAULTHANDLER=1
+
+ at rem `--capture=sys` avoids clobbering faulthandler tracebacks on crash
+set PYTEST=py.test --capture=sys
+
+ at rem %PYTEST% -v --runslow dask\dataframe\tests\test_groupby.py
+%PYTEST% dask
diff --git a/continuous_integration/run_with_env.cmd b/continuous_integration/run_with_env.cmd
new file mode 100644
index 0000000..3a56e3e
--- /dev/null
+++ b/continuous_integration/run_with_env.cmd
@@ -0,0 +1,90 @@
+:: From https://github.com/ogrisel/python-appveyor-demo
+::
+:: To build extensions for 64 bit Python 3, we need to configure environment
+:: variables to use the MSVC 2010 C++ compilers from GRMSDKX_EN_DVD.iso of:
+:: MS Windows SDK for Windows 7 and .NET Framework 4 (SDK v7.1)
+::
+:: To build extensions for 64 bit Python 2, we need to configure environment
+:: variables to use the MSVC 2008 C++ compilers from GRMSDKX_EN_DVD.iso of:
+:: MS Windows SDK for Windows 7 and .NET Framework 3.5 (SDK v7.0)
+::
+:: 32 bit builds, and 64-bit builds for 3.5 and beyond, do not require specific
+:: environment configurations.
+::
+:: Note: this script needs to be run with the /E:ON and /V:ON flags for the
+:: cmd interpreter, at least for (SDK v7.0)
+::
+:: More details at:
+:: https://github.com/cython/cython/wiki/64BitCythonExtensionsOnWindows
+:: http://stackoverflow.com/a/13751649/163740
+::
+:: Author: Olivier Grisel
+:: License: CC0 1.0 Universal: http://creativecommons.org/publicdomain/zero/1.0/
+::
+:: Notes about batch files for Python people:
+::
+:: Quotes in values are literally part of the values:
+:: SET FOO="bar"
+:: FOO is now five characters long: " b a r "
+:: If you don't want quotes, don't include them on the right-hand side.
+::
+:: The CALL lines at the end of this file look redundant, but if you move them
+:: outside of the IF clauses, they do not run properly in the SET_SDK_64==Y
+:: case, I don't know why.
+ at ECHO OFF
+
+SET COMMAND_TO_RUN=%*
+SET WIN_SDK_ROOT=C:\Program Files\Microsoft SDKs\Windows
+SET WIN_WDK=c:\Program Files (x86)\Windows Kits\10\Include\wdf
+
+:: Extract the major and minor versions, and allow for the minor version to be
+:: more than 9. This requires the version number to have two dots in it.
+SET MAJOR_PYTHON_VERSION=%PYTHON:~0,1%
+IF "%PYTHON:~3,1%" == "." (
+ SET MINOR_PYTHON_VERSION=%PYTHON:~2,1%
+) ELSE (
+ SET MINOR_PYTHON_VERSION=%PYTHON:~2,2%
+)
+
+:: Based on the Python version, determine what SDK version to use, and whether
+:: to set the SDK for 64-bit.
+IF %MAJOR_PYTHON_VERSION% == 2 (
+ SET WINDOWS_SDK_VERSION="v7.0"
+ SET SET_SDK_64=Y
+) ELSE (
+ IF %MAJOR_PYTHON_VERSION% == 3 (
+ SET WINDOWS_SDK_VERSION="v7.1"
+ IF %MINOR_PYTHON_VERSION% LEQ 4 (
+ SET SET_SDK_64=Y
+ ) ELSE (
+ SET SET_SDK_64=N
+ IF EXIST "%WIN_WDK%" (
+ :: See: https://connect.microsoft.com/VisualStudio/feedback/details/1610302/
+ REN "%WIN_WDK%" 0wdf
+ )
+ )
+ ) ELSE (
+ ECHO Unsupported Python version: "%MAJOR_PYTHON_VERSION%"
+ EXIT 1
+ )
+)
+
+IF %ARCH% == 64 (
+ IF %SET_SDK_64% == Y (
+ ECHO Configuring Windows SDK %WINDOWS_SDK_VERSION% for Python %MAJOR_PYTHON_VERSION% on a 64 bit architecture
+ SET DISTUTILS_USE_SDK=1
+ SET MSSdk=1
+ "%WIN_SDK_ROOT%\%WINDOWS_SDK_VERSION%\Setup\WindowsSdkVer.exe" -q -version:%WINDOWS_SDK_VERSION%
+ "%WIN_SDK_ROOT%\%WINDOWS_SDK_VERSION%\Bin\SetEnv.cmd" /x64 /release
+ ECHO Executing: %COMMAND_TO_RUN%
+ call %COMMAND_TO_RUN% || EXIT 1
+ ) ELSE (
+ ECHO Using default MSVC build environment for 64 bit architecture
+ ECHO Executing: %COMMAND_TO_RUN%
+ call %COMMAND_TO_RUN% || EXIT 1
+ )
+) ELSE (
+ ECHO Using default MSVC build environment for 32 bit architecture
+ ECHO Executing: %COMMAND_TO_RUN%
+ call %COMMAND_TO_RUN% || EXIT 1
+)
diff --git a/continuous_integration/setup_conda_environment.cmd b/continuous_integration/setup_conda_environment.cmd
new file mode 100644
index 0000000..b14e24b
--- /dev/null
+++ b/continuous_integration/setup_conda_environment.cmd
@@ -0,0 +1,37 @@
+ at rem The cmd /C hack circumvents a regression where conda installs a conda.bat
+ at rem script in non-root environments.
+set CONDA=cmd /C conda
+set CONDA_INSTALL=%CONDA% install -q -y
+set PIP_INSTALL=pip install -q
+
+ at echo on
+
+ at rem Deactivate any environment
+call deactivate
+ at rem Display root environment (for debugging)
+conda list
+ at rem Clean up any left-over from a previous build
+conda remove --all -q -y -n %CONDA_ENV%
+
+ at rem Create test environment
+ at rem (note: no cytoolz as it seems to prevent faulthandler tracebacks on crash)
+conda create -n %CONDA_ENV% -q -y python=%PYTHON% pytest toolz
+
+call activate %CONDA_ENV%
+
+ at rem Install optional dependencies for tests
+%CONDA_INSTALL% numpy=%NUMPY% pandas=%PANDAS% cloudpickle distributed
+%CONDA_INSTALL% s3fs psutil pytables bokeh bcolz scipy h5py ipython
+
+%PIP_INSTALL% git+https://github.com/mrocklin/partd --upgrade
+%PIP_INSTALL% git+https://github.com/mrocklin/cachey --upgrade
+%PIP_INSTALL% blosc --upgrade
+%PIP_INSTALL% moto
+
+if %PYTHON% LSS 3.0 (%PIP_INSTALL% git+https://github.com/Blosc/castra)
+if %PYTHON% LSS 3.0 (%PIP_INSTALL% backports.lzma mock)
+
+ at rem Display final environment (for reproducing)
+%CONDA% list
+%CONDA% list --explicit
+python -m site
diff --git a/dask/__init__.py b/dask/__init__.py
new file mode 100644
index 0000000..6b9a7dc
--- /dev/null
+++ b/dask/__init__.py
@@ -0,0 +1,17 @@
+from __future__ import absolute_import, division, print_function
+
+from .core import istask
+from .context import set_options
+from .async import get_sync as get
+try:
+ from .delayed import do, delayed, value
+except ImportError:
+ pass
+try:
+ from .base import visualize, compute
+except ImportError:
+ pass
+
+from ._version import get_versions
+__version__ = get_versions()['version']
+del get_versions
diff --git a/dask/_version.py b/dask/_version.py
new file mode 100644
index 0000000..159adee
--- /dev/null
+++ b/dask/_version.py
@@ -0,0 +1,484 @@
+
+# This file helps to compute a version number in source trees obtained from
+# git-archive tarball (such as those provided by githubs download-from-tag
+# feature). Distribution tarballs (built by setup.py sdist) and build
+# directories (produced by setup.py build) will contain a much shorter file
+# that just contains the computed version number.
+
+# This file is released into the public domain. Generated by
+# versioneer-0.16 (https://github.com/warner/python-versioneer)
+
+"""Git implementation of _version.py."""
+
+import errno
+import os
+import re
+import subprocess
+import sys
+
+
+def get_keywords():
+ """Get the keywords needed to look up the version information."""
+ # these strings will be replaced by git during git-archive.
+ # setup.py/versioneer.py will grep for the variable names, so they must
+ # each be defined on a line of their own. _version.py will just call
+ # get_keywords().
+ git_refnames = "$Format:%d$"
+ git_full = "$Format:%H$"
+ keywords = {"refnames": git_refnames, "full": git_full}
+ return keywords
+
+
+class VersioneerConfig:
+ """Container for Versioneer configuration parameters."""
+
+
+def get_config():
+ """Create, populate and return the VersioneerConfig() object."""
+ # these strings are filled in when 'setup.py versioneer' creates
+ # _version.py
+ cfg = VersioneerConfig()
+ cfg.VCS = "git"
+ cfg.style = "pep440"
+ cfg.tag_prefix = ""
+ cfg.parentdir_prefix = "dask-"
+ cfg.versionfile_source = "dask/_version.py"
+ cfg.verbose = False
+ return cfg
+
+
+class NotThisMethod(Exception):
+ """Exception raised if a method is not valid for the current scenario."""
+
+
+LONG_VERSION_PY = {}
+HANDLERS = {}
+
+
+def register_vcs_handler(vcs, method): # decorator
+ """Decorator to mark a method as the handler for a particular VCS."""
+ def decorate(f):
+ """Store f in HANDLERS[vcs][method]."""
+ if vcs not in HANDLERS:
+ HANDLERS[vcs] = {}
+ HANDLERS[vcs][method] = f
+ return f
+ return decorate
+
+
+def run_command(commands, args, cwd=None, verbose=False, hide_stderr=False):
+ """Call the given command(s)."""
+ assert isinstance(commands, list)
+ p = None
+ for c in commands:
+ try:
+ dispcmd = str([c] + args)
+ # remember shell=False, so use git.cmd on windows, not just git
+ p = subprocess.Popen([c] + args, cwd=cwd, stdout=subprocess.PIPE,
+ stderr=(subprocess.PIPE if hide_stderr
+ else None))
+ break
+ except EnvironmentError:
+ e = sys.exc_info()[1]
+ if e.errno == errno.ENOENT:
+ continue
+ if verbose:
+ print("unable to run %s" % dispcmd)
+ print(e)
+ return None
+ else:
+ if verbose:
+ print("unable to find command, tried %s" % (commands,))
+ return None
+ stdout = p.communicate()[0].strip()
+ if sys.version_info[0] >= 3:
+ stdout = stdout.decode()
+ if p.returncode != 0:
+ if verbose:
+ print("unable to run %s (error)" % dispcmd)
+ return None
+ return stdout
+
+
+def versions_from_parentdir(parentdir_prefix, root, verbose):
+ """Try to determine the version from the parent directory name.
+
+ Source tarballs conventionally unpack into a directory that includes
+ both the project name and a version string.
+ """
+ dirname = os.path.basename(root)
+ if not dirname.startswith(parentdir_prefix):
+ if verbose:
+ print("guessing rootdir is '%s', but '%s' doesn't start with "
+ "prefix '%s'" % (root, dirname, parentdir_prefix))
+ raise NotThisMethod("rootdir doesn't start with parentdir_prefix")
+ return {"version": dirname[len(parentdir_prefix):],
+ "full-revisionid": None,
+ "dirty": False, "error": None}
+
+
+ at register_vcs_handler("git", "get_keywords")
+def git_get_keywords(versionfile_abs):
+ """Extract version information from the given file."""
+ # the code embedded in _version.py can just fetch the value of these
+ # keywords. When used from setup.py, we don't want to import _version.py,
+ # so we do it with a regexp instead. This function is not used from
+ # _version.py.
+ keywords = {}
+ try:
+ f = open(versionfile_abs, "r")
+ for line in f.readlines():
+ if line.strip().startswith("git_refnames ="):
+ mo = re.search(r'=\s*"(.*)"', line)
+ if mo:
+ keywords["refnames"] = mo.group(1)
+ if line.strip().startswith("git_full ="):
+ mo = re.search(r'=\s*"(.*)"', line)
... 66552 lines suppressed ...
--
Alioth's /usr/local/bin/git-commit-notice on /srv/git.debian.org/git/python-modules/packages/dask.git
More information about the Python-modules-commits
mailing list