End of Life (EOL) for Python 2.7 is coming. Are you ready?

We all knew it was coming. Back in 2014 when Guido van Rossum, Python’s creator and principal author, made the announcement, January 1, 2020 seemed pretty far away. Now we’re less than a year out from Python 2.7’s sunset, after which there’ll be absolutely no more support from the core Python team.Many utilized projects pledge […]
Intake released on Conda-Forge

Intake is a package for cataloging, finding and loading your data. It has been developed recently by Anaconda, Inc., and continues to gain new features. To read general information about Intake and how to use it, please refer to the documentation. Until recently, Intake was only available for installation via Conda and the intake channel […]
Anaconda Distribution 2018.12 Released

We are changing versioning in Anaconda Distribution from a major/minor version scheme to a year.month scheme. We made this change to differentiate between the open source Anaconda Distribution and Anaconda Enterprise, our managed data science platform. Conda, will continue to use a major/minor versioning scheme. The number of 32-bit x86 Linux packages downloaded are a […]
Python Data Visualization 2018: Where Do We Go From Here?

This post is the third in a three-part series on the current state of Python data visualization and the trends that emerged from SciPy 2018. By James A. Bednar As we saw in Part I and Part II of this series, having so many separate Python visualization libraries to choose from often is confusing to […]
Intake for Cataloging Spark

Intake is an open source project for providing easy pythonic access to a wide variety of data formats, and a simple cataloging system for these data sources. Intake is a new project, and all are encouraged to try and comment on it. pySpark is the python interface to Apache Spark, a fast and general purpose cluster computing […]
Using Pip in a Conda Environment

Unfortunately, issues can arise when conda and pip are used together to create an environment, especially when the tools are used back-to-back multiple times, establishing a state that can be hard to reproduce. Most of these issues stem from that fact that conda, like other package managers, has limited abilities to control packages it did […]
Python Data Visualization 2018: Moving Toward Convergence

This post is the second in a three-part series on the current state of Python data visualization and the trends that emerged from SciPy 2018. In my previous post, I provided an overview of the myriad Python data visualization tools currently available, how they relate to each other, and their many differences. In this post […]
Understanding Conda and Pip

Conda and pip are often considered as being nearly identical. Although some of the functionality of these two tools overlap, they were designed and should be used for different purposes. Pip is the Python Packaging Authority’s recommended tool for installing packages from the Python Package Index, PyPI. Pip installs Python software packaged as wheels or […]
Deriving Business Value from Data Science Deployments

One of the biggest challenges facing organizations trying to derive value from data science and machine learning is deployment. In this post, we’l…
Python Data Visualization 2018: Why So Many Libraries?

This post is the first in a three-part series on the state of Python data visualization tools and the trends that emerged from SciPy 2018. By James A. Bednar At a special session of SciPy 2018 in Austin, representatives of a wide range of open-source Python visualization tools shared their visions for the future of […]