I just spent the last week giving birth to ecstasy, my first full-fledged, fully-distributed, fully-fancied-up Python project. It was an amazing adventure and I would like to share all the insights and knowledge I acquired over the last few days. Moreover, I’d like to spare you all the trouble I’ve had and headaches I got and tears I cried.
Description
It’s not that I hadn’t already written tons of stuff in Python and hadn’t already open-sourced multiple projects on Github, but this time I wanted all the fancy bells and whistles. With that I mean, and this is what I’ll show you how to do here:
- Automated unit tests with unittest/unittest2 and tox
- Easy distribution and installation with PyPI and pip
- Documentation with Sphinx and Napoleon, hosted on ReadTheDocs
- Continuous-Integration with Travis CI
- Code review with Code Climate
- Automated linting, code quality and style rating with Landscape.io
- Code coverage rating with Coveralls
- Gratipay (formerly Gittip)
- Gitter chat
- Badges and badges
- More badges
- Badges
I will be writing this article as if you hadn’t written anything yet, but had only an idea for your next project, which I’ll refer to as your_app. Nevertheless, it’s quite likely that you’re here right in the heat, looking for a solution to a sub-problem of this whole process and are thus not even reading this sentence but are somewhere below. So, since no-one is reading this sentence, I’ll just take this opportunity to say hi to my mom. Hi.
But on a more serious note, this article can be followed in a linear way, but needn’t be. Also note that I will be sharing valuable resources regarding certain topics that I used to acquire the knowledge I’m sharing with you and that you may also want to laugh at. I mean read. Regarding the "level", I'll try to be more explicit and explain more rather than not. You can always skip what you already know.
We’ll start with a basic directory structure and then add more and more layers of cream and cherries and chocolate, spray some sugar and love on it, spread caramel and marshmallow all over and lastly have it blessed by the almighty Octocat after which we’ll unite in a wild frenzy of excitement and exhilaration and dance around a fire. Sounds good?
Getting Started
A basic directory structure can look like follows:
your_app/
├── docs
├── (examples)
├── tests
└── your_app
- your_app/ is the top-level directory, containing your entire project.
- your_app/your_app/ contains your application’s source code.
- your_app/docs/ includes everything related to documentation.
- your_app/examples/ could include small examples of how to use your application. I put this in parentheses because I haven’t seen that many projects on Github with such an examples directory. Personally I quite like the idea of it, but you could also argue that you’d rather put examples into the documentation. Your call.
- Make sure you have pip so we can get any and all packages we want.
- Initialize a local git repository.
- Setup a virtual environment (virtualenv).
The Cheese Shop
pip is a command-line utility, more precisely a “package-manager”, for installing Python packages from PyPI — the Python Package Index (pronounced pie-pee-ie) — also referred to as The Butchery … no wait, that doesn’t seem right. Also referred to as The Cheese Shop. Yes, better. But if you’ve ever used npm for Node.js, RubyGems for Ruby or Bower for Javascript and all things web, then you know what’s up. Either way, we will need it for many, many things so make sure you have pip holstered and ready to fire.
Installing it is easy if your Python distribution’s version ($ python -- version) is 2.7.9 or 3.4.0 (and higher): you don’t have to. For Python 2 < 2.7.9 and Python 3 < 3.4, you can use this command to get it off the command-line:
$ curl --silent --show-error https://bootstrap.pypa.io/get-pip.py | sudo python
Resources:
- http://stackoverflow.com/questions/5585875/what-is-the-official-preferred-way-to-install-pip-and-virtualenv-systemwide
- http://www.dabapps.com/blog/introduction-to-pip-and-virtualenv-python/
Git
At this point, it's a good idea to initialize a local git repository in your application's root directory (your_app/) to record whatever we are doing and save us in case we mess anything up. I won't tell you when to commit anything for the remainder of this article, that's up to you, but remember: commit early and often.
your_app$ git init
At the same time, grab a suitable .gitignore file for Python off Github:
your_app$ curl https://raw.githubusercontent.com/github/gitignore/master/Python.gitignore > .gitignore
Virtualenv
We'll want to run everything we do inside a virtual environment. What's that? A virtual environment is an isolated container or environment that we can create with the virtualenv command, that acts just like your normal, global system environment with one important difference: whatever we install with pip will only be installed for the virtual environment (which is really a directory) and not globally -- i.e. what happens in a virtualenv stays in a virtualenv. Thereby, we can have different requirements for different projects, not pollute our global environment and also keep track of exactly what we need to execute and test our application. This last bit is important, because whatever we need to install in our virtual environment, users and contributors will also have to install. Thus, in summary, a virtualenv, in combination with pip, will greatly ease tracking our project's requirements.
To install virtualenv, use pip:
$ sudo pip install virtualenv
To then create a virtual environment called env for our project, use this command inside your_app's directory:
your_app$ virtualenv env
This creates a virtualenv directory env (already ignored by .gitignore). To activate the virtualenv, you would:
your_app$ source env/bin/activate
Once you're done, you can deactivate the virtual environment via:
your_app$ deactivate
And at last, a trick, coz ain't nobody got time for typing source env/bin/activate, add this to your .bash_profile or .bash_rc:
venv() { source ${1:-env}/bin/activate; }
So that, now, you can just type venv to activate your virtualenv called "env", and else call venv <env_name> if you decide to give your virtual environment a different name.
Resources:
Tests
I'll be taking the well-established approach of test-driven-development (TDD) here and first setup tests. In this section, I'll show you how to setup the tests directory, write a simple test module and test it with unittest2 (I'll explain why not the built-in unittest). We'll also use coverage to see how much of our code our tests really cover (execute). Regarding the actual source code: you have yours of course, but for the purpose of this tutorial I'll be writing an application that squares numbers. Yes I'd like 1 Nobel prize please and some fries, thanks.
Unit Tests
First of all, let's turn the tests/ directory into a package by touching an empty __init__.py file (note that we're in the virtual environment now):
(env)your_app$ touch tests/__init__.py
Then, we'll have one test module test_code.py that will test your_app's code, which we'll later on write. For this, we need unittest2. Unittest2 is a testing framework built on Python's own unittest package, but that is backported to Python 2.6. We'll use it because why not have that extra bit of support with no semantic changes (Note: there are many other testing frameworks out there, such as py.test or doctest, which are equally capable but have quite different semantics and flavors, which you can also check out). Get unittest2 with pip:
(env)your_app$ pip install unittest2
Then you can write your unit tests, for example tests/test_code.py (testing methods must start with 'test'):
import os import sys import unittest2 sys.path.insert(0, os.path.abspath("..")) import your_app.code class TestCode(unittest2.TestCase): def test_squares_positive_numbers_correctly(self): self.assertEqual(your_app.code.square(2), 4) def test_squares_negative_numbers_correctly(self): self.assertEqual(your_app.code.square(-3), 9)
Note that we have to make the parent directory along with our source package your_app visible to our test modules' path, which you can see in line 5. Now, to run our tests, we can use the unit2 command that came with unittest2 (which performs test discovery and will find our tests from the root your_app directory). The -v switch causes more verbose output and shows which tests run successfully and which fail. Note that you can also call unittest2 as an executable Python module, like you would unittest (if you've used it before), i.e.:
(env)your_app$ python -m unittest2
In any case, here's what unit2 does:
(env)your_app$ unit2 -v ====================================================================== ERROR: tests.test_code (unittest2.loader._FailedTest) ---------------------------------------------------------------------- Traceback (most recent call last): ImportError: Failed to import test module: tests.test_code Traceback (most recent call last): File "/Users/petergoldsborough/Documents/Projects/your_app/env/lib/python2.7/site-packages/unittest2/loader.py", line 456, in _find_test_path module = self._get_module_from_name(name) File "/Users/petergoldsborough/Documents/Projects/your_app/env/lib/python2.7/site-packages/unittest2/loader.py", line 395, in _get_module_from_name __import__(name) File "/Users/petergoldsborough/Documents/Projects/your_app/tests/test_code.py", line 7, in <module> import your_app.code ImportError: No module named code ---------------------------------------------------------------------- Ran 1 test in 0.001s FAILED (errors=1)
If all goes well, everything should fail. Of course: we haven't written any code yet (there isn't any module named your_app.code yet). Let's do that. After also creating an __init__.py file in your_app/your_app, the package that will contain all of our application's source code, we can write our main module your_app/code.py:
#!/usr/bin/env python # -*- coding: utf-8 -*- def square(x): return x * x
Now running the testing command from above should give us happy messages and make us smile:
(env)your_app$ unit2 -v test_squares_negative_numbers_correctly (tests.test_code.TestCode) ... ok test_squares_positive_numbers_correctly (tests.test_code.TestCode) ... ok ---------------------------------------------------------------------- Ran 2 tests in 0.000s OK
Coverage
It's a good idea to know about our code's test coverage, i.e. how much of our project's code we're testing (and thereby avoiding bugs). For this, we use the coverage utility, which describes itself as "[...] a tool for measuring code coverage of Python programs. It monitors your program, noting which parts of the code have been executed, then analyzes the source to identify code that could have been executed but was not." Grab coverage with pip:
(env)your_app$ pip install coverage
You use coverage by simply running your Python modules like you would with the python command, but substituting python with coverage run. For our tests, that means changing
(env)your_app$ python -m unittest2
to
(env)your_app$ coverage run -m unittest2
This will run the same code as the previous command, but now we can call coverage report (add the -m switch for uncovered line numbers). We'll want to specify which modules we want to have included in our report, because, at least for me, coverage also reports on coverage of modules in Python's site-packages directory. For this, we add the --include="./*" option. Now, this should happen:
(env)your_app$ coverage report -m --include="./*" Name Stmts Miss Cover Missing ------------------------------------------------- tests/__init__ 0 0 100% tests/test_code 10 0 100% your_app/__init__ 5 0 100% your_app/code 3 0 100% ------------------------------------------------- TOTAL 18 0 100%
Looks like we did a good job writing tests! What an achievement for such a large codebase!
Resources:
That's it for this section, here's what your current directory should look like if you're following this tutorial in a linear way:
├── docs
├── examples
├── temp
│ └── code.py
├── tests
│ ├── __init__.py
│ └── test_code.py
└── your_app
├── __init__.py
└── code.py
Documentation
Now that we have finished writing our enchanting code and also have our tests running perfectly, we can work on documentation. This section will cover setting up the docs/ directory, writing our code's docstrings ready to be parsed by the Sphinx documentation generation tool, using a special super-legible syntax enabled by the Napoleon extension to Sphinx, and lastly writing additional documentation in reStructuredText that will later go on ReadTheDocs. You can find the documentation I generated for my project here, to have a look what you'll get: http://ecstasy.readthedocs.org/
Sphinx
What is Sphinx? According to sphinx-doc.org: "Sphinx is a tool that makes it easy to create intelligent and beautiful documentation". Basically, Sphinx is the go-to solution for Python documentation. Anything to be parsed by Sphinx will have to be written in reStructuredText, a markup language like Markdown, but with, of course, an entirely different syntax.
Let's setup Sphinx. Download it with pip:
(env)your_app$ pip install sphinx
Then change into the docs/ directory. Fortunately, Sphinx comes with a command-line utility called 'sphinx-quickstart' which will let you setup your basic Sphinx configuration by asking you a small set of questions about your favorite food and the number of cats you have. From what I remember. In any case, invoking this command (from within your_app/docs/):
docs$ sphinx-quickstart
will trigger a setup wizard. You can configure your project's documentation to your liking or just stick with the defaults (press enter), but do answer yes ('y') when it asks about splitting your build directory into a build and source folder (question 2) and also yes when it asks about autodoc. The latter will ensure that Sphinx reads our source code's documentation (from its docstring). You'll definitely want a Makefile too, so please don't disable it (it's 'yes' by default). Here's my setup:
docs$ sphinx-quickstart Welcome to the Sphinx 1.3.1 quickstart utility. Please enter values for the following settings (just press Enter to accept a default value, if one is given in brackets). Enter the root path for documentation. > Root path for the documentation [.]: You have two options for placing the build directory for Sphinx output. Either, you use a directory "_build" within the root path, or you separate "source" and "build" directories within the root path. > Separate source and build directories (y/n) [n]: y Inside the root directory, two more directories will be created; "_templates" for custom HTML templates and "_static" for custom stylesheets and other static files. You can enter another prefix (such as ".") to replace the underscore. > Name prefix for templates and static dir [_]: The project name will occur in several places in the built documentation. > Project name: your_app > Author name(s): you Sphinx has the notion of a "version" and a "release" for the software. Each version can have multiple releases. For example, for Python the version is something like 2.5 or 3.0, while the release is something like 2.5.1 or 3.0a1. If you don't need this dual structure, just set both to the same value. > Project version: 0.1.0 > Project release [0.1.0]: If the documents are to be written in a language other than English, you can select a language here by its language code. Sphinx will then translate text that it generates into that language. For a list of supported codes, see http://sphinx-doc.org/config.html#confval-language. > Project language [en]: The file name suffix for source files. Commonly, this is either ".txt" or ".rst". Only files with this suffix are considered documents. > Source file suffix [.rst]: One document is special in that it is considered the top node of the "contents tree", that is, it is the root of the hierarchical structure of the documents. Normally, this is "index", but if your "index" document is a custom template, you can also set this to another filename. > Name of your master document (without suffix) [index]: Sphinx can also add configuration for epub output: > Do you want to use the epub builder (y/n) [n]: Please indicate if you want to use one of the following Sphinx extensions: > autodoc: automatically insert docstrings from modules (y/n) [n]: y > doctest: automatically test code snippets in doctest blocks (y/n) [n]: n > intersphinx: link between Sphinx documentation of different projects (y/n) [n]: n > todo: write "todo" entries that can be shown or hidden on build (y/n) [n]: y > coverage: checks for documentation coverage (y/n) [n]: y > pngmath: include math, rendered as PNG images (y/n) [n]: n > mathjax: include math, rendered in the browser by MathJax (y/n) [n]: y > ifconfig: conditional inclusion of content based on config values (y/n) [n]: y > viewcode: include links to the source code of documented Python objects (y/n) [n]: y A Makefile and a Windows command file can be generated for you so that you only have to run e.g. `make html' instead of invoking sphinx-build directly. > Create Makefile? (y/n) [y]: y > Create Windows command file? (y/n) [y]: n ...
Now, you'll want to fully enable Sphinx's ability to grab documentation out of your source code by editing the conf.py it created. Open it in your_app/docs/source/conf.py. At the very top, after the import statements, you'll see a commented line saying "# If extensions (or modules to document with autodoc) are in another directory ...". Comment out the single line of Python code below (with sys.path) and add the path to your_app's root directory, which is two relative directories up conf.py (see below).
While we're already in conf.py, let's do two more things. First of all, we'll want a minimum Sphinx version of 1.3 (second configuration option, needs_sphinx). Then, add 'sphinx.ext.napoleon' to the list of extension you see right below. I'll explain what that is in just a bit, but first confirm your conf.py looks like this now (the extensions list may vary according to the settings you selected, but napoleon should be there):
# ... sys.path.insert(0, os.path.abspath('../../')) # -- General configuration ------------------------------------------------ # If your documentation needs a minimal Sphinx version, state it here. needs_sphinx = '1.0' # ... extensions = [ 'sphinx.ext.autodoc', 'sphinx.ext.todo', 'sphinx.ext.coverage', 'sphinx.ext.mathjax', 'sphinx.ext.ifconfig', 'sphinx.ext.viewcode', 'sphinx.ext.napoleon' ]
Now, we can add documentation to our source code. Usually, the standard reStructuredText + Sphinx way of adding method, class, module, parameter and return value descriptions, exception specifications and other elements of documentation would look something like this (for your_app/your_app/code.py):
#!/usr/bin/env python # -*- coding: utf-8 -*- """ Your application's badass source code. """ def square(x): """ Squares a value. Equivalent to calling x**2. :param x: The value to square. :type x: int|float :returns: The squared value. :rtype: int|float """ return x * x
Look at that mess in square()'s docstring. The default reStructuredText syntax for docstrings is ugly and makes me cry. That's where napoleon comes in (the extension we added above). It lets us transform the above into the below:
#!/usr/bin/env python # -*- coding: utf-8 -*- """ Your application's badass source code. """ def square(x): """ Squares a value. Equivalent to calling x**2. Arguments: x (int|float): The value to square. Returns: The squared value. """ return x * x
Yes! Beauty! Of course, the output in Sphinx and ReadTheDocs will be the same, but if there's one thing I love then that's cats and source code documentation that looks good in the source code and in the output. You can read all about Napoleon and how to document your source code with it here. Note that this is not specifically "Napoleon Style", but actually the docstring layout recommended by the Google Style Guide for Python.
Next, we can have Sphinx generate documentation for us with the sphinx-apidoc command:
(env)docs$ sphinx-apidoc -o source ../your_app/
To the -o flag we must pass the output directory as its positional argument, in this case the 'source' directory Sphinx generated earlier. The last argument is then the package containing your source code, where Sphinx will pick up the documentation from your modules' docstrings. This will generate one .rst file per module in source. Then, you can use the Makefile to generate the actual HTML documentation. After building, you'll find an index.html file under your_app/docs/build/html.
(env)docs$ make html sphinx-build -b html -d build/doctrees source build/html Running Sphinx v1.3.1 loading pickled environment... done building [mo]: targets for 0 po files that are out of date building [html]: targets for 0 source files that are out of date updating environment: 0 added, 0 changed, 0 removed looking for now-outdated files... none found no targets are out of date. build succeeded. Build finished. The HTML pages are in build/html. (env)docs$ open build/html/index.html
Yey! Now it's really up to you regarding documentation -- the sky's the limit. I'll add some dummy text from www.lebowskiipsum.com and a picture of my cat to index.rst, just so you have an idea what could follow. In this case, I'll also delete your_app.rst and move its contents into modules.rst, because it's not worth having the extra file. Here are the now two remaining relevant files for your_app's documentation:
your_app/docs/source/index.rst:
your_app ======== \ Contents ======== * `Introduction`_ * `Unbaking a cake with your_app`_ * `kbyethx`_ * `Source Code`_ \ Introduction ============ Lebowski ipsum stay out of Malibu, Lebowski! Dolor sit amet, consectetur adipiscing elit praesent ac magna justo pellentesque ac lectus. Vee could do things you only dreamed of, Lebowski. Quis elit blandit fringilla a ut. Darkness warshed over the Dude— darker'n a black steer's tookus on a moonless prairie night. There was no bottom. Turpis praesent felis ligula, malesuada suscipit malesuada non, ultrices non. We want that money, Lebowski. Urna sed orci ipsum. Unbaking a cake with your_app ============================= #. Take cake out of oven #. Run your_app on it (may require root access rights) #. Extract eggs, flour, sugar etc. #. Cure cancer #. Profit kbyethx ======= .. image:: https://goo.gl/IpUmJn :alt: <3 :align: center Source Code =========== .. toctree:: :maxdepth: 2 modules
your_app/docs/source/modules.rst:
your_app ======== your_app.code ------------- .. automodule:: your_app.code :members: :undoc-members: :show-inheritance:
One last thing to mention is that Sphinx supports quite a few themes. The default one, which you should be seeing if you've followed the previous steps, is called "alabaster". Others include the one from ReadTheDocs and also some using Bootstrap. They're quite easy to setup and play around with. Note also that alabaster is very, very customizable (see more on its Github page).
That's it for documentation (we'll send it to ReadTheDocs once we have a GitHub repo). In the next section, we'll setup a few more files before we go live, first on PyPI and then GitHub and everywhere else.
Resources:
- http://gisellezeno.com/tutorials/sphinx-for-python-documentation.html
- http://thomas-cokelaer.info/tutorials/sphinx/rest_syntax.html#internal-and-external-links
- http://docutils.sourceforge.net/docs/ref/rst/restructuredtext.html
Additional Files
A proper repository needs a bunch of more files to be ready for distribution and publicity and fame.
These are:
- README.rst
- LICENSE
- HISTORY.rst
README.rst
The README should be clear. It is the main entry point to your project and should make clear what problem you are solving and how you are bettering the world; a basic usage example if your project happens to be a library; how to install your application; where to find documentation; the license you're using and whatever else you'd like and that makes sense. Have a look at these resources prior to writing a README:
- http://stackoverflow.com/questions/2304863/how-to-write-a-good-readme
- http://programmers.stackexchange.com/questions/199405/what-information-should-be-in-the-github-readme-md
- https://orchestrate.io/blog/2014/07/16/how-to-write-a-readme-worth-reading/
Note that your README must be in reStructuredText and not in Markdown, because PyPI will only parse reStructuredText.
LICENSE
A project without a license is implicitly closed-source. We don't want that and thus need a license. There are many options, though I always choose the MIT-license. It's permissive and lovely. One thing that's especially wonderful is that you can get your own, personal, free and ever-online MIT-license hosted under mit-license.org. See this repo for more information. Here's my license. Note that the license file doesn't need a file extension (common practice not to add one).
HISTORY.rst
This should be your changelog file (often called CHANGELOG.rst or CHANGE.rst), where you would add a list of your most important commits for every new version release.
Getting Ready for Distribution
We are very close to being able to upload our project to PyPI and then Github. However, now, we'll need a few more more files. These are very important:
- setup.py
- setup.cfg
- MANIFEST.in
- requirements.txt
- Makefile
# -*- coding: utf-8 -*- __title__ = 'your_app' __version__ = '0.1.0' __author__ = '<you>' __license__ = 'MIT' __copyright__ = 'Copyright 2015 <you>'
Setup.py
#!/usr/bin/env python # -*- coding: utf-8 -*- """ setup.py script for setuptools. """ import re from setuptools import setup, find_packages version = '' with open('your_app/__init__.py') as init: version = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', init.read(), re.MULTILINE).group(1) with open('README.rst') as readme: long_description = readme.read() requirements = [ # Your project's requirements ] test_requirements = [ "unittest2==1.1.0" ] setup( name='your_app', version=version, description='Your Python application.', long_description=long_description, author='You', author_email='your@email', license='MIT', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Topic :: Software Development', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], keywords='your_app', include_package_data=True, packages=find_packages(exclude=['contrib', 'docs', 'tests*']), install_requires=requirements, test_suite="tests", tests_require=test_requirements )
- Now, first of all, in setup.py, we must import the setup method from the setuptools module, which takes a great number of keyword arguments that enable us to configure our project's metadata. We'll also need setuptools find_packages to make it easy for us to specify which packages setup.py should handle.
- Then we grab our project's version string by doing a regular expression search in our __init__.py under the your_app/your_app directory (for this we need the re module). Note that we cannot directly import our module your_app and just retrieve its __version__ directly, as someone who just got your project and has not yet executed the setup procedure may not have all your project's dependencies. Therefore, certain imports may fail.
- For PyPI, we'll want a long_description, which is simply our README.
- Requirements/dependencies to run your project's application code successfully should be put into the requirements list. your_app doesn't have any, but I'm leaving it here for you to see. When someone then setup.py installs or pip installs your project, this will also install the dependencies.
- Test requirements go into the appropriately named list. your_app, for example, requires unittest2, version 1.1.0. Note that each element in these lists of requirements are commands for pip install, i.e. when setting up your project, pip install <element> will be called for each element in these lists.
- Inside setup(), you'll see a lot of meta-data arguments, which you can fill out for your persona and application
- Regarding classifiers, you can see the list of available classifiers you can choose from here.
- By setting include_package_data to True, we can later on specify which additional data files we want to include in our distribution. setup.py will then look for a manifest file (MANIFEST.in) which we'll write in a bit. By default, Python files are included, your README is included and a few other files.
- For the packages we want to distribute, we let setuptools' find_packages function handle that, as can be seen above.
- We specify install and test requirements with install_requires and test_requires.
- By handing over the path to our test directory ("tests") to the test_suite argument, we will later on be able to run our tests with unittest2 directly from the setup file (see below).
Once we're done, you'll be able to build your distribution files with:
(env)your_app$ python setup.py build sdist bdist_wheel
which builds a source and pre-built distribution (using the Python wheel format), and/or install your project with:
(env)your_app$ python setup.py install
and/or test your project with:
(env)your_app$ python setup.py test
and/or do many other things, such as registering your application with PyPI (more on that soon), which you can see by issuing:
(env)your_app$ python setup.py --help
Resources:
- https://hynek.me/articles/sharing-your-labor-of-love-pypi-quick-and-dirty/
- https://github.com/audreyr/cookiecutter-pypackage/blob/master/%7B%7Bcookiecutter.repo_name%7D%7D/setup.py
- https://packaging.python.org/en/latest/installing.html#installing-cached-wheels
setup.cfg
As quickly touched upon above, there are two main variants of distribution files for your project, the first is a source distribution (sdist) -- the standard distribution --- and the second is a wheel distribution (bdist_wheel). According to the documentation for wheels: "A wheel is a ZIP-format archive with a specially formatted filename and the .whl extension. It is designed to contain all the files for a PEP 376 compatible install in a way that is very close to the on-disk format. ". The main advantage of a wheel is that it is quicker to install than a source distribution. Note that you must have the wheel package installed to use wheels, but you should already have it if you're in a virtual environment. In any case, we can provide additional instructions on how to build wheels with a setup.cfg file. Most importantly, we'll want only one single wheel distribution for all Python versions. For this, setup.cfg should hold these two lines:
[wheel] universal=1
MANIFEST.in
In our setup.py script, we set include_package_data to True. As a consequence, setup.py will look for a MANIFEST.in file, where we can specify additional data files such as README.rst, LICENSE and others, that we want included in our distribution. It can look like this:
include *.rst include LICENSE include Makefile recursive-include tests * recursive-exclude * __pycache__ recursive-exclude * *.py[co] recursive-include docs *.rst conf.py Makefile
Resources:
- https://docs.python.org/3.4/distutils/sourcedist.html#specifying-the-files-to-distribute
- https://docs.python.org/3.4/distutils/sourcedist.html#the-manifest-in-template
- https://pythonhosted.org/setuptools/setuptools.html#including-data-files
requirements.txt
This file contains a list of pip packages needed to run your code successfully, explicitly put here alongside having them specified in setup.py. Each line in requirements.txt is a command to pip install. For example, we'll need unittest2 version 1.1.0 to run our tests, so requirements.txt would contain at least this one line: "unittest2==1.1.0". If you then run
$ pip install -r requirements.txt
pip will execute each line in requirements.txt as if you had called pip install <line>. The best way of getting all requirements for your project is to use:
(env)your_app$ pip freeze
which prints a list of packages installed with pip, either directly by us or as dependencies of those packages. Note that the value of the virtual environment really comes to show here, as pip freeze will only show what we installed for this specific project, whereas it would show an amalgamation of packages for different projects with different requirements if we hadn't used a virtual environment (and also not for other projects). In any case, for your_app, pip freeze shows this:
(env)your_app$ pip freeze alabaster==0.7.6 Babel==2.0 docutils==0.12 Jinja2==2.8 linecache2==1.0.0 MarkupSafe==0.23 pluggy==0.3.0 py==1.4.30 Pygments==2.0.2 pytz==2015.4 six==1.9.0 snowballstemmer==1.2.0 Sphinx==1.3.1 sphinx-rtd-theme==0.1.8 traceback2==1.4.0 unittest2==1.1.0 virtualenv==13.1.0 wheel==0.24.0
Note that most of these packages were installed as dependencies of the few packages we installed (unittest2, virtualenv and Sphinx -- wheel is installed in a virtualenv by default). We can then generate a requirements.txt file by directing pip freeze's output into it:
(env)your_app$ pip freeze > requirements.txt
Resources:
Makefile
We'll want a Makefile to make life easy for us. Mainly taken from here, but modified for our setup:
.PHONY: clean-pyc clean-build docs clean help: @echo "clean - remove all build, test, coverage and Python artifacts" @echo "clean-build - remove build artifacts" @echo "clean-pyc - remove Python file artifacts" @echo "clean-test - remove test and coverage artifacts" @echo "lint - check style with flake8" @echo "test - run tests quickly with the default Python" @echo "coverage - check code coverage quickly with the default Python" @echo "docs - generate Sphinx HTML documentation, including API docs" @echo "release - package and upload a release" @echo "dist - package" @echo "install - install the package to the active Python's site-packages" clean: clean-build clean-pyc clean-test clean-build: rm -rf build/ rm -rf dist/ rm -rf .eggs/ find . -name '*.egg-info' -exec rm -rf {} + find . -name '*.egg' -exec rm -f {} + clean-pyc: find . -name '*.pyc' -exec rm -f {} + find . -name '*.pyo' -exec rm -f {} + find . -name '*~' -exec rm -f {} + find . -name '__pycache__' -exec rm -rf {} + clean-test: rm -rf .tox/ rm -f .coverage rm -rf htmlcov/ lint: pylint ecstasy test: python setup.py test coverage: coverage run --include="./*" -m unittest2 coverage report -m coverage html open htmlcov/index.html docs: $(MAKE) -C docs clean $(MAKE) -C docs html open docs/build/html/index.html release: clean twine upload sdist twine upload bdist_wheel dist: clean python setup.py sdist python setup.py bdist_wheel install: clean python setup.py install
Tox
I'll just talk about one more thing before we send your_app online: tox. The tox automation project, according to itself, "aims to automate and standardize testing in Python". What's great about it is that it lets us test whether or not our project can be built and distributed correctly across multiple Python versions, without having to setup virtual environments ourselves (done by tox). I've waited till now for this, because we needed a setup.py script. Install tox with pip:
(env)your_app$ pip install tox
You'll then need a tox.ini configuration file, in which we can specify which python versions we'll want to test our distribution for, how to install requirements and what command to use to run tests. Note that what Python versions you can use tox with depends on which Python versions you have installed on your system, though you always grab more Python versions with pyenv or brew. tox.ini could look like this:
[tox] envlist = py26, py27, py31, py34 [testenv] deps = -rrequirements.txt commands = python setup.py test
Now you run tox to see if all is well for your application in multiple environments:
(env)your_app$ tox ... _________________________ summary __________________________ py26: commands succeeded py27: commands succeeded py31: commands succeeded py34: commands succeeded congratulations :)
PyPI
We are now ready to send our application to the cheese shop (PyPI). First, we'll send it to the test (staging) server and then to the real one. Either way, we'll want to build our project. Run:
(env)your_app$ python setup.py build sdist bdist_wheel
This creates the appropriate distributions in the your_app/dist/ directory. You may want to execute
your_app$ python setup.py check
before doing anything next, to make sure your project is ready (but don't care about the url for now, we'll add it later).
TestPyPI
Sign up for the staging server of PyPi here. Note that the staging server is cleaned periodically, so your user profile and repository may vanish from the staging server (not the real one after) at one point in the future. In any case, next, create a .pypirc file in your root directory. ~/pypirc should contain, for now (replace the username and password):
[distutils] index-servers = test [test] repository = https://testpypi.python.org/pypi username = <your_username> password = <your_password>
Then, you can register your application on the test server (note that this will warn about a missing url, we'll add it later):
(env)your_app$ python setup.py register -r test
Go have a look on testpypi.python.org if all looks well for your application on the testing server. If so, we can then upload the distribution files (it's only registered for now). For this, you can either run (-r stands for repository)
(env)your_app$ python setup.py sdist bdist_wheel upload -r test
or use twine. Twine is like the command you just saw, but more secure. You can even use PGP encryption. Get twine with pip
(env)your_app$ pip install twine
Then issue this command to upload your files (pass the -s flag to sign it with PGP):
twine upload -r test dist/*
Either way, you could now create a new virtual environment and test installing your package from the test server:
(env)your_app$ virtualenv test Using real prefix '/usr/local/Cellar/python/2.7.10_2/Frameworks/Python.framework/Versions/2.7' New python executable in test/bin/python2.7 Also creating executable in test/bin/python Installing setuptools, pip, wheel...done. (env)your_app$ venv test (test)your_app$ pip install -i https://testpypi.python.org/pypi your_app Collecting your-app Using cached https://testpypi.python.org/packages/2.7/y/your_app/your_app-0.1.0-py2.py3-none-any.whl Installing collected packages: your-app Successfully installed your-app-0.1.0 (test)your_app$ deactivate your_app$ rm -r test your_app$ venv env
PyPI
We can now go for the the real thing. Register on PyPI, if you haven't yet (separate registration from the test server). Then add change ~/.pypirc so that it looks like this:
[distutils] index-servers = pypi test [pypi] repository = https://pypi.python.org/pypi username = <your_username> password = <your_password> [test] repository = https://testpypi.python.org/pypi username = <your_username> password = <your_password>
And repeat the process for the live server:
your_app$ python setup.py register ... Registering your_app to https://pypi.python.org/pypi Server response (200): OK
and then either
your_app$ python setup.py sdist bdist_wheel upload
or
your_app$ twine upload dist/*
Voilà , c'est on PyPI, as my French dog would always say. If I had a French dog. Who could speak.
Resources:
- https://gist.github.com/audreyr/5990987
- https://hynek.me/articles/sharing-your-labor-of-love-pypi-quick-and-dirty/
Github
Register your project on Github now -- I'm sure you know how to do it. Once you have a URL, add it to your_app/your_app/__init__.py and the setup.py script. The url for this tutorial repository is https://github.com/goldsborough/your_app and I'll use it as a placeholder for your project's url from now on.
your_app/your_app/__init__.py:
... __title__ = 'your_app' __url__ = "https://github.com/goldsborough/your_app" __version__ = '0.1.0' ...
setup.py:
setup( ... author='You', author_email='your@email', url="https://github.com/goldsborough/your_app", license='MIT', ... )
Now that we have a little bit more meta-data, we can re-register the project on PyPI. This will only re-configure the meta data:
your_app$ python setup.py register
Adding the PyPI badge
First, we can add a "PyPI version" badge to our project. For this, visit this web page. There, you can enter the name of your application and retrieve the reStructuredText code to include in your README.rst file. For your_app, that would be:
.. image:: https://badge.fury.io/py/your_app.svg :target: http://badge.fury.io/py/your_app
On Github, your README will then show a badge like this:
In the target command, you would replace goldsborough with your name for mit-license.org. That gives you this badge:
You can change the color by making your own badge on http://shields.io.
Gratipay, formerly known under the, IMHO, much nicer name, Gittip, is a way for people to tip you for your great work. Sign up here: https://gratipay.com/
And either go to https://gratipay.com/~<your_user_name>/widgets/ or just change this code for you:
.. image:: http://img.shields.io/gratipay/goldsborough.svg
:target: https://gratipay.com/~goldsborough/
This will give you this badge (feel free to click and donate <3):
Landscape.io is a web tool for automated linting, code correctness and style rating. Everytime you push a commit to your repo, landscape.io will run pylint on it. Register for free and add your repository, no additional configuration required. On your repository's overview page on landscape.io, you'll see the badge in the top right corner. Clicking on it redirects you to badge heaven. The badge looks like this (with your code health):
Code Climate is another web application that performs code rating, or more specifically "static analysis". It differs from Landscape.io in that it really analyzes your code for quality in terms of how well it is structured and how maintainable is, rather than only checking for bugs and style as Landscape.io does. For example, for my latest project, ecstasy, it's giving me a pretty low GPA (2.8) because I have "high cyclomatic complexity" in certain areas. At the same, your_app gets a 4.0 (hehe). Sign up on their website and add your repository. You can get this kind of badge:
by visiting Settings > Badges in your Code Climate dashboard and adding the reStructuredText to your README.rst.
At the start of this article, I talked about code coverage. We examined our code's coverage with the command-line utility of the same name to understand how much of our code is really being executed when running our tests. Coveralls.io is the web counterpart for this -- sign up there and add your repository. Once you're done, you can see the badge being offered at the top. It looks like this:
However, there's a little extra configuration needed here. First of all, download python-coveralls with pip (and do a pip freeze > requirements.txt again, and also add it to your test_requirements in setup.py):
This utility will let us interact with coveralls from the command-line. Then, add this .coveralls.yml file to your repository (we'll setup Travis-CI in a bit):
your_app/.coveralls.yml:
Once you've added your repository on coveralls.io, you'll see setup-instructions for coveralls and Ruby. We don't need the instructions, but you'll see your repo_token in the instructions. Copy it to your clipboard and issue this command, replacing <your_repo_token> with your actual repo token:
That'll and you should see results once we've setup Travis CI and pushed a commit afterwards.
Resources:
Travis CI (Continuous Integration) is an awesomely awesome distributed build system for continuous integration in the open-source community, used for running your tests online. It is very, very helpful when working in teams or on open-source projects with multiple contributors working on improvements and features and fixes simultaneously. Sign up on travis-ci.org, the free version for open-source projects, and not on travis-ci.com, their paid counterpart. After adding your project, you'll see the badge in your overview page. Clicking on it and switching to 'rst' yields the reStructuredText code you can add to your README.rst to get this badge:
More configuration. Add this .travis.yml file to your root directory.
your_app/.travis.yml:
Here, we specify
There's more you can do with this configuration file. The full life cycle of a Travis "job" is described by these YAML statements:
See here: http://docs.travis-ci.com/user/customizing-the-build/#The-Build-Lifecycle
Maybe you'd also like a command-line client: https://github.com/travis-ci/travis.rb
In any case, you should now have Travis-CI and coveralls.io ready, so commit and push and check everything is wonderful on travis-ci.org and on coveralls.io.
Resources:
Adding your MIT License badge
If you're awesome and have your own mit-license.org subdomain (I talked about it earlier), you can also get a badge for that, by including the following code in your README.rst:
.. image:: https://img.shields.io/github/license/mashape/apistatus.svg :target: http://goldsborough.mit-license.org
In the target command, you would replace goldsborough with your name for mit-license.org. That gives you this badge:
You can change the color by making your own badge on http://shields.io.
Adding the Gratipay badge
Gratipay, formerly known under the, IMHO, much nicer name, Gittip, is a way for people to tip you for your great work. Sign up here: https://gratipay.com/
And either go to https://gratipay.com/~<your_user_name>/widgets/ or just change this code for you:
.. image:: http://img.shields.io/gratipay/goldsborough.svg
:target: https://gratipay.com/~goldsborough/
This will give you this badge (feel free to click and donate <3):
Setting up Landscape.io
Landscape.io is a web tool for automated linting, code correctness and style rating. Everytime you push a commit to your repo, landscape.io will run pylint on it. Register for free and add your repository, no additional configuration required. On your repository's overview page on landscape.io, you'll see the badge in the top right corner. Clicking on it redirects you to badge heaven. The badge looks like this (with your code health):
Setting up Code Climate
Code Climate is another web application that performs code rating, or more specifically "static analysis". It differs from Landscape.io in that it really analyzes your code for quality in terms of how well it is structured and how maintainable is, rather than only checking for bugs and style as Landscape.io does. For example, for my latest project, ecstasy, it's giving me a pretty low GPA (2.8) because I have "high cyclomatic complexity" in certain areas. At the same, your_app gets a 4.0 (hehe). Sign up on their website and add your repository. You can get this kind of badge:
by visiting Settings > Badges in your Code Climate dashboard and adding the reStructuredText to your README.rst.
Setting up Coveralls
At the start of this article, I talked about code coverage. We examined our code's coverage with the command-line utility of the same name to understand how much of our code is really being executed when running our tests. Coveralls.io is the web counterpart for this -- sign up there and add your repository. Once you're done, you can see the badge being offered at the top. It looks like this:
However, there's a little extra configuration needed here. First of all, download python-coveralls with pip (and do a pip freeze > requirements.txt again, and also add it to your test_requirements in setup.py):
(env)your_app$ pip install python-coveralls
This utility will let us interact with coveralls from the command-line. Then, add this .coveralls.yml file to your repository (we'll setup Travis-CI in a bit):
your_app/.coveralls.yml:
service_name: travis-ci
Once you've added your repository on coveralls.io, you'll see setup-instructions for coveralls and Ruby. We don't need the instructions, but you'll see your repo_token in the instructions. Copy it to your clipboard and issue this command, replacing <your_repo_token> with your actual repo token:
(env)your_app$ COVERALLS_REPO_TOKEN=<your_repo_token> coveralls
That'll and you should see results once we've setup Travis CI and pushed a commit afterwards.
Resources:
Setting up Travis-CI
Travis CI (Continuous Integration) is an awesomely awesome distributed build system for continuous integration in the open-source community, used for running your tests online. It is very, very helpful when working in teams or on open-source projects with multiple contributors working on improvements and features and fixes simultaneously. Sign up on travis-ci.org, the free version for open-source projects, and not on travis-ci.com, their paid counterpart. After adding your project, you'll see the badge in your overview page. Clicking on it and switching to 'rst' yields the reStructuredText code you can add to your README.rst to get this badge:
More configuration. Add this .travis.yml file to your root directory.
your_app/.travis.yml:
language: python python: - "2.6" - "2.7" - "3.2" - "3.3" - "3.4" install: - "pip install ." - "pip install -r requirements.txt" script: coverage run --include="./*" -m unittest2 notifications: email: - petergoldsborough@hotmail.com after_success: - coveralls
Here, we specify
- language: The language of our project ("python")
- python: Against which Python versions to test (e.g. "3.4")
- install: What shell commands to run to install the project
- script: What shell commands to run to execute our tests. Notice that we use coverage here like I showed you above. This is important for the after_success step and integration with coveralls.io. If you don't want coverage, you can just execute the "unit2" (make sure unittest2 is in your dependencies).
- notifications: How and where to notify you of failures and changes (you can change this on travis-ci.org)
- after_success: What to do when all our tests have passed (there's also after_failure). We want to call coveralls here, which will send the report generated by the coverage report to coveralls.io (python-coveralls interacts with the coveralls.io API).
There's more you can do with this configuration file. The full life cycle of a Travis "job" is described by these YAML statements:
- before_install
- install
- after_install
- before_script
- script
- after_script
- after_success or after_failure
Maybe you'd also like a command-line client: https://github.com/travis-ci/travis.rb
In any case, you should now have Travis-CI and coveralls.io ready, so commit and push and check everything is wonderful on travis-ci.org and on coveralls.io.
Resources:
Setting up Gitter
The last thing we'll want to do for our kick-ass project is to add Gitter. Gitter is a chatroom for Github repositories where you can talk with fellow contributors about your project. Sign up on gitter.im and click on "create a room" in the bottom left corner, then on "repository" and lastly on your Github repo. This will create a room on Gitter. Say hi on your_app's: https://gitter.im/goldsborough/your_app
Now you'll want to setup Gitter and integrate other services. If you're in your room on Gitter, it'll say "Configure your integrations" on the right sidebar. Click that and then integrate GitHub, Coveralls and Travis CI. There are nice instructions for all of that.
Once you're done you have to patience yourself a little, but Gitter will send you a Pull Request on GitHub to suggest adding its marvelous badge, which looks like so (click on it to shoot me a message in your_app's chatroom):
Setting up ReadTheDocs
The very last thing we'll do, and I almost forgot about it, is setup ReadTheDocs. Sign up on https://readthedocs.org and click on "Add Project" under your user profile. You can then select to "Import from Github" and, after possibly syncing your repos, you can select your Github repo. You'll then have to "Build" it and ReadTheDocs will automagically search your docs/ folder for everything it needs. After, you can click on "View Docs" and voilà , your Sphinx docs. Here are those from your_app: http://your-app.readthedocs.org/
For every <thing>.readthedocs.org there's also an alternative domain under <thing>.rtfd.org (read the fabulous docs -- right?), like your_app.rtfd.org
Farewell
I hope I could share with you what I learnt while setting up my own project. Creating and distributing such a project can be quite tough and there's so, so much to do when setting up a project other than coding (unfortunately). There's most likely many things I did wrong here, too, so please feel free to offer suggestions.
One last resource I'd like to share with you is cookiecutter, which automates much of the process I just outlined.
But otherwise, thanks for reading and farewell!