Back in November, I posted a detailed entry about the environments I used to develop, test, and deploy my Django applications. Since then, I've made a lot of changes to my configuration that I feel have helped boost my productivity and effectiveness. I wanted to talk a little bit about these changes here, and invite readers to share ideas below in the comments about how they do things differently.
This article was written to build upon the original entry and highlight the changes. You may want to read that entry first to better understand what I refer to here.
I still use Ubuntu exclusively on my desktop, laptop, and servers. I've never been more convinced that the era of the Microsoft monopoly is coming to a close, and I never cease to be amazed by the work that the Ubuntu developer community puts out. I also still use the latest version of Eclipse to develop with, using Aptana, PyDev, and Subclipse. Another plugin that I've incorporated, however, is MercurialEclipse.
Yes, I've been won over by the dark side of distributed version control systems. There are several reasons that I've decided to switch from Subversion to Mercurial for my internal source control. One of the biggest reasons is speed. I was already noticing sluggishness in my big all-in-one SVN repo. I started researching alternatives, and found a lot of anecdotal evidence suggesting that distributed systems were much faster. This article by Robert Fendt came out after I switched, but it further reinforces that point. SVN lags behind the three major DVCS systems in most operations.
Another advantage that I wanted was the ability to make multiple "personal" commits to your local copy of the repository before pushing major changesets to the central repository (which itself is really just a convention in DVCS's).
Mercurial was the obvious choice for me because it was based on Python, the language I am hopelessly biased towards. (Edit: See comments below - Bazaar is also based on Python.) Creating a new repository in Mercurial takes all of a couple seconds. This turned out to be a major benefit to me. As I mentioned in my previous environment entry, I was organizing all of my projects under one SVN repo. With Mercurial, it was very easy to create separate repositories for each project which are automatically picked up by the Web interface.
I no longer use the
trunk convention because I've found it's easier for me without them. When I need to test something, I push my development repo to the my central server. I then pull from the testing server, update, and test. When I'm ready to deploy, I push any changes back to the central server and then use Fabric to pull the changes to the production server and update. I'll talk more on Fabric later in this post.
I've tried a couple of different ways to manage settings between the development, testing, and production environments. I've finally settled on a way using Python's config file parser that makes it completely automatic and invisible to me during regular use. First, I filled in my
settings.py file with all of my production settings. Then, I created a directory in my home folder on each of my environments called
.adoleo. Within the config file, I created a category called
environment and a setting called
type. For my development machines, I set the
dev. For testing, I set it to
test. For production, I set it to
prod. That way, I can read the config files with Python and automatically override the production settings with appropriate settings for my other environments. Here's an example:
The top of my config file contains my production settings like any other Django project would.
DEBUG = False TEMPLATE_DEBUG = DEBUG ADMINS = ( ('Brandon Konkle', 'firstname.lastname@example.org'), ) MANAGERS = ADMINS DATABASE_ENGINE = 'my_db_engine' DATABASE_NAME = 'my_db_name' DATABASE_USER = 'my_db_user' DATABASE_PASSWORD = 'my_db_password' DATABASE_HOST = '' DATABASE_PORT = ''
And continuing on as usual until I get down to the bottom of the settings file:
# Import config to determine environment, and then override prod settings import os import sys import ConfigParser config_loc = os.path.expanduser('~') + '/.adoleo/config.file' env_type = False if os.path.exists(config_loc): config = ConfigParser.SafeConfigParser() config.read(config_loc) try: env_type = config.get('environment', 'type') except: pass if env_type == 'dev': DEBUG = True DATABASE_ENGINE = 'my_dev_db_engine' DATABASE_NAME = 'my_dev_db_name' DATABASE_USER = 'my_dev_db_user' DATABASE_PASSWORD = 'my_dev_db_password' MEDIA_ROOT = '/my/dev/media/root' TEMPLATE_DIRS = ( '/my/dev/templates', ) elif env_type == 'test': DATABASE_USER = 'my_test_db_user' DATABASE_PASSWORD = 'my_test_db_password' MEDIA_ROOT = '/my/test/media/root/' TEMPLATE_DIRS = ( '/my/test/templates', )
This way, if the config file contains a
test indicator, I override the needed settings with values for that particular environment. If the value is
prod, or there is a problem reading the config file, it defaults to the production settings.
Development Database Storage
For awhile I ran a full-scale PostgreSQL server on each of my development machines. As I mentioned in my previous post, I had to run pg_dumps every time I needed to switch back and forth between one of the development machines. Also, if I needed to wipe and reinstall a dev machine, I had to recreate the PostgreSQL environment completely. Fixtures are one way to address this problem, but I ran into some issues that made them a bit tricky.
My solution was to use SQLite for development. I don't know why I didn't think of it earlier - it's so easy and simple! I go ahead and commit my local database file to the Mercurial repository, and then if I need to switch machines I can do a simple
hg pull && hg up on the new machine and can instantly use the current dev database.
Deploying to Testing and Production with Fabric
I've recently started using the excellent Fabric tool to deploy my projects to testing and production. Working with Mercurial was a challenge at first until I discovered the
-R option, which designates the repository to work with. In my
fabfile.py script I first set a few details based on the environment:
repo_user = 'mercurialuser' repo_pass = 'mercurialpassword' # Could also use getpass() repo_url = 'central.mercurialserver.com/reponame' def test(): config.fab_hosts = ['testing.server.com'] repo_loc = '/mercurial/test/repo/location' def production(): config.fab_hosts = ['production.server.com'] repo_loc = '/mercurial/prod/repo/location'
Then I use commands like this to deploy:
def deploy(): run('hg -R %s pull http://%s:%s@%s' % (repo_loc, repo_user, repo_pass, repo_url)) run('hg -R %s up' % repo_loc)
That's it for now. What tricks do you use for managing your Django environments?