NewsBlur/fabfile.py

134 lines
3.3 KiB
Python
Raw Normal View History

from fabric.api import abort, cd, env, get, hide, hosts, local, prompt
from fabric.api import put, require, roles, run, runs_once, settings, show, sudo, warn
from fabric.colors import red, green, blue, cyan, magenta, white, yellow
from boto.s3.connection import S3Connection
from boto.s3.key import Key
from fabric.contrib import django
import os, sys
django.settings_module('settings')
from django.conf import settings as django_settings
# =========
# = Roles =
# =========
env.user = 'conesus'
# env.hosts = ['www.newsblur.com', 'db01.newsblur.com', 'db02.newsblur.com', 'db03.newsblur.com']
env.roledefs ={
'app': ['www.newsblur.com'],
'db': ['db01.newsblur.com'],
'task': ['db02.newsblur.com', 'db03.newsblur.com'],
}
2010-09-08 18:30:33 -07:00
"""
Base configuration
"""
2010-09-08 18:30:33 -07:00
"""
Environments
"""
def app():
env.roles = ['app']
def db():
env.roles = ['db']
def task():
env.roles = ['task']
2010-09-08 18:30:33 -07:00
# ==========
# = Deploy =
# ==========
@roles('app')
def deploy():
with cd('~/newsblur'):
run('git pull')
2011-02-05 19:39:30 -05:00
run('./utils/restart')
@roles('app')
def deploy_full():
with cd('~/newsblur'):
run('git pull')
run('./manage.py migrate')
run('sudo supervisorctl restart gunicorn')
2010-12-16 16:25:13 -05:00
@roles('app')
def staging():
with cd('~/staging'):
run('git pull')
2011-01-07 18:44:36 -05:00
run('kill -HUP `cat /var/run/gunicorn/gunicorn_staging.pid`')
2010-12-16 16:25:13 -05:00
@roles('app')
def staging_full():
with cd('~/staging'):
run('git pull')
run('./manage.py migrate')
run('kill -HUP `cat /var/run/gunicorn/gunicorn_staging.pid`')
@roles('task')
def celery():
with cd('~/newsblur'):
run('git pull')
run('sudo supervisorctl restart celery')
run('tail logs/newsblur.log')
@roles('task')
def force_celery():
with cd('~/newsblur'):
run('git pull')
run('ps aux | grep celeryd | egrep -v grep | awk \'{print $2}\' | sudo xargs kill -9')
# run('sudo supervisorctl start celery && tail logs/newsblur.log')
# ===========
# = Backups =
# ===========
@roles('app')
def backup_mongo():
with cd('~/newsblur/utils/backups'):
run('./mongo_backup.sh')
@roles('db')
def backup_postgresql():
with cd('~/newsblur/utils/backups'):
run('./postgresql_backup.sh')
# ======
# = S3 =
# ======
2010-09-08 18:30:33 -07:00
ACCESS_KEY = django_settings.S3_ACCESS_KEY
SECRET = django_settings.S3_SECRET
BUCKET_NAME = django_settings.S3_BACKUP_BUCKET # Note that you need to create this bucket first
2010-09-08 18:30:33 -07:00
def save_file_in_s3(filename):
conn = S3Connection(ACCESS_KEY, SECRET)
bucket = conn.get_bucket(BUCKET_NAME)
k = Key(bucket)
k.key = filename
k.set_contents_from_filename(filename)
def get_file_from_s3(filename):
conn = S3Connection(ACCESS_KEY, SECRET)
bucket = conn.get_bucket(BUCKET_NAME)
k = Key(bucket)
k.key = filename
2010-09-08 18:30:33 -07:00
k.get_contents_to_filename(filename)
2010-09-08 18:30:33 -07:00
def list_backup_in_s3():
conn = S3Connection(ACCESS_KEY, SECRET)
bucket = conn.get_bucket(BUCKET_NAME)
2010-09-08 18:30:33 -07:00
for i, key in enumerate(bucket.get_all_keys()):
print "[%s] %s" % (i, key.name)
2010-09-08 18:30:33 -07:00
def delete_all_backups():
#FIXME: validate filename exists
conn = S3Connection(ACCESS_KEY, SECRET)
bucket = conn.get_bucket(BUCKET_NAME)
2010-09-08 18:30:33 -07:00
for i, key in enumerate(bucket.get_all_keys()):
print "deleting %s" % (key.name)
key.delete()