Skip to content


Subversion checkout URL

You can clone with
Download ZIP
Fetching contributors…
Cannot retrieve contributors at this time
227 lines (178 sloc) 7.65 KB
#!/usr/bin/env python
# sample usage:
import pickle, os, sys, logging, time, urllib2, re
from optparse import OptionParser, OptionValueError
from smtplib import SMTP
from getpass import getuser
from socket import gethostname
def generate_email_alerter(to_addrs, from_addr=None, use_gmail=False,
username=None, password=None, hostname=None, port=25):
if not from_addr:
from_addr = getuser() + "@" + gethostname()
if use_gmail:
if username and password:
server = SMTP('', 587)
raise OptionValueError('You must provide a username and password to use GMail')
if hostname:
server = SMTP(hostname, port)
server = SMTP()
if username and password:
server.login(username, password)
def email_alerter(message, subject='You have an alert'):
server.sendmail(from_addr, to_addrs, 'To: %s\r\nFrom: %s\r\nSubject: %s\r\n\r\n%s' % (", ".join(to_addrs), from_addr, subject, message))
return email_alerter, server.quit
def get_site_status(url):
urlfile = urllib2.urlopen(url);
status_code = urlfile.code
if status_code in (200,302):
return 'up', urlfile
return 'down', None
def get_headers(url):
'''Gets all headers from URL request and returns'''
return urllib2.urlopen(url).info().headers
return 'Headers unavailable'
def compare_site_status(prev_results, alerter):
'''Report changed status based on previous results'''
def is_status_changed(url):
startTime = time.time()
status, urlfile = get_site_status(url)
endTime = time.time()
elapsedTime = endTime - startTime
msg = "%s took %s" % (url,elapsedTime)
friendly_status = '%s is %s' % (url, status)
print friendly_status
if url in prev_results and prev_results[url]['status'] != status:
# Email status messages
alerter(str(get_headers(url)), friendly_status)
# Create dictionary for url if one doesn't exist (first time url was checked)
if url not in prev_results:
prev_results[url] = {}
# Save results for later pickling and utility use
prev_results[url]['status'] = status
prev_results[url]['headers'] = None if urlfile == None else
prev_results[url]['rtime'] = elapsedTime
return is_status_changed
def is_internet_reachable():
'''Checks Google then Yahoo just in case one is down'''
statusGoogle, urlfileGoogle = get_site_status('')
statusYahoo, urlfileYahoo = get_site_status('')
if statusGoogle == 'down' and statusYahoo == 'down':
return False
return True
def load_old_results(file_path):
'''Attempts to load most recent results'''
pickledata = {}
if os.path.isfile(file_path):
picklefile = open(file_path, 'rb')
pickledata = pickle.load(picklefile)
return pickledata
def store_results(file_path, data):
'''Pickles results to compare on next run'''
output = open(file_path, 'wb')
pickle.dump(data, output)
def normalize_url(url):
'''If a url doesn't have a http/https prefix, add http://'''
if not re.match('^http[s]?://', url):
url = 'http://' + url
return url
def get_urls_from_file(filename):
f = open(filename, 'r')
filecontents = f.readlines()
results = []
for line in filecontents:
foo = line.strip('\n')
return results
logging.error('Unable to read %s' % filename)
return []
def get_command_line_options():
'''Sets up optparse and command line options'''
usage = "Usage: %prog [options] url"
parser = OptionParser(usage=usage)
parser.add_option("-t","--log-response-time", action="store_true",
help="Turn on logging for response times")
parser.add_option("-r","--alert-on-slow-response", action="store_true",
help="Turn on alerts for response times")
parser.add_option("-g","--use-gmail", action="store_true", dest="use_gmail",
help="Send email with Gmail. Must also specify username and password")
parser.add_option("--smtp-hostname", dest="smtp_hostname",
help="Set the stmp server host.")
parser.add_option("--smtp-port", dest="smtp_port", type="int",
help="Set the smtp server port.")
parser.add_option("-u","--smtp-username", dest="smtp_username",
help="Set the smtp username.")
parser.add_option("-p","--smtp-password", dest="smtp_password",
help="Set the smtp password.")
parser.add_option("-s","--from-addr", dest="from_addr",
help="Set the from email.")
parser.add_option("-d","--to-addrs", dest="to_addrs", action="append",
help="List of email addresses to send alerts to.")
parser.add_option("-f","--from-file", dest="from_file",
help="Import urls from a text file. Separated by newline.")
return parser.parse_args()
def main():
# Get argument flags and command options
(options,args) = get_command_line_options()
# Print out usage if no arguments are present
if len(args) == 0 and options.from_file == None:
print 'Usage:'
print "\tPlease specify a url like:"
print "\tNote: The http:// is not necessary"
print 'More Help:'
print "\tFor more help use the --help flag"
# If the -f flag is set we get urls from a file, otherwise we get them from the command line.
if options.from_file:
urls = get_urls_from_file(options.from_file)
urls = args
urls = map(normalize_url, urls)
# Change logging from WARNING to INFO when logResponseTime option is set
# so we can log response times as well as status changes.
if options.log_response_time:
logging.basicConfig(level=logging.INFO, filename='checksites.log',
format='%(asctime)s %(levelname)s: %(message)s',
datefmt='%Y-%m-%d %H:%M:%S')
logging.basicConfig(level=logging.WARNING, filename='checksites.log',
format='%(asctime)s %(levelname)s: %(message)s',
datefmt='%Y-%m-%d %H:%M:%S')
# Load previous data
pickle_file = 'data.pkl'
pickledata = load_old_results(pickle_file)
# Add some metadata to pickle
pickledata['meta'] = {} # Intentionally overwrite past metadata
pickledata['meta']['lastcheck'] = time.strftime('%Y-%m-%d %H:%M:%S')
# create an alerter
alerter, quiter = generate_email_alerter(options.to_addrs, from_addr=options.from_addr,
username=options.smtp_username, password=options.smtp_password,
hostname=options.smtp_hostname, port=options.smtp_port)
# Check sites only if Internet is_available
if is_internet_reachable():
status_checker = compare_site_status(pickledata, alerter)
map(status_checker, urls)
logging.error('Either the world ended or we are not connected to the net.')
# Store results in pickle file
store_results(pickle_file, pickledata)
if __name__ == '__main__':
# First arg is script name, skip it
Jump to Line
Something went wrong with that request. Please try again.