sipp11
10 years ago
commit
35da056e68
15 changed files with 896 additions and 0 deletions
@ -0,0 +1,4 @@
|
||||
Tornado & Flask together |
||||
================================ |
||||
|
||||
How to set up them to work together |
@ -0,0 +1,23 @@
|
||||
#!/usr/bin/env python |
||||
import tornado.httpserver |
||||
import tornado.ioloop |
||||
import tornado.web |
||||
from tornado.options import options |
||||
|
||||
from settings import settings |
||||
from urls import url_patterns |
||||
|
||||
|
||||
class TornardoWebApp(tornado.web.Application): |
||||
def __init__(self): |
||||
tornado.web.Application.__init__(self, url_patterns, **settings) |
||||
|
||||
|
||||
def main(): |
||||
app = TornardoWebApp() |
||||
http_server = tornado.httpserver.HTTPServer(app) |
||||
http_server.listen(options.port) |
||||
tornado.ioloop.IOLoop.instance().start() |
||||
|
||||
if __name__ == "__main__": |
||||
main() |
@ -0,0 +1,64 @@
|
||||
from flask import Flask, make_response, request |
||||
from simplejson import dumps |
||||
from pymongo import MongoClient |
||||
import datetime |
||||
import dateutil.parser |
||||
import bson |
||||
|
||||
app = Flask(__name__) |
||||
client = MongoClient(**{'host': 'localhost', 'port': 27017}) |
||||
db = client.showtimes |
||||
|
||||
miscObjHandler = lambda obj: ( |
||||
obj.isoformat() if isinstance(obj, datetime.datetime) |
||||
or isinstance(obj, datetime.date) |
||||
else str(obj) if isinstance(obj, bson.objectid.ObjectId) else None) |
||||
|
||||
|
||||
@app.route('/flask/') |
||||
def hello_world(): |
||||
return 'This comes from Flask ^_^' |
||||
|
||||
|
||||
@app.route('/groups/', methods=['GET']) |
||||
def groups(): |
||||
known_groups = ['sf', 'major'] |
||||
r = make_response(dumps(known_groups)) |
||||
r.mimetype = 'application/json' |
||||
return r |
||||
|
||||
|
||||
@app.route('/theaters/', methods=['GET']) |
||||
@app.route('/theaters/<group>/', methods=['GET']) |
||||
def list_theaters(group=None): |
||||
if not group: |
||||
result = db.theater.find() |
||||
else: |
||||
result = db.theater.find({'group': group}) |
||||
items = [i for i in result] |
||||
r = make_response(dumps(items, default=miscObjHandler)) |
||||
r.mimetype = 'application/json' |
||||
return r |
||||
|
||||
|
||||
@app.route('/showtimes/<group>/', methods=['GET']) |
||||
@app.route('/showtimes/<group>/<code>/', methods=['GET']) |
||||
def list_showtimes(group=None, code=None): |
||||
day = request.args.get('d', '') |
||||
q = {} |
||||
if day: |
||||
q['date'] = dateutil.parser.parse(day) |
||||
if group: |
||||
q['group'] = group |
||||
if code: |
||||
q['theater'] = code |
||||
|
||||
result = db.showtimes.find(q) |
||||
items = [i for i in result] |
||||
r = make_response(dumps(items, default=miscObjHandler)) |
||||
r.mimetype = 'application/json' |
||||
return r |
||||
|
||||
|
||||
if __name__ == '__main__': |
||||
app.run(debug=True) |
@ -0,0 +1,46 @@
|
||||
import json |
||||
import tornado.web |
||||
|
||||
import logging |
||||
logger = logging.getLogger('stockintel.' + __name__) |
||||
|
||||
|
||||
class BaseHandler(tornado.web.RequestHandler): |
||||
"""A class to collect common handler methods - all other handlers should |
||||
subclass this one. |
||||
""" |
||||
|
||||
def load_json(self): |
||||
"""Load JSON from the request body and store them in |
||||
self.request.arguments, like Tornado does by default for POSTed form |
||||
parameters. |
||||
|
||||
If JSON cannot be decoded, raises an HTTPError with status 400. |
||||
""" |
||||
try: |
||||
self.request.arguments = json.loads(self.request.body) |
||||
except ValueError: |
||||
msg = "Could not decode JSON: %s" % self.request.body |
||||
logger.debug(msg) |
||||
raise tornado.web.HTTPError(400, msg) |
||||
|
||||
def get_json_argument(self, name, default=None): |
||||
"""Find and return the argument with key 'name' from JSON request data. |
||||
Similar to Tornado's get_argument() method. |
||||
""" |
||||
if default is None: |
||||
default = self._ARG_DEFAULT |
||||
if not self.request.arguments: |
||||
self.load_json() |
||||
if name not in self.request.arguments: |
||||
if default is self._ARG_DEFAULT: |
||||
msg = "Missing argument '%s'" % name |
||||
logger.debug(msg) |
||||
raise tornado.web.HTTPError(400, msg) |
||||
logger.debug( |
||||
"Returning default argument %s, as we couldn't find " |
||||
"'%s' in %s" % (default, name, self.request.arguments)) |
||||
return default |
||||
arg = self.request.arguments[name] |
||||
logger.debug("Found '%s': %s in JSON arguments" % (name, arg)) |
||||
return arg |
@ -0,0 +1,10 @@
|
||||
from handlers.base import BaseHandler |
||||
|
||||
import logging |
||||
logger = logging.getLogger('stockintel.' + __name__) |
||||
|
||||
|
||||
class FooHandler(BaseHandler): |
||||
def get(self): |
||||
logger.info("200 OK base.html") |
||||
self.render("base.html") |
@ -0,0 +1,2 @@
|
||||
from __future__ import absolute_import |
||||
from logconfig.logconfig import initialize_logging |
@ -0,0 +1,545 @@
|
||||
# This is a copy of the Python logging.config.dictconfig module. It is provided |
||||
# here for backwards compatibility for Python versions prior to 2.7. |
||||
# |
||||
# Copyright 2009-2010 by Vinay Sajip. All Rights Reserved. |
||||
# |
||||
# Permission to use, copy, modify, and distribute this software and its |
||||
# documentation for any purpose and without fee is hereby granted, provided that |
||||
# the above copyright notice appear in all copies and that both that copyright |
||||
# notice and this permission notice appear in supporting documentation, and that |
||||
# the name of Vinay Sajip not be used in advertising or publicity pertaining to |
||||
# distribution of the software without specific, written prior permission. |
||||
# VINAY SAJIP DISCLAIMS ALL WARRANTIES WITH REGARD TO THIS SOFTWARE, INCLUDING |
||||
# ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL VINAY |
||||
# SAJIP BE LIABLE FOR ANY SPECIAL, INDIRECT OR CONSEQUENTIAL DAMAGES OR ANY |
||||
# DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN |
||||
# ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN |
||||
# CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. |
||||
|
||||
import logging.handlers |
||||
import re |
||||
import sys |
||||
import types |
||||
|
||||
IDENTIFIER = re.compile('^[a-z_][a-z0-9_]*$', re.I) |
||||
|
||||
def valid_ident(s): |
||||
m = IDENTIFIER.match(s) |
||||
if not m: |
||||
raise ValueError('Not a valid Python identifier: %r' % s) |
||||
return True |
||||
|
||||
# |
||||
# This function is defined in logging only in recent versions of Python |
||||
# |
||||
try: |
||||
from logging import _checkLevel |
||||
except ImportError: |
||||
def _checkLevel(level): |
||||
if isinstance(level, int): |
||||
rv = level |
||||
elif str(level) == level: |
||||
if level not in logging._levelNames: |
||||
raise ValueError('Unknown level: %r' % level) |
||||
rv = logging._levelNames[level] |
||||
else: |
||||
raise TypeError('Level not an integer or a ' |
||||
'valid string: %r' % level) |
||||
return rv |
||||
|
||||
# The ConvertingXXX classes are wrappers around standard Python containers, |
||||
# and they serve to convert any suitable values in the container. The |
||||
# conversion converts base dicts, lists and tuples to their wrapped |
||||
# equivalents, whereas strings which match a conversion format are converted |
||||
# appropriately. |
||||
# |
||||
# Each wrapper should have a configurator attribute holding the actual |
||||
# configurator to use for conversion. |
||||
|
||||
class ConvertingDict(dict): |
||||
"""A converting dictionary wrapper.""" |
||||
|
||||
def __getitem__(self, key): |
||||
value = dict.__getitem__(self, key) |
||||
result = self.configurator.convert(value) |
||||
#If the converted value is different, save for next time |
||||
if value is not result: |
||||
self[key] = result |
||||
if type(result) in (ConvertingDict, ConvertingList, |
||||
ConvertingTuple): |
||||
result.parent = self |
||||
result.key = key |
||||
return result |
||||
|
||||
def get(self, key, default=None): |
||||
value = dict.get(self, key, default) |
||||
result = self.configurator.convert(value) |
||||
#If the converted value is different, save for next time |
||||
if value is not result: |
||||
self[key] = result |
||||
if type(result) in (ConvertingDict, ConvertingList, |
||||
ConvertingTuple): |
||||
result.parent = self |
||||
result.key = key |
||||
return result |
||||
|
||||
def pop(self, key, default=None): |
||||
value = dict.pop(self, key, default) |
||||
result = self.configurator.convert(value) |
||||
if value is not result: |
||||
if type(result) in (ConvertingDict, ConvertingList, |
||||
ConvertingTuple): |
||||
result.parent = self |
||||
result.key = key |
||||
return result |
||||
|
||||
class ConvertingList(list): |
||||
"""A converting list wrapper.""" |
||||
def __getitem__(self, key): |
||||
value = list.__getitem__(self, key) |
||||
result = self.configurator.convert(value) |
||||
#If the converted value is different, save for next time |
||||
if value is not result: |
||||
self[key] = result |
||||
if type(result) in (ConvertingDict, ConvertingList, |
||||
ConvertingTuple): |
||||
result.parent = self |
||||
result.key = key |
||||
return result |
||||
|
||||
def pop(self, idx=-1): |
||||
value = list.pop(self, idx) |
||||
result = self.configurator.convert(value) |
||||
if value is not result: |
||||
if type(result) in (ConvertingDict, ConvertingList, |
||||
ConvertingTuple): |
||||
result.parent = self |
||||
return result |
||||
|
||||
class ConvertingTuple(tuple): |
||||
"""A converting tuple wrapper.""" |
||||
def __getitem__(self, key): |
||||
value = tuple.__getitem__(self, key) |
||||
result = self.configurator.convert(value) |
||||
if value is not result: |
||||
if type(result) in (ConvertingDict, ConvertingList, |
||||
ConvertingTuple): |
||||
result.parent = self |
||||
result.key = key |
||||
return result |
||||
|
||||
class BaseConfigurator(object): |
||||
""" |
||||
The configurator base class which defines some useful defaults. |
||||
""" |
||||
|
||||
CONVERT_PATTERN = re.compile(r'^(?P<prefix>[a-z]+)://(?P<suffix>.*)$') |
||||
|
||||
WORD_PATTERN = re.compile(r'^\s*(\w+)\s*') |
||||
DOT_PATTERN = re.compile(r'^\.\s*(\w+)\s*') |
||||
INDEX_PATTERN = re.compile(r'^\[\s*(\w+)\s*\]\s*') |
||||
DIGIT_PATTERN = re.compile(r'^\d+$') |
||||
|
||||
value_converters = { |
||||
'ext' : 'ext_convert', |
||||
'cfg' : 'cfg_convert', |
||||
} |
||||
|
||||
# We might want to use a different one, e.g. importlib |
||||
importer = __import__ |
||||
|
||||
def __init__(self, config): |
||||
self.config = ConvertingDict(config) |
||||
self.config.configurator = self |
||||
|
||||
def resolve(self, s): |
||||
""" |
||||
Resolve strings to objects using standard import and attribute |
||||
syntax. |
||||
""" |
||||
name = s.split('.') |
||||
used = name.pop(0) |
||||
found = self.importer(used) |
||||
for frag in name: |
||||
used += '.' + frag |
||||
try: |
||||
found = getattr(found, frag) |
||||
except AttributeError: |
||||
self.importer(used) |
||||
found = getattr(found, frag) |
||||
return found |
||||
|
||||
def ext_convert(self, value): |
||||
"""Default converter for the ext:// protocol.""" |
||||
return self.resolve(value) |
||||
|
||||
def cfg_convert(self, value): |
||||
"""Default converter for the cfg:// protocol.""" |
||||
rest = value |
||||
m = self.WORD_PATTERN.match(rest) |
||||
if m is None: |
||||
raise ValueError("Unable to convert %r" % value) |
||||
else: |
||||
rest = rest[m.end():] |
||||
d = self.config[m.groups()[0]] |
||||
#print d, rest |
||||
while rest: |
||||
m = self.DOT_PATTERN.match(rest) |
||||
if m: |
||||
d = d[m.groups()[0]] |
||||
else: |
||||
m = self.INDEX_PATTERN.match(rest) |
||||
if m: |
||||
idx = m.groups()[0] |
||||
if not self.DIGIT_PATTERN.match(idx): |
||||
d = d[idx] |
||||
else: |
||||
try: |
||||
n = int(idx) # try as number first (most likely) |
||||
d = d[n] |
||||
except TypeError: |
||||
d = d[idx] |
||||
if m: |
||||
rest = rest[m.end():] |
||||
else: |
||||
raise ValueError('Unable to convert ' |
||||
'%r at %r' % (value, rest)) |
||||
#rest should be empty |
||||
return d |
||||
|
||||
def convert(self, value): |
||||
""" |
||||
Convert values to an appropriate type. dicts, lists and tuples are |
||||
replaced by their converting alternatives. Strings are checked to |
||||
see if they have a conversion format and are converted if they do. |
||||
""" |
||||
if not isinstance(value, ConvertingDict) and isinstance(value, dict): |
||||
value = ConvertingDict(value) |
||||
value.configurator = self |
||||
elif not isinstance(value, ConvertingList) and isinstance(value, list): |
||||
value = ConvertingList(value) |
||||
value.configurator = self |
||||
elif not isinstance(value, ConvertingTuple) and\ |
||||
isinstance(value, tuple): |
||||
value = ConvertingTuple(value) |
||||
value.configurator = self |
||||
elif isinstance(value, str): # str for py3k |
||||
m = self.CONVERT_PATTERN.match(value) |
||||
if m: |
||||
d = m.groupdict() |
||||
prefix = d['prefix'] |
||||
converter = self.value_converters.get(prefix, None) |
||||
if converter: |
||||
suffix = d['suffix'] |
||||
converter = getattr(self, converter) |
||||
value = converter(suffix) |
||||
return value |
||||
|
||||
def configure_custom(self, config): |
||||
"""Configure an object with a user-supplied factory.""" |
||||
c = config.pop('()') |
||||
if not hasattr(c, '__call__') and hasattr(types, 'ClassType') and type(c) != types.ClassType: |
||||
c = self.resolve(c) |
||||
props = config.pop('.', None) |
||||
# Check for valid identifiers |
||||
kwargs = dict([(k, config[k]) for k in config if valid_ident(k)]) |
||||
result = c(**kwargs) |
||||
if props: |
||||
for name, value in props.items(): |
||||
setattr(result, name, value) |
||||
return result |
||||
|
||||
def as_tuple(self, value): |
||||
"""Utility function which converts lists to tuples.""" |
||||
if isinstance(value, list): |
||||
value = tuple(value) |
||||
return value |
||||
|
||||
class DictConfigurator(BaseConfigurator): |
||||
""" |
||||
Configure logging using a dictionary-like object to describe the |
||||
configuration. |
||||
""" |
||||
|
||||
def configure(self): |
||||
"""Do the configuration.""" |
||||
|
||||
config = self.config |
||||
if 'version' not in config: |
||||
raise ValueError("dictionary doesn't specify a version") |
||||
if config['version'] != 1: |
||||
raise ValueError("Unsupported version: %s" % config['version']) |
||||
incremental = config.pop('incremental', False) |
||||
EMPTY_DICT = {} |
||||
logging._acquireLock() |
||||
try: |
||||
if incremental: |
||||
handlers = config.get('handlers', EMPTY_DICT) |
||||
# incremental handler config only if handler name |
||||
# ties in to logging._handlers (Python 2.7) |
||||
if sys.version_info[:2] == (2, 7): |
||||
for name in handlers: |
||||
if name not in logging._handlers: |
||||
raise ValueError('No handler found with ' |
||||
'name %r' % name) |
||||
else: |
||||
try: |
||||
handler = logging._handlers[name] |
||||
handler_config = handlers[name] |
||||
level = handler_config.get('level', None) |
||||
if level: |
||||
handler.setLevel(_checkLevel(level)) |
||||
except StandardError as e: |
||||
raise ValueError('Unable to configure handler ' |
||||
'%r: %s' % (name, e)) |
||||
loggers = config.get('loggers', EMPTY_DICT) |
||||
for name in loggers: |
||||
try: |
||||
self.configure_logger(name, loggers[name], True) |
||||
except StandardError as e: |
||||
raise ValueError('Unable to configure logger ' |
||||
'%r: %s' % (name, e)) |
||||
root = config.get('root', None) |
||||
if root: |
||||
try: |
||||
self.configure_root(root, True) |
||||
except StandardError as e: |
||||
raise ValueError('Unable to configure root ' |
||||
'logger: %s' % e) |
||||
else: |
||||
disable_existing = config.pop('disable_existing_loggers', True) |
||||
|
||||
logging._handlers.clear() |
||||
del logging._handlerList[:] |
||||
|
||||
# Do formatters first - they don't refer to anything else |
||||
formatters = config.get('formatters', EMPTY_DICT) |
||||
for name in formatters: |
||||
try: |
||||
formatters[name] = self.configure_formatter( |
||||
formatters[name]) |
||||
except StandardError as e: |
||||
raise ValueError('Unable to configure ' |
||||
'formatter %r: %s' % (name, e)) |
||||
# Next, do filters - they don't refer to anything else, either |
||||
filters = config.get('filters', EMPTY_DICT) |
||||
for name in filters: |
||||
try: |
||||
filters[name] = self.configure_filter(filters[name]) |
||||
except StandardError as e: |
||||
raise ValueError('Unable to configure ' |
||||
'filter %r: %s' % (name, e)) |
||||
|
||||
# Next, do handlers - they refer to formatters and filters |
||||
# As handlers can refer to other handlers, sort the keys |
||||
# to allow a deterministic order of configuration |
||||
handlers = config.get('handlers', EMPTY_DICT) |
||||
for name in sorted(handlers): |
||||
try: |
||||
handler = self.configure_handler(handlers[name]) |
||||
handler.name = name |
||||
handlers[name] = handler |
||||
except StandardError as e: |
||||
raise ValueError('Unable to configure handler ' |
||||
'%r: %s' % (name, e)) |
||||
# Next, do loggers - they refer to handlers and filters |
||||
|
||||
#we don't want to lose the existing loggers, |
||||
#since other threads may have pointers to them. |
||||
#existing is set to contain all existing loggers, |
||||
#and as we go through the new configuration we |
||||
#remove any which are configured. At the end, |
||||
#what's left in existing is the set of loggers |
||||
#which were in the previous configuration but |
||||
#which are not in the new configuration. |
||||
root = logging.root |
||||
existing = root.manager.loggerDict.keys() |
||||
#The list needs to be sorted so that we can |
||||
#avoid disabling child loggers of explicitly |
||||
#named loggers. With a sorted list it is easier |
||||
#to find the child loggers. |
||||
sorted(existing) |
||||
#We'll keep the list of existing loggers |
||||
#which are children of named loggers here... |
||||
child_loggers = [] |
||||
#now set up the new ones... |
||||
loggers = config.get('loggers', EMPTY_DICT) |
||||
for name in loggers: |
||||
if name in existing: |
||||
i = existing.index(name) |
||||
prefixed = name + "." |
||||
pflen = len(prefixed) |
||||
num_existing = len(existing) |
||||
i = i + 1 # look at the entry after name |
||||
while (i < num_existing) and\ |
||||
(existing[i][:pflen] == prefixed): |
||||
child_loggers.append(existing[i]) |
||||
i = i + 1 |
||||
existing.remove(name) |
||||
try: |
||||
self.configure_logger(name, loggers[name]) |
||||
except StandardError as e: |
||||
raise ValueError('Unable to configure logger ' |
||||
'%r: %s' % (name, e)) |
||||
|
||||
#Disable any old loggers. There's no point deleting |
||||
#them as other threads may continue to hold references |
||||
#and by disabling them, you stop them doing any logging. |
||||
#However, don't disable children of named loggers, as that's |
||||
#probably not what was intended by the user. |
||||
for log in existing: |
||||
logger = root.manager.loggerDict[log] |
||||
if log in child_loggers: |
||||
logger.level = logging.NOTSET |
||||
logger.handlers = [] |
||||
logger.propagate = True |
||||
elif disable_existing: |
||||
logger.disabled = True |
||||
|
||||
# And finally, do the root logger |
||||
root = config.get('root', None) |
||||
if root: |
||||
try: |
||||
self.configure_root(root) |
||||
except StandardError as e: |
||||
raise ValueError('Unable to configure root ' |
||||
'logger: %s' % e) |
||||
finally: |
||||
logging._releaseLock() |
||||
|
||||
def configure_formatter(self, config): |
||||
"""Configure a formatter from a dictionary.""" |
||||
if '()' in config: |
||||
factory = config['()'] # for use in exception handler |
||||
try: |
||||
result = self.configure_custom(config) |
||||
except TypeError as te: |
||||
if "'format'" not in str(te): |
||||
raise |
||||
#Name of parameter changed from fmt to format. |
||||
#Retry with old name. |
||||
#This is so that code can be used with older Python versions |
||||
#(e.g. by Django) |
||||
config['fmt'] = config.pop('format') |
||||
config['()'] = factory |
||||
result = self.configure_custom(config) |
||||
else: |
||||
fmt = config.get('format', None) |
||||
dfmt = config.get('datefmt', None) |
||||
result = logging.Formatter(fmt, dfmt) |
||||
return result |
||||
|
||||
def configure_filter(self, config): |
||||
"""Configure a filter from a dictionary.""" |
||||
if '()' in config: |
||||
result = self.configure_custom(config) |
||||
else: |
||||
name = config.get('name', '') |
||||
result = logging.Filter(name) |
||||
return result |
||||
|
||||
def add_filters(self, filterer, filters): |
||||
"""Add filters to a filterer from a list of names.""" |
||||
for f in filters: |
||||
try: |
||||
filterer.addFilter(self.config['filters'][f]) |
||||
except StandardError as e: |
||||
raise ValueError('Unable to add filter %r: %s' % (f, e)) |
||||
|
||||
def configure_handler(self, config): |
||||
"""Configure a handler from a dictionary.""" |
||||
formatter = config.pop('formatter', None) |
||||
if formatter: |
||||
try: |
||||
formatter = self.config['formatters'][formatter] |
||||
except StandardError as e: |
||||
raise ValueError('Unable to set formatter ' |
||||
'%r: %s' % (formatter, e)) |
||||
level = config.pop('level', None) |
||||
filters = config.pop('filters', None) |
||||
if '()' in config: |
||||
c = config.pop('()') |
||||
if not hasattr(c, '__call__') and hasattr(types, 'ClassType') and type(c) != types.ClassType: |
||||
c = self.resolve(c) |
||||
factory = c |
||||
else: |
||||
klass = self.resolve(config.pop('class')) |
||||
#Special case for handler which refers to another handler |
||||
if issubclass(klass, logging.handlers.MemoryHandler) and\ |
||||
'target' in config: |
||||
try: |
||||
config['target'] = self.config['handlers'][config['target']] |
||||
except StandardError as e: |
||||
raise ValueError('Unable to set target handler ' |
||||
'%r: %s' % (config['target'], e)) |
||||
elif issubclass(klass, logging.handlers.SMTPHandler) and\ |
||||
'mailhost' in config: |
||||
config['mailhost'] = self.as_tuple(config['mailhost']) |
||||
elif issubclass(klass, logging.handlers.SysLogHandler) and\ |
||||
'address' in config: |
||||
config['address'] = self.as_tuple(config['address']) |
||||
factory = klass |
||||
kwargs = dict([(k, config[k]) for k in config if valid_ident(k)]) |
||||
try: |
||||
result = factory(**kwargs) |
||||
except TypeError as te: |
||||
if "'stream'" not in str(te): |
||||
raise |
||||
#The argument name changed from strm to stream |
||||
#Retry with old name. |
||||
#This is so that code can be used with older Python versions |
||||
#(e.g. by Django) |
||||
kwargs['strm'] = kwargs.pop('stream') |
||||
result = factory(**kwargs) |
||||
if formatter: |
||||
result.setFormatter(formatter) |
||||
if level is not None: |
||||
result.setLevel(_checkLevel(level)) |
||||
if filters: |
||||
self.add_filters(result, filters) |
||||
return result |
||||
|
||||
def add_handlers(self, logger, handlers): |
||||
"""Add handlers to a logger from a list of names.""" |
||||
for h in handlers: |
||||
try: |
||||
logger.addHandler(self.config['handlers'][h]) |
||||
except StandardError as e: |
||||
raise ValueError('Unable to add handler %r: %s' % (h, e)) |
||||
|
||||
def common_logger_config(self, logger, config, incremental=False): |
||||
""" |
||||
Perform configuration which is common to root and non-root loggers. |
||||
""" |
||||
level = config.get('level', None) |
||||
if level is not None: |
||||
logger.setLevel(_checkLevel(level)) |
||||
if not incremental: |
||||
#Remove any existing handlers |
||||
for h in logger.handlers[:]: |
||||
logger.removeHandler(h) |
||||
handlers = config.get('handlers', None) |
||||
if handlers: |
||||
self.add_handlers(logger, handlers) |
||||
filters = config.get('filters', None) |
||||
if filters: |
||||
self.add_filters(logger, filters) |
||||
|
||||
def configure_logger(self, name, config, incremental=False): |
||||
"""Configure a non-root logger from a dictionary.""" |
||||
logger = logging.getLogger(name) |
||||
self.common_logger_config(logger, config, incremental) |
||||
propagate = config.get('propagate', None) |
||||
if propagate is not None: |
||||
logger.propagate = propagate |
||||
|
||||
def configure_root(self, config, incremental=False): |
||||
"""Configure a root logger from a dictionary.""" |
||||
root = logging.getLogger() |
||||
self.common_logger_config(root, config, incremental) |
||||
|
||||
dictConfigClass = DictConfigurator |
||||
|
||||
def dictConfig(config): |
||||
"""Configure logging using a dictionary.""" |
||||
dictConfigClass(config).configure() |
@ -0,0 +1,104 @@
|
||||
"""An extended version of the log_settings module from zamboni: |
||||
https://github.com/jbalogh/zamboni/blob/master/log_settings.py |
||||
""" |
||||
from __future__ import absolute_import |
||||
|
||||
from tornado.log import LogFormatter as TornadoLogFormatter |
||||
import logging, logging.handlers |
||||
import os.path |
||||
import types |
||||
|
||||
from logconfig import dictconfig |
||||
|
||||
# Pulled from commonware.log we don't have to import that, which drags with |
||||
# it Django dependencies. |
||||
class RemoteAddressFormatter(logging.Formatter): |
||||
"""Formatter that makes sure REMOTE_ADDR is available.""" |
||||
|
||||
def format(self, record): |
||||
if ('%(REMOTE_ADDR)' in self._fmt |
||||
and 'REMOTE_ADDR' not in record.__dict__): |
||||
record.__dict__['REMOTE_ADDR'] = None |
||||
return logging.Formatter.format(self, record) |
||||
|
||||
class UTF8SafeFormatter(RemoteAddressFormatter): |
||||
def __init__(self, fmt=None, datefmt=None, encoding='utf-8'): |
||||
logging.Formatter.__init__(self, fmt, datefmt) |
||||
self.encoding = encoding |
||||
|
||||
def formatException(self, e): |
||||
r = logging.Formatter.formatException(self, e) |
||||
if type(r) in [types.StringType]: |
||||
r = r.decode(self.encoding, 'replace') # Convert to unicode |
||||
return r |
||||
|
||||
def format(self, record): |
||||
t = RemoteAddressFormatter.format(self, record) |
||||
if type(t) in [types.UnicodeType]: |
||||
t = t.encode(self.encoding, 'replace') |
||||
return t |
||||
|
||||
class NullHandler(logging.Handler): |
||||
def emit(self, record): |
||||
pass |
||||
|
||||
def initialize_logging(syslog_tag, syslog_facility, loggers, |
||||
log_level=logging.INFO, use_syslog=False): |
||||
if os.path.exists('/dev/log'): |
||||
syslog_device = '/dev/log' |
||||
elif os.path.exists('/var/run/syslog'): |
||||
syslog_device = '/var/run/syslog' |
||||
|
||||
base_fmt = ('%(name)s:%(levelname)s %(message)s:%(pathname)s:%(lineno)s') |
||||
|
||||
cfg = { |
||||
'version': 1, |
||||
'filters': {}, |
||||
'formatters': { |
||||
'debug': { |
||||
'()': UTF8SafeFormatter, |
||||
'datefmt': '%H:%M:%s', |
||||
'format': '%(asctime)s ' + base_fmt, |
||||
}, |
||||
'prod': { |
||||
'()': UTF8SafeFormatter, |
||||
'datefmt': '%H:%M:%s', |
||||
'format': '%s: [%%(REMOTE_ADDR)s] %s' % (syslog_tag, base_fmt), |
||||
}, |
||||
'tornado': { |
||||
'()': TornadoLogFormatter, |
||||
'color': True |
||||
}, |
||||
}, |
||||
'handlers': { |
||||
'console': { |
||||
'()': logging.StreamHandler, |
||||
'formatter': 'tornado' |
||||
}, |
||||
'null': { |
||||
'()': NullHandler, |
||||
}, |
||||
'syslog': { |
||||
'()': logging.handlers.SysLogHandler, |
||||
'facility': syslog_facility, |
||||
'address': syslog_device, |
||||
'formatter': 'prod', |
||||
}, |
||||
}, |
||||
'loggers': { |
||||
} |
||||
} |
||||
|
||||
for key, value in loggers.items(): |
||||
cfg[key].update(value) |
||||
|
||||
# Set the level and handlers for all loggers. |
||||
for logger in cfg['loggers'].values(): |
||||
if 'handlers' not in logger: |
||||
logger['handlers'] = ['syslog' if use_syslog else 'console'] |
||||
if 'level' not in logger: |
||||
logger['level'] = log_level |
||||
if 'propagate' not in logger: |
||||
logger['propagate'] = False |
||||
|
||||
dictconfig.dictConfig(cfg) |
@ -0,0 +1,4 @@
|
||||
requests==2.4.0 |
||||
tornado>=4.0 |
||||
flask |
||||
python-dateutil |
@ -0,0 +1,72 @@
|
||||
import logging |
||||
import tornado |
||||
import tornado.template |
||||
from tornado.options import define, options |
||||
import os |
||||
import logconfig |
||||
|
||||
|
||||
# Make filepaths relative to settings. |
||||
path = lambda root, *a: os.path.join(root, *a) |
||||
ROOT = os.path.dirname(os.path.abspath(__file__)) |
||||
|
||||
define("port", default=8888, help="run on the given port", type=int) |
||||
define("config", default=None, help="tornado config file") |
||||
define("debug", default=False, help="debug mode") |
||||
tornado.options.parse_command_line() |
||||
|
||||
|
||||
MEDIA_ROOT = path(ROOT, 'media') |
||||
TEMPLATE_ROOT = path(ROOT, 'templates') |
||||
|
||||
|
||||
class DeploymentType: |
||||
# Deployment Configuration |
||||
PRODUCTION = "PRODUCTION" |
||||
DEV = "DEV" |
||||
SOLO = "SOLO" |
||||
STAGING = "STAGING" |
||||
dict = { |
||||
SOLO: 1, |
||||
PRODUCTION: 2, |
||||
DEV: 3, |
||||
STAGING: 4 |
||||
} |
||||
if 'DEPLOYMENT_TYPE' in os.environ: |
||||
DEPLOYMENT = os.environ['DEPLOYMENT_TYPE'].upper() |
||||
else: |
||||
DEPLOYMENT = DeploymentType.SOLO |
||||
|
||||
settings = {} |
||||
settings['debug'] = DEPLOYMENT != DeploymentType.PRODUCTION or options.debug |
||||
settings['static_path'] = MEDIA_ROOT |
||||
settings['cookie_secret'] = '*(&$#^@&*YOURWHATEVERSECRET_JDF)#()$#@()*ur892h899' |
||||
settings['xsrf_cookies'] = True |
||||
settings['template_loader'] = tornado.template.Loader(TEMPLATE_ROOT) |
||||
|
||||
|
||||
SYSLOG_TAG = "stockintel" |
||||
SYSLOG_FACILITY = logging.handlers.SysLogHandler.LOG_LOCAL2 |
||||
|
||||
# See PEP 391 and logconfig for formatting help. Each section of LOGGERS |
||||
# will get merged into the corresponding section of log_settings.py. |
||||
# Handlers and log levels are set up automatically based on LOG_LEVEL and DEBUG |
||||
# unless you set them here. Messages will not propagate through a logger |
||||
# unless propagate: True is set. |
||||
LOGGERS = { |
||||
'loggers': { |
||||
'stockintel': {}, |
||||
}, |
||||
} |
||||
|
||||
if settings['debug']: |
||||
LOG_LEVEL = logging.DEBUG |
||||
else: |
||||
LOG_LEVEL = logging.INFO |
||||
USE_SYSLOG = DEPLOYMENT != DeploymentType.SOLO |
||||
|
||||
logconfig.initialize_logging( |
||||
SYSLOG_TAG, SYSLOG_FACILITY, LOGGERS, LOG_LEVEL, USE_SYSLOG) |
||||
|
||||
if options.config: |
||||
tornado.options.parse_config_file(options.config) |
@ -0,0 +1,9 @@
|
||||
<html> |
||||
<head> |
||||
<title>Tornado Boilerplate</title> |
||||
</head> |
||||
|
||||
<body> |
||||
<h1>It worked!</h1> |
||||
</body> |
||||
</html> |
@ -0,0 +1,11 @@
|
||||
from tornado.wsgi import WSGIContainer |
||||
from flasky import app |
||||
from handlers.foo import FooHandler |
||||
from tornado.web import FallbackHandler |
||||
|
||||
flaskFallback = WSGIContainer(app) |
||||
|
||||
url_patterns = [ |
||||
(r"/foo", FooHandler), |
||||
(r".*", FallbackHandler, dict(fallback=flaskFallback)), |
||||
] |
Loading…
Reference in new issue