592 lines
16 KiB
Python
592 lines
16 KiB
Python
|
# -*- coding: utf-8 -*-
|
||
|
import copy
|
||
|
import marshal
|
||
|
import struct
|
||
|
import threading
|
||
|
import time
|
||
|
import traceback
|
||
|
|
||
|
from .._compat import (
|
||
|
PY2, exists, copyreg, implements_bool, iterkeys, itervalues, iteritems,
|
||
|
long
|
||
|
)
|
||
|
from .._globals import THREAD_LOCAL
|
||
|
from .serializers import serializers
|
||
|
|
||
|
|
||
|
class cachedprop(object):
|
||
|
#: a read-only @property that is only evaluated once.
|
||
|
def __init__(self, fget, doc=None):
|
||
|
self.fget = fget
|
||
|
self.__doc__ = doc or fget.__doc__
|
||
|
self.__name__ = fget.__name__
|
||
|
|
||
|
def __get__(self, obj, cls):
|
||
|
if obj is None:
|
||
|
return self
|
||
|
obj.__dict__[self.__name__] = result = self.fget(obj)
|
||
|
return result
|
||
|
|
||
|
|
||
|
@implements_bool
|
||
|
class BasicStorage(object):
|
||
|
def __init__(self, *args, **kwargs):
|
||
|
return self.__dict__.__init__(*args, **kwargs)
|
||
|
|
||
|
def __getitem__(self, key):
|
||
|
return self.__dict__.__getitem__(str(key))
|
||
|
|
||
|
__setitem__ = object.__setattr__
|
||
|
|
||
|
def __delitem__(self, key):
|
||
|
try:
|
||
|
delattr(self, key)
|
||
|
except AttributeError:
|
||
|
raise KeyError(key)
|
||
|
|
||
|
def __bool__(self):
|
||
|
return len(self.__dict__) > 0
|
||
|
|
||
|
__iter__ = lambda self: self.__dict__.__iter__()
|
||
|
|
||
|
__str__ = lambda self: self.__dict__.__str__()
|
||
|
|
||
|
__repr__ = lambda self: self.__dict__.__repr__()
|
||
|
|
||
|
has_key = __contains__ = lambda self, key: key in self.__dict__
|
||
|
|
||
|
def get(self, key, default=None):
|
||
|
return self.__dict__.get(key, default)
|
||
|
|
||
|
def update(self, *args, **kwargs):
|
||
|
return self.__dict__.update(*args, **kwargs)
|
||
|
|
||
|
def keys(self):
|
||
|
return self.__dict__.keys()
|
||
|
|
||
|
def iterkeys(self):
|
||
|
return iterkeys(self.__dict__)
|
||
|
|
||
|
def values(self):
|
||
|
return self.__dict__.values()
|
||
|
|
||
|
def itervalues(self):
|
||
|
return itervalues(self.__dict__)
|
||
|
|
||
|
def items(self):
|
||
|
return self.__dict__.items()
|
||
|
|
||
|
def iteritems(self):
|
||
|
return iteritems(self.__dict__)
|
||
|
|
||
|
pop = lambda self, *args, **kwargs: self.__dict__.pop(*args, **kwargs)
|
||
|
|
||
|
clear = lambda self, *args, **kwargs: self.__dict__.clear(*args, **kwargs)
|
||
|
|
||
|
copy = lambda self, *args, **kwargs: self.__dict__.copy(*args, **kwargs)
|
||
|
|
||
|
|
||
|
def pickle_basicstorage(s):
|
||
|
return BasicStorage, (dict(s),)
|
||
|
|
||
|
copyreg.pickle(BasicStorage, pickle_basicstorage)
|
||
|
|
||
|
|
||
|
class OpRow(object):
|
||
|
__slots__ = ('_table', '_fields', '_values')
|
||
|
|
||
|
def __init__(self, table):
|
||
|
object.__setattr__(self, '_table', table)
|
||
|
object.__setattr__(self, '_fields', {})
|
||
|
object.__setattr__(self, '_values', {})
|
||
|
|
||
|
def set_value(self, key, value, field=None):
|
||
|
self._values[key] = value
|
||
|
self._fields[key] = self._fields.get(key, field or self._table[key])
|
||
|
|
||
|
def del_value(self, key):
|
||
|
del self._values[key]
|
||
|
del self._fields[key]
|
||
|
|
||
|
def __getitem__(self, key):
|
||
|
return self._values[key]
|
||
|
|
||
|
def __setitem__(self, key, value):
|
||
|
return self.set_value(key, value)
|
||
|
|
||
|
def __delitem__(self, key):
|
||
|
return self.del_value(key)
|
||
|
|
||
|
def __getattr__(self, key):
|
||
|
try:
|
||
|
return self[key]
|
||
|
except KeyError:
|
||
|
raise AttributeError
|
||
|
|
||
|
def __setattr__(self, key, value):
|
||
|
return self.set_value(key, value)
|
||
|
|
||
|
def __delattr__(self, key):
|
||
|
return self.del_value(key)
|
||
|
|
||
|
def __iter__(self):
|
||
|
return self._values.__iter__()
|
||
|
|
||
|
def __contains__(self, key):
|
||
|
return key in self._values
|
||
|
|
||
|
def get(self, key, default=None):
|
||
|
try:
|
||
|
rv = self[key]
|
||
|
except KeyError:
|
||
|
rv = default
|
||
|
return rv
|
||
|
|
||
|
def keys(self):
|
||
|
return self._values.keys()
|
||
|
|
||
|
def iterkeys(self):
|
||
|
return iterkeys(self._values)
|
||
|
|
||
|
def values(self):
|
||
|
return self._values.values()
|
||
|
|
||
|
def itervalues(self):
|
||
|
return itervalues(self._values)
|
||
|
|
||
|
def items(self):
|
||
|
return self._values.items()
|
||
|
|
||
|
def iteritems(self):
|
||
|
return iteritems(self._values)
|
||
|
|
||
|
def op_values(self):
|
||
|
return [
|
||
|
(self._fields[key], value)
|
||
|
for key, value in iteritems(self._values)
|
||
|
]
|
||
|
|
||
|
def __repr__(self):
|
||
|
return '<OpRow %s>' % repr(self._values)
|
||
|
|
||
|
|
||
|
class ConnectionConfigurationMixin(object):
|
||
|
def _mock_reconnect(self):
|
||
|
self._reconnect_lock = threading.RLock()
|
||
|
self._connection_reconnect = self.reconnect
|
||
|
self.reconnect = self._reconnect_and_configure
|
||
|
self._reconnect_mocked = True
|
||
|
|
||
|
def _reconnect_and_configure(self):
|
||
|
self._connection_reconnect()
|
||
|
with self._reconnect_lock:
|
||
|
if self._reconnect_mocked:
|
||
|
self._configure_on_first_reconnect()
|
||
|
self.reconnect = self._connection_reconnect
|
||
|
self._reconnect_mocked = False
|
||
|
|
||
|
def _configure_on_first_reconnect(self):
|
||
|
pass
|
||
|
|
||
|
|
||
|
class Serializable(object):
|
||
|
def as_dict(self, flat=False, sanitize=True):
|
||
|
return self.__dict__
|
||
|
|
||
|
def as_xml(self, sanitize=True):
|
||
|
return serializers.xml(self.as_dict(flat=True, sanitize=sanitize))
|
||
|
|
||
|
def as_json(self, sanitize=True):
|
||
|
return serializers.json(self.as_dict(flat=True, sanitize=sanitize))
|
||
|
|
||
|
def as_yaml(self, sanitize=True):
|
||
|
return serializers.yaml(self.as_dict(flat=True, sanitize=sanitize))
|
||
|
|
||
|
|
||
|
class Reference(long):
|
||
|
def __allocate(self):
|
||
|
if not self._record:
|
||
|
self._record = self._table[long(self)]
|
||
|
if not self._record:
|
||
|
raise RuntimeError(
|
||
|
"Using a recursive select but encountered a broken " +
|
||
|
"reference: %s %d" % (self._table, long(self))
|
||
|
)
|
||
|
|
||
|
def __getattr__(self, key, default=None):
|
||
|
if key == 'id':
|
||
|
return long(self)
|
||
|
if key in self._table:
|
||
|
self.__allocate()
|
||
|
if self._record:
|
||
|
# to deal with case self.update_record()
|
||
|
return self._record.get(key, default)
|
||
|
else:
|
||
|
return None
|
||
|
|
||
|
def get(self, key, default=None):
|
||
|
return self.__getattr__(key, default)
|
||
|
|
||
|
def __setattr__(self, key, value):
|
||
|
if key.startswith('_'):
|
||
|
long.__setattr__(self, key, value)
|
||
|
return
|
||
|
self.__allocate()
|
||
|
self._record[key] = value
|
||
|
|
||
|
def __getitem__(self, key):
|
||
|
if key == 'id':
|
||
|
return long(self)
|
||
|
self.__allocate()
|
||
|
return self._record.get(key, None)
|
||
|
|
||
|
def __setitem__(self, key, value):
|
||
|
self.__allocate()
|
||
|
self._record[key] = value
|
||
|
|
||
|
|
||
|
def Reference_unpickler(data):
|
||
|
return marshal.loads(data)
|
||
|
|
||
|
|
||
|
def Reference_pickler(data):
|
||
|
try:
|
||
|
marshal_dump = marshal.dumps(long(data))
|
||
|
except AttributeError:
|
||
|
marshal_dump = 'i%s' % struct.pack('<i', long(data))
|
||
|
return (Reference_unpickler, (marshal_dump,))
|
||
|
|
||
|
copyreg.pickle(Reference, Reference_pickler, Reference_unpickler)
|
||
|
|
||
|
|
||
|
class SQLCallableList(list):
|
||
|
def __call__(self):
|
||
|
return copy.copy(self)
|
||
|
|
||
|
|
||
|
class SQLALL(object):
|
||
|
"""
|
||
|
Helper class providing a comma-separated string having all the field names
|
||
|
(prefixed by table name and '.')
|
||
|
|
||
|
normally only called from within gluon.dal
|
||
|
"""
|
||
|
|
||
|
def __init__(self, table):
|
||
|
self._table = table
|
||
|
|
||
|
def __str__(self):
|
||
|
return ', '.join([str(field) for field in self._table])
|
||
|
|
||
|
|
||
|
class SQLCustomType(object):
|
||
|
"""
|
||
|
Allows defining of custom SQL types
|
||
|
|
||
|
Args:
|
||
|
type: the web2py type (default = 'string')
|
||
|
native: the backend type
|
||
|
encoder: how to encode the value to store it in the backend
|
||
|
decoder: how to decode the value retrieved from the backend
|
||
|
validator: what validators to use ( default = None, will use the
|
||
|
default validator for type)
|
||
|
|
||
|
Example::
|
||
|
Define as:
|
||
|
|
||
|
decimal = SQLCustomType(
|
||
|
type ='double',
|
||
|
native ='integer',
|
||
|
encoder =(lambda x: int(float(x) * 100)),
|
||
|
decoder = (lambda x: Decimal("0.00") + Decimal(str(float(x)/100)) )
|
||
|
)
|
||
|
|
||
|
db.define_table(
|
||
|
'example',
|
||
|
Field('value', type=decimal)
|
||
|
)
|
||
|
|
||
|
"""
|
||
|
|
||
|
def __init__(self, type='string', native=None, encoder=None, decoder=None,
|
||
|
validator=None, _class=None, widget=None, represent=None):
|
||
|
self.type = type
|
||
|
self.native = native
|
||
|
self.encoder = encoder or (lambda x: x)
|
||
|
self.decoder = decoder or (lambda x: x)
|
||
|
self.validator = validator
|
||
|
self._class = _class or type
|
||
|
self.widget = widget
|
||
|
self.represent = represent
|
||
|
|
||
|
def startswith(self, text=None):
|
||
|
try:
|
||
|
return self.type.startswith(self, text)
|
||
|
except TypeError:
|
||
|
return False
|
||
|
|
||
|
def endswith(self, text=None):
|
||
|
try:
|
||
|
return self.type.endswith(self, text)
|
||
|
except TypeError:
|
||
|
return False
|
||
|
|
||
|
def __getslice__(self, a=0, b=100):
|
||
|
return None
|
||
|
|
||
|
def __getitem__(self, i):
|
||
|
return None
|
||
|
|
||
|
def __str__(self):
|
||
|
return self._class
|
||
|
|
||
|
|
||
|
class RecordOperator(object):
|
||
|
def __init__(self, colset, table, id):
|
||
|
self.colset, self.db, self.tablename, self.id = \
|
||
|
colset, table._db, table._tablename, id
|
||
|
|
||
|
def __call__(self):
|
||
|
pass
|
||
|
|
||
|
|
||
|
class RecordUpdater(RecordOperator):
|
||
|
def __call__(self, **fields):
|
||
|
colset, db, tablename, id = self.colset, self.db, self.tablename, \
|
||
|
self.id
|
||
|
table = db[tablename]
|
||
|
newfields = fields or dict(colset)
|
||
|
for fieldname in list(newfields.keys()):
|
||
|
if fieldname not in table.fields or table[fieldname].type == 'id':
|
||
|
del newfields[fieldname]
|
||
|
table._db(table._id == id, ignore_common_filters=True).update(
|
||
|
**newfields
|
||
|
)
|
||
|
colset.update(newfields)
|
||
|
return colset
|
||
|
|
||
|
|
||
|
class RecordDeleter(RecordOperator):
|
||
|
def __call__(self):
|
||
|
return self.db(self.db[self.tablename]._id == self.id).delete()
|
||
|
|
||
|
|
||
|
class MethodAdder(object):
|
||
|
def __init__(self, table):
|
||
|
self.table = table
|
||
|
|
||
|
def __call__(self):
|
||
|
return self.register()
|
||
|
|
||
|
def __getattr__(self, method_name):
|
||
|
return self.register(method_name)
|
||
|
|
||
|
def register(self, method_name=None):
|
||
|
def _decorated(f):
|
||
|
instance = self.table
|
||
|
import types
|
||
|
if PY2:
|
||
|
method = types.MethodType(f, instance, instance.__class__)
|
||
|
else:
|
||
|
method = types.MethodType(f, instance)
|
||
|
name = method_name or f.func_name
|
||
|
setattr(instance, name, method)
|
||
|
return f
|
||
|
return _decorated
|
||
|
|
||
|
|
||
|
class FakeCursor(object):
|
||
|
'''
|
||
|
The Python Database API Specification has a cursor() method, which
|
||
|
NoSql drivers generally don't support. If the exception in this
|
||
|
function is taken then it likely means that some piece of
|
||
|
functionality has not yet been implemented in the driver. And
|
||
|
something is using the cursor.
|
||
|
|
||
|
https://www.python.org/dev/peps/pep-0249/
|
||
|
'''
|
||
|
def warn_bad_usage(self, attr):
|
||
|
raise Exception("FakeCursor.%s is not implemented" % attr)
|
||
|
|
||
|
def __getattr__(self, attr):
|
||
|
self.warn_bad_usage(attr)
|
||
|
|
||
|
def __setattr__(self, attr, value):
|
||
|
self.warn_bad_usage(attr)
|
||
|
|
||
|
|
||
|
class NullCursor(FakeCursor):
|
||
|
lastrowid = 1
|
||
|
|
||
|
def __getattr__(self, attr):
|
||
|
return lambda *a, **b: []
|
||
|
|
||
|
|
||
|
class FakeDriver(BasicStorage):
|
||
|
def __init__(self, *args, **kwargs):
|
||
|
super(FakeDriver, self).__init__(*args, **kwargs)
|
||
|
self._build_cursor_()
|
||
|
|
||
|
def _build_cursor_(self):
|
||
|
self._fake_cursor_ = FakeCursor()
|
||
|
|
||
|
def cursor(self):
|
||
|
return self._fake_cursor_
|
||
|
|
||
|
def close(self):
|
||
|
return None
|
||
|
|
||
|
def commit(self):
|
||
|
return None
|
||
|
|
||
|
def __str__(self):
|
||
|
state = ["%s=%r" % (attribute, value)
|
||
|
for (attribute, value) in self.items()]
|
||
|
return '\n'.join(state)
|
||
|
|
||
|
|
||
|
class NullDriver(FakeDriver):
|
||
|
def _build_cursor_(self):
|
||
|
self._fake_cursor_ = NullCursor()
|
||
|
|
||
|
|
||
|
class ExecutionHandler(object):
|
||
|
def __init__(self, adapter):
|
||
|
self.adapter = adapter
|
||
|
|
||
|
def before_execute(self, command):
|
||
|
pass
|
||
|
|
||
|
def after_execute(self, command):
|
||
|
pass
|
||
|
|
||
|
|
||
|
class TimingHandler(ExecutionHandler):
|
||
|
MAXSTORAGE = 100
|
||
|
|
||
|
def _timings(self):
|
||
|
THREAD_LOCAL._pydal_timings_ = getattr(
|
||
|
THREAD_LOCAL, '_pydal_timings_', [])
|
||
|
return THREAD_LOCAL._pydal_timings_
|
||
|
|
||
|
@property
|
||
|
def timings(self):
|
||
|
return self._timings()
|
||
|
|
||
|
def before_execute(self, command):
|
||
|
self.t = time.time()
|
||
|
|
||
|
def after_execute(self, command):
|
||
|
dt = time.time() - self.t
|
||
|
self.timings.append((command, dt))
|
||
|
del self.timings[:-self.MAXSTORAGE]
|
||
|
|
||
|
|
||
|
class DatabaseStoredFile:
|
||
|
|
||
|
web2py_filesystems = set()
|
||
|
|
||
|
def escape(self, obj):
|
||
|
return self.db._adapter.escape(obj)
|
||
|
|
||
|
@staticmethod
|
||
|
def try_create_web2py_filesystem(db):
|
||
|
if db._uri not in DatabaseStoredFile.web2py_filesystems:
|
||
|
if db._adapter.dbengine == 'mysql':
|
||
|
sql = "CREATE TABLE IF NOT EXISTS web2py_filesystem (path VARCHAR(255), content LONGTEXT, PRIMARY KEY(path) ) ENGINE=InnoDB;"
|
||
|
elif db._adapter.dbengine in ('postgres', 'sqlite'):
|
||
|
sql = "CREATE TABLE IF NOT EXISTS web2py_filesystem (path VARCHAR(255), content TEXT, PRIMARY KEY(path));"
|
||
|
db.executesql(sql)
|
||
|
DatabaseStoredFile.web2py_filesystems.add(db._uri)
|
||
|
|
||
|
def __init__(self, db, filename, mode):
|
||
|
if db._adapter.dbengine not in ('mysql', 'postgres', 'sqlite'):
|
||
|
raise RuntimeError(
|
||
|
"only MySQL/Postgres/SQLite can store metadata .table files" +
|
||
|
" in database for now")
|
||
|
self.db = db
|
||
|
self.filename = filename
|
||
|
self.mode = mode
|
||
|
DatabaseStoredFile.try_create_web2py_filesystem(db)
|
||
|
self.p = 0
|
||
|
self.data = ''
|
||
|
if mode in ('r', 'rw', 'rb', 'a'):
|
||
|
query = "SELECT content FROM web2py_filesystem WHERE path='%s'" \
|
||
|
% filename
|
||
|
rows = self.db.executesql(query)
|
||
|
if rows:
|
||
|
self.data = rows[0][0]
|
||
|
elif exists(filename):
|
||
|
datafile = open(filename, 'r')
|
||
|
try:
|
||
|
self.data = datafile.read()
|
||
|
finally:
|
||
|
datafile.close()
|
||
|
elif mode in ('r', 'rw'):
|
||
|
raise RuntimeError("File %s does not exist" % filename)
|
||
|
|
||
|
def read(self, bytes=None):
|
||
|
if bytes is None:
|
||
|
bytes = len(self.data)
|
||
|
data = self.data[self.p:self.p+bytes]
|
||
|
self.p += len(data)
|
||
|
return data
|
||
|
|
||
|
def readline(self):
|
||
|
i = self.data.find('\n', self.p)+1
|
||
|
if i > 0:
|
||
|
data, self.p = self.data[self.p:i], i
|
||
|
else:
|
||
|
data, self.p = self.data[self.p:], len(self.data)
|
||
|
return data
|
||
|
|
||
|
def write(self, data):
|
||
|
self.data += data
|
||
|
|
||
|
def close_connection(self):
|
||
|
if self.db is not None:
|
||
|
self.db.executesql(
|
||
|
"DELETE FROM web2py_filesystem WHERE path='%s'" %
|
||
|
self.filename
|
||
|
)
|
||
|
query = "INSERT INTO web2py_filesystem(path,content) VALUES ('%s','%s')"\
|
||
|
% (self.filename, self.data.replace("'", "''"))
|
||
|
self.db.executesql(query)
|
||
|
self.db.commit()
|
||
|
self.db = None
|
||
|
|
||
|
def close(self):
|
||
|
self.close_connection()
|
||
|
|
||
|
@staticmethod
|
||
|
def is_operational_error(db, error):
|
||
|
if not hasattr(db._adapter.driver, "OperationalError"):
|
||
|
return None
|
||
|
return isinstance(error, db._adapter.driver.OperationalError)
|
||
|
|
||
|
@staticmethod
|
||
|
def is_programming_error(db, error):
|
||
|
if not hasattr(db._adapter.driver, "ProgrammingError"):
|
||
|
return None
|
||
|
return isinstance(error, db._adapter.driver.ProgrammingError)
|
||
|
|
||
|
@staticmethod
|
||
|
def exists(db, filename):
|
||
|
if exists(filename):
|
||
|
return True
|
||
|
|
||
|
DatabaseStoredFile.try_create_web2py_filesystem(db)
|
||
|
|
||
|
query = "SELECT path FROM web2py_filesystem WHERE path='%s'" % filename
|
||
|
try:
|
||
|
if db.executesql(query):
|
||
|
return True
|
||
|
except Exception as e:
|
||
|
if not (DatabaseStoredFile.is_operational_error(db, e) or
|
||
|
DatabaseStoredFile.is_programming_error(db, e)):
|
||
|
raise
|
||
|
# no web2py_filesystem found?
|
||
|
tb = traceback.format_exc()
|
||
|
db.logger.error("Could not retrieve %s\n%s" % (filename, tb))
|
||
|
return False
|