Project

General

Profile

1
# Database access
2

    
3
import copy
4
import re
5
import warnings
6

    
7
import exc
8
import dicts
9
import iters
10
import lists
11
from Proxy import Proxy
12
import rand
13
import strings
14
import util
15

    
16
##### Exceptions
17

    
18
def get_cur_query(cur):
19
    if hasattr(cur, 'query'): return cur.query
20
    elif hasattr(cur, '_last_executed'): return cur._last_executed
21
    else: return None
22

    
23
def _add_cursor_info(e, cur): exc.add_msg(e, 'query: '+get_cur_query(cur))
24

    
25
class DbException(exc.ExceptionWithCause):
26
    def __init__(self, msg, cause=None, cur=None):
27
        exc.ExceptionWithCause.__init__(self, msg, cause)
28
        if cur != None: _add_cursor_info(self, cur)
29

    
30
class NameException(DbException): pass
31

    
32
class ExceptionWithColumns(DbException):
33
    def __init__(self, cols, cause=None):
34
        DbException.__init__(self, 'columns: ' + ', '.join(cols), cause)
35
        self.cols = cols
36

    
37
class DuplicateKeyException(ExceptionWithColumns): pass
38

    
39
class NullValueException(ExceptionWithColumns): pass
40

    
41
class EmptyRowException(DbException): pass
42

    
43
##### Warnings
44

    
45
class DbWarning(UserWarning): pass
46

    
47
##### Result retrieval
48

    
49
def col_names(cur): return (col[0] for col in cur.description)
50

    
51
def rows(cur): return iter(lambda: cur.fetchone(), None)
52

    
53
def consume_rows(cur):
54
    '''Used to fetch all rows so result will be cached'''
55
    iters.consume_iter(rows(cur))
56

    
57
def next_row(cur): return rows(cur).next()
58

    
59
def row(cur):
60
    row_ = next_row(cur)
61
    consume_rows(cur)
62
    return row_
63

    
64
def next_value(cur): return next_row(cur)[0]
65

    
66
def value(cur): return row(cur)[0]
67

    
68
def values(cur): return iters.func_iter(lambda: next_value(cur))
69

    
70
def value_or_none(cur):
71
    try: return value(cur)
72
    except StopIteration: return None
73

    
74
##### Database connections
75

    
76
db_config_names = ['engine', 'host', 'user', 'password', 'database']
77

    
78
db_engines = {
79
    'MySQL': ('MySQLdb', {'password': 'passwd', 'database': 'db'}),
80
    'PostgreSQL': ('psycopg2', {}),
81
}
82

    
83
DatabaseErrors_set = set([DbException])
84
DatabaseErrors = tuple(DatabaseErrors_set)
85

    
86
def _add_module(module):
87
    DatabaseErrors_set.add(module.DatabaseError)
88
    global DatabaseErrors
89
    DatabaseErrors = tuple(DatabaseErrors_set)
90

    
91
def db_config_str(db_config):
92
    return db_config['engine']+' database '+db_config['database']
93

    
94
def _query_lookup(query, params): return (query, dicts.make_hashable(params))
95

    
96
log_debug_none = lambda msg: None
97

    
98
class DbConn:
99
    def __init__(self, db_config, serializable=True, log_debug=log_debug_none,
100
        caching=True):
101
        self.db_config = db_config
102
        self.serializable = serializable
103
        self.log_debug = log_debug
104
        self.caching = caching
105
        
106
        self.__db = None
107
        self.query_results = {}
108
    
109
    def __getattr__(self, name):
110
        if name == '__dict__': raise Exception('getting __dict__')
111
        if name == 'db': return self._db()
112
        else: raise AttributeError()
113
    
114
    def __getstate__(self):
115
        state = copy.copy(self.__dict__) # shallow copy
116
        state['log_debug'] = None # don't pickle the debug callback
117
        state['_DbConn__db'] = None # don't pickle the connection
118
        return state
119
    
120
    def _db(self):
121
        if self.__db == None:
122
            # Process db_config
123
            db_config = self.db_config.copy() # don't modify input!
124
            module_name, mappings = db_engines[db_config.pop('engine')]
125
            module = __import__(module_name)
126
            _add_module(module)
127
            for orig, new in mappings.iteritems():
128
                try: util.rename_key(db_config, orig, new)
129
                except KeyError: pass
130
            
131
            # Connect
132
            self.__db = module.connect(**db_config)
133
            
134
            # Configure connection
135
            if self.serializable: run_raw_query(self,
136
                'SET TRANSACTION ISOLATION LEVEL SERIALIZABLE')
137
        
138
        return self.__db
139
    
140
    class DbCursor(Proxy):
141
        def __init__(self, outer):
142
            Proxy.__init__(self, outer.db.cursor())
143
            self.query_results = outer.query_results
144
            self.query_lookup = None
145
            self.result = []
146
        
147
        def execute(self, query, params=None):
148
            self._is_insert = query.upper().find('INSERT') >= 0
149
            self.query_lookup = _query_lookup(query, params)
150
            try: return_value = self.inner.execute(query, params)
151
            except Exception, e:
152
                self.result = e # cache the exception as the result
153
                self._cache_result()
154
                raise
155
            finally: self.query = get_cur_query(self.inner)
156
            # Fetch all rows so result will be cached
157
            if self.rowcount == 0 and not self._is_insert: consume_rows(self)
158
            return return_value
159
        
160
        def fetchone(self):
161
            row = self.inner.fetchone()
162
            if row != None: self.result.append(row)
163
            # otherwise, fetched all rows
164
            else: self._cache_result()
165
            return row
166
        
167
        def _cache_result(self):
168
            # For inserts, only cache exceptions since inserts are not
169
            # idempotent, but an invalid insert will always be invalid
170
            if self.query_results != None and (not self._is_insert
171
                or isinstance(self.result, Exception)):
172
                
173
                assert self.query_lookup != None
174
                self.query_results[self.query_lookup] = self.CacheCursor(
175
                    util.dict_subset(dicts.AttrsDictView(self),
176
                    ['query', 'result', 'rowcount', 'description']))
177
        
178
        class CacheCursor:
179
            def __init__(self, cached_result): self.__dict__ = cached_result
180
            
181
            def execute(self, *args, **kw_args):
182
                if isinstance(self.result, Exception): raise self.result
183
                # otherwise, result is a rows list
184
                self.iter = iter(self.result)
185
            
186
            def fetchone(self):
187
                try: return self.iter.next()
188
                except StopIteration: return None
189
    
190
    def run_query(self, query, params=None, cacheable=False):
191
        if not self.caching: cacheable = False
192
        used_cache = False
193
        try:
194
            # Get cursor
195
            if cacheable:
196
                query_lookup = _query_lookup(query, params)
197
                try:
198
                    cur = self.query_results[query_lookup]
199
                    used_cache = True
200
                except KeyError: cur = self.DbCursor(self)
201
            else: cur = self.db.cursor()
202
            
203
            # Run query
204
            try: cur.execute(query, params)
205
            except Exception, e:
206
                _add_cursor_info(e, cur)
207
                raise
208
        finally:
209
            if self.log_debug != log_debug_none: # only compute msg if needed
210
                if used_cache: cache_status = 'Cache hit'
211
                elif cacheable: cache_status = 'Cache miss'
212
                else: cache_status = 'Non-cacheable'
213
                self.log_debug(cache_status+': '
214
                    +strings.one_line(get_cur_query(cur)))
215
        
216
        return cur
217
    
218
    def is_cached(self, query, params=None):
219
        return _query_lookup(query, params) in self.query_results
220

    
221
connect = DbConn
222

    
223
##### Input validation
224

    
225
def check_name(name):
226
    if re.search(r'\W', name) != None: raise NameException('Name "'+name
227
        +'" may contain only alphanumeric characters and _')
228

    
229
def esc_name_by_module(module, name, preserve_case=True):
230
    if module == 'psycopg2':
231
        if preserve_case: quote = '"'
232
        # Don't enclose in quotes because this disables case-insensitivity
233
        else:
234
            check_name(name)
235
            return name
236
    elif module == 'MySQLdb': quote = '`'
237
    else: raise NotImplementedError("Can't escape name for "+module+' database')
238
    return quote + name.replace(quote, '') + quote
239

    
240
def esc_name_by_engine(engine, name, **kw_args):
241
    return esc_name_by_module(db_engines[engine][0], name, **kw_args)
242

    
243
def esc_name(db, name, **kw_args):
244
    return esc_name_by_module(util.root_module(db.db), name, **kw_args)
245

    
246
def qual_name(db, schema, table):
247
    def esc_name_(name): return esc_name(db, name, preserve_case=True)
248
    table = esc_name_(table)
249
    if schema != None: return esc_name_(schema)+'.'+table
250
    else: return table
251

    
252
##### Querying
253

    
254
def run_raw_query(db, *args, **kw_args):
255
    '''For args, see DbConn.run_query()'''
256
    return db.run_query(*args, **kw_args)
257

    
258
##### Recoverable querying
259

    
260
def with_savepoint(db, func):
261
    savepoint = 'savepoint_'+str(rand.rand_int()) # must be unique
262
    run_raw_query(db, 'SAVEPOINT '+savepoint)
263
    try: return_val = func()
264
    except:
265
        run_raw_query(db, 'ROLLBACK TO SAVEPOINT '+savepoint)
266
        raise
267
    else:
268
        run_raw_query(db, 'RELEASE SAVEPOINT '+savepoint)
269
        return return_val
270

    
271
def run_query(db, query, params=None, recover=None, cacheable=False):
272
    if recover == None: recover = False
273
    
274
    def run(): return run_raw_query(db, query, params, cacheable)
275
    if recover and not db.is_cached(query, params):
276
        return with_savepoint(db, run)
277
    else: return run() # don't need savepoint if cached
278

    
279
##### Basic queries
280

    
281
def mk_select(db, table, fields=None, conds=None, limit=None, start=None,
282
    table_is_esc=False):
283
    '''
284
    @param fields Use None to select all fields in the table
285
    @param table_is_esc Whether the table name has already been escaped
286
    @return tuple(query, params)
287
    '''
288
    def esc_name_(name): return esc_name(db, name, preserve_case=True)
289
    
290
    if conds == None: conds = {}
291
    assert limit == None or type(limit) == int
292
    assert start == None or type(start) == int
293
    if not table_is_esc: table = esc_name_(table)
294
    
295
    params = []
296
    
297
    def parse_col(field):
298
        '''Parses fields'''
299
        if isinstance(field, tuple): # field is literal values
300
            value, col = field
301
            sql_ = '%s'
302
            params.append(value)
303
            if col != None: sql_ += ' AS '+esc_name_(col)
304
        else: sql_ = esc_name_(field) # field is col name
305
        return sql_
306
    def cond(entry):
307
        '''Parses conditions'''
308
        col, value = entry
309
        cond_ = esc_name_(col)+' '
310
        if value == None: cond_ += 'IS'
311
        else: cond_ += '='
312
        cond_ += ' %s'
313
        return cond_
314
    
315
    query = 'SELECT '
316
    if fields == None: query += '*'
317
    else: query += ', '.join(map(parse_col, fields))
318
    query += ' FROM '+table
319
    
320
    missing = True
321
    if conds != {}:
322
        query += ' WHERE '+' AND '.join(map(cond, conds.iteritems()))
323
        params += conds.values()
324
        missing = False
325
    if limit != None: query += ' LIMIT '+str(limit); missing = False
326
    if start != None:
327
        if start != 0: query += ' OFFSET '+str(start)
328
        missing = False
329
    if missing: warnings.warn(DbWarning(
330
        'SELECT statement missing a WHERE, LIMIT, or OFFSET clause: '+query))
331
    
332
    return (query, params)
333

    
334
def select(db, *args, **kw_args):
335
    '''For params, see mk_select() and run_query()'''
336
    recover = kw_args.pop('recover', None)
337
    cacheable = kw_args.pop('cacheable', True)
338
    
339
    query, params = mk_select(db, *args, **kw_args)
340
    return run_query(db, query, params, recover, cacheable)
341

    
342
default = object() # tells insert() to use the default value for a column
343

    
344
def insert(db, table, row, returning=None, recover=None, cacheable=True,
345
    table_is_esc=False):
346
    '''
347
    @param returning str|None An inserted column (such as pkey) to return
348
    @param table_is_esc Whether the table name has already been escaped
349
    '''
350
    if not table_is_esc: check_name(table)
351
    if lists.is_seq(row): cols = None
352
    else:
353
        cols = row.keys()
354
        row = row.values()
355
        map(check_name, cols)
356
    row = list(row) # ensure that "!= []" works
357
    
358
    # Check for special values
359
    labels = []
360
    values = []
361
    for value in row:
362
        if value == default: labels.append('DEFAULT')
363
        else:
364
            labels.append('%s')
365
            values.append(value)
366
    
367
    # Build query
368
    query = 'INSERT INTO '+table
369
    if values != []:
370
        if cols != None: query += ' ('+', '.join(cols)+')'
371
        query += ' VALUES ('+(', '.join(labels))+')'
372
    else: query += ' DEFAULT VALUES'
373
    
374
    if returning != None:
375
        check_name(returning)
376
        query += ' RETURNING '+returning
377
    
378
    return run_query(db, query, values, recover, cacheable)
379

    
380
def last_insert_id(db):
381
    module = util.root_module(db.db)
382
    if module == 'psycopg2': return value(run_query(db, 'SELECT lastval()'))
383
    elif module == 'MySQLdb': return db.insert_id()
384
    else: return None
385

    
386
def truncate(db, table, schema='public'):
387
    return run_query(db, 'TRUNCATE '+qual_name(db, schema, table)+' CASCADE')
388

    
389
##### Database structure queries
390

    
391
def pkey(db, table, recover=None):
392
    '''Assumed to be first column in table'''
393
    check_name(table)
394
    return col_names(select(db, table, limit=0, recover=recover)).next()
395

    
396
def index_cols(db, table, index):
397
    '''Can also use this for UNIQUE constraints, because a UNIQUE index is
398
    automatically created. When you don't know whether something is a UNIQUE
399
    constraint or a UNIQUE index, use this function.'''
400
    check_name(table)
401
    check_name(index)
402
    module = util.root_module(db.db)
403
    if module == 'psycopg2':
404
        return list(values(run_query(db, '''\
405
SELECT attname
406
FROM
407
(
408
        SELECT attnum, attname
409
        FROM pg_index
410
        JOIN pg_class index ON index.oid = indexrelid
411
        JOIN pg_class table_ ON table_.oid = indrelid
412
        JOIN pg_attribute ON attrelid = indrelid AND attnum = ANY (indkey)
413
        WHERE
414
            table_.relname = %(table)s
415
            AND index.relname = %(index)s
416
    UNION
417
        SELECT attnum, attname
418
        FROM
419
        (
420
            SELECT
421
                indrelid
422
                , (regexp_matches(indexprs, E':varattno (\\\\d+)', 'g'))[1]::int
423
                    AS indkey
424
            FROM pg_index
425
            JOIN pg_class index ON index.oid = indexrelid
426
            JOIN pg_class table_ ON table_.oid = indrelid
427
            WHERE
428
                table_.relname = %(table)s
429
                AND index.relname = %(index)s
430
        ) s
431
        JOIN pg_attribute ON attrelid = indrelid AND attnum = indkey
432
) s
433
ORDER BY attnum
434
''',
435
            {'table': table, 'index': index}, cacheable=True)))
436
    else: raise NotImplementedError("Can't list index columns for "+module+
437
        ' database')
438

    
439
def constraint_cols(db, table, constraint):
440
    check_name(table)
441
    check_name(constraint)
442
    module = util.root_module(db.db)
443
    if module == 'psycopg2':
444
        return list(values(run_query(db, '''\
445
SELECT attname
446
FROM pg_constraint
447
JOIN pg_class ON pg_class.oid = conrelid
448
JOIN pg_attribute ON attrelid = conrelid AND attnum = ANY (conkey)
449
WHERE
450
    relname = %(table)s
451
    AND conname = %(constraint)s
452
ORDER BY attnum
453
''',
454
            {'table': table, 'constraint': constraint})))
455
    else: raise NotImplementedError("Can't list constraint columns for "+module+
456
        ' database')
457

    
458
def tables(db, schema='public', table_like='%'):
459
    module = util.root_module(db.db)
460
    params = {'schema': schema, 'table_like': table_like}
461
    if module == 'psycopg2':
462
        return values(run_query(db, '''\
463
SELECT tablename
464
FROM pg_tables
465
WHERE
466
    schemaname = %(schema)s
467
    AND tablename LIKE %(table_like)s
468
ORDER BY tablename
469
''',
470
            params, cacheable=True))
471
    elif module == 'MySQLdb':
472
        return values(run_query(db, 'SHOW TABLES LIKE %(table_like)s', params,
473
            cacheable=True))
474
    else: raise NotImplementedError("Can't list tables for "+module+' database')
475

    
476
##### Database management
477

    
478
def empty_db(db, schema='public', **kw_args):
479
    '''For kw_args, see tables()'''
480
    for table in tables(db, schema, **kw_args): truncate(db, table, schema)
481

    
482
##### Heuristic queries
483

    
484
def try_insert(db, table, row, returning=None):
485
    '''Recovers from errors'''
486
    try: return insert(db, table, row, returning, recover=True)
487
    except Exception, e:
488
        msg = str(e)
489
        match = re.search(r'duplicate key value violates unique constraint '
490
            r'"(([^\W_]+)_[^"]+)"', msg)
491
        if match:
492
            constraint, table = match.groups()
493
            try: cols = index_cols(db, table, constraint)
494
            except NotImplementedError: raise e
495
            else: raise DuplicateKeyException(cols, e)
496
        match = re.search(r'null value in column "(\w+)" violates not-null '
497
            'constraint', msg)
498
        if match: raise NullValueException([match.group(1)], e)
499
        raise # no specific exception raised
500

    
501
def put(db, table, row, pkey, row_ct_ref=None):
502
    '''Recovers from errors.
503
    Only works under PostgreSQL (uses `INSERT ... RETURNING`)'''
504
    try:
505
        cur = try_insert(db, table, row, pkey)
506
        if row_ct_ref != None and cur.rowcount >= 0:
507
            row_ct_ref[0] += cur.rowcount
508
        return value(cur)
509
    except DuplicateKeyException, e:
510
        return value(select(db, table, [pkey],
511
            util.dict_subset_right_join(row, e.cols), recover=True))
512

    
513
def get(db, table, row, pkey, row_ct_ref=None, create=False):
514
    '''Recovers from errors'''
515
    try: return value(select(db, table, [pkey], row, 1, recover=True))
516
    except StopIteration:
517
        if not create: raise
518
        return put(db, table, row, pkey, row_ct_ref) # insert new row
(22-22/33)