Project

General

Profile

1
# Database access
2

    
3
import copy
4
import re
5
import warnings
6

    
7
import exc
8
import dicts
9
import iters
10
import lists
11
from Proxy import Proxy
12
import rand
13
import strings
14
import util
15

    
16
##### Exceptions
17

    
18
def get_cur_query(cur):
19
    if hasattr(cur, 'query'): return cur.query
20
    elif hasattr(cur, '_last_executed'): return cur._last_executed
21
    else: return None
22

    
23
def _add_cursor_info(e, cur): exc.add_msg(e, 'query: '+get_cur_query(cur))
24

    
25
class DbException(exc.ExceptionWithCause):
26
    def __init__(self, msg, cause=None, cur=None):
27
        exc.ExceptionWithCause.__init__(self, msg, cause)
28
        if cur != None: _add_cursor_info(self, cur)
29

    
30
class NameException(DbException): pass
31

    
32
class ExceptionWithColumns(DbException):
33
    def __init__(self, cols, cause=None):
34
        DbException.__init__(self, 'columns: ' + ', '.join(cols), cause)
35
        self.cols = cols
36

    
37
class DuplicateKeyException(ExceptionWithColumns): pass
38

    
39
class NullValueException(ExceptionWithColumns): pass
40

    
41
class EmptyRowException(DbException): pass
42

    
43
##### Warnings
44

    
45
class DbWarning(UserWarning): pass
46

    
47
##### Result retrieval
48

    
49
def col_names(cur): return (col[0] for col in cur.description)
50

    
51
def rows(cur): return iter(lambda: cur.fetchone(), None)
52

    
53
def consume_rows(cur):
54
    '''Used to fetch all rows so result will be cached'''
55
    iters.consume_iter(rows(cur))
56

    
57
def next_row(cur): return rows(cur).next()
58

    
59
def row(cur):
60
    row_ = next_row(cur)
61
    consume_rows(cur)
62
    return row_
63

    
64
def next_value(cur): return next_row(cur)[0]
65

    
66
def value(cur): return row(cur)[0]
67

    
68
def values(cur): return iters.func_iter(lambda: next_value(cur))
69

    
70
def value_or_none(cur):
71
    try: return value(cur)
72
    except StopIteration: return None
73

    
74
##### Database connections
75

    
76
db_config_names = ['engine', 'host', 'user', 'password', 'database']
77

    
78
db_engines = {
79
    'MySQL': ('MySQLdb', {'password': 'passwd', 'database': 'db'}),
80
    'PostgreSQL': ('psycopg2', {}),
81
}
82

    
83
DatabaseErrors_set = set([DbException])
84
DatabaseErrors = tuple(DatabaseErrors_set)
85

    
86
def _add_module(module):
87
    DatabaseErrors_set.add(module.DatabaseError)
88
    global DatabaseErrors
89
    DatabaseErrors = tuple(DatabaseErrors_set)
90

    
91
def db_config_str(db_config):
92
    return db_config['engine']+' database '+db_config['database']
93

    
94
def _query_lookup(query, params): return (query, dicts.make_hashable(params))
95

    
96
log_debug_none = lambda msg: None
97

    
98
class DbConn:
99
    def __init__(self, db_config, serializable=True, log_debug=log_debug_none,
100
        caching=True):
101
        self.db_config = db_config
102
        self.serializable = serializable
103
        self.log_debug = log_debug
104
        self.caching = caching
105
        
106
        self.__db = None
107
        self.query_results = {}
108
    
109
    def __getattr__(self, name):
110
        if name == '__dict__': raise Exception('getting __dict__')
111
        if name == 'db': return self._db()
112
        else: raise AttributeError()
113
    
114
    def __getstate__(self):
115
        state = copy.copy(self.__dict__) # shallow copy
116
        state['log_debug'] = None # don't pickle the debug callback
117
        state['_DbConn__db'] = None # don't pickle the connection
118
        return state
119
    
120
    def _db(self):
121
        if self.__db == None:
122
            # Process db_config
123
            db_config = self.db_config.copy() # don't modify input!
124
            module_name, mappings = db_engines[db_config.pop('engine')]
125
            module = __import__(module_name)
126
            _add_module(module)
127
            for orig, new in mappings.iteritems():
128
                try: util.rename_key(db_config, orig, new)
129
                except KeyError: pass
130
            
131
            # Connect
132
            self.__db = module.connect(**db_config)
133
            
134
            # Configure connection
135
            if self.serializable: run_raw_query(self,
136
                'SET TRANSACTION ISOLATION LEVEL SERIALIZABLE')
137
        
138
        return self.__db
139
    
140
    class DbCursor(Proxy):
141
        def __init__(self, outer):
142
            Proxy.__init__(self, outer.db.cursor())
143
            self.query_results = outer.query_results
144
            self.query_lookup = None
145
            self.result = []
146
        
147
        def execute(self, query, params=None):
148
            self._is_insert = query.upper().find('INSERT') >= 0
149
            self.query_lookup = _query_lookup(query, params)
150
            try: return_value = self.inner.execute(query, params)
151
            except Exception, e:
152
                self.result = e # cache the exception as the result
153
                self._cache_result()
154
                raise
155
            finally: self.query = get_cur_query(self.inner)
156
            # Fetch all rows so result will be cached
157
            if self.rowcount == 0 and not self._is_insert: consume_rows(self)
158
            return return_value
159
        
160
        def fetchone(self):
161
            row = self.inner.fetchone()
162
            if row != None: self.result.append(row)
163
            # otherwise, fetched all rows
164
            else: self._cache_result()
165
            return row
166
        
167
        def _cache_result(self):
168
            # For inserts, only cache exceptions since inserts are not
169
            # idempotent, but an invalid insert will always be invalid
170
            if self.query_results != None and (not self._is_insert
171
                or isinstance(self.result, Exception)):
172
                
173
                assert self.query_lookup != None
174
                self.query_results[self.query_lookup] = self.CacheCursor(
175
                    util.dict_subset(dicts.AttrsDictView(self),
176
                    ['query', 'result', 'rowcount', 'description']))
177
        
178
        class CacheCursor:
179
            def __init__(self, cached_result): self.__dict__ = cached_result
180
            
181
            def execute(self, *args, **kw_args):
182
                if isinstance(self.result, Exception): raise self.result
183
                # otherwise, result is a rows list
184
                self.iter = iter(self.result)
185
            
186
            def fetchone(self):
187
                try: return self.iter.next()
188
                except StopIteration: return None
189
    
190
    def run_query(self, query, params=None, cacheable=False):
191
        if not self.caching: cacheable = False
192
        used_cache = False
193
        try:
194
            # Get cursor
195
            if cacheable:
196
                query_lookup = _query_lookup(query, params)
197
                try:
198
                    cur = self.query_results[query_lookup]
199
                    used_cache = True
200
                except KeyError: cur = self.DbCursor(self)
201
            else: cur = self.db.cursor()
202
            
203
            # Run query
204
            try: cur.execute(query, params)
205
            except Exception, e:
206
                _add_cursor_info(e, cur)
207
                raise
208
        finally:
209
            if self.log_debug != log_debug_none: # only compute msg if needed
210
                if used_cache: cache_status = 'Cache hit'
211
                elif cacheable: cache_status = 'Cache miss'
212
                else: cache_status = 'Non-cacheable'
213
                self.log_debug(cache_status+': '
214
                    +strings.one_line(get_cur_query(cur)))
215
        
216
        return cur
217
    
218
    def is_cached(self, query, params=None):
219
        return _query_lookup(query, params) in self.query_results
220

    
221
connect = DbConn
222

    
223
##### Input validation
224

    
225
def check_name(name):
226
    if re.search(r'\W', name) != None: raise NameException('Name "'+name
227
        +'" may contain only alphanumeric characters and _')
228

    
229
def esc_name_by_module(module, name, ignore_case=False):
230
    if module == 'psycopg2':
231
        if ignore_case:
232
            # Don't enclose in quotes because this disables case-insensitivity
233
            check_name(name)
234
            return name
235
        else: quote = '"'
236
    elif module == 'MySQLdb': quote = '`'
237
    else: raise NotImplementedError("Can't escape name for "+module+' database')
238
    return quote + name.replace(quote, '') + quote
239

    
240
def esc_name_by_engine(engine, name, **kw_args):
241
    return esc_name_by_module(db_engines[engine][0], name, **kw_args)
242

    
243
def esc_name(db, name, **kw_args):
244
    return esc_name_by_module(util.root_module(db.db), name, **kw_args)
245

    
246
def qual_name(db, schema, table):
247
    def esc_name_(name): return esc_name(db, name)
248
    table = esc_name_(table)
249
    if schema != None: return esc_name_(schema)+'.'+table
250
    else: return table
251

    
252
##### Querying
253

    
254
def run_raw_query(db, *args, **kw_args):
255
    '''For args, see DbConn.run_query()'''
256
    return db.run_query(*args, **kw_args)
257

    
258
def mogrify(db, query, params):
259
    module = util.root_module(db.db)
260
    if module == 'psycopg2': return db.db.cursor().mogrify(query, params)
261
    else: raise NotImplementedError("Can't mogrify query for "+module+
262
        ' database')
263

    
264
##### Recoverable querying
265

    
266
def with_savepoint(db, func):
267
    savepoint = 'savepoint_'+str(rand.rand_int()) # must be unique
268
    run_raw_query(db, 'SAVEPOINT '+savepoint)
269
    try: return_val = func()
270
    except:
271
        run_raw_query(db, 'ROLLBACK TO SAVEPOINT '+savepoint)
272
        raise
273
    else:
274
        run_raw_query(db, 'RELEASE SAVEPOINT '+savepoint)
275
        return return_val
276

    
277
def run_query(db, query, params=None, recover=None, cacheable=False):
278
    if recover == None: recover = False
279
    
280
    def run(): return run_raw_query(db, query, params, cacheable)
281
    if recover and not db.is_cached(query, params):
282
        return with_savepoint(db, run)
283
    else: return run() # don't need savepoint if cached
284

    
285
##### Basic queries
286

    
287
def mk_select(db, table, fields=None, conds=None, limit=None, start=None,
288
    table_is_esc=False):
289
    '''
290
    @param fields Use None to select all fields in the table
291
    @param table_is_esc Whether the table name has already been escaped
292
    @return tuple(query, params)
293
    '''
294
    def esc_name_(name): return esc_name(db, name)
295
    
296
    if conds == None: conds = {}
297
    assert limit == None or type(limit) == int
298
    assert start == None or type(start) == int
299
    if not table_is_esc: table = esc_name_(table)
300
    
301
    params = []
302
    
303
    def parse_col(field):
304
        '''Parses fields'''
305
        if isinstance(field, tuple): # field is literal values
306
            value, col = field
307
            sql_ = '%s'
308
            params.append(value)
309
            if col != None: sql_ += ' AS '+esc_name_(col)
310
        else: sql_ = esc_name_(field) # field is col name
311
        return sql_
312
    def cond(entry):
313
        '''Parses conditions'''
314
        col, value = entry
315
        cond_ = esc_name_(col)+' '
316
        if value == None: cond_ += 'IS'
317
        else: cond_ += '='
318
        cond_ += ' %s'
319
        return cond_
320
    
321
    query = 'SELECT '
322
    if fields == None: query += '*'
323
    else: query += ', '.join(map(parse_col, fields))
324
    query += ' FROM '+table
325
    
326
    missing = True
327
    if conds != {}:
328
        query += ' WHERE '+' AND '.join(map(cond, conds.iteritems()))
329
        params += conds.values()
330
        missing = False
331
    if limit != None: query += ' LIMIT '+str(limit); missing = False
332
    if start != None:
333
        if start != 0: query += ' OFFSET '+str(start)
334
        missing = False
335
    if missing: warnings.warn(DbWarning(
336
        'SELECT statement missing a WHERE, LIMIT, or OFFSET clause: '+query))
337
    
338
    return (query, params)
339

    
340
def select(db, *args, **kw_args):
341
    '''For params, see mk_select() and run_query()'''
342
    recover = kw_args.pop('recover', None)
343
    cacheable = kw_args.pop('cacheable', True)
344
    
345
    query, params = mk_select(db, *args, **kw_args)
346
    return run_query(db, query, params, recover, cacheable)
347

    
348
def mk_insert_select(db, table, cols=None, select_query=None, params=None,
349
    returning=None, table_is_esc=False):
350
    '''
351
    @param returning str|None An inserted column (such as pkey) to return
352
    @param table_is_esc Whether the table name has already been escaped
353
    '''
354
    if select_query == None: select_query = 'DEFAULT VALUES'
355
    if cols == []: cols = None # no cols (all defaults) = unknown col names
356
    if not table_is_esc: check_name(table)
357
    
358
    # Build query
359
    query = 'INSERT INTO '+table
360
    if cols != None:
361
        map(check_name, cols)
362
        query += ' ('+', '.join(cols)+')'
363
    query += ' '+select_query
364
    
365
    if returning != None:
366
        check_name(returning)
367
        query += ' RETURNING '+returning
368
    
369
    return (query, params)
370

    
371
def insert_select(db, *args, **kw_args):
372
    '''For params, see mk_insert_select() and run_query()'''
373
    recover = kw_args.pop('recover', None)
374
    cacheable = kw_args.pop('cacheable', True)
375
    
376
    query, params = mk_insert_select(db, *args, **kw_args)
377
    return run_query(db, query, params, recover, cacheable)
378

    
379
default = object() # tells insert() to use the default value for a column
380

    
381
def insert(db, table, row, *args, **kw_args):
382
    '''For args, see insert_select()'''
383
    if lists.is_seq(row): cols = None
384
    else:
385
        cols = row.keys()
386
        row = row.values()
387
    row = list(row) # ensure that "!= []" works
388
    
389
    # Check for special values
390
    labels = []
391
    values = []
392
    for value in row:
393
        if value == default: labels.append('DEFAULT')
394
        else:
395
            labels.append('%s')
396
            values.append(value)
397
    
398
    # Build query
399
    if values != []: query = ' VALUES ('+(', '.join(labels))+')'
400
    else: query = None
401
    
402
    return insert_select(db, table, cols, query, values, *args, **kw_args)
403

    
404
def last_insert_id(db):
405
    module = util.root_module(db.db)
406
    if module == 'psycopg2': return value(run_query(db, 'SELECT lastval()'))
407
    elif module == 'MySQLdb': return db.insert_id()
408
    else: return None
409

    
410
def truncate(db, table, schema='public'):
411
    return run_query(db, 'TRUNCATE '+qual_name(db, schema, table)+' CASCADE')
412

    
413
##### Database structure queries
414

    
415
def pkey(db, table, recover=None):
416
    '''Assumed to be first column in table'''
417
    check_name(table)
418
    return col_names(select(db, table, limit=0, recover=recover)).next()
419

    
420
def index_cols(db, table, index):
421
    '''Can also use this for UNIQUE constraints, because a UNIQUE index is
422
    automatically created. When you don't know whether something is a UNIQUE
423
    constraint or a UNIQUE index, use this function.'''
424
    check_name(table)
425
    check_name(index)
426
    module = util.root_module(db.db)
427
    if module == 'psycopg2':
428
        return list(values(run_query(db, '''\
429
SELECT attname
430
FROM
431
(
432
        SELECT attnum, attname
433
        FROM pg_index
434
        JOIN pg_class index ON index.oid = indexrelid
435
        JOIN pg_class table_ ON table_.oid = indrelid
436
        JOIN pg_attribute ON attrelid = indrelid AND attnum = ANY (indkey)
437
        WHERE
438
            table_.relname = %(table)s
439
            AND index.relname = %(index)s
440
    UNION
441
        SELECT attnum, attname
442
        FROM
443
        (
444
            SELECT
445
                indrelid
446
                , (regexp_matches(indexprs, E':varattno (\\\\d+)', 'g'))[1]::int
447
                    AS indkey
448
            FROM pg_index
449
            JOIN pg_class index ON index.oid = indexrelid
450
            JOIN pg_class table_ ON table_.oid = indrelid
451
            WHERE
452
                table_.relname = %(table)s
453
                AND index.relname = %(index)s
454
        ) s
455
        JOIN pg_attribute ON attrelid = indrelid AND attnum = indkey
456
) s
457
ORDER BY attnum
458
''',
459
            {'table': table, 'index': index}, cacheable=True)))
460
    else: raise NotImplementedError("Can't list index columns for "+module+
461
        ' database')
462

    
463
def constraint_cols(db, table, constraint):
464
    check_name(table)
465
    check_name(constraint)
466
    module = util.root_module(db.db)
467
    if module == 'psycopg2':
468
        return list(values(run_query(db, '''\
469
SELECT attname
470
FROM pg_constraint
471
JOIN pg_class ON pg_class.oid = conrelid
472
JOIN pg_attribute ON attrelid = conrelid AND attnum = ANY (conkey)
473
WHERE
474
    relname = %(table)s
475
    AND conname = %(constraint)s
476
ORDER BY attnum
477
''',
478
            {'table': table, 'constraint': constraint})))
479
    else: raise NotImplementedError("Can't list constraint columns for "+module+
480
        ' database')
481

    
482
def tables(db, schema='public', table_like='%'):
483
    module = util.root_module(db.db)
484
    params = {'schema': schema, 'table_like': table_like}
485
    if module == 'psycopg2':
486
        return values(run_query(db, '''\
487
SELECT tablename
488
FROM pg_tables
489
WHERE
490
    schemaname = %(schema)s
491
    AND tablename LIKE %(table_like)s
492
ORDER BY tablename
493
''',
494
            params, cacheable=True))
495
    elif module == 'MySQLdb':
496
        return values(run_query(db, 'SHOW TABLES LIKE %(table_like)s', params,
497
            cacheable=True))
498
    else: raise NotImplementedError("Can't list tables for "+module+' database')
499

    
500
##### Database management
501

    
502
def empty_db(db, schema='public', **kw_args):
503
    '''For kw_args, see tables()'''
504
    for table in tables(db, schema, **kw_args): truncate(db, table, schema)
505

    
506
##### Heuristic queries
507

    
508
def try_insert(db, table, row, returning=None):
509
    '''Recovers from errors'''
510
    try: return insert(db, table, row, returning, recover=True)
511
    except Exception, e:
512
        msg = str(e)
513
        match = re.search(r'duplicate key value violates unique constraint '
514
            r'"(([^\W_]+)_[^"]+)"', msg)
515
        if match:
516
            constraint, table = match.groups()
517
            try: cols = index_cols(db, table, constraint)
518
            except NotImplementedError: raise e
519
            else: raise DuplicateKeyException(cols, e)
520
        match = re.search(r'null value in column "(\w+)" violates not-null '
521
            'constraint', msg)
522
        if match: raise NullValueException([match.group(1)], e)
523
        raise # no specific exception raised
524

    
525
def put(db, table, row, pkey, row_ct_ref=None):
526
    '''Recovers from errors.
527
    Only works under PostgreSQL (uses `INSERT ... RETURNING`)'''
528
    try:
529
        cur = try_insert(db, table, row, pkey)
530
        if row_ct_ref != None and cur.rowcount >= 0:
531
            row_ct_ref[0] += cur.rowcount
532
        return value(cur)
533
    except DuplicateKeyException, e:
534
        return value(select(db, table, [pkey],
535
            util.dict_subset_right_join(row, e.cols), recover=True))
536

    
537
def get(db, table, row, pkey, row_ct_ref=None, create=False):
538
    '''Recovers from errors'''
539
    try: return value(select(db, table, [pkey], row, 1, recover=True))
540
    except StopIteration:
541
        if not create: raise
542
        return put(db, table, row, pkey, row_ct_ref) # insert new row
(22-22/33)