Project

General

Profile

1 5169 aaronmk
# TNRS
2 4990 aaronmk
3 13464 aaronmk
import os.path
4 4990 aaronmk
import re
5
import sys
6
import time
7
import urllib2
8
9 5149 aaronmk
import csvs
10 5107 aaronmk
import exc
11 5120 aaronmk
import profiling
12 4990 aaronmk
import streams
13 5144 aaronmk
import strings
14 4990 aaronmk
15
# Config
16
initial_pause = 0.35 # sec
17
pause_growth_factor = 1.3
18 5125 aaronmk
max_pause = 30*60 # sec; = 30 min; TNRS sometimes freezes for ~10 min
19 4990 aaronmk
assert initial_pause <= max_pause
20 13548 aaronmk
#max_names = 5000 #according to http://tnrs.iplantcollaborative.org/TNRSapp.html
21 13636 aaronmk
max_names = 500 # the maximum above crashes the live and dev TNRS servers
22 4990 aaronmk
23
# Protocol params
24 13464 aaronmk
server = streams.file_get_contents(os.path.dirname(__file__)+"/tnrs.url")
25
#server = 'tnrs.iplantcollaborative.org' # live server
26 13462 aaronmk
url_base = 'http://'+server+'/tnrsdemo/'
27 4990 aaronmk
url = url_base+'search'
28
initial_headers = {
29
    'Content-Type': 'text/x-gwt-rpc; charset=utf-8',
30
    'X-GWT-Module-Base': url_base,
31
    'X-GWT-Permutation': '574AA16D15D917C7704646FD92AFF6B3',
32
}
33
submission_request_template = ('7|0|7|'+url_base+
34 9912 aaronmk
'|1E87C78041CEFBF0992F46BDF84D7D60|org.iplantc.tnrs.demo.client.SearchService|\
35 13544 aaronmk
doSearch|java.lang.String/2004016611|{"sources":"gcc,tpl,tropicos,usda", \
36 9912 aaronmk
"names":"[names]", "type":"matching", "taxonomic":"true", \
37
"classification":"tropicos", "match_to_rank":"true"}|0.05|1|2|3|4|2|5|5|6|7|')
38 4990 aaronmk
submission_response_pattern = r'^//OK\[1,\["(\w+)"\],0,7\]$'
39
retrieval_request_template = ('7|0|15|'+url_base+
40
'|1E87C78041CEFBF0992F46BDF84D7D60|org.iplantc.tnrs.demo.client.SearchService\
41
|getRemoteData|com.extjs.gxt.ui.client.data.PagingLoadConfig|\
42
java.lang.String/2004016611|com.extjs.gxt.ui.client.data.BasePagingLoadConfig/\
43
2011366567|com.extjs.gxt.ui.client.data.RpcMap/3441186752|sortField|sortDir|\
44
com.extjs.gxt.ui.client.Style$SortDir/640452531|offset|java.lang.Integer/\
45
3438268394|limit|{"email":"tnrs@lka5jjs.orv", "key":"[key]", \
46 13597 aaronmk
"taxonomic_constraint":"true", "source_sorting":"true", "first":"false"}\
47 4990 aaronmk
|1|2|3|4|2|5|6|7|0|1|8|4|9|0|10|11|0|12|13|0|14|13|100|15|')
48 13436 aaronmk
	# taxonomic_constraint (Constrain by Higher Taxonomy): selects lower ranks
49
	#  only from within the matched higher ranks. must be turned on, to ensure
50
	#  that higher ranks are always authoritative.
51 13833 aaronmk
	# source_sorting (Constrain by Source):
52
	#  has different behavior depending on the match mode:
53
	#  - in all-matches mode (`"mode":"All"` @ download_request_template):
54
	#    always puts matches in the order of the sources. although the 1st match
55
	#    won't necessarily be the best one, the 1st match *with Selected=true*
56
	#    will be (according to the source sort order). must be turned on, so
57
	#    that the selected matches are sorted by source.
58
	#  - in best-match mode (`"mode":"Best"` @ download_request_template):
59
	#    always selects a match from the first source in the list, no matter how
60
	#    low the match score. must be turned *off* (unlike in all-matches mode),
61
	#    to avoid worse matches being selected instead of better ones.
62
	#    *however*, since this is currently broken and always on, we turn it on
63
	#    so that the download settings reflect what TNRS actually used.
64 4990 aaronmk
retrieval_response_pattern = '^//OK\[.*?\["com.extjs.gxt.ui.client.data.\
65
BasePagingLoadResult/496878394","java.util.ArrayList/4159755760","org.iplantc.\
66
tnrs.demo.shared.BeanTNRSEntry/1039545748",".*"\],0,7\]$'
67
retrieval_response_info_pattern = r'(?ms).*^Set-Cookie: JSESSIONID=(\w+);'
68
download_request_template = ('7|0|6|'+url_base+
69
'|1E87C78041CEFBF0992F46BDF84D7D60|org.iplantc.tnrs.demo.client.SearchService|\
70
downloadRemoteResults|java.lang.String/2004016611|{"name":"tnrs_results.txt", \
71 13591 aaronmk
"mode":"All", "type":"Detailed", "encoding":"utf8", "dirty":"true", \
72 9911 aaronmk
"sources":"false", "taxonomic":"true", "email":"tnrs@lka5jjs.orv", \
73 4990 aaronmk
"key":"[key]"}|1|2|3|4|1|5|6|')
74
download_response_pattern = '^//OK\[1,\["(.*)"\],0,7\]$'
75
download_url_suffix = '&name=tnrs_results.txt&encoding=utf8'
76
77 5083 aaronmk
class InvalidResponse(Exception): pass
78
79 5144 aaronmk
def gwt_encode(str_):
80
    return strings.esc_quotes(strings.json_encode(str_), '|', quote_esc='\!')
81 4990 aaronmk
82 5149 aaronmk
def make_spliced_decode_map(decode_map):
83
    return [(r'(?: |(?<=\t)|^)'+re.escape(from_.strip())+r'(?: |(?=\t)|$)',
84
        strings.regexp_repl_esc(to)) for from_, to in decode_map]
85
86 5154 aaronmk
padding = ' !pad ' # prepend to empty and whitespace-only strings
87 5149 aaronmk
encode_map = [
88 5167 aaronmk
    ('!', ' !exc '), # our escape char
89
    ('\t', ' !tab '), # TNRS replaces with " "
90
    ('\n', ' !nl '), # used to separate multiple names
91
    ('\r', ' !cr '), # used to separate multiple names
92 5168 aaronmk
    ('"', ' !quo '), # TNRS removes it when at the beginning or end
93 5167 aaronmk
    ('%', ' !pct '), # TNRS URL-decodes it in matched fields
94 5168 aaronmk
    ("'", ' !apo '), # TNRS removes it when at the beginning or end
95 5167 aaronmk
    (';', ' !sem '), # changes TNRS response format
96
    ('\\', ' !bsl '), # TNRS removes it
97 5165 aaronmk
    ('_', ' !und '), # TNRS replaces with " "
98 5171 aaronmk
    ('­', ' !sub '), # TNRS removes it
99 5169 aaronmk
    ('×', ' !mul '), # TNRS replaces with "x"
100 5149 aaronmk
]
101
decode_map = strings.flip_map(encode_map)
102 5154 aaronmk
decode_map.append((padding, ''))
103 5149 aaronmk
spliced_decode_map = make_spliced_decode_map(decode_map)
104
105 5154 aaronmk
def encode(str_):
106
    str_ = strings.replace_all(encode_map, str_)
107
    # Empty and whitespace-only strings are ignored by TNRS (no response row)
108
    if str_.strip() == '': str_ = padding+str_
109
    return str_
110 5149 aaronmk
111
def decode(str_): return strings.replace_all_re(spliced_decode_map, str_)
112
113
decode_for_tsv_map = make_spliced_decode_map([(from_, strings.replace_all(
114
    csvs.tsv_encode_map, to)) for from_, to in decode_map])
115
116
def decode_for_tsv(str_):
117
    return strings.replace_all_re(decode_for_tsv_map, str_)
118
119
class TnrsOutputStream(streams.FilterStream):
120
    '''Decodes a TNRS response whose names were encoded with encode()'''
121
    def __init__(self, stream):
122
        streams.FilterStream.__init__(self, decode_for_tsv, stream)
123
124 5106 aaronmk
def parse_response(name, pattern, str_, response, response_info):
125
    match = re.match(pattern, str_)
126
    if not match:
127
        raise InvalidResponse('Invalid '+name+' response:\n'+response_info+'\n'
128
            +response)
129 4990 aaronmk
    return match.groups()
130
131 9525 aaronmk
def single_tnrs_request(names, debug=False, cumulative_profiler=None):
132 5127 aaronmk
    '''
133
    Note that names containing only whitespace characters (after gwt_encode())
134
    are ignored by TNRS and do not receive a response row. Thus, you should
135
    always match up the Name_submitted returned by TNRS with the actual
136
    submitted name to determine the corresponding TNRS response row.
137
    '''
138 5121 aaronmk
    name_ct = len(names)
139
    assert name_ct <= max_names
140 4990 aaronmk
141
    # Logging
142
    def debug_log(label, str_=''):
143
        if debug: sys.stderr.write('\n'+label+':\n'+str_+'\n')
144
145
    ## HTTP
146 5119 aaronmk
    headers = initial_headers.copy() # don't modify global constant!
147 4990 aaronmk
148 5005 aaronmk
    def do_request(request):
149 4990 aaronmk
        debug_log('request', str(request))
150
        response = urllib2.urlopen(urllib2.Request(url, request, headers))
151
        response_str = streams.read_all(response)
152
        response_info = str(response.info())
153
        debug_log('response info', response_info)
154
        debug_log('response str', response_str)
155
        return response_str, response_info
156
157
    def do_repeated_request(request):
158
        pause = initial_pause
159
        total_pause = 0
160
        while True:
161
            total_pause += pause
162
            if total_pause > max_pause: raise # error is not temporary
163
            debug_log('total_pause', str(total_pause)+'s')
164
            time.sleep(pause) # wait for job to complete
165
166
            try: return do_request(request)
167
            except urllib2.HTTPError: pass # try again
168
            pause *= pause_growth_factor
169
170 5120 aaronmk
    profiler = profiling.ItersProfiler(start_now=True, iter_text='name')
171
    try:
172
        debug_log('Submit')
173 5121 aaronmk
        request = submission_request_template.replace('[names]',
174 5150 aaronmk
            gwt_encode('\n'.join(map(encode, names))))
175 5120 aaronmk
        response, response_info = do_request(request)
176 5151 aaronmk
        key, = parse_response('submission', submission_response_pattern,
177
            response, response, response_info)
178 5120 aaronmk
        debug_log('key', key)
179
        key_enc = gwt_encode(key)
180
181
        debug_log('Retrieve')
182
        request = retrieval_request_template.replace('[key]', key_enc)
183
        response, response_info = do_repeated_request(request)
184 5151 aaronmk
        parse_response('retrieval', retrieval_response_pattern, response,
185
            response, response_info)
186
        session_id, = parse_response('retrieval info',
187
            retrieval_response_info_pattern, response_info, response,
188 5120 aaronmk
            response_info)
189
        debug_log('session_id', session_id)
190
        headers['Cookie'] = 'JSESSIONID='+session_id
191
192
        # The output of the retrieve step is unusable because the array has
193 5151 aaronmk
        # different lengths depending on the taxonomic ranks present in the
194
        # provided taxon name. The extra download step is therefore necessary.
195 5120 aaronmk
196
        debug_log('Prepare download')
197
        request = download_request_template.replace('[key]', key_enc)
198
        response, response_info = do_request(request)
199 5151 aaronmk
        csv_url, = parse_response('download', download_response_pattern,
200
            response, response, response_info)
201 5120 aaronmk
        csv_url += download_url_suffix
202
        debug_log('csv_url', csv_url)
203
204
        debug_log('Download')
205
        response = urllib2.urlopen(urllib2.Request(csv_url))
206
        debug_log('response info', str(response.info()))
207 5150 aaronmk
        return TnrsOutputStream(response)
208 5120 aaronmk
    finally:
209 5121 aaronmk
        profiler.stop(name_ct)
210 5120 aaronmk
        sys.stderr.write(profiler.msg()+'\n')
211 9525 aaronmk
212
        if cumulative_profiler != None:
213
            cumulative_profiler.add_subprofiler(profiler)
214
            sys.stderr.write('Cumulatively: '+cumulative_profiler.msg()+'\n')
215 5088 aaronmk
216 9520 aaronmk
def tnrs_request(names, debug=False, **kw_args):
217 5108 aaronmk
    for try_num in xrange(2):
218 9519 aaronmk
        try: return single_tnrs_request(names, debug, **kw_args)
219 5160 aaronmk
        except (urllib2.HTTPError, InvalidResponse), e:
220 5107 aaronmk
            exc.print_ex(e, detail=False)
221 5108 aaronmk
            debug = True
222
            # try again with debug turned on
223 5088 aaronmk
    raise # error is not temporary