1 |
4918
|
aaronmk
|
#!/usr/bin/env python
|
2 |
|
|
# A TNRS client
|
3 |
4920
|
aaronmk
|
# When using xargs to pass many names, note that xargs will by default split its
|
4 |
|
|
# arguments into chunks of 5000. You can change this using the -n option.
|
5 |
4918
|
aaronmk
|
# Note that obtaining an actual CSV requires four (!) steps: submit, retrieve,
|
6 |
|
|
# prepare download, and download. The output of the retrieve step is unusable
|
7 |
|
|
# because the array has different lengths depending on the taxonomic ranks
|
8 |
|
|
# present in the provided taxon name.
|
9 |
|
|
|
10 |
|
|
import os.path
|
11 |
|
|
import re
|
12 |
|
|
import sys
|
13 |
|
|
import time
|
14 |
|
|
import urllib2
|
15 |
|
|
|
16 |
|
|
sys.path.append(os.path.dirname(__file__)+"/../lib")
|
17 |
|
|
|
18 |
|
|
import opts
|
19 |
|
|
import streams
|
20 |
|
|
#import util
|
21 |
|
|
|
22 |
|
|
# Config
|
23 |
|
|
initial_pause = 0.35 # sec
|
24 |
|
|
pause_growth_factor = 1.3
|
25 |
4920
|
aaronmk
|
max_pause = 600 # sec; = 10 min
|
26 |
4918
|
aaronmk
|
assert initial_pause <= max_pause
|
27 |
4920
|
aaronmk
|
max_taxons = 5000# according to http://tnrs.iplantcollaborative.org/TNRSapp.html
|
28 |
4918
|
aaronmk
|
|
29 |
|
|
# Protocol params
|
30 |
|
|
url_base = 'http://tnrs.iplantcollaborative.org/tnrsdemo/'
|
31 |
|
|
url = url_base+'search'
|
32 |
|
|
initial_headers = {
|
33 |
|
|
'Content-Type': 'text/x-gwt-rpc; charset=utf-8',
|
34 |
|
|
'X-GWT-Module-Base': url_base,
|
35 |
|
|
'X-GWT-Permutation': '574AA16D15D917C7704646FD92AFF6B3',
|
36 |
|
|
}
|
37 |
|
|
submission_request_template = ('7|0|7|'+url_base+
|
38 |
|
|
'||org.iplantc.tnrs.demo.client.SearchService|doSearch|\
|
39 |
4920
|
aaronmk
|
java.lang.String/2004016611|{"sources":"gcc,tropicos,usda", "names":"[taxons]"\
|
40 |
|
|
, "type":"matching", "taxonomic":"true", "classification":"tropicos", \
|
41 |
4918
|
aaronmk
|
"match_to_rank":"true"}|0.05|1|2|3|4|2|5|5|6|7|')
|
42 |
|
|
submission_response_pattern = r'^//OK\[1,\["(\w+)"\],0,7\]$'
|
43 |
|
|
retrieval_request_template = ('7|0|15|'+url_base+
|
44 |
|
|
'|1E87C78041CEFBF0992F46BDF84D7D60|org.iplantc.tnrs.demo.client.SearchService\
|
45 |
|
|
|getRemoteData|com.extjs.gxt.ui.client.data.PagingLoadConfig|\
|
46 |
|
|
java.lang.String/2004016611|com.extjs.gxt.ui.client.data.BasePagingLoadConfig/\
|
47 |
|
|
2011366567|com.extjs.gxt.ui.client.data.RpcMap/3441186752|sortField|sortDir|\
|
48 |
|
|
com.extjs.gxt.ui.client.Style$SortDir/640452531|offset|java.lang.Integer/\
|
49 |
|
|
3438268394|limit|{"email":"tnrs@lka5jjs.orv", "key":"[key]", \
|
50 |
|
|
"taxonomic_constraint":"false", "source_sorting":"false", "first":"false"}\
|
51 |
|
|
|1|2|3|4|2|5|6|7|0|1|8|4|9|0|10|11|0|12|13|0|14|13|100|15|')
|
52 |
|
|
retrieval_response_pattern = '^//OK\[.*?\["com.extjs.gxt.ui.client.data.\
|
53 |
|
|
BasePagingLoadResult/496878394","java.util.ArrayList/4159755760","org.iplantc.\
|
54 |
|
|
tnrs.demo.shared.BeanTNRSEntry/1039545748",".*"\],0,7\]$'
|
55 |
|
|
retrieval_response_info_pattern = r'(?ms).*^Set-Cookie: JSESSIONID=(\w+);'
|
56 |
|
|
download_request_template = ('7|0|6|'+url_base+
|
57 |
|
|
'|1E87C78041CEFBF0992F46BDF84D7D60|org.iplantc.tnrs.demo.client.SearchService|\
|
58 |
|
|
downloadRemoteResults|java.lang.String/2004016611|{"name":"tnrs_results.txt", \
|
59 |
|
|
"mode":"Best", "type":"Detailed", "encoding":"utf8", "dirty":"false", \
|
60 |
|
|
"sources":"false", "taxonomic":"false", "email":"tnrs@lka5jjs.orv", \
|
61 |
|
|
"key":"[key]"}|1|2|3|4|1|5|6|')
|
62 |
|
|
download_response_pattern = '^//OK\[1,\["(.*)"\],0,7\]$'
|
63 |
|
|
download_url_suffix = '&name=tnrs_results.txt&encoding=utf8'
|
64 |
|
|
|
65 |
|
|
def gwt_encode(str_): return re.sub(r'[^\w.() -]+', r' ', str_)
|
66 |
|
|
|
67 |
|
|
def parse_response(name, pattern, response):
|
68 |
|
|
match = re.match(pattern, response)
|
69 |
|
|
if not match: raise SystemExit('Invalid '+name+' response:\n'+response)
|
70 |
|
|
return match.groups()
|
71 |
|
|
|
72 |
|
|
def main():
|
73 |
4920
|
aaronmk
|
# Input
|
74 |
4918
|
aaronmk
|
env_names = []
|
75 |
|
|
debug = opts.env_flag('debug', False, env_names)
|
76 |
4920
|
aaronmk
|
taxons = sys.argv[1:]
|
77 |
|
|
if not taxons: raise SystemExit('Usage: '+opts.env_usage(env_names, True)
|
78 |
|
|
+' '+sys.argv[0]+' taxon... >out 2>>log')
|
79 |
|
|
assert len(taxons) <= max_taxons
|
80 |
4918
|
aaronmk
|
|
81 |
|
|
# Logging
|
82 |
|
|
def debug_log(label, str_=''):
|
83 |
|
|
if debug: sys.stderr.write('\n'+label+':\n'+str_+'\n')
|
84 |
|
|
|
85 |
|
|
## HTTP
|
86 |
|
|
headers = initial_headers
|
87 |
|
|
|
88 |
|
|
def do_request(request=None, url=url, headers=headers):
|
89 |
|
|
debug_log('request', str(request))
|
90 |
|
|
response = urllib2.urlopen(urllib2.Request(url, request, headers))
|
91 |
|
|
response_str = streams.read_all(response)
|
92 |
|
|
response_info = str(response.info())
|
93 |
|
|
debug_log('response info', response_info)
|
94 |
|
|
debug_log('response str', response_str)
|
95 |
|
|
return response_str, response_info
|
96 |
|
|
|
97 |
|
|
def do_repeated_request(request):
|
98 |
|
|
pause = initial_pause
|
99 |
|
|
total_pause = 0
|
100 |
|
|
while True:
|
101 |
|
|
total_pause += pause
|
102 |
|
|
if total_pause > max_pause: raise # error is not temporary
|
103 |
|
|
debug_log('total_pause', str(total_pause)+'s')
|
104 |
|
|
time.sleep(pause) # wait for job to complete
|
105 |
|
|
|
106 |
|
|
try: return do_request(request)
|
107 |
|
|
except urllib2.HTTPError: pass # try again
|
108 |
|
|
pause *= pause_growth_factor
|
109 |
|
|
|
110 |
|
|
debug_log('Submit')
|
111 |
4920
|
aaronmk
|
request = submission_request_template.replace('[taxons]',
|
112 |
|
|
r'\\n'.join(map(gwt_encode, taxons))) # double-escape \n
|
113 |
4918
|
aaronmk
|
response, response_info = do_request(request)
|
114 |
|
|
key, = parse_response('submission', submission_response_pattern, response)
|
115 |
|
|
debug_log('key', key)
|
116 |
|
|
key_enc = gwt_encode(key)
|
117 |
|
|
|
118 |
|
|
debug_log('Retrieve')
|
119 |
|
|
request = retrieval_request_template.replace('[key]', key_enc)
|
120 |
|
|
response, response_info = do_repeated_request(request)
|
121 |
|
|
parse_response('retrieval', retrieval_response_pattern, response)
|
122 |
|
|
session_id, = parse_response('retrieval info',
|
123 |
|
|
retrieval_response_info_pattern, response_info)
|
124 |
|
|
debug_log('session_id', session_id)
|
125 |
|
|
headers['Cookie'] = 'JSESSIONID='+session_id
|
126 |
|
|
|
127 |
|
|
debug_log('Prepare download')
|
128 |
|
|
request = download_request_template.replace('[key]', key_enc)
|
129 |
|
|
response, response_info = do_request(request)
|
130 |
|
|
csv_url, = parse_response('download', download_response_pattern, response)
|
131 |
|
|
csv_url += download_url_suffix
|
132 |
|
|
debug_log('csv_url', csv_url)
|
133 |
|
|
|
134 |
|
|
debug_log('Download')
|
135 |
|
|
response, response_info = do_request(url=csv_url, headers={})
|
136 |
|
|
sys.stdout.write(response)
|
137 |
|
|
|
138 |
|
|
main()
|