2015-05-30 18:11:16 -04:00
|
|
|
"""
|
|
|
|
Process raw qstr file and output qstr data with length, hash and data bytes.
|
|
|
|
|
|
|
|
This script works with Python 2.6, 2.7, 3.3 and 3.4.
|
|
|
|
"""
|
|
|
|
|
2014-03-10 03:07:35 -04:00
|
|
|
from __future__ import print_function
|
|
|
|
|
2014-01-21 16:40:13 -05:00
|
|
|
import re
|
2014-03-08 10:03:25 -05:00
|
|
|
import sys
|
2014-01-24 17:22:00 -05:00
|
|
|
|
2018-07-31 19:53:54 -04:00
|
|
|
import collections
|
|
|
|
import gettext
|
|
|
|
|
2018-08-15 21:32:37 -04:00
|
|
|
sys.path.append("../../tools/huffman")
|
|
|
|
|
|
|
|
import huffman
|
|
|
|
|
2016-04-14 09:37:04 -04:00
|
|
|
# Python 2/3 compatibility:
|
|
|
|
# - iterating through bytes is different
|
|
|
|
# - codepoint2name lives in a different module
|
2014-01-24 17:22:00 -05:00
|
|
|
import platform
|
|
|
|
if platform.python_version_tuple()[0] == '2':
|
2016-09-02 00:32:47 -04:00
|
|
|
bytes_cons = lambda val, enc=None: bytearray(val)
|
2014-01-24 17:22:00 -05:00
|
|
|
from htmlentitydefs import codepoint2name
|
|
|
|
elif platform.python_version_tuple()[0] == '3':
|
2016-09-02 00:32:47 -04:00
|
|
|
bytes_cons = bytes
|
2014-01-24 17:22:00 -05:00
|
|
|
from html.entities import codepoint2name
|
2016-09-02 00:32:47 -04:00
|
|
|
# end compatibility code
|
|
|
|
|
2014-04-12 22:28:46 -04:00
|
|
|
codepoint2name[ord('-')] = 'hyphen';
|
2014-01-21 16:40:13 -05:00
|
|
|
|
2014-02-15 06:34:50 -05:00
|
|
|
# add some custom names to map characters that aren't in HTML
|
2015-01-11 09:16:24 -05:00
|
|
|
codepoint2name[ord(' ')] = 'space'
|
|
|
|
codepoint2name[ord('\'')] = 'squot'
|
|
|
|
codepoint2name[ord(',')] = 'comma'
|
2014-02-15 06:34:50 -05:00
|
|
|
codepoint2name[ord('.')] = 'dot'
|
2014-02-17 17:06:37 -05:00
|
|
|
codepoint2name[ord(':')] = 'colon'
|
2016-04-13 17:12:39 -04:00
|
|
|
codepoint2name[ord(';')] = 'semicolon'
|
2014-02-17 17:06:37 -05:00
|
|
|
codepoint2name[ord('/')] = 'slash'
|
2014-04-15 07:42:52 -04:00
|
|
|
codepoint2name[ord('%')] = 'percent'
|
2014-04-15 07:50:21 -04:00
|
|
|
codepoint2name[ord('#')] = 'hash'
|
2015-01-11 09:16:24 -05:00
|
|
|
codepoint2name[ord('(')] = 'paren_open'
|
|
|
|
codepoint2name[ord(')')] = 'paren_close'
|
|
|
|
codepoint2name[ord('[')] = 'bracket_open'
|
|
|
|
codepoint2name[ord(']')] = 'bracket_close'
|
2014-04-15 17:03:55 -04:00
|
|
|
codepoint2name[ord('{')] = 'brace_open'
|
|
|
|
codepoint2name[ord('}')] = 'brace_close'
|
2014-04-27 14:23:46 -04:00
|
|
|
codepoint2name[ord('*')] = 'star'
|
2015-01-11 09:16:24 -05:00
|
|
|
codepoint2name[ord('!')] = 'bang'
|
2015-04-01 18:09:24 -04:00
|
|
|
codepoint2name[ord('\\')] = 'backslash'
|
2015-08-30 17:20:38 -04:00
|
|
|
codepoint2name[ord('+')] = 'plus'
|
2016-04-13 17:12:39 -04:00
|
|
|
codepoint2name[ord('$')] = 'dollar'
|
|
|
|
codepoint2name[ord('=')] = 'equals'
|
|
|
|
codepoint2name[ord('?')] = 'question'
|
|
|
|
codepoint2name[ord('@')] = 'at_sign'
|
|
|
|
codepoint2name[ord('^')] = 'caret'
|
|
|
|
codepoint2name[ord('|')] = 'pipe'
|
|
|
|
codepoint2name[ord('~')] = 'tilde'
|
2014-02-15 06:34:50 -05:00
|
|
|
|
2018-08-09 18:58:45 -04:00
|
|
|
C_ESCAPES = {
|
|
|
|
"\a": "\\a",
|
|
|
|
"\b": "\\b",
|
|
|
|
"\f": "\\f",
|
2018-08-10 19:17:03 -04:00
|
|
|
"\n": "\\n",
|
2018-08-09 18:58:45 -04:00
|
|
|
"\r": "\\r",
|
|
|
|
"\t": "\\t",
|
|
|
|
"\v": "\\v",
|
|
|
|
"\'": "\\'",
|
|
|
|
"\"": "\\\""
|
|
|
|
}
|
|
|
|
|
2014-01-21 16:40:13 -05:00
|
|
|
# this must match the equivalent function in qstr.c
|
2015-07-20 07:03:13 -04:00
|
|
|
def compute_hash(qstr, bytes_hash):
|
2014-03-25 11:27:15 -04:00
|
|
|
hash = 5381
|
2016-09-02 00:32:47 -04:00
|
|
|
for b in qstr:
|
|
|
|
hash = (hash * 33) ^ b
|
2014-06-06 16:55:27 -04:00
|
|
|
# Make sure that valid hash is never zero, zero means "hash not computed"
|
2015-07-20 07:03:13 -04:00
|
|
|
return (hash & ((1 << (8 * bytes_hash)) - 1)) or 1
|
2014-01-21 16:40:13 -05:00
|
|
|
|
2018-07-31 19:53:54 -04:00
|
|
|
def translate(translation_file, i18ns):
|
|
|
|
with open(translation_file, "rb") as f:
|
|
|
|
table = gettext.GNUTranslations(f)
|
|
|
|
|
2018-08-09 18:58:45 -04:00
|
|
|
translations = []
|
|
|
|
for original in i18ns:
|
|
|
|
unescaped = original
|
|
|
|
for s in C_ESCAPES:
|
|
|
|
unescaped = unescaped.replace(C_ESCAPES[s], s)
|
2018-08-15 21:32:37 -04:00
|
|
|
translation = table.gettext(unescaped)
|
|
|
|
# Add in carriage returns to work in terminals
|
|
|
|
translation = translation.replace("\n", "\r\n")
|
|
|
|
translations.append((original, translation))
|
2018-08-09 18:58:45 -04:00
|
|
|
return translations
|
2018-07-31 19:53:54 -04:00
|
|
|
|
2018-08-15 21:32:37 -04:00
|
|
|
def compute_huffman_coding(translations, qstrs, compression_filename):
|
|
|
|
all_strings = [x[1] for x in translations]
|
|
|
|
|
|
|
|
# go through each qstr and print it out
|
|
|
|
for _, _, qstr in qstrs.values():
|
|
|
|
all_strings.append(qstr)
|
|
|
|
all_strings_concat = "".join(all_strings).encode("utf-8")
|
|
|
|
counts = collections.Counter(all_strings_concat)
|
|
|
|
# add other values
|
|
|
|
for i in range(256):
|
|
|
|
if i not in counts:
|
|
|
|
counts[i] = 0
|
|
|
|
cb = huffman.codebook(counts.items())
|
|
|
|
values = bytearray()
|
|
|
|
length_count = {}
|
|
|
|
renumbered = 0
|
|
|
|
last_l = None
|
|
|
|
canonical = {}
|
|
|
|
for ch, code in sorted(cb.items(), key=lambda x: (len(x[1]), x[0])):
|
|
|
|
values.append(ch)
|
|
|
|
l = len(code)
|
|
|
|
if l not in length_count:
|
|
|
|
length_count[l] = 0
|
|
|
|
length_count[l] += 1
|
|
|
|
if last_l:
|
|
|
|
renumbered <<= (l - last_l)
|
|
|
|
canonical[ch] = '{0:0{width}b}'.format(renumbered, width=l)
|
|
|
|
if chr(ch) in C_ESCAPES:
|
|
|
|
s = C_ESCAPES[chr(ch)]
|
|
|
|
else:
|
|
|
|
s = chr(ch)
|
|
|
|
print("//", ch, s, counts[ch], canonical[ch], renumbered)
|
|
|
|
renumbered += 1
|
|
|
|
last_l = l
|
|
|
|
lengths = bytearray()
|
|
|
|
for i in range(1, max(length_count) + 1):
|
|
|
|
lengths.append(length_count.get(i, 0))
|
|
|
|
print("//", values, lengths)
|
|
|
|
with open(compression_filename, "w") as f:
|
|
|
|
f.write("const uint8_t lengths[] = {{ {} }};\n".format(", ".join(map(str, lengths))))
|
|
|
|
f.write("const uint8_t values[256] = {{ {} }};\n".format(", ".join(map(str, values))))
|
|
|
|
return values, lengths
|
|
|
|
|
|
|
|
def decompress(encoding_table, length, encoded):
|
|
|
|
values, lengths = encoding_table
|
|
|
|
#print(l, encoded)
|
|
|
|
dec = bytearray(length)
|
|
|
|
this_byte = 0
|
|
|
|
this_bit = 7
|
|
|
|
b = encoded[this_byte]
|
|
|
|
for i in range(length):
|
|
|
|
bits = 0
|
|
|
|
bit_length = 0
|
|
|
|
max_code = lengths[0]
|
|
|
|
searched_length = lengths[0]
|
|
|
|
while True:
|
|
|
|
bits <<= 1
|
|
|
|
if 0x80 & b:
|
|
|
|
bits |= 1
|
|
|
|
|
|
|
|
b <<= 1
|
|
|
|
bit_length += 1
|
|
|
|
if this_bit == 0:
|
|
|
|
this_bit = 7
|
|
|
|
this_byte += 1
|
|
|
|
if this_byte < len(encoded):
|
|
|
|
b = encoded[this_byte]
|
|
|
|
else:
|
|
|
|
this_bit -= 1
|
|
|
|
if max_code > 0 and bits < max_code:
|
|
|
|
#print('{0:0{width}b}'.format(bits, width=bit_length))
|
|
|
|
break
|
|
|
|
max_code = (max_code << 1) + lengths[bit_length]
|
|
|
|
searched_length += lengths[bit_length]
|
|
|
|
|
|
|
|
v = values[searched_length + bits - max_code]
|
|
|
|
dec[i] = v
|
|
|
|
return dec
|
|
|
|
|
|
|
|
def compress(encoding_table, decompressed):
|
|
|
|
if not isinstance(decompressed, bytes):
|
|
|
|
raise TypeError()
|
|
|
|
values, lengths = encoding_table
|
|
|
|
enc = bytearray(len(decompressed))
|
|
|
|
#print(decompressed)
|
|
|
|
#print(lengths)
|
|
|
|
current_bit = 7
|
|
|
|
current_byte = 0
|
|
|
|
for c in decompressed:
|
|
|
|
#print()
|
|
|
|
#print("char", c, values.index(c))
|
|
|
|
start = 0
|
|
|
|
end = lengths[0]
|
|
|
|
bits = 1
|
|
|
|
compressed = None
|
|
|
|
code = 0
|
|
|
|
while compressed is None:
|
|
|
|
s = start
|
|
|
|
e = end
|
|
|
|
#print("{0:0{width}b}".format(code, width=bits))
|
|
|
|
# Binary search!
|
|
|
|
while e > s:
|
|
|
|
midpoint = (s + e) // 2
|
|
|
|
#print(s, e, midpoint)
|
|
|
|
if values[midpoint] == c:
|
|
|
|
compressed = code + (midpoint - start)
|
|
|
|
#print("found {0:0{width}b}".format(compressed, width=bits))
|
|
|
|
break
|
|
|
|
elif c < values[midpoint]:
|
|
|
|
e = midpoint
|
|
|
|
else:
|
|
|
|
s = midpoint + 1
|
|
|
|
code += end - start
|
|
|
|
code <<= 1
|
|
|
|
start = end
|
|
|
|
end += lengths[bits]
|
|
|
|
bits += 1
|
|
|
|
#print("next bit", bits)
|
|
|
|
|
|
|
|
for i in range(bits - 1, 0, -1):
|
|
|
|
if compressed & (1 << (i - 1)):
|
|
|
|
enc[current_byte] |= 1 << current_bit
|
|
|
|
if current_bit == 0:
|
|
|
|
current_bit = 7
|
|
|
|
#print("packed {0:0{width}b}".format(enc[current_byte], width=8))
|
|
|
|
current_byte += 1
|
|
|
|
else:
|
|
|
|
current_bit -= 1
|
|
|
|
if current_bit != 7:
|
|
|
|
current_byte += 1
|
|
|
|
return enc[:current_byte]
|
|
|
|
|
2016-01-31 07:59:59 -05:00
|
|
|
def qstr_escape(qst):
|
2016-04-13 17:12:39 -04:00
|
|
|
def esc_char(m):
|
|
|
|
c = ord(m.group(0))
|
|
|
|
try:
|
|
|
|
name = codepoint2name[c]
|
|
|
|
except KeyError:
|
|
|
|
name = '0x%02x' % c
|
|
|
|
return "_" + name + '_'
|
|
|
|
return re.sub(r'[^A-Za-z0-9_]', esc_char, qst)
|
2016-01-31 07:59:59 -05:00
|
|
|
|
|
|
|
def parse_input_headers(infiles):
|
2014-01-21 16:40:13 -05:00
|
|
|
# read the qstrs in from the input files
|
2015-01-11 12:52:45 -05:00
|
|
|
qcfgs = {}
|
2014-01-23 17:22:00 -05:00
|
|
|
qstrs = {}
|
2018-07-31 19:53:54 -04:00
|
|
|
i18ns = set()
|
2014-01-21 16:40:13 -05:00
|
|
|
for infile in infiles:
|
|
|
|
with open(infile, 'rt') as f:
|
|
|
|
for line in f:
|
2015-01-11 12:52:45 -05:00
|
|
|
line = line.strip()
|
|
|
|
|
|
|
|
# is this a config line?
|
|
|
|
match = re.match(r'^QCFG\((.+), (.+)\)', line)
|
|
|
|
if match:
|
|
|
|
value = match.group(2)
|
|
|
|
if value[0] == '(' and value[-1] == ')':
|
|
|
|
# strip parenthesis from config value
|
|
|
|
value = value[1:-1]
|
|
|
|
qcfgs[match.group(1)] = value
|
|
|
|
continue
|
|
|
|
|
2018-07-31 19:53:54 -04:00
|
|
|
|
|
|
|
match = re.match(r'^TRANSLATE\("(.*)"\)$', line)
|
|
|
|
if match:
|
|
|
|
i18ns.add(match.group(1))
|
|
|
|
continue
|
|
|
|
|
2014-05-02 15:10:47 -04:00
|
|
|
# is this a QSTR line?
|
2015-01-11 12:52:45 -05:00
|
|
|
match = re.match(r'^Q\((.*)\)$', line)
|
2014-05-02 15:10:47 -04:00
|
|
|
if not match:
|
2014-04-13 08:16:51 -04:00
|
|
|
continue
|
2014-01-21 16:40:13 -05:00
|
|
|
|
|
|
|
# get the qstr value
|
|
|
|
qstr = match.group(1)
|
2016-04-14 10:22:36 -04:00
|
|
|
|
|
|
|
# special case to specify control characters
|
|
|
|
if qstr == '\\n':
|
|
|
|
qstr = '\n'
|
|
|
|
|
|
|
|
# work out the corresponding qstr name
|
2016-01-31 07:59:59 -05:00
|
|
|
ident = qstr_escape(qstr)
|
2014-01-21 16:40:13 -05:00
|
|
|
|
|
|
|
# don't add duplicates
|
2014-01-23 17:22:00 -05:00
|
|
|
if ident in qstrs:
|
2014-01-21 16:40:13 -05:00
|
|
|
continue
|
|
|
|
|
2014-01-24 17:22:00 -05:00
|
|
|
# add the qstr to the list, with order number to retain original order in file
|
2017-10-21 04:06:32 -04:00
|
|
|
order = len(qstrs)
|
|
|
|
# but put special method names like __add__ at the top of list, so
|
|
|
|
# that their id's fit into a byte
|
|
|
|
if ident == "":
|
|
|
|
# Sort empty qstr above all still
|
|
|
|
order = -200000
|
2018-05-10 09:10:46 -04:00
|
|
|
elif ident == "__dir__":
|
|
|
|
# Put __dir__ after empty qstr for builtin dir() to work
|
|
|
|
order = -190000
|
2017-10-21 04:06:32 -04:00
|
|
|
elif ident.startswith("__"):
|
|
|
|
order -= 100000
|
|
|
|
qstrs[ident] = (order, ident, qstr)
|
2014-01-21 16:40:13 -05:00
|
|
|
|
2018-07-31 19:53:54 -04:00
|
|
|
if not qcfgs and qstrs:
|
2015-10-11 04:09:57 -04:00
|
|
|
sys.stderr.write("ERROR: Empty preprocessor output - check for errors above\n")
|
|
|
|
sys.exit(1)
|
|
|
|
|
2018-07-31 19:53:54 -04:00
|
|
|
return qcfgs, qstrs, i18ns
|
2016-01-31 07:59:59 -05:00
|
|
|
|
|
|
|
def make_bytes(cfg_bytes_len, cfg_bytes_hash, qstr):
|
2016-09-02 00:32:47 -04:00
|
|
|
qbytes = bytes_cons(qstr, 'utf8')
|
|
|
|
qlen = len(qbytes)
|
|
|
|
qhash = compute_hash(qbytes, cfg_bytes_hash)
|
2016-05-23 10:18:55 -04:00
|
|
|
if all(32 <= ord(c) <= 126 and c != '\\' and c != '"' for c in qstr):
|
2016-04-14 09:20:25 -04:00
|
|
|
# qstr is all printable ASCII so render it as-is (for easier debugging)
|
|
|
|
qdata = qstr
|
|
|
|
else:
|
|
|
|
# qstr contains non-printable codes so render entire thing as hex pairs
|
2016-09-02 00:32:47 -04:00
|
|
|
qdata = ''.join(('\\x%02x' % b) for b in qbytes)
|
2016-01-31 07:59:59 -05:00
|
|
|
if qlen >= (1 << (8 * cfg_bytes_len)):
|
|
|
|
print('qstr is too long:', qstr)
|
|
|
|
assert False
|
|
|
|
qlen_str = ('\\x%02x' * cfg_bytes_len) % tuple(((qlen >> (8 * i)) & 0xff) for i in range(cfg_bytes_len))
|
|
|
|
qhash_str = ('\\x%02x' * cfg_bytes_hash) % tuple(((qhash >> (8 * i)) & 0xff) for i in range(cfg_bytes_hash))
|
|
|
|
return '(const byte*)"%s%s" "%s"' % (qhash_str, qlen_str, qdata)
|
|
|
|
|
2018-08-15 21:32:37 -04:00
|
|
|
def print_qstr_data(encoding_table, qcfgs, qstrs, i18ns):
|
2015-01-11 17:27:30 -05:00
|
|
|
# get config variables
|
|
|
|
cfg_bytes_len = int(qcfgs['BYTES_IN_LEN'])
|
2015-07-20 07:03:13 -04:00
|
|
|
cfg_bytes_hash = int(qcfgs['BYTES_IN_HASH'])
|
2015-01-11 17:27:30 -05:00
|
|
|
|
2015-07-31 07:57:36 -04:00
|
|
|
# print out the starter of the generated C header file
|
2014-01-21 16:40:13 -05:00
|
|
|
print('// This file was automatically generated by makeqstrdata.py')
|
2014-01-21 18:28:27 -05:00
|
|
|
print('')
|
2015-01-11 17:27:30 -05:00
|
|
|
|
2015-01-11 12:52:45 -05:00
|
|
|
# add NULL qstr with no hash or data
|
2015-07-20 07:03:13 -04:00
|
|
|
print('QDEF(MP_QSTR_NULL, (const byte*)"%s%s" "")' % ('\\x00' * cfg_bytes_hash, '\\x00' * cfg_bytes_len))
|
2015-01-11 17:27:30 -05:00
|
|
|
|
2018-07-31 19:53:54 -04:00
|
|
|
total_qstr_size = 0
|
2018-08-15 21:32:37 -04:00
|
|
|
total_qstr_compressed_size = 0
|
2015-01-11 17:27:30 -05:00
|
|
|
# go through each qstr and print it out
|
2014-04-11 13:36:08 -04:00
|
|
|
for order, ident, qstr in sorted(qstrs.values(), key=lambda x: x[0]):
|
2016-01-31 07:59:59 -05:00
|
|
|
qbytes = make_bytes(cfg_bytes_len, cfg_bytes_hash, qstr)
|
|
|
|
print('QDEF(MP_QSTR_%s, %s)' % (ident, qbytes))
|
2018-07-31 19:53:54 -04:00
|
|
|
total_qstr_size += len(qstr)
|
|
|
|
|
|
|
|
total_text_size = 0
|
2018-08-15 21:32:37 -04:00
|
|
|
total_text_compressed_size = 0
|
2018-07-31 19:53:54 -04:00
|
|
|
for original, translation in i18ns:
|
2018-08-15 21:32:37 -04:00
|
|
|
translation_encoded = translation.encode("utf-8")
|
|
|
|
compressed = compress(encoding_table, translation_encoded)
|
|
|
|
total_text_compressed_size += len(compressed)
|
|
|
|
decompressed = decompress(encoding_table, len(translation_encoded), compressed).decode("utf-8")
|
|
|
|
for c in C_ESCAPES:
|
|
|
|
decompressed.replace(c, C_ESCAPES[c])
|
|
|
|
#print("// \"{}\"".format(translation))
|
|
|
|
print("TRANSLATION(\"{}\", {}, {{ {} }}) // {}".format(original, len(translation_encoded)+1, ", ".join(["0x{:02x}".format(x) for x in compressed]), decompressed))
|
|
|
|
total_text_size += len(translation.encode("utf-8"))
|
2016-01-31 07:59:59 -05:00
|
|
|
|
2018-07-31 19:53:54 -04:00
|
|
|
print()
|
|
|
|
print("// {} bytes worth of qstr".format(total_qstr_size))
|
|
|
|
print("// {} bytes worth of translations".format(total_text_size))
|
2018-08-15 21:32:37 -04:00
|
|
|
print("// {} bytes worth of translations compressed".format(total_text_compressed_size))
|
|
|
|
print("// {} bytes saved".format(total_text_size - total_text_compressed_size))
|
2018-07-31 19:53:54 -04:00
|
|
|
|
|
|
|
def print_qstr_enums(qstrs):
|
|
|
|
# print out the starter of the generated C header file
|
|
|
|
print('// This file was automatically generated by makeqstrdata.py')
|
|
|
|
print('')
|
|
|
|
|
|
|
|
# add NULL qstr with no hash or data
|
|
|
|
print('QENUM(MP_QSTR_NULL)')
|
|
|
|
|
|
|
|
# go through each qstr and print it out
|
|
|
|
for order, ident, qstr in sorted(qstrs.values(), key=lambda x: x[0]):
|
|
|
|
print('QENUM(MP_QSTR_%s)' % (ident,))
|
2014-01-21 16:40:13 -05:00
|
|
|
|
|
|
|
if __name__ == "__main__":
|
2018-07-31 19:53:54 -04:00
|
|
|
import argparse
|
|
|
|
|
|
|
|
parser = argparse.ArgumentParser(description='Process QSTR definitions into headers for compilation')
|
|
|
|
parser.add_argument('infiles', metavar='N', type=str, nargs='+',
|
|
|
|
help='an integer for the accumulator')
|
|
|
|
parser.add_argument('--translation', default=None, type=str,
|
|
|
|
help='translations for i18n() items')
|
2018-08-15 21:32:37 -04:00
|
|
|
parser.add_argument('--compression_filename', default=None, type=str,
|
|
|
|
help='header for compression info')
|
2018-07-31 19:53:54 -04:00
|
|
|
|
|
|
|
args = parser.parse_args()
|
|
|
|
|
|
|
|
qcfgs, qstrs, i18ns = parse_input_headers(args.infiles)
|
|
|
|
if args.translation:
|
|
|
|
translations = translate(args.translation, i18ns)
|
2018-08-15 21:32:37 -04:00
|
|
|
encoding_table = compute_huffman_coding(translations, qstrs, args.compression_filename)
|
|
|
|
print_qstr_data(encoding_table, qcfgs, qstrs, translations)
|
2018-07-31 19:53:54 -04:00
|
|
|
else:
|
|
|
|
print_qstr_enums(qstrs)
|