| #!/usr/bin/python |
| |
| import sys |
| |
| if len (sys.argv) != 5: |
| print >>sys.stderr, "usage: ./gen-use-table.py IndicSyllabicCategory.txt IndicPositionalCategory.txt UnicodeData.txt Blocks.txt" |
| sys.exit (1) |
| |
| BLACKLISTED_BLOCKS = ["Thai", "Lao", "Tibetan"] |
| |
| files = [file (x) for x in sys.argv[1:]] |
| |
| headers = [[f.readline () for i in range (2)] for j,f in enumerate(files) if j != 2] |
| headers.append (["UnicodeData.txt does not have a header."]) |
| |
| data = [{} for f in files] |
| values = [{} for f in files] |
| for i, f in enumerate (files): |
| for line in f: |
| |
| j = line.find ('#') |
| if j >= 0: |
| line = line[:j] |
| |
| fields = [x.strip () for x in line.split (';')] |
| if len (fields) == 1: |
| continue |
| |
| uu = fields[0].split ('..') |
| start = int (uu[0], 16) |
| if len (uu) == 1: |
| end = start |
| else: |
| end = int (uu[1], 16) |
| |
| t = fields[1 if i != 2 else 2] |
| |
| for u in range (start, end + 1): |
| data[i][u] = t |
| values[i][t] = values[i].get (t, 0) + end - start + 1 |
| |
| # Merge data into one dict: |
| defaults = ('Other', 'Not_Applicable', 'Cn', 'No_Block') |
| for i,v in enumerate (defaults): |
| values[i][v] = values[i].get (v, 0) + 1 |
| combined = {} |
| for i,d in enumerate (data): |
| for u,v in d.items (): |
| if i >= 2 and not u in combined: |
| continue |
| if not u in combined: |
| combined[u] = list (defaults) |
| combined[u][i] = v |
| combined = {k:v for k,v in combined.items() if v[3] not in BLACKLISTED_BLOCKS} |
| data = combined |
| del combined |
| num = len (data) |
| |
| |
| property_names = [ |
| # General_Category |
| 'Cc', 'Cf', 'Cn', 'Co', 'Cs', 'Ll', 'Lm', 'Lo', 'Lt', 'Lu', 'Mc', |
| 'Me', 'Mn', 'Nd', 'Nl', 'No', 'Pc', 'Pd', 'Pe', 'Pf', 'Pi', 'Po', |
| 'Ps', 'Sc', 'Sk', 'Sm', 'So', 'Zl', 'Zp', 'Zs', |
| # Indic_Syllabic_Category |
| 'Bindu', |
| 'Visarga', |
| 'Avagraha', |
| 'Nukta', |
| 'Virama', |
| 'Pure_Killer', |
| 'Invisible_Stacker', |
| 'Vowel_Independent', |
| 'Vowel_Dependent', |
| 'Vowel', |
| 'Consonant_Placeholder', |
| 'Consonant', |
| 'Consonant_Dead', |
| 'Consonant_With_Stacker', |
| 'Consonant_Prefixed', |
| 'Consonant_Preceding_Repha', |
| 'Consonant_Succeeding_Repha', |
| 'Consonant_Subjoined', |
| 'Consonant_Medial', |
| 'Consonant_Final', |
| 'Consonant_Head_Letter', |
| 'Modifying_Letter', |
| 'Tone_Letter', |
| 'Tone_Mark', |
| 'Gemination_Mark', |
| 'Cantillation_Mark', |
| 'Register_Shifter', |
| 'Syllable_Modifier', |
| 'Consonant_Killer', |
| 'Non_Joiner', |
| 'Joiner', |
| 'Number_Joiner', |
| 'Number', |
| 'Brahmi_Joining_Number', |
| # Indic_Positional_Category |
| 'Right', |
| 'Left', |
| 'Visual_Order_Left', |
| 'Left_And_Right', |
| 'Top', |
| 'Bottom', |
| 'Top_And_Bottom', |
| 'Top_And_Right', |
| 'Top_And_Left', |
| 'Top_And_Left_And_Right', |
| 'Bottom_And_Right', |
| 'Top_And_Bottom_And_Right', |
| 'Overstruck', |
| ] |
| |
| class PropertyValue(object): |
| def __init__(self, name_): |
| self.name = name_ |
| |
| property_values = {} |
| |
| for name in property_names: |
| value = PropertyValue(name) |
| assert value not in property_values |
| assert value not in globals() |
| property_values[name] = value |
| globals().update(property_values) |
| |
| |
| def is_BASE(U, UISC, UGC): |
| return (UISC in [Number, Consonant, Consonant_Head_Letter, Consonant_Placeholder, Tone_Letter] or |
| (UGC == Lo and UISC in [Avagraha, Bindu, Consonant_Final, Consonant_Medial, |
| Consonant_Subjoined, Vowel, Vowel_Dependent])) |
| def is_BASE_VOWEL(U, UISC, UGC): |
| return UISC == Vowel_Independent |
| def is_BASE_IND(U, UISC, UGC): |
| return (UISC in [Consonant_Dead, Modifying_Letter] or UGC == Po) |
| def is_BASE_NUM(U, UISC, UGC): |
| return UISC == Brahmi_Joining_Number |
| def is_BASE_OTHER(U, UISC, UGC): |
| return U in [0x00A0, 0x00D7, 0x2015, 0x2022, 0x25CC, |
| 0x25FB, 0x25FC, 0x25FD, 0x25FE] |
| def is_CGJ(U, UISC, UGC): |
| return U == 0x034F |
| def is_CONS_FINAL(U, UISC, UGC): |
| return ((UISC == Consonant_Final and UGC != Lo) or |
| UISC == Consonant_Succeeding_Repha) |
| def is_CONS_FINAL_MOD(U, UISC, UGC): |
| return UISC in [Consonant_Final_Modifier, Syllable_Modifier] |
| def is_CONS_MED(U, UISC, UGC): |
| return UISC == Consonant_Medial and UGC != Lo |
| def is_CONS_MOD(U, UISC, UGC): |
| return UISC in [Nukta, Gemination_Mark, Consonant_Killer] |
| def is_CONS_SUB(U, UISC, UGC): |
| return UISC == Consonant_Subjoined |
| def is_HALANT(U, UISC, UGC): |
| return UISC in [Virama, Invisible_Stacker] |
| def is_HALANT_NUM(U, UISC, UGC): |
| return UISC == Number_Joiner |
| def is_ZWNJ(U, UISC, UGC): |
| return UISC == Non_Joiner |
| def is_ZWJ(U, UISC, UGC): |
| return UISC == Joiner |
| def is_Word_Joiner(U, UISC, UGC): |
| return U == 0x2060 |
| def is_OTHER(U, UISC, UGC): |
| return UGC == Zs # or any other SCRIPT_COMMON characters |
| def is_Reserved(U, UISC, UGC): |
| return UGC == 'Cn' |
| def is_REPHA(U, UISC, UGC): |
| return UISC == Consonant_Preceding_Repha |
| def is_SYM(U, UISC, UGC): |
| return UGC in [So, Sc] or UISC == Symbol_Letter |
| def is_SYM_MOD(U, UISC, UGC): |
| return U in [0x1B6B, 0x1B6C, 0x1B6D, 0x1B6E, 0x1B6F, 0x1B70, 0x1B71, 0x1B72, 0x1B73] |
| def is_VARIATION_SELECTOR(U, UISC, UGC): |
| return 0xFE00 <= U <= 0xFE0F |
| def is_VOWEL(U, UISC, UGC): |
| return (UISC == Pure_Killer or |
| (UGC != Lo and UISC in [Vowel, Vowel_Dependent])) |
| def is_VOWEL_MOD(U, UISC, UGC): |
| return (UISC in [Tone_Mark, Cantillation_Mark, Register_Shifter, Visarga] or |
| (UGC != Lo and UISC == Bindu)) |
| |
| use_mapping = { |
| 'B': is_BASE, |
| 'IV': is_BASE_VOWEL, |
| 'IND': is_BASE_IND, |
| 'N': is_BASE_NUM, |
| 'GB': is_BASE_OTHER, |
| 'CGJ': is_CGJ, |
| 'F': is_CONS_FINAL, |
| 'FM': is_CONS_FINAL_MOD, |
| 'M': is_CONS_MED, |
| 'CM': is_CONS_MOD, |
| 'SUB': is_CONS_SUB, |
| 'H': is_HALANT, |
| 'HN': is_HALANT_NUM, |
| 'ZWNJ': is_ZWNJ, |
| 'ZWJ': is_ZWJ, |
| 'WJ': is_Word_Joiner, |
| 'O': is_OTHER, |
| 'Rsv': is_Reserved, |
| 'R': is_REPHA, |
| 'S': is_SYM, |
| 'SM': is_SYM_MOD, |
| 'VS': is_VARIATION_SELECTOR, |
| 'V': is_VOWEL, |
| 'VM': is_VOWEL_MOD, |
| } |
| |
| #data = map_to_use(data) |
| |
| # Remove the outliers |
| singles = {} |
| for u in [0x25CC, 0x1107F]: |
| singles[u] = data[u] |
| del data[u] |
| |
| print "/* == Start of generated table == */" |
| print "/*" |
| print " * The following table is generated by running:" |
| print " *" |
| print " * ./gen-use-table.py IndicSyllabicCategory.txt IndicPositionalCategory.txt UnicodeData.txt Blocks.txt" |
| print " *" |
| print " * on files with these headers:" |
| print " *" |
| for h in headers: |
| for l in h: |
| print " * %s" % (l.strip()) |
| print " */" |
| print |
| print '#include "hb-ot-shape-complex-use-private.hh"' |
| print |
| |
| # Shorten values |
| short = [{ |
| "Bindu": 'Bi', |
| "Cantillation_Mark": 'Ca', |
| "Joiner": 'ZWJ', |
| "Non_Joiner": 'ZWNJ', |
| "Number": 'Nd', |
| "Visarga": 'Vs', |
| "Vowel": 'Vo', |
| "Vowel_Dependent": 'M', |
| "Other": 'x', |
| "Consonant_Placeholder":'GB', |
| },{ |
| "Not_Applicable": 'x', |
| }] |
| all_shorts = [{},{}] |
| |
| # Add some of the values, to make them more readable, and to avoid duplicates |
| |
| |
| for i in range (2): |
| for v,s in short[i].items (): |
| all_shorts[i][s] = v |
| |
| what = ["INDIC_SYLLABIC_CATEGORY", "INDIC_POSITIONAL_CATEGORY"] |
| what_short = ["SC", "PC"] |
| for i in range (2): |
| print |
| vv = values[i].keys () |
| vv.sort () |
| for v in vv: |
| v_no_and = v.replace ('_And_', '_') |
| if v in short[i]: |
| s = short[i][v] |
| else: |
| s = ''.join ([c for c in v_no_and if ord ('A') <= ord (c) <= ord ('Z')]) |
| if s in all_shorts[i]: |
| raise Exception ("Duplicate short value alias", v, all_shorts[i][s]) |
| all_shorts[i][s] = v |
| short[i][v] = s |
| print "#define %s_%s %s_%s %s/* %3d chars; %s */" % \ |
| (what_short[i], s, what[i], v.upper (), \ |
| ' '* ((56-1 - len (what[i]) - 1 - len (v)) / 8), \ |
| values[i][v], v) |
| print |
| print "#define _(S,M) USE_COMBINE_CATEGORIES (SC_##S, PC_##M)" |
| print |
| print |
| |
| total = 0 |
| used = 0 |
| last_block = None |
| def print_block (block, start, end, data): |
| global total, used, last_block |
| if block and block != last_block: |
| print |
| print |
| print " /* %s */" % block |
| num = 0 |
| assert start % 8 == 0 |
| assert (end+1) % 8 == 0 |
| for u in range (start, end+1): |
| if u % 8 == 0: |
| print |
| print " /* %04X */" % u, |
| if u in data: |
| num += 1 |
| d = data.get (u, defaults) |
| sys.stdout.write ("%9s" % ("_(%s,%s)," % (short[0][d[0]], short[1][d[1]]))) |
| |
| total += end - start + 1 |
| used += num |
| if block: |
| last_block = block |
| |
| uu = data.keys () |
| uu.sort () |
| |
| last = -100000 |
| num = 0 |
| offset = 0 |
| starts = [] |
| ends = [] |
| print "static const USE_TABLE_ELEMENT_TYPE use_table[] = {" |
| for u in uu: |
| if u <= last: |
| continue |
| block = data[u][3] |
| |
| start = u//8*8 |
| end = start+1 |
| while end in uu and block == data[end][3]: |
| end += 1 |
| end = (end-1)//8*8 + 7 |
| |
| if start != last + 1: |
| if start - last <= 1+16*3: |
| print_block (None, last+1, start-1, data) |
| last = start-1 |
| else: |
| if last >= 0: |
| ends.append (last + 1) |
| offset += ends[-1] - starts[-1] |
| print |
| print |
| print "#define use_offset_0x%04xu %d" % (start, offset) |
| starts.append (start) |
| |
| print_block (block, start, end, data) |
| last = end |
| ends.append (last + 1) |
| offset += ends[-1] - starts[-1] |
| print |
| print |
| occupancy = used * 100. / total |
| page_bits = 12 |
| print "}; /* Table items: %d; occupancy: %d%% */" % (offset, occupancy) |
| print |
| print "USE_TABLE_ELEMENT_TYPE" |
| print "hb_use_get_categories (hb_codepoint_t u)" |
| print "{" |
| print " switch (u >> %d)" % page_bits |
| print " {" |
| pages = set([u>>page_bits for u in starts+ends+singles.keys()]) |
| for p in sorted(pages): |
| print " case 0x%0Xu:" % p |
| for (start,end) in zip (starts, ends): |
| if p not in [start>>page_bits, end>>page_bits]: continue |
| offset = "use_offset_0x%04xu" % start |
| print " if (hb_in_range (u, 0x%04Xu, 0x%04Xu)) return use_table[u - 0x%04Xu + %s];" % (start, end-1, start, offset) |
| for u,d in singles.items (): |
| if p != u>>page_bits: continue |
| print " if (unlikely (u == 0x%04Xu)) return _(%s,%s);" % (u, short[0][d[0]], short[1][d[1]]) |
| print " break;" |
| print "" |
| print " default:" |
| print " break;" |
| print " }" |
| print " return _(x,x);" |
| print "}" |
| print |
| print "#undef _" |
| for i in range (2): |
| print |
| vv = values[i].keys () |
| vv.sort () |
| for v in vv: |
| print "#undef %s_%s" % \ |
| (what_short[i], short[i][v]) |
| print |
| print "/* == End of generated table == */" |
| |
| # Maintain at least 30% occupancy in the table */ |
| if occupancy < 30: |
| raise Exception ("Table too sparse, please investigate: ", occupancy) |