| #!/usr/bin/env python3 |
| # flake8: noqa: F821 |
| |
| import logging |
| logging.basicConfig(format='%(levelname)s: %(message)s', level=logging.INFO) |
| |
| """usage: ./gen-use-table.py IndicSyllabicCategory.txt IndicPositionalCategory.txt ArabicShaping.txt DerivedCoreProperties.txt UnicodeData.txt Blocks.txt Scripts.txt IndicSyllabicCategory-Additional.txt IndicPositionalCategory-Additional.txt |
| |
| Input files: |
| * https://unicode.org/Public/UCD/latest/ucd/IndicSyllabicCategory.txt |
| * https://unicode.org/Public/UCD/latest/ucd/IndicPositionalCategory.txt |
| * https://unicode.org/Public/UCD/latest/ucd/ArabicShaping.txt |
| * https://unicode.org/Public/UCD/latest/ucd/DerivedCoreProperties.txt |
| * https://unicode.org/Public/UCD/latest/ucd/UnicodeData.txt |
| * https://unicode.org/Public/UCD/latest/ucd/Blocks.txt |
| * https://unicode.org/Public/UCD/latest/ucd/Scripts.txt |
| * ms-use/IndicSyllabicCategory-Additional.txt |
| * ms-use/IndicPositionalCategory-Additional.txt |
| """ |
| |
| import sys |
| |
| if len (sys.argv) != 10: |
| sys.exit (__doc__) |
| |
| DISABLED_SCRIPTS = { |
| 'Arabic', |
| 'Lao', |
| 'Samaritan', |
| 'Syriac', |
| 'Thai', |
| } |
| |
| files = [open (x, encoding='utf-8') for x in sys.argv[1:]] |
| |
| headers = [[f.readline () for i in range (2)] for j,f in enumerate(files) if j != 4] |
| for j in range(7, 9): |
| for line in files[j]: |
| line = line.rstrip() |
| if not line: |
| break |
| headers[j - 1].append(line) |
| headers.append (["UnicodeData.txt does not have a header."]) |
| |
| unicode_data = [{} for _ in files] |
| values = [{} for _ in files] |
| for i, f in enumerate (files): |
| for line in f: |
| |
| j = line.find ('#') |
| if j >= 0: |
| line = line[:j] |
| |
| fields = [x.strip () for x in line.split (';')] |
| if len (fields) == 1: |
| continue |
| |
| uu = fields[0].split ('..') |
| start = int (uu[0], 16) |
| if len (uu) == 1: |
| end = start |
| else: |
| end = int (uu[1], 16) |
| |
| t = fields[1 if i not in [2, 4] else 2] |
| |
| if i == 2: |
| t = 'jt_' + t |
| elif i == 3 and t != 'Default_Ignorable_Code_Point': |
| continue |
| elif i == 7 and t == 'Consonant_Final_Modifier': |
| # TODO: https://github.com/MicrosoftDocs/typography-issues/issues/336 |
| t = 'Syllable_Modifier' |
| elif i == 8 and t == 'NA': |
| t = 'Not_Applicable' |
| |
| i0 = i if i < 7 else i - 7 |
| for u in range (start, end + 1): |
| unicode_data[i0][u] = t |
| values[i0][t] = values[i0].get (t, 0) + end - start + 1 |
| |
| defaults = ('Other', 'Not_Applicable', 'jt_X', '', 'Cn', 'No_Block', 'Unknown') |
| |
| # Merge data into one dict: |
| for i,v in enumerate (defaults): |
| values[i][v] = values[i].get (v, 0) + 1 |
| combined = {} |
| for i,d in enumerate (unicode_data): |
| for u,v in d.items (): |
| if not u in combined: |
| if i >= 4: |
| continue |
| combined[u] = list (defaults) |
| combined[u][i] = v |
| combined = {k: v for k, v in combined.items() if v[6] not in DISABLED_SCRIPTS} |
| |
| |
| property_names = [ |
| # General_Category |
| 'Cc', 'Cf', 'Cn', 'Co', 'Cs', 'Ll', 'Lm', 'Lo', 'Lt', 'Lu', 'Mc', |
| 'Me', 'Mn', 'Nd', 'Nl', 'No', 'Pc', 'Pd', 'Pe', 'Pf', 'Pi', 'Po', |
| 'Ps', 'Sc', 'Sk', 'Sm', 'So', 'Zl', 'Zp', 'Zs', |
| # Indic_Syllabic_Category |
| 'Other', |
| 'Bindu', |
| 'Visarga', |
| 'Avagraha', |
| 'Nukta', |
| 'Virama', |
| 'Pure_Killer', |
| 'Invisible_Stacker', |
| 'Vowel_Independent', |
| 'Vowel_Dependent', |
| 'Vowel', |
| 'Consonant_Placeholder', |
| 'Consonant', |
| 'Consonant_Dead', |
| 'Consonant_With_Stacker', |
| 'Consonant_Prefixed', |
| 'Consonant_Preceding_Repha', |
| 'Consonant_Succeeding_Repha', |
| 'Consonant_Subjoined', |
| 'Consonant_Medial', |
| 'Consonant_Final', |
| 'Consonant_Head_Letter', |
| 'Consonant_Initial_Postfixed', |
| 'Modifying_Letter', |
| 'Tone_Letter', |
| 'Tone_Mark', |
| 'Gemination_Mark', |
| 'Cantillation_Mark', |
| 'Register_Shifter', |
| 'Syllable_Modifier', |
| 'Consonant_Killer', |
| 'Non_Joiner', |
| 'Joiner', |
| 'Number_Joiner', |
| 'Number', |
| 'Brahmi_Joining_Number', |
| 'Symbol_Modifier', |
| 'Hieroglyph', |
| 'Hieroglyph_Joiner', |
| 'Hieroglyph_Segment_Begin', |
| 'Hieroglyph_Segment_End', |
| # Indic_Positional_Category |
| 'Not_Applicable', |
| 'Right', |
| 'Left', |
| 'Visual_Order_Left', |
| 'Left_And_Right', |
| 'Top', |
| 'Bottom', |
| 'Top_And_Bottom', |
| 'Top_And_Bottom_And_Left', |
| 'Top_And_Right', |
| 'Top_And_Left', |
| 'Top_And_Left_And_Right', |
| 'Bottom_And_Left', |
| 'Bottom_And_Right', |
| 'Top_And_Bottom_And_Right', |
| 'Overstruck', |
| # Joining_Type |
| 'jt_C', |
| 'jt_D', |
| 'jt_L', |
| 'jt_R', |
| 'jt_T', |
| 'jt_U', |
| 'jt_X', |
| ] |
| |
| class PropertyValue(object): |
| def __init__(self, name_): |
| self.name = name_ |
| def __str__(self): |
| return self.name |
| def __eq__(self, other): |
| return self.name == (other if isinstance(other, str) else other.name) |
| def __ne__(self, other): |
| return not (self == other) |
| def __hash__(self): |
| return hash(str(self)) |
| |
| property_values = {} |
| |
| for name in property_names: |
| value = PropertyValue(name) |
| assert value not in property_values |
| assert value not in globals() |
| property_values[name] = value |
| globals().update(property_values) |
| |
| |
| def is_BASE(U, UISC, UDI, UGC, AJT): |
| return (UISC in [Number, Consonant, Consonant_Head_Letter, |
| Tone_Letter, |
| Vowel_Independent, |
| ] or |
| # TODO: https://github.com/MicrosoftDocs/typography-issues/issues/484 |
| AJT in [jt_C, jt_D, jt_L, jt_R] and UISC != Joiner or |
| (UGC == Lo and UISC in [Avagraha, Bindu, Consonant_Final, Consonant_Medial, |
| Consonant_Subjoined, Vowel, Vowel_Dependent])) |
| def is_BASE_NUM(U, UISC, UDI, UGC, AJT): |
| return UISC == Brahmi_Joining_Number |
| def is_BASE_OTHER(U, UISC, UDI, UGC, AJT): |
| if UISC == Consonant_Placeholder: return True |
| return U in [0x2015, 0x2022, 0x25FB, 0x25FC, 0x25FD, 0x25FE] |
| def is_CGJ(U, UISC, UDI, UGC, AJT): |
| # Also includes VARIATION_SELECTOR and ZWJ |
| return UISC == Joiner or UDI and UGC in [Mc, Me, Mn] |
| def is_CONS_FINAL(U, UISC, UDI, UGC, AJT): |
| return ((UISC == Consonant_Final and UGC != Lo) or |
| UISC == Consonant_Succeeding_Repha) |
| def is_CONS_FINAL_MOD(U, UISC, UDI, UGC, AJT): |
| return UISC == Syllable_Modifier |
| def is_CONS_MED(U, UISC, UDI, UGC, AJT): |
| # Consonant_Initial_Postfixed is new in Unicode 11; not in the spec. |
| return (UISC == Consonant_Medial and UGC != Lo or |
| UISC == Consonant_Initial_Postfixed) |
| def is_CONS_MOD(U, UISC, UDI, UGC, AJT): |
| return UISC in [Nukta, Gemination_Mark, Consonant_Killer] |
| def is_CONS_SUB(U, UISC, UDI, UGC, AJT): |
| return UISC == Consonant_Subjoined and UGC != Lo |
| def is_CONS_WITH_STACKER(U, UISC, UDI, UGC, AJT): |
| return UISC == Consonant_With_Stacker |
| def is_HALANT(U, UISC, UDI, UGC, AJT): |
| return UISC == Virama and not is_HALANT_OR_VOWEL_MODIFIER(U, UISC, UDI, UGC, AJT) |
| def is_HALANT_OR_VOWEL_MODIFIER(U, UISC, UDI, UGC, AJT): |
| # Split off of HALANT |
| return U == 0x0DCA |
| def is_HALANT_NUM(U, UISC, UDI, UGC, AJT): |
| return UISC == Number_Joiner |
| def is_HIEROGLYPH(U, UISC, UDI, UGC, AJT): |
| return UISC == Hieroglyph |
| def is_HIEROGLYPH_JOINER(U, UISC, UDI, UGC, AJT): |
| return UISC == Hieroglyph_Joiner |
| def is_HIEROGLYPH_SEGMENT_BEGIN(U, UISC, UDI, UGC, AJT): |
| return UISC == Hieroglyph_Segment_Begin |
| def is_HIEROGLYPH_SEGMENT_END(U, UISC, UDI, UGC, AJT): |
| return UISC == Hieroglyph_Segment_End |
| def is_INVISIBLE_STACKER(U, UISC, UDI, UGC, AJT): |
| # Split off of HALANT |
| return (UISC == Invisible_Stacker |
| and not is_SAKOT(U, UISC, UDI, UGC, AJT) |
| ) |
| def is_ZWNJ(U, UISC, UDI, UGC, AJT): |
| return UISC == Non_Joiner |
| def is_OTHER(U, UISC, UDI, UGC, AJT): |
| # Also includes BASE_IND and SYM |
| return ((UGC == Po or UISC in [Consonant_Dead, Joiner, Modifying_Letter, Other]) |
| and not is_BASE(U, UISC, UDI, UGC, AJT) |
| and not is_BASE_OTHER(U, UISC, UDI, UGC, AJT) |
| and not is_CGJ(U, UISC, UDI, UGC, AJT) |
| and not is_SYM_MOD(U, UISC, UDI, UGC, AJT) |
| and not is_Word_Joiner(U, UISC, UDI, UGC, AJT) |
| ) |
| def is_REPHA(U, UISC, UDI, UGC, AJT): |
| return UISC in [Consonant_Preceding_Repha, Consonant_Prefixed] |
| def is_SAKOT(U, UISC, UDI, UGC, AJT): |
| # Split off of HALANT |
| return U == 0x1A60 |
| def is_SYM_MOD(U, UISC, UDI, UGC, AJT): |
| return UISC == Symbol_Modifier |
| def is_VOWEL(U, UISC, UDI, UGC, AJT): |
| return (UISC == Pure_Killer or |
| UGC != Lo and UISC in [Vowel, Vowel_Dependent]) |
| def is_VOWEL_MOD(U, UISC, UDI, UGC, AJT): |
| return (UISC in [Tone_Mark, Cantillation_Mark, Register_Shifter, Visarga] or |
| UGC != Lo and UISC == Bindu) |
| def is_Word_Joiner(U, UISC, UDI, UGC, AJT): |
| # Also includes Rsv |
| return (UDI and U not in [0x115F, 0x1160, 0x3164, 0xFFA0, 0x1BCA0, 0x1BCA1, 0x1BCA2, 0x1BCA3] |
| and UISC == Other |
| and not is_CGJ(U, UISC, UDI, UGC, AJT) |
| ) or UGC == Cn |
| |
| use_mapping = { |
| 'B': is_BASE, |
| 'N': is_BASE_NUM, |
| 'GB': is_BASE_OTHER, |
| 'CGJ': is_CGJ, |
| 'F': is_CONS_FINAL, |
| 'FM': is_CONS_FINAL_MOD, |
| 'M': is_CONS_MED, |
| 'CM': is_CONS_MOD, |
| 'SUB': is_CONS_SUB, |
| 'CS': is_CONS_WITH_STACKER, |
| 'H': is_HALANT, |
| 'HVM': is_HALANT_OR_VOWEL_MODIFIER, |
| 'HN': is_HALANT_NUM, |
| 'IS': is_INVISIBLE_STACKER, |
| 'G': is_HIEROGLYPH, |
| 'J': is_HIEROGLYPH_JOINER, |
| 'SB': is_HIEROGLYPH_SEGMENT_BEGIN, |
| 'SE': is_HIEROGLYPH_SEGMENT_END, |
| 'ZWNJ': is_ZWNJ, |
| 'O': is_OTHER, |
| 'R': is_REPHA, |
| 'Sk': is_SAKOT, |
| 'SM': is_SYM_MOD, |
| 'V': is_VOWEL, |
| 'VM': is_VOWEL_MOD, |
| 'WJ': is_Word_Joiner, |
| } |
| |
| use_positions = { |
| 'F': { |
| 'Abv': [Top], |
| 'Blw': [Bottom], |
| 'Pst': [Right], |
| }, |
| 'M': { |
| 'Abv': [Top], |
| 'Blw': [Bottom, Bottom_And_Left, Bottom_And_Right], |
| 'Pst': [Right], |
| 'Pre': [Left, Top_And_Bottom_And_Left], |
| }, |
| 'CM': { |
| 'Abv': [Top], |
| 'Blw': [Bottom, Overstruck], |
| }, |
| 'V': { |
| 'Abv': [Top, Top_And_Bottom, Top_And_Bottom_And_Right, Top_And_Right], |
| 'Blw': [Bottom, Overstruck, Bottom_And_Right], |
| 'Pst': [Right], |
| 'Pre': [Left, Top_And_Left, Top_And_Left_And_Right, Left_And_Right], |
| }, |
| 'VM': { |
| 'Abv': [Top], |
| 'Blw': [Bottom, Overstruck], |
| 'Pst': [Right], |
| 'Pre': [Left], |
| }, |
| 'SM': { |
| 'Abv': [Top], |
| 'Blw': [Bottom], |
| }, |
| 'H': None, |
| 'HVM': None, |
| 'IS': None, |
| 'B': None, |
| 'FM': { |
| 'Abv': [Top], |
| 'Blw': [Bottom], |
| 'Pst': [Not_Applicable], |
| }, |
| 'R': None, |
| 'SUB': None, |
| } |
| |
| def map_to_use(data): |
| out = {} |
| items = use_mapping.items() |
| for U, (UISC, UIPC, AJT, UDI, UGC, UBlock, _) in data.items(): |
| |
| # Resolve Indic_Syllabic_Category |
| |
| # TODO: These don't have UISC assigned in Unicode 13.0.0, but have UIPC |
| if 0x1CE2 <= U <= 0x1CE8: UISC = Cantillation_Mark |
| |
| # Tibetan: |
| # TODO: These don't have UISC assigned in Unicode 13.0.0, but have UIPC |
| if 0x0F18 <= U <= 0x0F19 or 0x0F3E <= U <= 0x0F3F: UISC = Vowel_Dependent |
| |
| # TODO: U+1CED should only be allowed after some of |
| # the nasalization marks, maybe only for U+1CE9..U+1CF1. |
| if U == 0x1CED: UISC = Tone_Mark |
| |
| values = [k for k,v in items if v(U, UISC, UDI, UGC, AJT)] |
| assert len(values) == 1, "%s %s %s %s %s %s" % (hex(U), UISC, UDI, UGC, AJT, values) |
| USE = values[0] |
| |
| # Resolve Indic_Positional_Category |
| |
| # TODO: https://github.com/harfbuzz/harfbuzz/pull/1037 |
| # and https://github.com/harfbuzz/harfbuzz/issues/1631 |
| if U in [0x11302, 0x11303, 0x114C1]: UIPC = Top |
| |
| assert (UIPC in [Not_Applicable, Visual_Order_Left] or U == 0x0F7F or |
| USE in use_positions), "%s %s %s %s %s %s %s" % (hex(U), UIPC, USE, UISC, UDI, UGC, AJT) |
| |
| pos_mapping = use_positions.get(USE, None) |
| if pos_mapping: |
| values = [k for k,v in pos_mapping.items() if v and UIPC in v] |
| assert len(values) == 1, "%s %s %s %s %s %s %s %s" % (hex(U), UIPC, USE, UISC, UDI, UGC, AJT, values) |
| USE = USE + values[0] |
| |
| out[U] = (USE, UBlock) |
| return out |
| |
| use_data = map_to_use(combined) |
| |
| print ("/* == Start of generated table == */") |
| print ("/*") |
| print (" * The following table is generated by running:") |
| print (" *") |
| print (" * {} IndicSyllabicCategory.txt IndicPositionalCategory.txt ArabicShaping.txt DerivedCoreProperties.txt UnicodeData.txt Blocks.txt Scripts.txt IndicSyllabicCategory-Additional.txt IndicPositionalCategory-Additional.txt".format (sys.argv[0])) |
| print (" *") |
| print (" * on files with these headers:") |
| print (" *") |
| for h in headers: |
| for l in h: |
| print (" * %s" % (l.strip())) |
| print (" */") |
| print () |
| print ("#ifndef HB_OT_SHAPER_USE_TABLE_HH") |
| print ("#define HB_OT_SHAPER_USE_TABLE_HH") |
| print () |
| print ('#include "hb.hh"') |
| print () |
| print ('#include "hb-ot-shaper-use-machine.hh"') |
| print () |
| |
| total = 0 |
| used = 0 |
| last_block = None |
| def print_block (block, start, end, use_data): |
| global total, used, last_block |
| if block and block != last_block: |
| print () |
| print () |
| print (" /* %s */" % block) |
| if start % 16: |
| print (' ' * (20 + (start % 16 * 6)), end='') |
| num = 0 |
| assert start % 8 == 0 |
| assert (end+1) % 8 == 0 |
| for u in range (start, end+1): |
| if u % 16 == 0: |
| print () |
| print (" /* %04X */" % u, end='') |
| if u in use_data: |
| num += 1 |
| d = use_data.get (u) |
| if d is not None: |
| d = d[0] |
| elif u in unicode_data[4]: |
| d = 'O' |
| else: |
| d = 'WJ' |
| print ("%6s," % d, end='') |
| |
| total += end - start + 1 |
| used += num |
| if block: |
| last_block = block |
| |
| uu = sorted (use_data.keys ()) |
| |
| last = -100000 |
| num = 0 |
| offset = 0 |
| starts = [] |
| ends = [] |
| print ('#pragma GCC diagnostic push') |
| print ('#pragma GCC diagnostic ignored "-Wunused-macros"') |
| for k,v in sorted(use_mapping.items()): |
| if k in use_positions and use_positions[k]: continue |
| print ("#define %s USE(%s) /* %s */" % (k, k, v.__name__[3:])) |
| for k,v in sorted(use_positions.items()): |
| if not v: continue |
| for suf in v.keys(): |
| tag = k + suf |
| print ("#define %s USE(%s)" % (tag, tag)) |
| print ('#pragma GCC diagnostic pop') |
| print ("") |
| |
| |
| import packTab |
| data = {u:v[0] for u,v in use_data.items()} |
| |
| DEFAULT = 5 |
| COMPACT = 9 |
| for compression in (DEFAULT, COMPACT): |
| |
| logging.info(' Compression=%d:' % compression) |
| print() |
| if compression == DEFAULT: |
| print('#ifndef HB_OPTIMIZE_SIZE') |
| elif compression == COMPACT: |
| print('#else') |
| else: |
| assert False |
| print() |
| |
| code = packTab.Code('hb_use') |
| sol = packTab.pack_table(data, compression=compression, default='O') |
| logging.info(' FullCost=%d' % (sol.fullCost)) |
| sol.genCode(code, f'get_category') |
| code.print_c(linkage='static inline') |
| print () |
| |
| print('#endif') |
| |
| print () |
| for k in sorted(use_mapping.keys()): |
| if k in use_positions and use_positions[k]: continue |
| print ("#undef %s" % k) |
| for k,v in sorted(use_positions.items()): |
| if not v: continue |
| for suf in v.keys(): |
| tag = k + suf |
| print ("#undef %s" % tag) |
| print () |
| print () |
| print ("#endif /* HB_OT_SHAPER_USE_TABLE_HH */") |
| print ("/* == End of generated table == */") |