#!/usr/bin/python import sys if len (sys.argv) != 5: print >>sys.stderr, "usage: ./gen-use-table.py IndicSyllabicCategory.txt IndicPositionalCategory.txt UnicodeData.txt Blocks.txt" sys.exit (1) BLACKLISTED_BLOCKS = ["Thai", "Lao", "Tibetan"] files = [file (x) for x in sys.argv[1:]] headers = [[f.readline () for i in range (2)] for j,f in enumerate(files) if j != 2] headers.append (["UnicodeData.txt does not have a header."]) data = [{} for f in files] values = [{} for f in files] for i, f in enumerate (files): for line in f: j = line.find ('#') if j >= 0: line = line[:j] fields = [x.strip () for x in line.split (';')] if len (fields) == 1: continue uu = fields[0].split ('..') start = int (uu[0], 16) if len (uu) == 1: end = start else: end = int (uu[1], 16) t = fields[1 if i != 2 else 2] for u in range (start, end + 1): data[i][u] = t values[i][t] = values[i].get (t, 0) + end - start + 1 # Merge data into one dict: defaults = ('Other', 'Not_Applicable', 'Cn', 'No_Block') for i,v in enumerate (defaults): values[i][v] = values[i].get (v, 0) + 1 combined = {} for i,d in enumerate (data): for u,v in d.items (): if i >= 2 and not u in combined: continue if not u in combined: combined[u] = list (defaults) combined[u][i] = v combined = {k:v for k,v in combined.items() if v[3] not in BLACKLISTED_BLOCKS} data = combined del combined num = len (data) property_names = [ # General_Category 'Cc', 'Cf', 'Cn', 'Co', 'Cs', 'Ll', 'Lm', 'Lo', 'Lt', 'Lu', 'Mc', 'Me', 'Mn', 'Nd', 'Nl', 'No', 'Pc', 'Pd', 'Pe', 'Pf', 'Pi', 'Po', 'Ps', 'Sc', 'Sk', 'Sm', 'So', 'Zl', 'Zp', 'Zs', # Indic_Syllabic_Category 'Other', 'Bindu', 'Visarga', 'Avagraha', 'Nukta', 'Virama', 'Pure_Killer', 'Invisible_Stacker', 'Vowel_Independent', 'Vowel_Dependent', 'Vowel', 'Consonant_Placeholder', 'Consonant', 'Consonant_Dead', 'Consonant_With_Stacker', 'Consonant_Prefixed', 'Consonant_Preceding_Repha', 'Consonant_Succeeding_Repha', 'Consonant_Subjoined', 'Consonant_Medial', 'Consonant_Final', 'Consonant_Head_Letter', 'Modifying_Letter', 'Tone_Letter', 'Tone_Mark', 'Gemination_Mark', 'Cantillation_Mark', 'Register_Shifter', 'Syllable_Modifier', 'Consonant_Killer', 'Non_Joiner', 'Joiner', 'Number_Joiner', 'Number', 'Brahmi_Joining_Number', # Indic_Positional_Category 'Not_Applicable', 'Right', 'Left', 'Visual_Order_Left', 'Left_And_Right', 'Top', 'Bottom', 'Top_And_Bottom', 'Top_And_Right', 'Top_And_Left', 'Top_And_Left_And_Right', 'Bottom_And_Right', 'Top_And_Bottom_And_Right', 'Overstruck', ] class PropertyValue(object): def __init__(self, name_): self.name = name_ def __str__(self): return self.name def __eq__(self, other): return self.name == (other if isinstance(other, basestring) else other.name) def __ne__(self, other): return not (self == other) property_values = {} for name in property_names: value = PropertyValue(name) assert value not in property_values assert value not in globals() property_values[name] = value globals().update(property_values) def is_BASE(U, UISC, UGC): return (UISC in [Number, Consonant, Consonant_Head_Letter, #SPEC-OUTDATED Consonant_Placeholder, Tone_Letter] or (UGC == Lo and UISC in [Avagraha, Bindu, Consonant_Final, Consonant_Medial, Consonant_Subjoined, Vowel, Vowel_Dependent])) def is_BASE_VOWEL(U, UISC, UGC): return UISC == Vowel_Independent def is_BASE_IND(U, UISC, UGC): #SPEC-BROKEN return (UISC in [Consonant_Dead, Modifying_Letter] or UGC == Po) return (UISC in [Consonant_Dead, Modifying_Letter] or (UGC == Po and not is_BASE_OTHER(U, UISC, UGC))) # for 104E def is_BASE_NUM(U, UISC, UGC): return UISC == Brahmi_Joining_Number def is_BASE_OTHER(U, UISC, UGC): if UISC == Consonant_Placeholder: return True #SPEC-OUTDATED return U in [0x00A0, 0x00D7, 0x2015, 0x2022, 0x25CC, 0x25FB, 0x25FC, 0x25FD, 0x25FE] def is_CGJ(U, UISC, UGC): return U == 0x034F def is_CONS_FINAL(U, UISC, UGC): return ((UISC == Consonant_Final and UGC != Lo) or UISC == Consonant_Succeeding_Repha) def is_CONS_FINAL_MOD(U, UISC, UGC): #SPEC-OUTDATED return UISC in [Consonant_Final_Modifier, Syllable_Modifier] return UISC == Syllable_Modifier def is_CONS_MED(U, UISC, UGC): return UISC == Consonant_Medial and UGC != Lo def is_CONS_MOD(U, UISC, UGC): return UISC in [Nukta, Gemination_Mark, Consonant_Killer] def is_CONS_SUB(U, UISC, UGC): #SPEC-OUTDATED return UISC == Consonant_Subjoined return UISC == Consonant_Subjoined and UGC != Lo def is_HALANT(U, UISC, UGC): return UISC in [Virama, Invisible_Stacker] def is_HALANT_NUM(U, UISC, UGC): return UISC == Number_Joiner def is_ZWNJ(U, UISC, UGC): return UISC == Non_Joiner def is_ZWJ(U, UISC, UGC): return UISC == Joiner def is_Word_Joiner(U, UISC, UGC): return U == 0x2060 def is_OTHER(U, UISC, UGC): #SPEC-OUTDATED return UGC == Zs # or any other SCRIPT_COMMON characters return UISC == Other and not is_SYM_MOD(U, UISC, UGC) def is_Reserved(U, UISC, UGC): return UGC == 'Cn' def is_REPHA(U, UISC, UGC): #return UISC == Consonant_Preceding_Repha #SPEC-OUTDATED hack to categorize Consonant_With_Stacker and Consonant_Prefixed return UISC in [Consonant_Preceding_Repha, Consonant_With_Stacker, Consonant_Prefixed] def is_SYM(U, UISC, UGC): if U == 0x25CC: return False #SPEC-OUTDATED #SPEC-OUTDATED return UGC in [So, Sc] or UISC == Symbol_Letter return UGC in [So, Sc] def is_SYM_MOD(U, UISC, UGC): return U in [0x1B6B, 0x1B6C, 0x1B6D, 0x1B6E, 0x1B6F, 0x1B70, 0x1B71, 0x1B72, 0x1B73] def is_VARIATION_SELECTOR(U, UISC, UGC): return 0xFE00 <= U <= 0xFE0F def is_VOWEL(U, UISC, UGC): return (UISC == Pure_Killer or (UGC != Lo and UISC in [Vowel, Vowel_Dependent])) def is_VOWEL_MOD(U, UISC, UGC): return (UISC in [Tone_Mark, Cantillation_Mark, Register_Shifter, Visarga] or (UGC != Lo and UISC == Bindu)) use_mapping = { 'B': is_BASE, 'IV': is_BASE_VOWEL, 'IND': is_BASE_IND, 'N': is_BASE_NUM, 'GB': is_BASE_OTHER, 'CGJ': is_CGJ, 'F': is_CONS_FINAL, 'FM': is_CONS_FINAL_MOD, 'M': is_CONS_MED, 'CM': is_CONS_MOD, 'SUB': is_CONS_SUB, 'H': is_HALANT, 'HN': is_HALANT_NUM, 'ZWNJ': is_ZWNJ, 'ZWJ': is_ZWJ, 'WJ': is_Word_Joiner, 'O': is_OTHER, 'Rsv': is_Reserved, 'R': is_REPHA, 'S': is_SYM, 'SM': is_SYM_MOD, 'VS': is_VARIATION_SELECTOR, 'V': is_VOWEL, 'VM': is_VOWEL_MOD, } use_positions = { 'F': { 'Abv': [Top], 'Blw': [Bottom], 'Pst': [Right], }, 'M': { 'Abv': [Top], 'Blw': [Bottom], 'Pst': [Right], 'Pre': [Left], }, 'CM': { 'Abv': [Top], 'Blw': [Bottom], }, 'V': { 'Abv': [Top, Top_And_Bottom, Top_And_Bottom_And_Right, Top_And_Right], 'Blw': [Bottom, Overstruck, Bottom_And_Right], 'Pst': [Right], 'Pre': [Left, Top_And_Left, Top_And_Left_And_Right, Left_And_Right], }, 'VM': { 'Abv': [Top], 'Blw': [Bottom, Overstruck], 'Pst': [Right], 'Pre': [Left], }, 'SM': { 'Abv': [Top], 'Blw': [Bottom], }, 'H': None, 'B': None, 'FM': None, 'SUB': None, } def map_to_use(data): out = {} items = use_mapping.items() for U,(UISC,UIPC,UGC,UBlock) in data.items(): # Resolve Indic_Syllabic_Category # TODO: These don't have UISC assigned in Unicode 8.0, but # have UIPC if U == 0x17DD: UISC = Vowel_Dependent if 0x1CE2 <= U <= 0x1CE8: UISC = Cantillation_Mark # TODO: U+1CED should only be allowed after some of # the nasalization marks, maybe only for U+1CE9..U+1CF1. if U == 0x1CED: UISC = Tone_Mark evals = [(k, v(U,UISC,UGC)) for k,v in items] values = [k for k,v in evals if v] assert len(values) == 1, "%s %s %s %s" % (hex(U), UISC, UGC, values) USE = values[0] # Resolve Indic_Positional_Category # TODO: Not in Unicode 8.0 yet, but in spec. if U == 0x1B6C: UIPC = Bottom # TODO: These should die, but have UIPC in Unicode 8.0 if U in [0x953, 0x954]: UIPC = Not_Applicable # TODO: In USE's override list but not in Unicode 8.0 if U == 0x103C: UIPC = Left # TODO: These are not in USE's override list that we have, nor are they in Unicode 8.0 if 0xA926 <= U <= 0xA92A: UIPC = Top if U == 0x111CA: UIPC = Bottom if U == 0x11300: UIPC = Top if U == 0x1133C: UIPC = Bottom if U == 0x1171E: UIPC = Left # Correct?! if 0x1CF2 <= U <= 0x1CF3: UIPC = Right if 0x1CF8 <= U <= 0x1CF9: UIPC = Top assert (UIPC in [Not_Applicable, Visual_Order_Left] or USE in use_positions), "%s %s %s %s %s" % (hex(U), UIPC, USE, UISC, UGC) pos_mapping = use_positions.get(USE, None) if pos_mapping: values = [k for k,v in pos_mapping.items() if v and UIPC in v] assert len(values) == 1, "%s %s %s %s %s %s" % (hex(U), UIPC, USE, UISC, UGC, values) USE = USE + values[0] out[U] = (USE, UBlock) return out defaults = ('O', 'No_Block') data = map_to_use(data) # Remove the outliers singles = {} for u in [0x25CC, 0x1107F]: singles[u] = data[u] del data[u] print "/* == Start of generated table == */" print "/*" print " * The following table is generated by running:" print " *" print " * ./gen-use-table.py IndicSyllabicCategory.txt IndicPositionalCategory.txt UnicodeData.txt Blocks.txt" print " *" print " * on files with these headers:" print " *" for h in headers: for l in h: print " * %s" % (l.strip()) print " */" print print '#include "hb-ot-shape-complex-use-private.hh"' print total = 0 used = 0 last_block = None def print_block (block, start, end, data): global total, used, last_block if block and block != last_block: print print print " /* %s */" % block if start % 16: print ' ' * (20 + (start % 16 * 6)), num = 0 assert start % 8 == 0 assert (end+1) % 8 == 0 for u in range (start, end+1): if u % 16 == 0: print print " /* %04X */" % u, if u in data: num += 1 d = data.get (u, defaults) sys.stdout.write ("%6s," % d[0]) total += end - start + 1 used += num if block: last_block = block uu = data.keys () uu.sort () last = -100000 num = 0 offset = 0 starts = [] ends = [] for k,v in sorted(use_mapping.items()): if k in use_positions and use_positions[k]: continue print "#define %s USE_%s /* %s */" % (k, k, v.__name__[3:]) for k,v in sorted(use_positions.items()): if not v: continue for suf in v.keys(): tag = k + suf print "#define %s USE_%s" % (tag, tag) print "" print "static const USE_TABLE_ELEMENT_TYPE use_table[] = {" for u in uu: if u <= last: continue block = data[u][1] start = u//8*8 end = start+1 while end in uu and block == data[end][1]: end += 1 end = (end-1)//8*8 + 7 if start != last + 1: if start - last <= 1+16*3: print_block (None, last+1, start-1, data) last = start-1 else: if last >= 0: ends.append (last + 1) offset += ends[-1] - starts[-1] print print print "#define use_offset_0x%04xu %d" % (start, offset) starts.append (start) print_block (block, start, end, data) last = end ends.append (last + 1) offset += ends[-1] - starts[-1] print print occupancy = used * 100. / total page_bits = 12 print "}; /* Table items: %d; occupancy: %d%% */" % (offset, occupancy) print for k in sorted(use_mapping.keys()): if k in use_positions and use_positions[k]: continue print "#undef %s" % k for k,v in sorted(use_positions.items()): if not v: continue for suf in v.keys(): tag = k + suf print "#undef %s" % tag print print "USE_TABLE_ELEMENT_TYPE" print "hb_use_get_categories (hb_codepoint_t u)" print "{" print " switch (u >> %d)" % page_bits print " {" pages = set([u>>page_bits for u in starts+ends+singles.keys()]) for p in sorted(pages): print " case 0x%0Xu:" % p for (start,end) in zip (starts, ends): if p not in [start>>page_bits, end>>page_bits]: continue offset = "use_offset_0x%04xu" % start print " if (hb_in_range (u, 0x%04Xu, 0x%04Xu)) return use_table[u - 0x%04Xu + %s];" % (start, end-1, start, offset) for u,d in singles.items (): if p != u>>page_bits: continue print " if (unlikely (u == 0x%04Xu)) return USE_%s;" % (u, d[0]) print " break;" print "" print " default:" print " break;" print " }" print " return USE_O;" print "}" print print "/* == End of generated table == */" # Maintain at least 50% occupancy in the table */ if occupancy < 50: raise Exception ("Table too sparse, please investigate: ", occupancy)