508 lines
14 KiB
Python
Executable File
508 lines
14 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
# flake8: noqa: F821
|
|
|
|
import logging
|
|
logging.basicConfig(format='%(levelname)s: %(message)s', level=logging.INFO)
|
|
|
|
"""usage: ./gen-use-table.py IndicSyllabicCategory.txt IndicPositionalCategory.txt ArabicShaping.txt DerivedCoreProperties.txt UnicodeData.txt Blocks.txt Scripts.txt IndicSyllabicCategory-Additional.txt IndicPositionalCategory-Additional.txt
|
|
|
|
Input files:
|
|
* https://unicode.org/Public/UCD/latest/ucd/IndicSyllabicCategory.txt
|
|
* https://unicode.org/Public/UCD/latest/ucd/IndicPositionalCategory.txt
|
|
* https://unicode.org/Public/UCD/latest/ucd/ArabicShaping.txt
|
|
* https://unicode.org/Public/UCD/latest/ucd/DerivedCoreProperties.txt
|
|
* https://unicode.org/Public/UCD/latest/ucd/UnicodeData.txt
|
|
* https://unicode.org/Public/UCD/latest/ucd/Blocks.txt
|
|
* https://unicode.org/Public/UCD/latest/ucd/Scripts.txt
|
|
* ms-use/IndicSyllabicCategory-Additional.txt
|
|
* ms-use/IndicPositionalCategory-Additional.txt
|
|
"""
|
|
|
|
import sys
|
|
|
|
if len (sys.argv) != 10:
|
|
sys.exit (__doc__)
|
|
|
|
DISABLED_SCRIPTS = {
|
|
'Arabic',
|
|
'Lao',
|
|
'Samaritan',
|
|
'Syriac',
|
|
'Thai',
|
|
}
|
|
|
|
files = [open (x, encoding='utf-8') for x in sys.argv[1:]]
|
|
|
|
headers = [[f.readline () for i in range (2)] for j,f in enumerate(files) if j != 4]
|
|
for j in range(7, 9):
|
|
for line in files[j]:
|
|
line = line.rstrip()
|
|
if not line:
|
|
break
|
|
headers[j - 1].append(line)
|
|
headers.append (["UnicodeData.txt does not have a header."])
|
|
|
|
unicode_data = [{} for _ in files]
|
|
values = [{} for _ in files]
|
|
for i, f in enumerate (files):
|
|
for line in f:
|
|
|
|
j = line.find ('#')
|
|
if j >= 0:
|
|
line = line[:j]
|
|
|
|
fields = [x.strip () for x in line.split (';')]
|
|
if len (fields) == 1:
|
|
continue
|
|
|
|
uu = fields[0].split ('..')
|
|
start = int (uu[0], 16)
|
|
if len (uu) == 1:
|
|
end = start
|
|
else:
|
|
end = int (uu[1], 16)
|
|
|
|
t = fields[1 if i not in [2, 4] else 2]
|
|
|
|
if i == 2:
|
|
t = 'jt_' + t
|
|
elif i == 3 and t != 'Default_Ignorable_Code_Point':
|
|
continue
|
|
elif i == 7 and t == 'Consonant_Final_Modifier':
|
|
# TODO: https://github.com/MicrosoftDocs/typography-issues/issues/336
|
|
t = 'Syllable_Modifier'
|
|
elif i == 8 and t == 'NA':
|
|
t = 'Not_Applicable'
|
|
|
|
i0 = i if i < 7 else i - 7
|
|
for u in range (start, end + 1):
|
|
unicode_data[i0][u] = t
|
|
values[i0][t] = values[i0].get (t, 0) + end - start + 1
|
|
|
|
defaults = ('Other', 'Not_Applicable', 'jt_X', '', 'Cn', 'No_Block', 'Unknown')
|
|
|
|
# Merge data into one dict:
|
|
for i,v in enumerate (defaults):
|
|
values[i][v] = values[i].get (v, 0) + 1
|
|
combined = {}
|
|
for i,d in enumerate (unicode_data):
|
|
for u,v in d.items ():
|
|
if not u in combined:
|
|
if i >= 4:
|
|
continue
|
|
combined[u] = list (defaults)
|
|
combined[u][i] = v
|
|
combined = {k: v for k, v in combined.items() if v[6] not in DISABLED_SCRIPTS}
|
|
|
|
|
|
property_names = [
|
|
# General_Category
|
|
'Cc', 'Cf', 'Cn', 'Co', 'Cs', 'Ll', 'Lm', 'Lo', 'Lt', 'Lu', 'Mc',
|
|
'Me', 'Mn', 'Nd', 'Nl', 'No', 'Pc', 'Pd', 'Pe', 'Pf', 'Pi', 'Po',
|
|
'Ps', 'Sc', 'Sk', 'Sm', 'So', 'Zl', 'Zp', 'Zs',
|
|
# Indic_Syllabic_Category
|
|
'Other',
|
|
'Bindu',
|
|
'Visarga',
|
|
'Avagraha',
|
|
'Nukta',
|
|
'Virama',
|
|
'Pure_Killer',
|
|
'Invisible_Stacker',
|
|
'Vowel_Independent',
|
|
'Vowel_Dependent',
|
|
'Vowel',
|
|
'Consonant_Placeholder',
|
|
'Consonant',
|
|
'Consonant_Dead',
|
|
'Consonant_With_Stacker',
|
|
'Consonant_Prefixed',
|
|
'Consonant_Preceding_Repha',
|
|
'Consonant_Succeeding_Repha',
|
|
'Consonant_Subjoined',
|
|
'Consonant_Medial',
|
|
'Consonant_Final',
|
|
'Consonant_Head_Letter',
|
|
'Consonant_Initial_Postfixed',
|
|
'Modifying_Letter',
|
|
'Tone_Letter',
|
|
'Tone_Mark',
|
|
'Gemination_Mark',
|
|
'Cantillation_Mark',
|
|
'Register_Shifter',
|
|
'Syllable_Modifier',
|
|
'Consonant_Killer',
|
|
'Non_Joiner',
|
|
'Joiner',
|
|
'Number_Joiner',
|
|
'Number',
|
|
'Brahmi_Joining_Number',
|
|
'Symbol_Modifier',
|
|
'Hieroglyph',
|
|
'Hieroglyph_Joiner',
|
|
'Hieroglyph_Segment_Begin',
|
|
'Hieroglyph_Segment_End',
|
|
# Indic_Positional_Category
|
|
'Not_Applicable',
|
|
'Right',
|
|
'Left',
|
|
'Visual_Order_Left',
|
|
'Left_And_Right',
|
|
'Top',
|
|
'Bottom',
|
|
'Top_And_Bottom',
|
|
'Top_And_Bottom_And_Left',
|
|
'Top_And_Right',
|
|
'Top_And_Left',
|
|
'Top_And_Left_And_Right',
|
|
'Bottom_And_Left',
|
|
'Bottom_And_Right',
|
|
'Top_And_Bottom_And_Right',
|
|
'Overstruck',
|
|
# Joining_Type
|
|
'jt_C',
|
|
'jt_D',
|
|
'jt_L',
|
|
'jt_R',
|
|
'jt_T',
|
|
'jt_U',
|
|
'jt_X',
|
|
]
|
|
|
|
class PropertyValue(object):
|
|
def __init__(self, name_):
|
|
self.name = name_
|
|
def __str__(self):
|
|
return self.name
|
|
def __eq__(self, other):
|
|
return self.name == (other if isinstance(other, str) else other.name)
|
|
def __ne__(self, other):
|
|
return not (self == other)
|
|
def __hash__(self):
|
|
return hash(str(self))
|
|
|
|
property_values = {}
|
|
|
|
for name in property_names:
|
|
value = PropertyValue(name)
|
|
assert value not in property_values
|
|
assert value not in globals()
|
|
property_values[name] = value
|
|
globals().update(property_values)
|
|
|
|
|
|
def is_BASE(U, UISC, UDI, UGC, AJT):
|
|
return (UISC in [Number, Consonant, Consonant_Head_Letter,
|
|
Tone_Letter,
|
|
Vowel_Independent,
|
|
] or
|
|
# TODO: https://github.com/MicrosoftDocs/typography-issues/issues/484
|
|
AJT in [jt_C, jt_D, jt_L, jt_R] and UISC != Joiner or
|
|
(UGC == Lo and UISC in [Avagraha, Bindu, Consonant_Final, Consonant_Medial,
|
|
Consonant_Subjoined, Vowel, Vowel_Dependent]))
|
|
def is_BASE_NUM(U, UISC, UDI, UGC, AJT):
|
|
return UISC == Brahmi_Joining_Number
|
|
def is_BASE_OTHER(U, UISC, UDI, UGC, AJT):
|
|
if UISC == Consonant_Placeholder: return True
|
|
return U in [0x2015, 0x2022, 0x25FB, 0x25FC, 0x25FD, 0x25FE]
|
|
def is_CGJ(U, UISC, UDI, UGC, AJT):
|
|
# Also includes VARIATION_SELECTOR and ZWJ
|
|
return UISC == Joiner or UDI and UGC in [Mc, Me, Mn]
|
|
def is_CONS_FINAL(U, UISC, UDI, UGC, AJT):
|
|
return ((UISC == Consonant_Final and UGC != Lo) or
|
|
UISC == Consonant_Succeeding_Repha)
|
|
def is_CONS_FINAL_MOD(U, UISC, UDI, UGC, AJT):
|
|
return UISC == Syllable_Modifier
|
|
def is_CONS_MED(U, UISC, UDI, UGC, AJT):
|
|
# Consonant_Initial_Postfixed is new in Unicode 11; not in the spec.
|
|
return (UISC == Consonant_Medial and UGC != Lo or
|
|
UISC == Consonant_Initial_Postfixed)
|
|
def is_CONS_MOD(U, UISC, UDI, UGC, AJT):
|
|
return UISC in [Nukta, Gemination_Mark, Consonant_Killer]
|
|
def is_CONS_SUB(U, UISC, UDI, UGC, AJT):
|
|
return UISC == Consonant_Subjoined and UGC != Lo
|
|
def is_CONS_WITH_STACKER(U, UISC, UDI, UGC, AJT):
|
|
return UISC == Consonant_With_Stacker
|
|
def is_HALANT(U, UISC, UDI, UGC, AJT):
|
|
return UISC == Virama and not is_HALANT_OR_VOWEL_MODIFIER(U, UISC, UDI, UGC, AJT)
|
|
def is_HALANT_OR_VOWEL_MODIFIER(U, UISC, UDI, UGC, AJT):
|
|
# Split off of HALANT
|
|
return U == 0x0DCA
|
|
def is_HALANT_NUM(U, UISC, UDI, UGC, AJT):
|
|
return UISC == Number_Joiner
|
|
def is_HIEROGLYPH(U, UISC, UDI, UGC, AJT):
|
|
return UISC == Hieroglyph
|
|
def is_HIEROGLYPH_JOINER(U, UISC, UDI, UGC, AJT):
|
|
return UISC == Hieroglyph_Joiner
|
|
def is_HIEROGLYPH_SEGMENT_BEGIN(U, UISC, UDI, UGC, AJT):
|
|
return UISC == Hieroglyph_Segment_Begin
|
|
def is_HIEROGLYPH_SEGMENT_END(U, UISC, UDI, UGC, AJT):
|
|
return UISC == Hieroglyph_Segment_End
|
|
def is_INVISIBLE_STACKER(U, UISC, UDI, UGC, AJT):
|
|
# Split off of HALANT
|
|
return (UISC == Invisible_Stacker
|
|
and not is_SAKOT(U, UISC, UDI, UGC, AJT)
|
|
)
|
|
def is_ZWNJ(U, UISC, UDI, UGC, AJT):
|
|
return UISC == Non_Joiner
|
|
def is_OTHER(U, UISC, UDI, UGC, AJT):
|
|
# Also includes BASE_IND and SYM
|
|
return ((UGC == Po or UISC in [Consonant_Dead, Joiner, Modifying_Letter, Other])
|
|
and not is_BASE(U, UISC, UDI, UGC, AJT)
|
|
and not is_BASE_OTHER(U, UISC, UDI, UGC, AJT)
|
|
and not is_CGJ(U, UISC, UDI, UGC, AJT)
|
|
and not is_SYM_MOD(U, UISC, UDI, UGC, AJT)
|
|
and not is_Word_Joiner(U, UISC, UDI, UGC, AJT)
|
|
)
|
|
def is_REPHA(U, UISC, UDI, UGC, AJT):
|
|
return UISC in [Consonant_Preceding_Repha, Consonant_Prefixed]
|
|
def is_SAKOT(U, UISC, UDI, UGC, AJT):
|
|
# Split off of HALANT
|
|
return U == 0x1A60
|
|
def is_SYM_MOD(U, UISC, UDI, UGC, AJT):
|
|
return UISC == Symbol_Modifier
|
|
def is_VOWEL(U, UISC, UDI, UGC, AJT):
|
|
return (UISC == Pure_Killer or
|
|
UGC != Lo and UISC in [Vowel, Vowel_Dependent])
|
|
def is_VOWEL_MOD(U, UISC, UDI, UGC, AJT):
|
|
return (UISC in [Tone_Mark, Cantillation_Mark, Register_Shifter, Visarga] or
|
|
UGC != Lo and UISC == Bindu)
|
|
def is_Word_Joiner(U, UISC, UDI, UGC, AJT):
|
|
# Also includes Rsv
|
|
return (UDI and U not in [0x115F, 0x1160, 0x3164, 0xFFA0, 0x1BCA0, 0x1BCA1, 0x1BCA2, 0x1BCA3]
|
|
and UISC == Other
|
|
and not is_CGJ(U, UISC, UDI, UGC, AJT)
|
|
) or UGC == Cn
|
|
|
|
use_mapping = {
|
|
'B': is_BASE,
|
|
'N': is_BASE_NUM,
|
|
'GB': is_BASE_OTHER,
|
|
'CGJ': is_CGJ,
|
|
'F': is_CONS_FINAL,
|
|
'FM': is_CONS_FINAL_MOD,
|
|
'M': is_CONS_MED,
|
|
'CM': is_CONS_MOD,
|
|
'SUB': is_CONS_SUB,
|
|
'CS': is_CONS_WITH_STACKER,
|
|
'H': is_HALANT,
|
|
'HVM': is_HALANT_OR_VOWEL_MODIFIER,
|
|
'HN': is_HALANT_NUM,
|
|
'IS': is_INVISIBLE_STACKER,
|
|
'G': is_HIEROGLYPH,
|
|
'J': is_HIEROGLYPH_JOINER,
|
|
'SB': is_HIEROGLYPH_SEGMENT_BEGIN,
|
|
'SE': is_HIEROGLYPH_SEGMENT_END,
|
|
'ZWNJ': is_ZWNJ,
|
|
'O': is_OTHER,
|
|
'R': is_REPHA,
|
|
'Sk': is_SAKOT,
|
|
'SM': is_SYM_MOD,
|
|
'V': is_VOWEL,
|
|
'VM': is_VOWEL_MOD,
|
|
'WJ': is_Word_Joiner,
|
|
}
|
|
|
|
use_positions = {
|
|
'F': {
|
|
'Abv': [Top],
|
|
'Blw': [Bottom],
|
|
'Pst': [Right],
|
|
},
|
|
'M': {
|
|
'Abv': [Top],
|
|
'Blw': [Bottom, Bottom_And_Left, Bottom_And_Right],
|
|
'Pst': [Right],
|
|
'Pre': [Left, Top_And_Bottom_And_Left],
|
|
},
|
|
'CM': {
|
|
'Abv': [Top],
|
|
'Blw': [Bottom, Overstruck],
|
|
},
|
|
'V': {
|
|
'Abv': [Top, Top_And_Bottom, Top_And_Bottom_And_Right, Top_And_Right],
|
|
'Blw': [Bottom, Overstruck, Bottom_And_Right],
|
|
'Pst': [Right],
|
|
'Pre': [Left, Top_And_Left, Top_And_Left_And_Right, Left_And_Right],
|
|
},
|
|
'VM': {
|
|
'Abv': [Top],
|
|
'Blw': [Bottom, Overstruck],
|
|
'Pst': [Right],
|
|
'Pre': [Left],
|
|
},
|
|
'SM': {
|
|
'Abv': [Top],
|
|
'Blw': [Bottom],
|
|
},
|
|
'H': None,
|
|
'HVM': None,
|
|
'IS': None,
|
|
'B': None,
|
|
'FM': {
|
|
'Abv': [Top],
|
|
'Blw': [Bottom],
|
|
'Pst': [Not_Applicable],
|
|
},
|
|
'R': None,
|
|
'SUB': None,
|
|
}
|
|
|
|
def map_to_use(data):
|
|
out = {}
|
|
items = use_mapping.items()
|
|
for U, (UISC, UIPC, AJT, UDI, UGC, UBlock, _) in data.items():
|
|
|
|
# Resolve Indic_Syllabic_Category
|
|
|
|
# TODO: These don't have UISC assigned in Unicode 13.0.0, but have UIPC
|
|
if 0x1CE2 <= U <= 0x1CE8: UISC = Cantillation_Mark
|
|
|
|
# Tibetan:
|
|
# TODO: These don't have UISC assigned in Unicode 13.0.0, but have UIPC
|
|
if 0x0F18 <= U <= 0x0F19 or 0x0F3E <= U <= 0x0F3F: UISC = Vowel_Dependent
|
|
|
|
# TODO: U+1CED should only be allowed after some of
|
|
# the nasalization marks, maybe only for U+1CE9..U+1CF1.
|
|
if U == 0x1CED: UISC = Tone_Mark
|
|
|
|
values = [k for k,v in items if v(U, UISC, UDI, UGC, AJT)]
|
|
assert len(values) == 1, "%s %s %s %s %s %s" % (hex(U), UISC, UDI, UGC, AJT, values)
|
|
USE = values[0]
|
|
|
|
# Resolve Indic_Positional_Category
|
|
|
|
# TODO: https://github.com/harfbuzz/harfbuzz/pull/1037
|
|
# and https://github.com/harfbuzz/harfbuzz/issues/1631
|
|
if U in [0x11302, 0x11303, 0x114C1]: UIPC = Top
|
|
|
|
assert (UIPC in [Not_Applicable, Visual_Order_Left] or U == 0x0F7F or
|
|
USE in use_positions), "%s %s %s %s %s %s %s" % (hex(U), UIPC, USE, UISC, UDI, UGC, AJT)
|
|
|
|
pos_mapping = use_positions.get(USE, None)
|
|
if pos_mapping:
|
|
values = [k for k,v in pos_mapping.items() if v and UIPC in v]
|
|
assert len(values) == 1, "%s %s %s %s %s %s %s %s" % (hex(U), UIPC, USE, UISC, UDI, UGC, AJT, values)
|
|
USE = USE + values[0]
|
|
|
|
out[U] = (USE, UBlock)
|
|
return out
|
|
|
|
use_data = map_to_use(combined)
|
|
|
|
print ("/* == Start of generated table == */")
|
|
print ("/*")
|
|
print (" * The following table is generated by running:")
|
|
print (" *")
|
|
print (" * {} IndicSyllabicCategory.txt IndicPositionalCategory.txt ArabicShaping.txt DerivedCoreProperties.txt UnicodeData.txt Blocks.txt Scripts.txt IndicSyllabicCategory-Additional.txt IndicPositionalCategory-Additional.txt".format (sys.argv[0]))
|
|
print (" *")
|
|
print (" * on files with these headers:")
|
|
print (" *")
|
|
for h in headers:
|
|
for l in h:
|
|
print (" * %s" % (l.strip()))
|
|
print (" */")
|
|
print ()
|
|
print ("#ifndef HB_OT_SHAPER_USE_TABLE_HH")
|
|
print ("#define HB_OT_SHAPER_USE_TABLE_HH")
|
|
print ()
|
|
print ('#include "hb.hh"')
|
|
print ()
|
|
print ('#include "hb-ot-shaper-use-machine.hh"')
|
|
print ()
|
|
|
|
total = 0
|
|
used = 0
|
|
last_block = None
|
|
def print_block (block, start, end, use_data):
|
|
global total, used, last_block
|
|
if block and block != last_block:
|
|
print ()
|
|
print ()
|
|
print (" /* %s */" % block)
|
|
if start % 16:
|
|
print (' ' * (20 + (start % 16 * 6)), end='')
|
|
num = 0
|
|
assert start % 8 == 0
|
|
assert (end+1) % 8 == 0
|
|
for u in range (start, end+1):
|
|
if u % 16 == 0:
|
|
print ()
|
|
print (" /* %04X */" % u, end='')
|
|
if u in use_data:
|
|
num += 1
|
|
d = use_data.get (u)
|
|
if d is not None:
|
|
d = d[0]
|
|
elif u in unicode_data[4]:
|
|
d = 'O'
|
|
else:
|
|
d = 'WJ'
|
|
print ("%6s," % d, end='')
|
|
|
|
total += end - start + 1
|
|
used += num
|
|
if block:
|
|
last_block = block
|
|
|
|
uu = sorted (use_data.keys ())
|
|
|
|
last = -100000
|
|
num = 0
|
|
offset = 0
|
|
starts = []
|
|
ends = []
|
|
print ('#pragma GCC diagnostic push')
|
|
print ('#pragma GCC diagnostic ignored "-Wunused-macros"')
|
|
for k,v in sorted(use_mapping.items()):
|
|
if k in use_positions and use_positions[k]: continue
|
|
print ("#define %s USE(%s) /* %s */" % (k, k, v.__name__[3:]))
|
|
for k,v in sorted(use_positions.items()):
|
|
if not v: continue
|
|
for suf in v.keys():
|
|
tag = k + suf
|
|
print ("#define %s USE(%s)" % (tag, tag))
|
|
print ('#pragma GCC diagnostic pop')
|
|
print ("")
|
|
|
|
|
|
import packTab
|
|
data = {u:v[0] for u,v in use_data.items()}
|
|
|
|
DEFAULT = 5
|
|
COMPACT = 9
|
|
for compression in (DEFAULT, COMPACT):
|
|
|
|
logging.info(' Compression=%d:' % compression)
|
|
print()
|
|
if compression == DEFAULT:
|
|
print('#ifndef HB_OPTIMIZE_SIZE')
|
|
elif compression == COMPACT:
|
|
print('#else')
|
|
else:
|
|
assert False
|
|
print()
|
|
|
|
code = packTab.Code('hb_use')
|
|
sol = packTab.pack_table(data, compression=compression, default='O')
|
|
logging.info(' FullCost=%d' % (sol.fullCost))
|
|
sol.genCode(code, f'get_category')
|
|
code.print_c(linkage='static inline')
|
|
print ()
|
|
|
|
print('#endif')
|
|
|
|
print ()
|
|
for k in sorted(use_mapping.keys()):
|
|
if k in use_positions and use_positions[k]: continue
|
|
print ("#undef %s" % k)
|
|
for k,v in sorted(use_positions.items()):
|
|
if not v: continue
|
|
for suf in v.keys():
|
|
tag = k + suf
|
|
print ("#undef %s" % tag)
|
|
print ()
|
|
print ()
|
|
print ("#endif /* HB_OT_SHAPER_USE_TABLE_HH */")
|
|
print ("/* == End of generated table == */")
|