harfbuzz/src/hb-ot-layout-gsubgpos.hh

3419 lines
107 KiB
C++
Raw Normal View History

/*
2011-04-21 23:14:28 +02:00
* Copyright © 2007,2008,2009,2010 Red Hat, Inc.
2012-04-24 04:26:13 +02:00
* Copyright © 2010,2012 Google, Inc.
*
2010-04-22 06:11:43 +02:00
* This is part of HarfBuzz, a text shaping library.
*
* Permission is hereby granted, without written agreement and without
* license or royalty fees, to use, copy, modify, and distribute this
* software and its documentation for any purpose, provided that the
* above copyright notice and the following two paragraphs appear in
* all copies of this software.
*
* IN NO EVENT SHALL THE COPYRIGHT HOLDER BE LIABLE TO ANY PARTY FOR
* DIRECT, INDIRECT, SPECIAL, INCIDENTAL, OR CONSEQUENTIAL DAMAGES
* ARISING OUT OF THE USE OF THIS SOFTWARE AND ITS DOCUMENTATION, EVEN
* IF THE COPYRIGHT HOLDER HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH
* DAMAGE.
*
* THE COPYRIGHT HOLDER SPECIFICALLY DISCLAIMS ANY WARRANTIES, INCLUDING,
* BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND
* FITNESS FOR A PARTICULAR PURPOSE. THE SOFTWARE PROVIDED HEREUNDER IS
* ON AN "AS IS" BASIS, AND THE COPYRIGHT HOLDER HAS NO OBLIGATION TO
* PROVIDE MAINTENANCE, SUPPORT, UPDATES, ENHANCEMENTS, OR MODIFICATIONS.
*
* Red Hat Author(s): Behdad Esfahbod
* Google Author(s): Behdad Esfahbod
*/
#ifndef HB_OT_LAYOUT_GSUBGPOS_HH
#define HB_OT_LAYOUT_GSUBGPOS_HH
#include "hb.hh"
#include "hb-buffer.hh"
#include "hb-map.hh"
#include "hb-set.hh"
#include "hb-ot-map.hh"
#include "hb-ot-layout-common.hh"
#include "hb-ot-layout-gdef-table.hh"
namespace OT {
2010-07-23 21:11:18 +02:00
struct hb_intersects_context_t :
hb_dispatch_context_t<hb_intersects_context_t, bool, 0>
{
const char *get_name () { return "INTERSECTS"; }
template <typename T>
return_t dispatch (const T &obj) { return obj.intersects (this->glyphs); }
static return_t default_return_value () { return false; }
bool stop_sublookup_iteration (return_t r) const { return r; }
const hb_set_t *glyphs;
unsigned int debug_depth;
hb_intersects_context_t (const hb_set_t *glyphs_) :
glyphs (glyphs_),
debug_depth (0) {}
};
2015-10-09 18:20:58 +02:00
struct hb_closure_context_t :
2019-05-11 05:43:26 +02:00
hb_dispatch_context_t<hb_closure_context_t, hb_empty_t, 0>
{
const char *get_name () { return "CLOSURE"; }
typedef return_t (*recurse_func_t) (hb_closure_context_t *c, unsigned int lookup_index);
template <typename T>
2019-05-11 05:43:26 +02:00
return_t dispatch (const T &obj) { obj.closure (this); return hb_empty_t (); }
static return_t default_return_value () { return hb_empty_t (); }
void recurse (unsigned int lookup_index)
{
2012-11-23 23:55:40 +01:00
if (unlikely (nesting_level_left == 0 || !recurse_func))
return;
nesting_level_left--;
recurse_func (this, lookup_index);
nesting_level_left++;
}
bool lookup_limit_exceeded ()
2020-03-26 07:45:09 +01:00
{ return lookup_count > HB_MAX_LOOKUP_INDICES; }
bool should_visit_lookup (unsigned int lookup_index)
{
if (lookup_count++ > HB_MAX_LOOKUP_INDICES)
return false;
if (is_lookup_done (lookup_index))
return false;
done_lookups->set (lookup_index, glyphs->get_population ());
return true;
}
bool is_lookup_done (unsigned int lookup_index)
{
/* Have we visited this lookup with the current set of glyphs? */
return done_lookups->get (lookup_index) == glyphs->get_population ();
}
hb_face_t *face;
2012-04-24 04:23:17 +02:00
hb_set_t *glyphs;
2019-01-09 19:45:53 +01:00
hb_set_t output[1];
recurse_func_t recurse_func;
unsigned int nesting_level_left;
unsigned int debug_depth;
hb_closure_context_t (hb_face_t *face_,
2012-04-24 04:23:17 +02:00
hb_set_t *glyphs_,
2018-11-15 20:40:56 +01:00
hb_map_t *done_lookups_,
unsigned int nesting_level_left_ = HB_MAX_NESTING_LEVEL) :
face (face_),
glyphs (glyphs_),
2017-10-15 12:11:08 +02:00
recurse_func (nullptr),
nesting_level_left (nesting_level_left_),
debug_depth (0),
done_lookups (done_lookups_),
lookup_count (0)
{}
2012-11-23 23:55:40 +01:00
~hb_closure_context_t () { flush (); }
2012-11-23 23:55:40 +01:00
void set_recurse_func (recurse_func_t func) { recurse_func = func; }
void flush ()
{
hb_set_del_range (output, face->get_num_glyphs (), hb_set_get_max (output)); /* Remove invalid glyphs. */
2019-01-09 19:45:53 +01:00
hb_set_union (glyphs, output);
hb_set_clear (output);
}
private:
hb_map_t *done_lookups;
unsigned int lookup_count;
};
2019-10-23 01:00:43 +02:00
struct hb_closure_lookups_context_t :
hb_dispatch_context_t<hb_closure_lookups_context_t, hb_empty_t, 0>
{
const char *get_name () { return "CLOSURE_LOOKUPS"; }
typedef return_t (*recurse_func_t) (hb_closure_lookups_context_t *c, unsigned lookup_index);
template <typename T>
return_t dispatch (const T &obj) { obj.closure_lookups (this); return hb_empty_t (); }
static return_t default_return_value () { return hb_empty_t (); }
void recurse (unsigned lookup_index)
{
if (unlikely (nesting_level_left == 0 || !recurse_func))
return;
/* Return if new lookup was recursed to before. */
if (is_lookup_visited (lookup_index))
return;
set_lookup_visited (lookup_index);
nesting_level_left--;
recurse_func (this, lookup_index);
nesting_level_left++;
}
void set_lookup_visited (unsigned lookup_index)
{ visited_lookups->add (lookup_index); }
void set_lookup_inactive (unsigned lookup_index)
{ inactive_lookups->add (lookup_index); }
bool lookup_limit_exceeded ()
2020-03-26 07:45:09 +01:00
{ return lookup_count > HB_MAX_LOOKUP_INDICES; }
2019-10-23 01:00:43 +02:00
bool is_lookup_visited (unsigned lookup_index)
{
if (lookup_count++ > HB_MAX_LOOKUP_INDICES)
return true;
return visited_lookups->has (lookup_index);
}
2019-10-23 01:00:43 +02:00
hb_face_t *face;
const hb_set_t *glyphs;
recurse_func_t recurse_func;
unsigned int nesting_level_left;
unsigned int debug_depth;
hb_closure_lookups_context_t (hb_face_t *face_,
const hb_set_t *glyphs_,
hb_set_t *visited_lookups_,
hb_set_t *inactive_lookups_,
unsigned nesting_level_left_ = HB_MAX_NESTING_LEVEL) :
face (face_),
glyphs (glyphs_),
recurse_func (nullptr),
nesting_level_left (nesting_level_left_),
debug_depth (0),
visited_lookups (visited_lookups_),
inactive_lookups (inactive_lookups_),
lookup_count (0) {}
2019-10-23 01:00:43 +02:00
void set_recurse_func (recurse_func_t func) { recurse_func = func; }
private:
hb_set_t *visited_lookups;
hb_set_t *inactive_lookups;
unsigned int lookup_count;
2019-10-23 01:00:43 +02:00
};
2015-10-09 18:20:58 +02:00
struct hb_would_apply_context_t :
hb_dispatch_context_t<hb_would_apply_context_t, bool, 0>
{
const char *get_name () { return "WOULD_APPLY"; }
2012-11-22 22:47:53 +01:00
template <typename T>
return_t dispatch (const T &obj) { return obj.would_apply (this); }
static return_t default_return_value () { return false; }
bool stop_sublookup_iteration (return_t r) const { return r; }
2012-11-22 22:47:53 +01:00
hb_face_t *face;
const hb_codepoint_t *glyphs;
unsigned int len;
bool zero_context;
unsigned int debug_depth;
hb_would_apply_context_t (hb_face_t *face_,
const hb_codepoint_t *glyphs_,
unsigned int len_,
2012-09-04 21:15:19 +02:00
bool zero_context_) :
face (face_),
glyphs (glyphs_),
len (len_),
zero_context (zero_context_),
debug_depth (0) {}
};
2015-10-09 18:20:58 +02:00
struct hb_collect_glyphs_context_t :
2019-05-11 05:43:26 +02:00
hb_dispatch_context_t<hb_collect_glyphs_context_t, hb_empty_t, 0>
{
const char *get_name () { return "COLLECT_GLYPHS"; }
2012-11-24 00:13:48 +01:00
typedef return_t (*recurse_func_t) (hb_collect_glyphs_context_t *c, unsigned int lookup_index);
2012-11-22 22:47:53 +01:00
template <typename T>
2019-05-11 05:43:26 +02:00
return_t dispatch (const T &obj) { obj.collect_glyphs (this); return hb_empty_t (); }
static return_t default_return_value () { return hb_empty_t (); }
void recurse (unsigned int lookup_index)
2012-11-22 22:47:53 +01:00
{
2012-11-24 00:13:48 +01:00
if (unlikely (nesting_level_left == 0 || !recurse_func))
return;
2012-11-24 00:13:48 +01:00
2017-10-15 12:11:08 +02:00
/* Note that GPOS sets recurse_func to nullptr already, so it doesn't get
* past the previous check. For GSUB, we only want to collect the output
* glyphs in the recursion. If output is not requested, we can go home now.
*
* Note further, that the above is not exactly correct. A recursed lookup
* is allowed to match input that is not matched in the context, but that's
* not how most fonts are built. It's possible to relax that and recurse
* with all sets here if it proves to be an issue.
*/
if (output == hb_set_get_empty ())
return;
/* Return if new lookup was recursed to before. */
2017-10-22 23:48:06 +02:00
if (recursed_lookups->has (lookup_index))
return;
hb_set_t *old_before = before;
hb_set_t *old_input = input;
hb_set_t *old_after = after;
before = input = after = hb_set_get_empty ();
2012-11-24 00:13:48 +01:00
nesting_level_left--;
recurse_func (this, lookup_index);
2012-11-24 00:13:48 +01:00
nesting_level_left++;
before = old_before;
input = old_input;
after = old_after;
2017-10-22 23:48:06 +02:00
recursed_lookups->add (lookup_index);
2012-11-22 22:47:53 +01:00
}
hb_face_t *face;
hb_set_t *before;
hb_set_t *input;
hb_set_t *after;
hb_set_t *output;
2012-11-24 00:13:48 +01:00
recurse_func_t recurse_func;
2017-10-22 23:48:06 +02:00
hb_set_t *recursed_lookups;
2012-11-24 00:13:48 +01:00
unsigned int nesting_level_left;
unsigned int debug_depth;
hb_collect_glyphs_context_t (hb_face_t *face_,
2018-10-13 12:30:05 +02:00
hb_set_t *glyphs_before, /* OUT. May be NULL */
hb_set_t *glyphs_input, /* OUT. May be NULL */
hb_set_t *glyphs_after, /* OUT. May be NULL */
hb_set_t *glyphs_output, /* OUT. May be NULL */
unsigned int nesting_level_left_ = HB_MAX_NESTING_LEVEL) :
face (face_),
before (glyphs_before ? glyphs_before : hb_set_get_empty ()),
input (glyphs_input ? glyphs_input : hb_set_get_empty ()),
after (glyphs_after ? glyphs_after : hb_set_get_empty ()),
output (glyphs_output ? glyphs_output : hb_set_get_empty ()),
2017-10-15 12:11:08 +02:00
recurse_func (nullptr),
recursed_lookups (hb_set_create ()),
2012-11-24 00:13:48 +01:00
nesting_level_left (nesting_level_left_),
debug_depth (0) {}
~hb_collect_glyphs_context_t () { hb_set_destroy (recursed_lookups); }
2012-11-24 00:13:48 +01:00
void set_recurse_func (recurse_func_t func) { recurse_func = func; }
};
template <typename set_t>
struct hb_collect_coverage_context_t :
hb_dispatch_context_t<hb_collect_coverage_context_t<set_t>, const Coverage &, HB_DEBUG_GET_COVERAGE>
{
const char *get_name () { return "GET_COVERAGE"; }
typedef const Coverage &return_t;
template <typename T>
return_t dispatch (const T &obj) { return obj.get_coverage (); }
static return_t default_return_value () { return Null (Coverage); }
bool stop_sublookup_iteration (return_t r) const
{
r.collect_coverage (set);
return false;
}
2012-11-22 22:47:53 +01:00
hb_collect_coverage_context_t (set_t *set_) :
set (set_),
debug_depth (0) {}
set_t *set;
unsigned int debug_depth;
};
struct hb_ot_apply_context_t :
hb_dispatch_context_t<hb_ot_apply_context_t, bool, HB_DEBUG_APPLY>
{
struct matcher_t
2012-01-17 04:05:08 +01:00
{
matcher_t () :
lookup_props (0),
[Indic-like] Disable automatic joiner handling for basic shaping features Not for Arabic, but for Indic-like scripts. ZWJ/ZWNJ have special meanings in those scripts, so let font lookups take full control. This undoes the regression caused by automatic-joiners handling introduced two commits ago. We only disable automatic joiner handling for the "basic shaping features" of Indic, Myanmar, and SEAsian shapers. The "presentation forms" and other features are still applied with automatic-joiner handling. This change also changes the test suite failure statistics, such that a few scripts show more "failures". The most affected is Kannada. However, upon inspection, we believe that in most, if not all, of the new failures, we are producing results superior to Uniscribe. Hard to count those! Here's an example of what is fixed by the recent joiner-handling changes: https://bugs.freedesktop.org/show_bug.cgi?id=58714 New numbers, for future reference: BENGALI: 353892 out of 354188 tests passed. 296 failed (0.0835714%) DEVANAGARI: 707336 out of 707394 tests passed. 58 failed (0.00819911%) GUJARATI: 366262 out of 366457 tests passed. 195 failed (0.0532122%) GURMUKHI: 60706 out of 60747 tests passed. 41 failed (0.067493%) KANNADA: 950680 out of 951913 tests passed. 1233 failed (0.129529%) KHMER: 299074 out of 299124 tests passed. 50 failed (0.0167155%) LAO: 53611 out of 53644 tests passed. 33 failed (0.0615167%) MALAYALAM: 1047983 out of 1048334 tests passed. 351 failed (0.0334817%) ORIYA: 42320 out of 42329 tests passed. 9 failed (0.021262%) SINHALA: 271539 out of 271847 tests passed. 308 failed (0.113299%) TAMIL: 1091753 out of 1091754 tests passed. 1 failed (9.15957e-05%) TELUGU: 970555 out of 970573 tests passed. 18 failed (0.00185457%) TIBETAN: 208469 out of 208469 tests passed. 0 failed (0%)
2013-02-14 16:40:12 +01:00
ignore_zwnj (false),
ignore_zwj (false),
mask (-1),
#define arg1(arg) (arg) /* Remove the macro to see why it's needed! */
syllable arg1(0),
#undef arg1
2017-10-15 12:11:08 +02:00
match_func (nullptr),
match_data (nullptr) {}
typedef bool (*match_func_t) (hb_codepoint_t glyph_id, const HBUINT16 &value, const void *data);
void set_ignore_zwnj (bool ignore_zwnj_) { ignore_zwnj = ignore_zwnj_; }
void set_ignore_zwj (bool ignore_zwj_) { ignore_zwj = ignore_zwj_; }
void set_lookup_props (unsigned int lookup_props_) { lookup_props = lookup_props_; }
void set_mask (hb_mask_t mask_) { mask = mask_; }
void set_syllable (uint8_t syllable_) { syllable = syllable_; }
void set_match_func (match_func_t match_func_,
const void *match_data_)
{ match_func = match_func_; match_data = match_data_; }
enum may_match_t {
MATCH_NO,
MATCH_YES,
MATCH_MAYBE
};
may_match_t may_match (const hb_glyph_info_t &info,
2019-04-24 16:15:59 +02:00
const HBUINT16 *glyph_data) const
2012-01-17 04:05:08 +01:00
{
if (!(info.mask & mask) ||
(syllable && syllable != info.syllable ()))
return MATCH_NO;
if (match_func)
2018-11-15 20:40:56 +01:00
return match_func (info.codepoint, *glyph_data, match_data) ? MATCH_YES : MATCH_NO;
return MATCH_MAYBE;
2012-01-17 04:05:08 +01:00
}
enum may_skip_t {
SKIP_NO,
SKIP_YES,
SKIP_MAYBE
};
may_skip_t may_skip (const hb_ot_apply_context_t *c,
const hb_glyph_info_t &info) const
2013-02-13 17:22:42 +01:00
{
2014-07-16 19:44:01 +02:00
if (!c->check_glyph_property (&info, lookup_props))
return SKIP_YES;
if (unlikely (_hb_glyph_info_is_default_ignorable_and_not_hidden (&info) &&
(ignore_zwnj || !_hb_glyph_info_is_zwnj (&info)) &&
(ignore_zwj || !_hb_glyph_info_is_zwj (&info))))
return SKIP_MAYBE;
return SKIP_NO;
2013-02-13 17:22:42 +01:00
}
protected:
unsigned int lookup_props;
bool ignore_zwnj;
bool ignore_zwj;
hb_mask_t mask;
uint8_t syllable;
match_func_t match_func;
const void *match_data;
};
struct skipping_iterator_t
{
void init (hb_ot_apply_context_t *c_, bool context_match = false)
2012-01-17 04:05:08 +01:00
{
c = c_;
2017-10-15 12:11:08 +02:00
match_glyph_data = nullptr;
matcher.set_match_func (nullptr, nullptr);
matcher.set_lookup_props (c->lookup_props);
/* Ignore ZWNJ if we are matching GPOS, or matching GSUB context and asked to. */
matcher.set_ignore_zwnj (c->table_index == 1 || (context_match && c->auto_zwnj));
/* Ignore ZWJ if we are matching context, or asked to. */
matcher.set_ignore_zwj (context_match || c->auto_zwj);
matcher.set_mask (context_match ? -1 : c->lookup_mask);
}
void set_lookup_props (unsigned int lookup_props)
{
matcher.set_lookup_props (lookup_props);
2012-01-17 04:05:08 +01:00
}
void set_match_func (matcher_t::match_func_t match_func_,
const void *match_data_,
const HBUINT16 glyph_data[])
{
matcher.set_match_func (match_func_, match_data_);
match_glyph_data = glyph_data;
}
void reset (unsigned int start_index_,
2020-04-24 21:15:17 +02:00
unsigned int num_items_)
{
idx = start_index_;
num_items = num_items_;
end = c->buffer->len;
matcher.set_syllable (start_index_ == c->buffer->idx ? c->buffer->cur().syllable () : 0);
}
void reject ()
{
num_items++;
if (match_glyph_data) match_glyph_data--;
}
matcher_t::may_skip_t
may_skip (const hb_glyph_info_t &info) const
{ return matcher.may_skip (c, info); }
bool next ()
2012-01-17 04:05:08 +01:00
{
assert (num_items > 0);
while (idx + num_items < end)
2012-01-17 04:05:08 +01:00
{
2012-01-18 00:08:41 +01:00
idx++;
const hb_glyph_info_t &info = c->buffer->info[idx];
2013-02-21 20:51:40 +01:00
matcher_t::may_skip_t skip = matcher.may_skip (c, info);
if (unlikely (skip == matcher_t::SKIP_YES))
continue;
matcher_t::may_match_t match = matcher.may_match (info, match_glyph_data);
if (match == matcher_t::MATCH_YES ||
(match == matcher_t::MATCH_MAYBE &&
skip == matcher_t::SKIP_NO))
{
num_items--;
if (match_glyph_data) match_glyph_data++;
return true;
}
if (skip == matcher_t::SKIP_NO)
return false;
}
return false;
2012-01-17 04:05:08 +01:00
}
bool prev ()
2012-01-17 04:05:08 +01:00
{
assert (num_items > 0);
while (idx > num_items - 1)
2012-01-17 04:05:08 +01:00
{
idx--;
const hb_glyph_info_t &info = c->buffer->out_info[idx];
2013-02-21 20:51:40 +01:00
matcher_t::may_skip_t skip = matcher.may_skip (c, info);
if (unlikely (skip == matcher_t::SKIP_YES))
continue;
matcher_t::may_match_t match = matcher.may_match (info, match_glyph_data);
if (match == matcher_t::MATCH_YES ||
(match == matcher_t::MATCH_MAYBE &&
skip == matcher_t::SKIP_NO))
{
num_items--;
if (match_glyph_data) match_glyph_data++;
return true;
}
if (skip == matcher_t::SKIP_NO)
return false;
}
return false;
2012-01-17 04:05:08 +01:00
}
unsigned int idx;
protected:
hb_ot_apply_context_t *c;
matcher_t matcher;
const HBUINT16 *match_glyph_data;
2012-01-17 04:05:08 +01:00
unsigned int num_items;
unsigned int end;
2012-01-17 04:05:08 +01:00
};
2015-01-29 13:32:05 +01:00
const char *get_name () { return "APPLY"; }
typedef return_t (*recurse_func_t) (hb_ot_apply_context_t *c, unsigned int lookup_index);
2015-01-29 13:32:05 +01:00
template <typename T>
return_t dispatch (const T &obj) { return obj.apply (this); }
static return_t default_return_value () { return false; }
2015-01-29 13:32:05 +01:00
bool stop_sublookup_iteration (return_t r) const { return r; }
return_t recurse (unsigned int sub_lookup_index)
2015-01-29 13:32:05 +01:00
{
if (unlikely (nesting_level_left == 0 || !recurse_func || buffer->max_ops-- <= 0))
2015-01-29 13:32:05 +01:00
return default_return_value ();
nesting_level_left--;
bool ret = recurse_func (this, sub_lookup_index);
2015-01-29 13:32:05 +01:00
nesting_level_left++;
return ret;
}
skipping_iterator_t iter_input, iter_context;
2015-01-29 13:32:05 +01:00
hb_font_t *font;
hb_face_t *face;
hb_buffer_t *buffer;
recurse_func_t recurse_func;
const GDEF &gdef;
2016-09-10 09:22:24 +02:00
const VariationStore &var_store;
hb_direction_t direction;
hb_mask_t lookup_mask;
unsigned int table_index; /* GSUB/GPOS */
unsigned int lookup_index;
unsigned int lookup_props;
unsigned int nesting_level_left;
2015-01-29 13:32:05 +01:00
unsigned int debug_depth;
2018-09-10 16:24:52 +02:00
bool has_glyph_classes;
bool auto_zwnj;
bool auto_zwj;
2018-01-25 20:22:03 +01:00
bool random;
2018-09-10 16:24:52 +02:00
2018-09-11 10:57:48 +02:00
uint32_t random_state;
2015-01-29 13:32:05 +01:00
hb_ot_apply_context_t (unsigned int table_index_,
2015-01-29 13:32:05 +01:00
hb_font_t *font_,
hb_buffer_t *buffer_) :
iter_input (), iter_context (),
2015-01-29 13:32:05 +01:00
font (font_), face (font->face), buffer (buffer_),
2017-10-15 12:11:08 +02:00
recurse_func (nullptr),
gdef (
#ifndef HB_NO_OT_LAYOUT
*face->table.GDEF->table
#else
Null (GDEF)
#endif
),
var_store (gdef.get_var_store ()),
direction (buffer_->props.direction),
lookup_mask (1),
table_index (table_index_),
lookup_index ((unsigned int) -1),
lookup_props (0),
nesting_level_left (HB_MAX_NESTING_LEVEL),
debug_depth (0),
2018-09-10 16:24:52 +02:00
has_glyph_classes (gdef.has_glyph_classes ()),
auto_zwnj (true),
auto_zwj (true),
2018-01-25 20:22:03 +01:00
random (false),
2018-09-25 00:30:50 +02:00
random_state (1) { init_iters (); }
2015-01-29 13:32:05 +01:00
void init_iters ()
{
iter_input.init (this, false);
iter_context.init (this, true);
}
2015-01-29 13:32:05 +01:00
void set_lookup_mask (hb_mask_t mask) { lookup_mask = mask; init_iters (); }
void set_auto_zwj (bool auto_zwj_) { auto_zwj = auto_zwj_; init_iters (); }
void set_auto_zwnj (bool auto_zwnj_) { auto_zwnj = auto_zwnj_; init_iters (); }
void set_random (bool random_) { random = random_; }
void set_recurse_func (recurse_func_t func) { recurse_func = func; }
void set_lookup_index (unsigned int lookup_index_) { lookup_index = lookup_index_; }
void set_lookup_props (unsigned int lookup_props_) { lookup_props = lookup_props_; init_iters (); }
uint32_t random_number ()
2018-09-11 10:51:19 +02:00
{
2018-09-11 10:57:48 +02:00
/* http://www.cplusplus.com/reference/random/minstd_rand/ */
random_state = random_state * 48271 % 2147483647;
return random_state;
2018-09-11 10:51:19 +02:00
}
bool match_properties_mark (hb_codepoint_t glyph,
unsigned int glyph_props,
unsigned int match_props) const
2012-07-31 01:47:53 +02:00
{
/* If using mark filtering sets, the high short of
* match_props has the set index.
2012-07-31 01:47:53 +02:00
*/
if (match_props & LookupFlag::UseMarkFilteringSet)
return gdef.mark_set_covers (match_props >> 16, glyph);
2012-07-31 01:47:53 +02:00
/* The second byte of match_props has the meaning
2012-07-31 01:47:53 +02:00
* "ignore marks of attachment type different than
* the attachment type specified."
*/
if (match_props & LookupFlag::MarkAttachmentType)
return (match_props & LookupFlag::MarkAttachmentType) == (glyph_props & LookupFlag::MarkAttachmentType);
2012-07-31 01:47:53 +02:00
return true;
}
bool check_glyph_property (const hb_glyph_info_t *info,
unsigned int match_props) const
2012-07-31 01:47:53 +02:00
{
2014-07-16 19:44:01 +02:00
hb_codepoint_t glyph = info->codepoint;
unsigned int glyph_props = _hb_glyph_info_get_glyph_props (info);
2012-07-31 01:47:53 +02:00
/* Not covered, if, for example, glyph class is ligature and
* match_props includes LookupFlags::IgnoreLigatures
2012-07-31 01:47:53 +02:00
*/
if (glyph_props & match_props & LookupFlag::IgnoreFlags)
2012-07-31 01:47:53 +02:00
return false;
if (unlikely (glyph_props & HB_OT_LAYOUT_GLYPH_PROPS_MARK))
return match_properties_mark (glyph, glyph_props, match_props);
2012-07-31 01:47:53 +02:00
return true;
}
void _set_glyph_class (hb_codepoint_t glyph_index,
unsigned int class_guess = 0,
bool ligature = false,
bool component = false) const
{
unsigned int props = _hb_glyph_info_get_glyph_props (&buffer->cur());
props |= HB_OT_LAYOUT_GLYPH_PROPS_SUBSTITUTED;
if (ligature)
{
props |= HB_OT_LAYOUT_GLYPH_PROPS_LIGATED;
/* In the only place that the MULTIPLIED bit is used, Uniscribe
* seems to only care about the "last" transformation between
2018-10-19 17:49:21 +02:00
* Ligature and Multiple substitutions. Ie. if you ligate, expand,
* and ligate again, it forgives the multiplication and acts as
* if only ligation happened. As such, clear MULTIPLIED bit.
*/
props &= ~HB_OT_LAYOUT_GLYPH_PROPS_MULTIPLIED;
}
if (component)
props |= HB_OT_LAYOUT_GLYPH_PROPS_MULTIPLIED;
if (likely (has_glyph_classes))
props = (props & ~HB_OT_LAYOUT_GLYPH_PROPS_CLASS_MASK) | gdef.get_glyph_props (glyph_index);
else if (class_guess)
props = (props & ~HB_OT_LAYOUT_GLYPH_PROPS_CLASS_MASK) | class_guess;
_hb_glyph_info_set_glyph_props (&buffer->cur(), props);
}
2012-01-17 04:05:08 +01:00
void replace_glyph (hb_codepoint_t glyph_index) const
{
_set_glyph_class (glyph_index);
2013-10-18 00:06:30 +02:00
buffer->replace_glyph (glyph_index);
}
void replace_glyph_inplace (hb_codepoint_t glyph_index) const
{
_set_glyph_class (glyph_index);
2013-10-18 00:06:30 +02:00
buffer->cur().codepoint = glyph_index;
}
void replace_glyph_with_ligature (hb_codepoint_t glyph_index,
unsigned int class_guess) const
2013-10-18 00:06:30 +02:00
{
_set_glyph_class (glyph_index, class_guess, true);
buffer->replace_glyph (glyph_index);
}
void output_glyph_for_component (hb_codepoint_t glyph_index,
unsigned int class_guess) const
2012-07-31 00:36:42 +02:00
{
_set_glyph_class (glyph_index, class_guess, false, true);
2013-10-18 00:06:30 +02:00
buffer->output_glyph (glyph_index);
2012-07-31 00:36:42 +02:00
}
};
struct hb_get_subtables_context_t :
2019-05-11 05:43:26 +02:00
hb_dispatch_context_t<hb_get_subtables_context_t, hb_empty_t, HB_DEBUG_APPLY>
{
template <typename Type>
static inline bool apply_to (const void *obj, OT::hb_ot_apply_context_t *c)
{
const Type *typed_obj = (const Type *) obj;
return typed_obj->apply (c);
}
typedef bool (*hb_apply_func_t) (const void *obj, OT::hb_ot_apply_context_t *c);
struct hb_applicable_t
{
template <typename T>
void init (const T &obj_, hb_apply_func_t apply_func_)
{
obj = &obj_;
apply_func = apply_func_;
digest.init ();
obj_.get_coverage ().collect_coverage (&digest);
}
bool apply (OT::hb_ot_apply_context_t *c) const
{
return digest.may_have (c->buffer->cur().codepoint) && apply_func (obj, c);
}
private:
const void *obj;
hb_apply_func_t apply_func;
hb_set_digest_t digest;
};
typedef hb_vector_t<hb_applicable_t> array_t;
/* Dispatch interface. */
const char *get_name () { return "GET_SUBTABLES"; }
template <typename T>
return_t dispatch (const T &obj)
{
hb_applicable_t *entry = array.push();
entry->init (obj, apply_to<T>);
2019-05-11 05:43:26 +02:00
return hb_empty_t ();
}
2019-05-11 05:43:26 +02:00
static return_t default_return_value () { return hb_empty_t (); }
hb_get_subtables_context_t (array_t &array_) :
array (array_),
debug_depth (0) {}
array_t &array;
unsigned int debug_depth;
};
2010-05-05 07:13:09 +02:00
typedef bool (*intersects_func_t) (const hb_set_t *glyphs, const HBUINT16 &value, const void *data);
typedef void (*collect_glyphs_func_t) (hb_set_t *glyphs, const HBUINT16 &value, const void *data);
typedef bool (*match_func_t) (hb_codepoint_t glyph_id, const HBUINT16 &value, const void *data);
struct ContextClosureFuncs
{
intersects_func_t intersects;
};
struct ContextCollectGlyphsFuncs
{
collect_glyphs_func_t collect;
};
struct ContextApplyFuncs
2009-05-20 05:58:54 +02:00
{
match_func_t match;
};
static inline bool intersects_glyph (const hb_set_t *glyphs, const HBUINT16 &value, const void *data HB_UNUSED)
{
return glyphs->has (value);
}
static inline bool intersects_class (const hb_set_t *glyphs, const HBUINT16 &value, const void *data)
{
const ClassDef &class_def = *reinterpret_cast<const ClassDef *>(data);
return class_def.intersects_class (glyphs, value);
}
static inline bool intersects_coverage (const hb_set_t *glyphs, const HBUINT16 &value, const void *data)
{
const OffsetTo<Coverage> &coverage = (const OffsetTo<Coverage>&)value;
return (data+coverage).intersects (glyphs);
}
static inline bool array_is_subset_of (const hb_set_t *glyphs,
unsigned int count,
const HBUINT16 values[],
intersects_func_t intersects_func,
const void *intersects_data)
{
for (const HBUINT16 &_ : + hb_iter (values, count))
if (!intersects_func (glyphs, _, intersects_data)) return false;
return true;
}
static inline void collect_glyph (hb_set_t *glyphs, const HBUINT16 &value, const void *data HB_UNUSED)
{
glyphs->add (value);
}
static inline void collect_class (hb_set_t *glyphs, const HBUINT16 &value, const void *data)
{
const ClassDef &class_def = *reinterpret_cast<const ClassDef *>(data);
2020-04-23 19:57:30 +02:00
class_def.collect_class (glyphs, value);
}
static inline void collect_coverage (hb_set_t *glyphs, const HBUINT16 &value, const void *data)
{
const OffsetTo<Coverage> &coverage = (const OffsetTo<Coverage>&)value;
(data+coverage).collect_coverage (glyphs);
}
2012-12-06 00:46:04 +01:00
static inline void collect_array (hb_collect_glyphs_context_t *c HB_UNUSED,
hb_set_t *glyphs,
unsigned int count,
const HBUINT16 values[],
collect_glyphs_func_t collect_func,
const void *collect_data)
{
2019-03-30 06:27:46 +01:00
return
+ hb_iter (values, count)
| hb_apply ([&] (const HBUINT16 &_) { collect_func (glyphs, _, collect_data); })
;
}
static inline bool match_glyph (hb_codepoint_t glyph_id, const HBUINT16 &value, const void *data HB_UNUSED)
2009-05-20 05:58:54 +02:00
{
return glyph_id == value;
}
static inline bool match_class (hb_codepoint_t glyph_id, const HBUINT16 &value, const void *data)
2009-05-20 05:58:54 +02:00
{
2009-08-04 17:38:50 +02:00
const ClassDef &class_def = *reinterpret_cast<const ClassDef *>(data);
return class_def.get_class (glyph_id) == value;
}
static inline bool match_coverage (hb_codepoint_t glyph_id, const HBUINT16 &value, const void *data)
2009-05-20 05:58:54 +02:00
{
const OffsetTo<Coverage> &coverage = (const OffsetTo<Coverage>&)value;
return (data+coverage).get_coverage (glyph_id) != NOT_COVERED;
}
static inline bool would_match_input (hb_would_apply_context_t *c,
unsigned int count, /* Including the first glyph (not matched) */
const HBUINT16 input[], /* Array of input values--start with second glyph */
match_func_t match_func,
const void *match_data)
{
if (count != c->len)
return false;
for (unsigned int i = 1; i < count; i++)
if (likely (!match_func (c->glyphs[i], input[i - 1], match_data)))
return false;
return true;
}
static inline bool match_input (hb_ot_apply_context_t *c,
unsigned int count, /* Including the first glyph (not matched) */
const HBUINT16 input[], /* Array of input values--start with second glyph */
match_func_t match_func,
2010-05-10 23:47:22 +02:00
const void *match_data,
2013-10-17 13:55:48 +02:00
unsigned int *end_offset,
unsigned int match_positions[HB_MAX_CONTEXT_LENGTH],
2017-10-15 12:11:08 +02:00
unsigned int *p_total_component_count = nullptr)
{
2017-10-15 12:11:08 +02:00
TRACE_APPLY (nullptr);
if (unlikely (count > HB_MAX_CONTEXT_LENGTH)) return_trace (false);
2013-10-17 13:58:31 +02:00
hb_buffer_t *buffer = c->buffer;
hb_ot_apply_context_t::skipping_iterator_t &skippy_iter = c->iter_input;
skippy_iter.reset (buffer->idx, count - 1);
skippy_iter.set_match_func (match_func, match_data, input);
/*
* This is perhaps the trickiest part of OpenType... Remarks:
*
* - If all components of the ligature were marks, we call this a mark ligature.
*
* - If there is no GDEF, and the ligature is NOT a mark ligature, we categorize
* it as a ligature glyph.
*
* - Ligatures cannot be formed across glyphs attached to different components
* of previous ligatures. Eg. the sequence is LAM,SHADDA,LAM,FATHA,HEH, and
* LAM,LAM,HEH form a ligature, leaving SHADDA,FATHA next to eachother.
* However, it would be wrong to ligate that SHADDA,FATHA sequence.
* There are a couple of exceptions to this:
*
* o If a ligature tries ligating with marks that belong to it itself, go ahead,
* assuming that the font designer knows what they are doing (otherwise it can
* break Indic stuff when a matra wants to ligate with a conjunct,
*
* o If two marks want to ligate and they belong to different components of the
* same ligature glyph, and said ligature glyph is to be ignored according to
* mark-filtering rules, then allow.
* https://github.com/harfbuzz/harfbuzz/issues/545
*/
unsigned int total_component_count = 0;
2013-10-18 00:02:43 +02:00
total_component_count += _hb_glyph_info_get_lig_num_comps (&buffer->cur());
2013-10-18 00:02:43 +02:00
unsigned int first_lig_id = _hb_glyph_info_get_lig_id (&buffer->cur());
unsigned int first_lig_comp = _hb_glyph_info_get_lig_comp (&buffer->cur());
enum {
LIGBASE_NOT_CHECKED,
LIGBASE_MAY_NOT_SKIP,
LIGBASE_MAY_SKIP
} ligbase = LIGBASE_NOT_CHECKED;
2013-10-17 13:58:31 +02:00
match_positions[0] = buffer->idx;
2012-01-16 23:03:55 +01:00
for (unsigned int i = 1; i < count; i++)
2009-05-20 05:58:54 +02:00
{
if (!skippy_iter.next ()) return_trace (false);
2013-10-17 13:55:48 +02:00
match_positions[i] = skippy_iter.idx;
2013-10-18 00:02:43 +02:00
unsigned int this_lig_id = _hb_glyph_info_get_lig_id (&buffer->info[skippy_iter.idx]);
unsigned int this_lig_comp = _hb_glyph_info_get_lig_comp (&buffer->info[skippy_iter.idx]);
if (first_lig_id && first_lig_comp)
{
/* If first component was attached to a previous ligature component,
* all subsequent components should be attached to the same ligature
* component, otherwise we shouldn't ligate them... */
if (first_lig_id != this_lig_id || first_lig_comp != this_lig_comp)
{
2018-11-15 20:40:56 +01:00
/* ...unless, we are attached to a base ligature and that base
* ligature is ignorable. */
2018-11-15 20:40:56 +01:00
if (ligbase == LIGBASE_NOT_CHECKED)
{
bool found = false;
2019-05-11 01:35:31 +02:00
const auto *out = buffer->out_info;
unsigned int j = buffer->out_len;
while (j && _hb_glyph_info_get_lig_id (&out[j - 1]) == first_lig_id)
{
if (_hb_glyph_info_get_lig_comp (&out[j - 1]) == 0)
{
j--;
found = true;
break;
}
j--;
}
if (found && skippy_iter.may_skip (out[j]) == hb_ot_apply_context_t::matcher_t::SKIP_YES)
ligbase = LIGBASE_MAY_SKIP;
else
ligbase = LIGBASE_MAY_NOT_SKIP;
}
2018-11-15 20:40:56 +01:00
if (ligbase == LIGBASE_MAY_NOT_SKIP)
return_trace (false);
}
}
else
{
/* If first component was NOT attached to a previous ligature component,
* all subsequent components should also NOT be attached to any ligature
* component, unless they are attached to the first component itself! */
if (this_lig_id && this_lig_comp && (this_lig_id != first_lig_id))
return_trace (false);
}
2013-10-18 00:02:43 +02:00
total_component_count += _hb_glyph_info_get_lig_num_comps (&buffer->info[skippy_iter.idx]);
}
2013-10-17 13:58:31 +02:00
*end_offset = skippy_iter.idx - buffer->idx + 1;
if (p_total_component_count)
*p_total_component_count = total_component_count;
return_trace (true);
}
static inline bool ligate_input (hb_ot_apply_context_t *c,
unsigned int count, /* Including the first glyph */
2018-10-20 03:09:52 +02:00
const unsigned int match_positions[HB_MAX_CONTEXT_LENGTH], /* Including the first glyph */
unsigned int match_length,
2012-08-29 05:18:22 +02:00
hb_codepoint_t lig_glyph,
unsigned int total_component_count)
{
2017-10-15 12:11:08 +02:00
TRACE_APPLY (nullptr);
2013-10-17 13:58:31 +02:00
hb_buffer_t *buffer = c->buffer;
buffer->merge_clusters (buffer->idx, buffer->idx + match_length);
/* - If a base and one or more marks ligate, consider that as a base, NOT
* ligature, such that all following marks can still attach to it.
* https://github.com/harfbuzz/harfbuzz/issues/1109
*
* - If all components of the ligature were marks, we call this a mark ligature.
2018-10-02 15:02:16 +02:00
* If it *is* a mark ligature, we don't allocate a new ligature id, and leave
2012-08-29 05:18:22 +02:00
* the ligature to keep its old ligature id. This will allow it to attach to
* a base ligature in GPOS. Eg. if the sequence is: LAM,LAM,SHADDA,FATHA,HEH,
2018-10-02 15:02:16 +02:00
* and LAM,LAM,HEH for a ligature, they will leave SHADDA and FATHA with a
2012-08-29 05:18:22 +02:00
* ligature id and component value of 2. Then if SHADDA,FATHA form a ligature
* later, we don't want them to lose their ligature id/component, otherwise
* GPOS will fail to correctly position the mark ligature on top of the
* LAM,LAM,HEH ligature. See:
* https://bugzilla.gnome.org/show_bug.cgi?id=676343
*
* - If a ligature is formed of components that some of which are also ligatures
* themselves, and those ligature components had marks attached to *their*
* components, we have to attach the marks to the new ligature component
* positions! Now *that*'s tricky! And these marks may be following the
* last component of the whole sequence, so we should loop forward looking
* for them and update them.
*
* Eg. the sequence is LAM,LAM,SHADDA,FATHA,HEH, and the font first forms a
* 'calt' ligature of LAM,HEH, leaving the SHADDA and FATHA with a ligature
* id and component == 1. Now, during 'liga', the LAM and the LAM-HEH ligature
* form a LAM-LAM-HEH ligature. We need to reassign the SHADDA and FATHA to
* the new ligature with a component value of 2.
*
* This in fact happened to a font... See:
* https://bugzilla.gnome.org/show_bug.cgi?id=437633
*/
bool is_base_ligature = _hb_glyph_info_is_base_glyph (&buffer->info[match_positions[0]]);
bool is_mark_ligature = _hb_glyph_info_is_mark (&buffer->info[match_positions[0]]);
for (unsigned int i = 1; i < count; i++)
2018-10-02 15:02:16 +02:00
if (!_hb_glyph_info_is_mark (&buffer->info[match_positions[i]]))
{
is_base_ligature = false;
2018-10-02 15:02:16 +02:00
is_mark_ligature = false;
break;
}
bool is_ligature = !is_base_ligature && !is_mark_ligature;
2018-10-02 15:02:16 +02:00
unsigned int klass = is_ligature ? HB_OT_LAYOUT_GLYPH_PROPS_LIGATURE : 0;
unsigned int lig_id = is_ligature ? _hb_allocate_lig_id (buffer) : 0;
2013-10-18 00:02:43 +02:00
unsigned int last_lig_id = _hb_glyph_info_get_lig_id (&buffer->cur());
unsigned int last_num_components = _hb_glyph_info_get_lig_num_comps (&buffer->cur());
2012-08-29 05:18:22 +02:00
unsigned int components_so_far = last_num_components;
if (is_ligature)
{
2013-10-18 00:02:43 +02:00
_hb_glyph_info_set_lig_props_for_ligature (&buffer->cur(), lig_id, total_component_count);
2013-10-17 13:58:31 +02:00
if (_hb_glyph_info_get_general_category (&buffer->cur()) == HB_UNICODE_GENERAL_CATEGORY_NON_SPACING_MARK)
{
2015-11-03 02:44:05 +01:00
_hb_glyph_info_set_general_category (&buffer->cur(), HB_UNICODE_GENERAL_CATEGORY_OTHER_LETTER);
}
}
2013-10-18 00:06:30 +02:00
c->replace_glyph_with_ligature (lig_glyph, klass);
2012-08-29 05:18:22 +02:00
for (unsigned int i = 1; i < count; i++)
{
while (buffer->idx < match_positions[i] && buffer->successful)
2012-08-29 05:18:22 +02:00
{
if (is_ligature)
{
2018-11-15 20:40:56 +01:00
unsigned int this_comp = _hb_glyph_info_get_lig_comp (&buffer->cur());
if (this_comp == 0)
this_comp = last_num_components;
2012-08-29 05:18:22 +02:00
unsigned int new_lig_comp = components_so_far - last_num_components +
hb_min (this_comp, last_num_components);
_hb_glyph_info_set_lig_props_for_mark (&buffer->cur(), lig_id, new_lig_comp);
2012-08-29 05:18:22 +02:00
}
2013-10-17 13:58:31 +02:00
buffer->next_glyph ();
2012-08-29 05:18:22 +02:00
}
2013-10-18 00:02:43 +02:00
last_lig_id = _hb_glyph_info_get_lig_id (&buffer->cur());
last_num_components = _hb_glyph_info_get_lig_num_comps (&buffer->cur());
2012-08-29 05:18:22 +02:00
components_so_far += last_num_components;
/* Skip the base glyph */
2013-10-17 13:58:31 +02:00
buffer->idx++;
2012-08-29 05:18:22 +02:00
}
if (!is_mark_ligature && last_lig_id)
{
2012-08-29 05:18:22 +02:00
/* Re-adjust components for any marks following. */
for (unsigned i = buffer->idx; i < buffer->len; ++i)
{
if (last_lig_id != _hb_glyph_info_get_lig_id (&buffer->info[i])) break;
unsigned this_comp = _hb_glyph_info_get_lig_comp (&buffer->info[i]);
if (!this_comp) break;
unsigned new_lig_comp = components_so_far - last_num_components +
hb_min (this_comp, last_num_components);
_hb_glyph_info_set_lig_props_for_mark (&buffer->info[i], lig_id, new_lig_comp);
2012-08-29 05:18:22 +02:00
}
}
return_trace (true);
2012-08-29 05:18:22 +02:00
}
static inline bool match_backtrack (hb_ot_apply_context_t *c,
2009-05-18 09:47:31 +02:00
unsigned int count,
const HBUINT16 backtrack[],
2009-05-18 09:47:31 +02:00
match_func_t match_func,
const void *match_data,
unsigned int *match_start)
2009-05-18 09:47:31 +02:00
{
2017-10-15 12:11:08 +02:00
TRACE_APPLY (nullptr);
hb_ot_apply_context_t::skipping_iterator_t &skippy_iter = c->iter_context;
skippy_iter.reset (c->buffer->backtrack_len (), count);
skippy_iter.set_match_func (match_func, match_data, backtrack);
2009-05-18 09:47:31 +02:00
for (unsigned int i = 0; i < count; i++)
2012-01-17 04:05:08 +01:00
if (!skippy_iter.prev ())
return_trace (false);
2009-05-18 09:47:31 +02:00
*match_start = skippy_iter.idx;
return_trace (true);
2009-05-18 09:47:31 +02:00
}
static inline bool match_lookahead (hb_ot_apply_context_t *c,
unsigned int count,
const HBUINT16 lookahead[],
match_func_t match_func,
2010-05-10 23:47:22 +02:00
const void *match_data,
unsigned int offset,
unsigned int *end_index)
{
2017-10-15 12:11:08 +02:00
TRACE_APPLY (nullptr);
hb_ot_apply_context_t::skipping_iterator_t &skippy_iter = c->iter_context;
skippy_iter.reset (c->buffer->idx + offset - 1, count);
skippy_iter.set_match_func (match_func, match_data, lookahead);
2012-01-16 23:03:55 +01:00
for (unsigned int i = 0; i < count; i++)
2012-01-17 04:05:08 +01:00
if (!skippy_iter.next ())
return_trace (false);
*end_index = skippy_iter.idx + 1;
return_trace (true);
}
2010-07-23 21:11:18 +02:00
2009-05-20 05:58:54 +02:00
struct LookupRecord
{
2019-10-23 01:00:43 +02:00
LookupRecord* copy (hb_serialize_context_t *c,
const hb_map_t *lookup_map) const
2019-10-23 01:00:43 +02:00
{
TRACE_SERIALIZE (this);
auto *out = c->embed (*this);
if (unlikely (!out)) return_trace (nullptr);
out->lookupListIndex = hb_map_get (lookup_map, lookupListIndex);
return_trace (out);
}
bool sanitize (hb_sanitize_context_t *c) const
{
2012-11-23 21:32:14 +01:00
TRACE_SANITIZE (this);
return_trace (c->check_struct (this));
2009-08-04 08:09:34 +02:00
}
HBUINT16 sequenceIndex; /* Index into current glyph
* sequence--first glyph = 0 */
HBUINT16 lookupListIndex; /* Lookup to apply to that
* position--zero--based */
2010-05-10 22:38:32 +02:00
public:
DEFINE_SIZE_STATIC (4);
};
template <typename context_t>
static inline void recurse_lookups (context_t *c,
unsigned int lookupCount,
const LookupRecord lookupRecord[] /* Array of LookupRecords--in design order */)
{
for (unsigned int i = 0; i < lookupCount; i++)
2013-07-23 01:07:53 +02:00
c->recurse (lookupRecord[i].lookupListIndex);
}
2010-07-23 21:11:18 +02:00
static inline bool apply_lookup (hb_ot_apply_context_t *c,
2009-05-18 09:47:31 +02:00
unsigned int count, /* Including the first glyph */
unsigned int match_positions[HB_MAX_CONTEXT_LENGTH], /* Including the first glyph */
2009-05-18 09:47:31 +02:00
unsigned int lookupCount,
const LookupRecord lookupRecord[], /* Array of LookupRecords--in design order */
unsigned int match_length)
{
2017-10-15 12:11:08 +02:00
TRACE_APPLY (nullptr);
2012-11-23 21:06:59 +01:00
hb_buffer_t *buffer = c->buffer;
int end;
2009-09-21 19:58:56 +02:00
/* All positions are distance from beginning of *output* buffer.
* Adjust. */
{
unsigned int bl = buffer->backtrack_len ();
end = bl + match_length;
int delta = bl - buffer->idx;
/* Convert positions to new indexing. */
for (unsigned int j = 0; j < count; j++)
match_positions[j] += delta;
}
for (unsigned int i = 0; i < lookupCount && buffer->successful; i++)
2009-05-20 05:58:54 +02:00
{
unsigned int idx = lookupRecord[i].sequenceIndex;
if (idx >= count)
continue;
/* Don't recurse to ourself at same position.
* Note that this test is too naive, it doesn't catch longer loops. */
if (idx == 0 && lookupRecord[i].lookupListIndex == c->lookup_index)
continue;
if (unlikely (!buffer->move_to (match_positions[idx])))
break;
if (unlikely (buffer->max_ops <= 0))
break;
unsigned int orig_len = buffer->backtrack_len () + buffer->lookahead_len ();
if (!c->recurse (lookupRecord[i].lookupListIndex))
continue;
unsigned int new_len = buffer->backtrack_len () + buffer->lookahead_len ();
int delta = new_len - orig_len;
if (!delta)
2018-11-15 20:40:56 +01:00
continue;
2017-03-05 22:51:01 +01:00
/* Recursed lookup changed buffer len. Adjust.
*
* TODO:
*
* Right now, if buffer length increased by n, we assume n new glyphs
* were added right after the current position, and if buffer length
* was decreased by n, we assume n match positions after the current
* one where removed. The former (buffer length increased) case is
* fine, but the decrease case can be improved in at least two ways,
* both of which are significant:
*
* - If recursed-to lookup is MultipleSubst and buffer length
* decreased, then it's current match position that was deleted,
* NOT the one after it.
*
* - If buffer length was decreased by n, it does not necessarily
* mean that n match positions where removed, as there might
* have been marks and default-ignorables in the sequence. We
* should instead drop match positions between current-position
* and current-position + n instead.
*
* It should be possible to construct tests for both of these cases.
*/
end += delta;
if (end <= int (match_positions[idx]))
{
/* End might end up being smaller than match_positions[idx] if the recursed
* lookup ended up removing many items, more than we have had matched.
* Just never rewind end back and get out of here.
* https://bugs.chromium.org/p/chromium/issues/detail?id=659496 */
end = match_positions[idx];
/* There can't be any further changes. */
break;
}
unsigned int next = idx + 1; /* next now is the position after the recursed lookup. */
if (delta > 0)
{
if (unlikely (delta + count > HB_MAX_CONTEXT_LENGTH))
break;
}
else
{
/* NOTE: delta is negative. */
delta = hb_max (delta, (int) next - (int) count);
next -= delta;
}
/* Shift! */
memmove (match_positions + next + delta, match_positions + next,
(count - next) * sizeof (match_positions[0]));
next += delta;
count += delta;
/* Fill in new entries. */
for (unsigned int j = idx + 1; j < next; j++)
match_positions[j] = match_positions[j - 1] + 1;
/* And fixup the rest. */
for (; next < count; next++)
match_positions[next] += delta;
}
buffer->move_to (end);
return_trace (true);
}
2010-07-23 21:11:18 +02:00
/* Contextual lookups */
struct ContextClosureLookupContext
{
ContextClosureFuncs funcs;
const void *intersects_data;
};
struct ContextCollectGlyphsLookupContext
{
ContextCollectGlyphsFuncs funcs;
const void *collect_data;
};
struct ContextApplyLookupContext
2009-05-20 05:58:54 +02:00
{
ContextApplyFuncs funcs;
2010-05-10 23:47:22 +02:00
const void *match_data;
};
static inline bool context_intersects (const hb_set_t *glyphs,
unsigned int inputCount, /* Including the first glyph (not matched) */
const HBUINT16 input[], /* Array of input values--start with second glyph */
ContextClosureLookupContext &lookup_context)
{
return array_is_subset_of (glyphs,
inputCount ? inputCount - 1 : 0, input,
lookup_context.funcs.intersects, lookup_context.intersects_data);
}
2012-04-24 05:03:12 +02:00
static inline void context_closure_lookup (hb_closure_context_t *c,
unsigned int inputCount, /* Including the first glyph (not matched) */
const HBUINT16 input[], /* Array of input values--start with second glyph */
unsigned int lookupCount,
const LookupRecord lookupRecord[],
ContextClosureLookupContext &lookup_context)
{
if (context_intersects (c->glyphs,
inputCount, input,
lookup_context))
recurse_lookups (c,
lookupCount, lookupRecord);
}
static inline void context_collect_glyphs_lookup (hb_collect_glyphs_context_t *c,
unsigned int inputCount, /* Including the first glyph (not matched) */
const HBUINT16 input[], /* Array of input values--start with second glyph */
unsigned int lookupCount,
const LookupRecord lookupRecord[],
ContextCollectGlyphsLookupContext &lookup_context)
{
collect_array (c, c->input,
inputCount ? inputCount - 1 : 0, input,
lookup_context.funcs.collect, lookup_context.collect_data);
recurse_lookups (c,
lookupCount, lookupRecord);
}
static inline bool context_would_apply_lookup (hb_would_apply_context_t *c,
unsigned int inputCount, /* Including the first glyph (not matched) */
const HBUINT16 input[], /* Array of input values--start with second glyph */
2012-12-06 00:46:04 +01:00
unsigned int lookupCount HB_UNUSED,
const LookupRecord lookupRecord[] HB_UNUSED,
ContextApplyLookupContext &lookup_context)
{
return would_match_input (c,
inputCount, input,
lookup_context.funcs.match, lookup_context.match_data);
}
static inline bool context_apply_lookup (hb_ot_apply_context_t *c,
unsigned int inputCount, /* Including the first glyph (not matched) */
const HBUINT16 input[], /* Array of input values--start with second glyph */
unsigned int lookupCount,
const LookupRecord lookupRecord[],
ContextApplyLookupContext &lookup_context)
{
unsigned int match_length = 0;
unsigned int match_positions[HB_MAX_CONTEXT_LENGTH];
2010-05-13 20:18:49 +02:00
return match_input (c,
2010-05-05 07:37:58 +02:00
inputCount, input,
lookup_context.funcs.match, lookup_context.match_data,
&match_length, match_positions)
&& (c->buffer->unsafe_to_break (c->buffer->idx, c->buffer->idx + match_length),
apply_lookup (c,
inputCount, match_positions,
lookupCount, lookupRecord,
match_length));
}
2009-05-20 05:58:54 +02:00
struct Rule
{
bool intersects (const hb_set_t *glyphs, ContextClosureLookupContext &lookup_context) const
{
return context_intersects (glyphs,
inputCount, inputZ.arrayZ,
lookup_context);
}
void closure (hb_closure_context_t *c, ContextClosureLookupContext &lookup_context) const
{
if (unlikely (c->lookup_limit_exceeded ())) return;
const UnsizedArrayOf<LookupRecord> &lookupRecord = StructAfter<UnsizedArrayOf<LookupRecord>>
2018-11-02 17:23:26 +01:00
(inputZ.as_array ((inputCount ? inputCount - 1 : 0)));
2012-04-24 05:03:12 +02:00
context_closure_lookup (c,
inputCount, inputZ.arrayZ,
lookupCount, lookupRecord.arrayZ,
2012-04-24 05:03:12 +02:00
lookup_context);
}
2019-10-23 01:00:43 +02:00
void closure_lookups (hb_closure_lookups_context_t *c) const
{
if (unlikely (c->lookup_limit_exceeded ())) return;
2019-10-23 01:00:43 +02:00
const UnsizedArrayOf<LookupRecord> &lookupRecord = StructAfter<UnsizedArrayOf<LookupRecord>>
(inputZ.as_array (inputCount ? inputCount - 1 : 0));
recurse_lookups (c, lookupCount, lookupRecord.arrayZ);
}
void collect_glyphs (hb_collect_glyphs_context_t *c,
ContextCollectGlyphsLookupContext &lookup_context) const
{
const UnsizedArrayOf<LookupRecord> &lookupRecord = StructAfter<UnsizedArrayOf<LookupRecord>>
2018-11-02 17:23:26 +01:00
(inputZ.as_array (inputCount ? inputCount - 1 : 0));
context_collect_glyphs_lookup (c,
inputCount, inputZ.arrayZ,
lookupCount, lookupRecord.arrayZ,
lookup_context);
}
bool would_apply (hb_would_apply_context_t *c,
ContextApplyLookupContext &lookup_context) const
{
const UnsizedArrayOf<LookupRecord> &lookupRecord = StructAfter<UnsizedArrayOf<LookupRecord>>
2018-11-02 17:23:26 +01:00
(inputZ.as_array (inputCount ? inputCount - 1 : 0));
return context_would_apply_lookup (c,
inputCount, inputZ.arrayZ,
lookupCount, lookupRecord.arrayZ,
lookup_context);
}
bool apply (hb_ot_apply_context_t *c,
ContextApplyLookupContext &lookup_context) const
2009-05-20 05:58:54 +02:00
{
2012-11-23 21:32:14 +01:00
TRACE_APPLY (this);
const UnsizedArrayOf<LookupRecord> &lookupRecord = StructAfter<UnsizedArrayOf<LookupRecord>>
2018-11-02 17:23:26 +01:00
(inputZ.as_array (inputCount ? inputCount - 1 : 0));
return_trace (context_apply_lookup (c, inputCount, inputZ.arrayZ, lookupCount, lookupRecord.arrayZ, lookup_context));
}
bool serialize (hb_serialize_context_t *c,
const hb_map_t *input_mapping, /* old->new glyphid or class mapping */
const hb_map_t *lookup_map) const
{
TRACE_SERIALIZE (this);
auto *out = c->start_embed (this);
if (unlikely (!c->extend_min (out))) return_trace (false);
out->inputCount = inputCount;
out->lookupCount = lookupCount;
const hb_array_t<const HBUINT16> input = inputZ.as_array (inputCount - 1);
for (const auto org : input)
{
HBUINT16 d;
d = input_mapping->get (org);
c->copy (d);
}
const UnsizedArrayOf<LookupRecord> &lookupRecord = StructAfter<UnsizedArrayOf<LookupRecord>>
(inputZ.as_array ((inputCount ? inputCount - 1 : 0)));
for (unsigned i = 0; i < (unsigned) lookupCount; i++)
c->copy (lookupRecord[i], lookup_map);
return_trace (true);
}
bool subset (hb_subset_context_t *c,
const hb_map_t *lookup_map,
const hb_map_t *klass_map = nullptr) const
{
TRACE_SUBSET (this);
const hb_array_t<const HBUINT16> input = inputZ.as_array ((inputCount ? inputCount - 1 : 0));
if (!input.length) return_trace (false);
const hb_map_t *mapping = klass_map == nullptr ? c->plan->glyph_map : klass_map;
if (!hb_all (input, mapping)) return_trace (false);
return_trace (serialize (c->serializer, mapping, lookup_map));
}
2009-08-04 06:58:28 +02:00
public:
bool sanitize (hb_sanitize_context_t *c) const
{
2012-11-23 21:32:14 +01:00
TRACE_SANITIZE (this);
return_trace (inputCount.sanitize (c) &&
lookupCount.sanitize (c) &&
c->check_range (inputZ.arrayZ,
inputZ.item_size * (inputCount ? inputCount - 1 : 0) +
LookupRecord::static_size * lookupCount));
2009-08-04 06:58:28 +02:00
}
protected:
HBUINT16 inputCount; /* Total number of glyphs in input
* glyph sequence--includes the first
* glyph */
HBUINT16 lookupCount; /* Number of LookupRecords */
UnsizedArrayOf<HBUINT16>
2019-12-31 13:23:02 +01:00
inputZ; /* Array of match inputs--start with
* second glyph */
/*UnsizedArrayOf<LookupRecord>
lookupRecordX;*/ /* Array of LookupRecords--in
* design order */
2010-05-10 22:38:32 +02:00
public:
DEFINE_SIZE_ARRAY (4, inputZ);
};
2009-05-20 05:58:54 +02:00
struct RuleSet
{
bool intersects (const hb_set_t *glyphs,
ContextClosureLookupContext &lookup_context) const
{
2019-03-30 06:27:46 +01:00
return
+ hb_iter (rule)
| hb_map (hb_add (this))
| hb_map ([&] (const Rule &_) { return _.intersects (glyphs, lookup_context); })
2019-03-30 06:27:46 +01:00
| hb_any
;
}
void closure (hb_closure_context_t *c,
ContextClosureLookupContext &lookup_context) const
{
if (unlikely (c->lookup_limit_exceeded ())) return;
2019-03-30 06:27:46 +01:00
return
+ hb_iter (rule)
| hb_map (hb_add (this))
| hb_apply ([&] (const Rule &_) { _.closure (c, lookup_context); })
2019-03-30 06:27:46 +01:00
;
}
2019-10-23 01:00:43 +02:00
void closure_lookups (hb_closure_lookups_context_t *c) const
{
if (unlikely (c->lookup_limit_exceeded ())) return;
2019-10-23 01:00:43 +02:00
return
+ hb_iter (rule)
| hb_map (hb_add (this))
| hb_apply ([&] (const Rule &_) { _.closure_lookups (c); })
;
}
void collect_glyphs (hb_collect_glyphs_context_t *c,
ContextCollectGlyphsLookupContext &lookup_context) const
{
2019-03-30 06:27:46 +01:00
return
+ hb_iter (rule)
| hb_map (hb_add (this))
| hb_apply ([&] (const Rule &_) { _.collect_glyphs (c, lookup_context); })
2019-03-30 06:27:46 +01:00
;
}
bool would_apply (hb_would_apply_context_t *c,
ContextApplyLookupContext &lookup_context) const
{
2019-03-30 06:27:46 +01:00
return
+ hb_iter (rule)
| hb_map (hb_add (this))
| hb_map ([&] (const Rule &_) { return _.would_apply (c, lookup_context); })
2019-03-30 06:27:46 +01:00
| hb_any
;
}
bool apply (hb_ot_apply_context_t *c,
ContextApplyLookupContext &lookup_context) const
2009-05-20 05:58:54 +02:00
{
2012-11-23 21:32:14 +01:00
TRACE_APPLY (this);
2019-03-30 06:27:46 +01:00
return_trace (
+ hb_iter (rule)
| hb_map (hb_add (this))
| hb_map ([&] (const Rule &_) { return _.apply (c, lookup_context); })
2019-03-30 06:27:46 +01:00
| hb_any
)
;
}
bool subset (hb_subset_context_t *c,
const hb_map_t *lookup_map,
const hb_map_t *klass_map = nullptr) const
{
TRACE_SUBSET (this);
auto snap = c->serializer->snapshot ();
auto *out = c->serializer->start_embed (*this);
if (unlikely (!c->serializer->extend_min (out))) return_trace (false);
for (const OffsetTo<Rule>& _ : rule)
{
if (!_) continue;
auto *o = out->rule.serialize_append (c->serializer);
if (unlikely (!o)) continue;
auto o_snap = c->serializer->snapshot ();
if (!o->serialize_subset (c, _, this, lookup_map, klass_map))
{
out->rule.pop ();
c->serializer->revert (o_snap);
}
}
bool ret = bool (out->rule);
if (!ret) c->serializer->revert (snap);
return_trace (ret);
}
bool sanitize (hb_sanitize_context_t *c) const
{
2012-11-23 21:32:14 +01:00
TRACE_SANITIZE (this);
return_trace (rule.sanitize (c, this));
2009-08-04 06:58:28 +02:00
}
protected:
OffsetArrayOf<Rule>
rule; /* Array of Rule tables
* ordered by preference */
2010-05-11 00:08:46 +02:00
public:
DEFINE_SIZE_ARRAY (2, rule);
};
2009-05-20 05:58:54 +02:00
struct ContextFormat1
{
bool intersects (const hb_set_t *glyphs) const
{
struct ContextClosureLookupContext lookup_context = {
{intersects_glyph},
nullptr
};
2019-03-30 06:40:13 +01:00
return
+ hb_zip (this+coverage, ruleSet)
| hb_filter (*glyphs, hb_first)
| hb_map (hb_second)
| hb_map (hb_add (this))
| hb_map ([&] (const RuleSet &_) { return _.intersects (glyphs, lookup_context); })
2019-03-30 06:40:13 +01:00
| hb_any
;
}
void closure (hb_closure_context_t *c) const
{
struct ContextClosureLookupContext lookup_context = {
{intersects_glyph},
2017-10-15 12:11:08 +02:00
nullptr
};
2019-03-30 06:40:13 +01:00
+ hb_zip (this+coverage, ruleSet)
| hb_filter (*c->glyphs, hb_first)
| hb_map (hb_second)
| hb_map (hb_add (this))
| hb_apply ([&] (const RuleSet &_) { _.closure (c, lookup_context); })
2019-03-30 06:40:13 +01:00
;
}
2019-10-23 01:00:43 +02:00
void closure_lookups (hb_closure_lookups_context_t *c) const
{
+ hb_iter (ruleSet)
| hb_map (hb_add (this))
| hb_apply ([&] (const RuleSet &_) { _.closure_lookups (c); })
;
}
void collect_variation_indices (hb_collect_variation_indices_context_t *c) const {}
void collect_glyphs (hb_collect_glyphs_context_t *c) const
2012-11-24 00:13:48 +01:00
{
(this+coverage).collect_coverage (c->input);
struct ContextCollectGlyphsLookupContext lookup_context = {
{collect_glyph},
2017-10-15 12:11:08 +02:00
nullptr
};
2019-03-30 06:40:13 +01:00
+ hb_iter (ruleSet)
| hb_map (hb_add (this))
| hb_apply ([&] (const RuleSet &_) { _.collect_glyphs (c, lookup_context); })
2019-03-30 06:40:13 +01:00
;
2012-11-24 00:13:48 +01:00
}
bool would_apply (hb_would_apply_context_t *c) const
{
const RuleSet &rule_set = this+ruleSet[(this+coverage).get_coverage (c->glyphs[0])];
struct ContextApplyLookupContext lookup_context = {
{match_glyph},
2017-10-15 12:11:08 +02:00
nullptr
};
return rule_set.would_apply (c, lookup_context);
}
const Coverage &get_coverage () const { return this+coverage; }
bool apply (hb_ot_apply_context_t *c) const
2009-05-20 05:58:54 +02:00
{
2012-11-23 21:32:14 +01:00
TRACE_APPLY (this);
unsigned int index = (this+coverage).get_coverage (c->buffer->cur().codepoint);
if (likely (index == NOT_COVERED))
return_trace (false);
const RuleSet &rule_set = this+ruleSet[index];
struct ContextApplyLookupContext lookup_context = {
{match_glyph},
2017-10-15 12:11:08 +02:00
nullptr
};
return_trace (rule_set.apply (c, lookup_context));
}
bool subset (hb_subset_context_t *c) const
{
TRACE_SUBSET (this);
const hb_set_t &glyphset = *c->plan->glyphset ();
const hb_map_t &glyph_map = *c->plan->glyph_map;
auto *out = c->serializer->start_embed (*this);
if (unlikely (!c->serializer->extend_min (out))) return_trace (false);
out->format = format;
const hb_map_t *lookup_map = c->table_tag == HB_OT_TAG_GSUB ? c->plan->gsub_lookups : c->plan->gpos_lookups;
hb_sorted_vector_t<hb_codepoint_t> new_coverage;
+ hb_zip (this+coverage, ruleSet)
| hb_filter (glyphset, hb_first)
| hb_filter (subset_offset_array (c, out->ruleSet, this, lookup_map), hb_second)
| hb_map (hb_first)
| hb_map (glyph_map)
| hb_sink (new_coverage)
;
out->coverage.serialize (c->serializer, out)
.serialize (c->serializer, new_coverage.iter ());
return_trace (bool (new_coverage));
}
bool sanitize (hb_sanitize_context_t *c) const
{
2012-11-23 21:32:14 +01:00
TRACE_SANITIZE (this);
return_trace (coverage.sanitize (c, this) && ruleSet.sanitize (c, this));
2009-08-04 06:58:28 +02:00
}
protected:
HBUINT16 format; /* Format identifier--format = 1 */
OffsetTo<Coverage>
coverage; /* Offset to Coverage table--from
* beginning of table */
OffsetArrayOf<RuleSet>
ruleSet; /* Array of RuleSet tables
* ordered by Coverage Index */
public:
DEFINE_SIZE_ARRAY (6, ruleSet);
};
2009-05-20 05:58:54 +02:00
struct ContextFormat2
{
bool intersects (const hb_set_t *glyphs) const
{
if (!(this+coverage).intersects (glyphs))
return false;
const ClassDef &class_def = this+classDef;
struct ContextClosureLookupContext lookup_context = {
{intersects_class},
&class_def
};
2019-03-30 06:48:38 +01:00
return
+ hb_enumerate (ruleSet)
2019-05-16 03:15:05 +02:00
| hb_map ([&] (const hb_pair_t<unsigned, const OffsetTo<RuleSet> &> p)
{ return class_def.intersects_class (glyphs, p.first) &&
(this+p.second).intersects (glyphs, lookup_context); })
2019-03-30 06:48:38 +01:00
| hb_any
;
}
void closure (hb_closure_context_t *c) const
{
if (!(this+coverage).intersects (c->glyphs))
2012-04-24 05:03:12 +02:00
return;
const ClassDef &class_def = this+classDef;
struct ContextClosureLookupContext lookup_context = {
{intersects_class},
&class_def
};
return
+ hb_enumerate (ruleSet)
2019-05-16 03:15:05 +02:00
| hb_filter ([&] (unsigned _)
{ return class_def.intersects_class (c->glyphs, _); },
hb_first)
| hb_map (hb_second)
| hb_map (hb_add (this))
| hb_apply ([&] (const RuleSet &_) { _.closure (c, lookup_context); })
;
}
2019-10-23 01:00:43 +02:00
void closure_lookups (hb_closure_lookups_context_t *c) const
{
+ hb_iter (ruleSet)
| hb_map (hb_add (this))
| hb_apply ([&] (const RuleSet &_) { _.closure_lookups (c); })
;
}
void collect_variation_indices (hb_collect_variation_indices_context_t *c) const {}
void collect_glyphs (hb_collect_glyphs_context_t *c) const
2012-11-24 00:13:48 +01:00
{
(this+coverage).collect_coverage (c->input);
const ClassDef &class_def = this+classDef;
struct ContextCollectGlyphsLookupContext lookup_context = {
{collect_class},
&class_def
};
2019-03-30 06:40:13 +01:00
+ hb_iter (ruleSet)
| hb_map (hb_add (this))
| hb_apply ([&] (const RuleSet &_) { _.collect_glyphs (c, lookup_context); })
2019-03-30 06:40:13 +01:00
;
2012-11-24 00:13:48 +01:00
}
bool would_apply (hb_would_apply_context_t *c) const
{
const ClassDef &class_def = this+classDef;
unsigned int index = class_def.get_class (c->glyphs[0]);
const RuleSet &rule_set = this+ruleSet[index];
struct ContextApplyLookupContext lookup_context = {
{match_class},
&class_def
};
return rule_set.would_apply (c, lookup_context);
}
const Coverage &get_coverage () const { return this+coverage; }
bool apply (hb_ot_apply_context_t *c) const
2009-05-20 05:58:54 +02:00
{
2012-11-23 21:32:14 +01:00
TRACE_APPLY (this);
unsigned int index = (this+coverage).get_coverage (c->buffer->cur().codepoint);
if (likely (index == NOT_COVERED)) return_trace (false);
const ClassDef &class_def = this+classDef;
index = class_def.get_class (c->buffer->cur().codepoint);
const RuleSet &rule_set = this+ruleSet[index];
struct ContextApplyLookupContext lookup_context = {
{match_class},
2010-05-10 23:47:22 +02:00
&class_def
};
return_trace (rule_set.apply (c, lookup_context));
}
bool subset (hb_subset_context_t *c) const
{
TRACE_SUBSET (this);
auto *out = c->serializer->start_embed (*this);
if (unlikely (!c->serializer->extend_min (out))) return_trace (false);
out->format = format;
if (unlikely (!out->coverage.serialize_subset (c, coverage, this)))
return_trace (false);
hb_map_t klass_map;
out->classDef.serialize_subset (c, classDef, this, &klass_map);
const hb_map_t *lookup_map = c->table_tag == HB_OT_TAG_GSUB ? c->plan->gsub_lookups : c->plan->gpos_lookups;
bool ret = true;
unsigned non_zero_index = 0, index = 0;
for (const hb_pair_t<unsigned, const OffsetTo<RuleSet>&> _ : + hb_enumerate (ruleSet)
| hb_filter (klass_map, hb_first))
{
auto *o = out->ruleSet.serialize_append (c->serializer);
if (unlikely (!o))
{
ret = false;
break;
}
if (o->serialize_subset (c, _.second, this, lookup_map, &klass_map))
non_zero_index = index;
index++;
}
if (!ret) return_trace (ret);
//prune empty trailing ruleSets
--index;
while (index > non_zero_index)
{
out->ruleSet.pop ();
index--;
}
return_trace (bool (out->ruleSet));
}
bool sanitize (hb_sanitize_context_t *c) const
{
2012-11-23 21:32:14 +01:00
TRACE_SANITIZE (this);
return_trace (coverage.sanitize (c, this) && classDef.sanitize (c, this) && ruleSet.sanitize (c, this));
2009-08-04 06:58:28 +02:00
}
protected:
HBUINT16 format; /* Format identifier--format = 2 */
OffsetTo<Coverage>
coverage; /* Offset to Coverage table--from
* beginning of table */
OffsetTo<ClassDef>
classDef; /* Offset to glyph ClassDef table--from
* beginning of table */
OffsetArrayOf<RuleSet>
ruleSet; /* Array of RuleSet tables
* ordered by class */
public:
DEFINE_SIZE_ARRAY (8, ruleSet);
};
2009-05-20 05:58:54 +02:00
struct ContextFormat3
{
bool intersects (const hb_set_t *glyphs) const
{
if (!(this+coverageZ[0]).intersects (glyphs))
return false;
struct ContextClosureLookupContext lookup_context = {
{intersects_coverage},
this
};
return context_intersects (glyphs,
glyphCount, (const HBUINT16 *) (coverageZ.arrayZ + 1),
lookup_context);
}
void closure (hb_closure_context_t *c) const
{
if (!(this+coverageZ[0]).intersects (c->glyphs))
2012-04-24 05:03:12 +02:00
return;
2018-11-02 17:23:26 +01:00
const LookupRecord *lookupRecord = &StructAfter<LookupRecord> (coverageZ.as_array (glyphCount));
struct ContextClosureLookupContext lookup_context = {
{intersects_coverage},
this
};
2012-04-24 05:03:12 +02:00
context_closure_lookup (c,
glyphCount, (const HBUINT16 *) (coverageZ.arrayZ + 1),
2012-04-24 05:03:12 +02:00
lookupCount, lookupRecord,
lookup_context);
}
2019-10-23 01:00:43 +02:00
void closure_lookups (hb_closure_lookups_context_t *c) const
{
const LookupRecord *lookupRecord = &StructAfter<LookupRecord> (coverageZ.as_array (glyphCount));
recurse_lookups (c, lookupCount, lookupRecord);
}
void collect_variation_indices (hb_collect_variation_indices_context_t *c) const {}
void collect_glyphs (hb_collect_glyphs_context_t *c) const
2012-11-24 00:13:48 +01:00
{
(this+coverageZ[0]).collect_coverage (c->input);
2018-11-02 17:23:26 +01:00
const LookupRecord *lookupRecord = &StructAfter<LookupRecord> (coverageZ.as_array (glyphCount));
struct ContextCollectGlyphsLookupContext lookup_context = {
{collect_coverage},
this
};
context_collect_glyphs_lookup (c,
glyphCount, (const HBUINT16 *) (coverageZ.arrayZ + 1),
lookupCount, lookupRecord,
lookup_context);
2012-11-24 00:13:48 +01:00
}
bool would_apply (hb_would_apply_context_t *c) const
{
2018-11-02 17:23:26 +01:00
const LookupRecord *lookupRecord = &StructAfter<LookupRecord> (coverageZ.as_array (glyphCount));
struct ContextApplyLookupContext lookup_context = {
{match_coverage},
this
};
return context_would_apply_lookup (c,
glyphCount, (const HBUINT16 *) (coverageZ.arrayZ + 1),
lookupCount, lookupRecord,
lookup_context);
}
const Coverage &get_coverage () const { return this+coverageZ[0]; }
bool apply (hb_ot_apply_context_t *c) const
2009-05-20 05:58:54 +02:00
{
2012-11-23 21:32:14 +01:00
TRACE_APPLY (this);
unsigned int index = (this+coverageZ[0]).get_coverage (c->buffer->cur().codepoint);
if (likely (index == NOT_COVERED)) return_trace (false);
2018-11-02 17:23:26 +01:00
const LookupRecord *lookupRecord = &StructAfter<LookupRecord> (coverageZ.as_array (glyphCount));
struct ContextApplyLookupContext lookup_context = {
{match_coverage},
2010-05-10 23:47:22 +02:00
this
};
return_trace (context_apply_lookup (c, glyphCount, (const HBUINT16 *) (coverageZ.arrayZ + 1), lookupCount, lookupRecord, lookup_context));
}
bool subset (hb_subset_context_t *c) const
{
TRACE_SUBSET (this);
auto *out = c->serializer->start_embed (this);
if (unlikely (!c->serializer->extend_min (out))) return_trace (false);
out->format = format;
out->glyphCount = glyphCount;
out->lookupCount = lookupCount;
const hb_array_t<const OffsetTo<Coverage>> coverages = coverageZ.as_array (glyphCount);
for (const OffsetTo<Coverage>& offset : coverages)
{
auto *o = c->serializer->allocate_size<OffsetTo<Coverage>> (OffsetTo<Coverage>::static_size);
if (unlikely (!o)) return_trace (false);
if (!o->serialize_subset (c, offset, this)) return_trace (false);
}
const LookupRecord *lookupRecord = &StructAfter<LookupRecord> (coverageZ.as_array (glyphCount));
const hb_map_t *lookup_map = c->table_tag == HB_OT_TAG_GSUB ? c->plan->gsub_lookups : c->plan->gpos_lookups;
for (unsigned i = 0; i < (unsigned) lookupCount; i++)
c->serializer->copy (lookupRecord[i], lookup_map);
return_trace (true);
}
bool sanitize (hb_sanitize_context_t *c) const
{
2012-11-23 21:32:14 +01:00
TRACE_SANITIZE (this);
if (!c->check_struct (this)) return_trace (false);
2009-08-04 06:58:28 +02:00
unsigned int count = glyphCount;
if (!count) return_trace (false); /* We want to access coverageZ[0] freely. */
if (!c->check_array (coverageZ.arrayZ, count)) return_trace (false);
2009-08-04 06:58:28 +02:00
for (unsigned int i = 0; i < count; i++)
if (!coverageZ[i].sanitize (c, this)) return_trace (false);
2018-11-02 17:23:26 +01:00
const LookupRecord *lookupRecord = &StructAfter<LookupRecord> (coverageZ.as_array (glyphCount));
return_trace (c->check_array (lookupRecord, lookupCount));
2009-08-04 06:58:28 +02:00
}
protected:
HBUINT16 format; /* Format identifier--format = 3 */
HBUINT16 glyphCount; /* Number of glyphs in the input glyph
* sequence */
HBUINT16 lookupCount; /* Number of LookupRecords */
UnsizedArrayOf<OffsetTo<Coverage>>
coverageZ; /* Array of offsets to Coverage
* table in glyph sequence order */
/*UnsizedArrayOf<LookupRecord>
lookupRecordX;*/ /* Array of LookupRecords--in
* design order */
2010-05-10 22:38:32 +02:00
public:
DEFINE_SIZE_ARRAY (6, coverageZ);
};
2009-05-20 05:58:54 +02:00
struct Context
{
template <typename context_t, typename ...Ts>
2019-05-08 05:58:43 +02:00
typename context_t::return_t dispatch (context_t *c, Ts&&... ds) const
{
2014-12-13 05:36:49 +01:00
TRACE_DISPATCH (this, u.format);
if (unlikely (!c->may_dispatch (this, &u.format))) return_trace (c->no_dispatch_return_value ());
switch (u.format) {
case 1: return_trace (c->dispatch (u.format1, hb_forward<Ts> (ds)...));
case 2: return_trace (c->dispatch (u.format2, hb_forward<Ts> (ds)...));
case 3: return_trace (c->dispatch (u.format3, hb_forward<Ts> (ds)...));
default:return_trace (c->default_return_value ());
}
}
protected:
union {
HBUINT16 format; /* Format identifier */
2010-05-11 01:45:41 +02:00
ContextFormat1 format1;
ContextFormat2 format2;
ContextFormat3 format3;
} u;
};
/* Chaining Contextual lookups */
struct ChainContextClosureLookupContext
2009-05-20 05:58:54 +02:00
{
ContextClosureFuncs funcs;
const void *intersects_data[3];
};
struct ChainContextCollectGlyphsLookupContext
{
ContextCollectGlyphsFuncs funcs;
const void *collect_data[3];
};
struct ChainContextApplyLookupContext
{
ContextApplyFuncs funcs;
2010-05-10 23:47:22 +02:00
const void *match_data[3];
};
static inline bool chain_context_intersects (const hb_set_t *glyphs,
unsigned int backtrackCount,
const HBUINT16 backtrack[],
unsigned int inputCount, /* Including the first glyph (not matched) */
const HBUINT16 input[], /* Array of input values--start with second glyph */
unsigned int lookaheadCount,
const HBUINT16 lookahead[],
ChainContextClosureLookupContext &lookup_context)
{
return array_is_subset_of (glyphs,
backtrackCount, backtrack,
lookup_context.funcs.intersects, lookup_context.intersects_data[0])
&& array_is_subset_of (glyphs,
inputCount ? inputCount - 1 : 0, input,
lookup_context.funcs.intersects, lookup_context.intersects_data[1])
&& array_is_subset_of (glyphs,
lookaheadCount, lookahead,
lookup_context.funcs.intersects, lookup_context.intersects_data[2]);
}
2012-04-24 05:03:12 +02:00
static inline void chain_context_closure_lookup (hb_closure_context_t *c,
unsigned int backtrackCount,
const HBUINT16 backtrack[],
unsigned int inputCount, /* Including the first glyph (not matched) */
const HBUINT16 input[], /* Array of input values--start with second glyph */
unsigned int lookaheadCount,
const HBUINT16 lookahead[],
unsigned int lookupCount,
const LookupRecord lookupRecord[],
ChainContextClosureLookupContext &lookup_context)
{
if (chain_context_intersects (c->glyphs,
backtrackCount, backtrack,
inputCount, input,
lookaheadCount, lookahead,
lookup_context))
recurse_lookups (c,
lookupCount, lookupRecord);
}
static inline void chain_context_collect_glyphs_lookup (hb_collect_glyphs_context_t *c,
2018-11-15 20:40:56 +01:00
unsigned int backtrackCount,
const HBUINT16 backtrack[],
unsigned int inputCount, /* Including the first glyph (not matched) */
const HBUINT16 input[], /* Array of input values--start with second glyph */
unsigned int lookaheadCount,
const HBUINT16 lookahead[],
unsigned int lookupCount,
const LookupRecord lookupRecord[],
ChainContextCollectGlyphsLookupContext &lookup_context)
{
collect_array (c, c->before,
backtrackCount, backtrack,
lookup_context.funcs.collect, lookup_context.collect_data[0]);
collect_array (c, c->input,
inputCount ? inputCount - 1 : 0, input,
lookup_context.funcs.collect, lookup_context.collect_data[1]);
collect_array (c, c->after,
lookaheadCount, lookahead,
lookup_context.funcs.collect, lookup_context.collect_data[2]);
recurse_lookups (c,
lookupCount, lookupRecord);
}
static inline bool chain_context_would_apply_lookup (hb_would_apply_context_t *c,
unsigned int backtrackCount,
const HBUINT16 backtrack[] HB_UNUSED,
unsigned int inputCount, /* Including the first glyph (not matched) */
const HBUINT16 input[], /* Array of input values--start with second glyph */
unsigned int lookaheadCount,
const HBUINT16 lookahead[] HB_UNUSED,
2012-12-06 00:46:04 +01:00
unsigned int lookupCount HB_UNUSED,
const LookupRecord lookupRecord[] HB_UNUSED,
ChainContextApplyLookupContext &lookup_context)
{
return (c->zero_context ? !backtrackCount && !lookaheadCount : true)
&& would_match_input (c,
inputCount, input,
lookup_context.funcs.match, lookup_context.match_data[1]);
}
static inline bool chain_context_apply_lookup (hb_ot_apply_context_t *c,
unsigned int backtrackCount,
const HBUINT16 backtrack[],
unsigned int inputCount, /* Including the first glyph (not matched) */
const HBUINT16 input[], /* Array of input values--start with second glyph */
unsigned int lookaheadCount,
const HBUINT16 lookahead[],
unsigned int lookupCount,
const LookupRecord lookupRecord[],
ChainContextApplyLookupContext &lookup_context)
{
unsigned int start_index = 0, match_length = 0, end_index = 0;
unsigned int match_positions[HB_MAX_CONTEXT_LENGTH];
return match_input (c,
2010-05-05 07:37:58 +02:00
inputCount, input,
lookup_context.funcs.match, lookup_context.match_data[1],
&match_length, match_positions)
&& match_backtrack (c,
backtrackCount, backtrack,
lookup_context.funcs.match, lookup_context.match_data[0],
&start_index)
2010-05-13 20:18:49 +02:00
&& match_lookahead (c,
2010-05-05 07:37:58 +02:00
lookaheadCount, lookahead,
lookup_context.funcs.match, lookup_context.match_data[2],
match_length, &end_index)
&& (c->buffer->unsafe_to_break_from_outbuffer (start_index, end_index),
2018-11-15 20:40:56 +01:00
apply_lookup (c,
inputCount, match_positions,
lookupCount, lookupRecord,
match_length));
}
2009-05-20 05:58:54 +02:00
struct ChainRule
{
bool intersects (const hb_set_t *glyphs, ChainContextClosureLookupContext &lookup_context) const
{
const HeadlessArrayOf<HBUINT16> &input = StructAfter<HeadlessArrayOf<HBUINT16>> (backtrack);
const ArrayOf<HBUINT16> &lookahead = StructAfter<ArrayOf<HBUINT16>> (input);
return chain_context_intersects (glyphs,
backtrack.len, backtrack.arrayZ,
input.lenP1, input.arrayZ,
lookahead.len, lookahead.arrayZ,
lookup_context);
}
void closure (hb_closure_context_t *c,
ChainContextClosureLookupContext &lookup_context) const
{
if (unlikely (c->lookup_limit_exceeded ())) return;
const HeadlessArrayOf<HBUINT16> &input = StructAfter<HeadlessArrayOf<HBUINT16>> (backtrack);
const ArrayOf<HBUINT16> &lookahead = StructAfter<ArrayOf<HBUINT16>> (input);
const ArrayOf<LookupRecord> &lookup = StructAfter<ArrayOf<LookupRecord>> (lookahead);
2012-04-24 05:03:12 +02:00
chain_context_closure_lookup (c,
2018-05-09 01:56:11 +02:00
backtrack.len, backtrack.arrayZ,
input.lenP1, input.arrayZ,
2018-05-09 01:56:11 +02:00
lookahead.len, lookahead.arrayZ,
lookup.len, lookup.arrayZ,
2012-04-24 05:03:12 +02:00
lookup_context);
}
2019-10-23 01:00:43 +02:00
void closure_lookups (hb_closure_lookups_context_t *c) const
{
if (unlikely (c->lookup_limit_exceeded ())) return;
2019-10-23 01:00:43 +02:00
const HeadlessArrayOf<HBUINT16> &input = StructAfter<HeadlessArrayOf<HBUINT16>> (backtrack);
const ArrayOf<HBUINT16> &lookahead = StructAfter<ArrayOf<HBUINT16>> (input);
const ArrayOf<LookupRecord> &lookup = StructAfter<ArrayOf<LookupRecord>> (lookahead);
recurse_lookups (c, lookup.len, lookup.arrayZ);
}
void collect_glyphs (hb_collect_glyphs_context_t *c,
ChainContextCollectGlyphsLookupContext &lookup_context) const
{
const HeadlessArrayOf<HBUINT16> &input = StructAfter<HeadlessArrayOf<HBUINT16>> (backtrack);
const ArrayOf<HBUINT16> &lookahead = StructAfter<ArrayOf<HBUINT16>> (input);
const ArrayOf<LookupRecord> &lookup = StructAfter<ArrayOf<LookupRecord>> (lookahead);
chain_context_collect_glyphs_lookup (c,
2018-05-09 01:56:11 +02:00
backtrack.len, backtrack.arrayZ,
input.lenP1, input.arrayZ,
2018-05-09 01:56:11 +02:00
lookahead.len, lookahead.arrayZ,
lookup.len, lookup.arrayZ,
lookup_context);
}
bool would_apply (hb_would_apply_context_t *c,
ChainContextApplyLookupContext &lookup_context) const
{
const HeadlessArrayOf<HBUINT16> &input = StructAfter<HeadlessArrayOf<HBUINT16>> (backtrack);
const ArrayOf<HBUINT16> &lookahead = StructAfter<ArrayOf<HBUINT16>> (input);
const ArrayOf<LookupRecord> &lookup = StructAfter<ArrayOf<LookupRecord>> (lookahead);
return chain_context_would_apply_lookup (c,
backtrack.len, backtrack.arrayZ,
input.lenP1, input.arrayZ,
lookahead.len, lookahead.arrayZ, lookup.len,
lookup.arrayZ, lookup_context);
}
bool apply (hb_ot_apply_context_t *c, ChainContextApplyLookupContext &lookup_context) const
2009-05-20 05:58:54 +02:00
{
2012-11-23 21:32:14 +01:00
TRACE_APPLY (this);
const HeadlessArrayOf<HBUINT16> &input = StructAfter<HeadlessArrayOf<HBUINT16>> (backtrack);
const ArrayOf<HBUINT16> &lookahead = StructAfter<ArrayOf<HBUINT16>> (input);
const ArrayOf<LookupRecord> &lookup = StructAfter<ArrayOf<LookupRecord>> (lookahead);
return_trace (chain_context_apply_lookup (c,
2018-05-09 01:56:11 +02:00
backtrack.len, backtrack.arrayZ,
input.lenP1, input.arrayZ,
2018-05-09 01:56:11 +02:00
lookahead.len, lookahead.arrayZ, lookup.len,
lookup.arrayZ, lookup_context));
}
template<typename Iterator,
hb_requires (hb_is_iterator (Iterator))>
void serialize_array (hb_serialize_context_t *c,
HBUINT16 len,
Iterator it) const
{
c->copy (len);
for (const auto g : it)
{
HBUINT16 gid;
gid = g;
c->copy (gid);
}
}
ChainRule* copy (hb_serialize_context_t *c,
const hb_map_t *lookup_map,
const hb_map_t *backtrack_map,
const hb_map_t *input_map = nullptr,
const hb_map_t *lookahead_map = nullptr) const
{
TRACE_SERIALIZE (this);
auto *out = c->start_embed (this);
if (unlikely (!out)) return_trace (nullptr);
const hb_map_t *mapping = backtrack_map;
serialize_array (c, backtrack.len, + backtrack.iter ()
| hb_map (mapping));
const HeadlessArrayOf<HBUINT16> &input = StructAfter<HeadlessArrayOf<HBUINT16>> (backtrack);
if (input_map) mapping = input_map;
serialize_array (c, input.lenP1, + input.iter ()
| hb_map (mapping));
const ArrayOf<HBUINT16> &lookahead = StructAfter<ArrayOf<HBUINT16>> (input);
if (lookahead_map) mapping = lookahead_map;
serialize_array (c, lookahead.len, + lookahead.iter ()
| hb_map (mapping));
const ArrayOf<LookupRecord> &lookupRecord = StructAfter<ArrayOf<LookupRecord>> (lookahead);
HBUINT16 lookupCount;
lookupCount = lookupRecord.len;
if (!c->copy (lookupCount)) return_trace (nullptr);
for (unsigned i = 0; i < (unsigned) lookupCount; i++)
if (!c->copy (lookupRecord[i], lookup_map)) return_trace (nullptr);
return_trace (out);
}
bool subset (hb_subset_context_t *c,
const hb_map_t *lookup_map,
const hb_map_t *backtrack_map = nullptr,
const hb_map_t *input_map = nullptr,
const hb_map_t *lookahead_map = nullptr) const
{
TRACE_SUBSET (this);
const HeadlessArrayOf<HBUINT16> &input = StructAfter<HeadlessArrayOf<HBUINT16>> (backtrack);
const ArrayOf<HBUINT16> &lookahead = StructAfter<ArrayOf<HBUINT16>> (input);
if (!backtrack_map)
{
const hb_set_t &glyphset = *c->plan->glyphset ();
if (!hb_all (backtrack, glyphset) ||
!hb_all (input, glyphset) ||
!hb_all (lookahead, glyphset))
return_trace (false);
copy (c->serializer, lookup_map, c->plan->glyph_map);
}
else
{
if (!hb_all (backtrack, backtrack_map) ||
!hb_all (input, input_map) ||
!hb_all (lookahead, lookahead_map))
return_trace (false);
copy (c->serializer, lookup_map, backtrack_map, input_map, lookahead_map);
}
return_trace (true);
}
bool sanitize (hb_sanitize_context_t *c) const
{
2012-11-23 21:32:14 +01:00
TRACE_SANITIZE (this);
if (!backtrack.sanitize (c)) return_trace (false);
const HeadlessArrayOf<HBUINT16> &input = StructAfter<HeadlessArrayOf<HBUINT16>> (backtrack);
if (!input.sanitize (c)) return_trace (false);
const ArrayOf<HBUINT16> &lookahead = StructAfter<ArrayOf<HBUINT16>> (input);
if (!lookahead.sanitize (c)) return_trace (false);
const ArrayOf<LookupRecord> &lookup = StructAfter<ArrayOf<LookupRecord>> (lookahead);
return_trace (lookup.sanitize (c));
2009-08-04 06:58:28 +02:00
}
protected:
ArrayOf<HBUINT16>
2009-05-18 07:49:57 +02:00
backtrack; /* Array of backtracking values
* (to be matched before the input
* sequence) */
HeadlessArrayOf<HBUINT16>
2009-05-18 08:03:58 +02:00
inputX; /* Array of input values (start with
* second glyph) */
ArrayOf<HBUINT16>
2009-05-18 07:49:57 +02:00
lookaheadX; /* Array of lookahead values's (to be
* matched after the input sequence) */
2009-05-18 07:49:57 +02:00
ArrayOf<LookupRecord>
lookupX; /* Array of LookupRecords--in
* design order) */
public:
2010-05-10 23:28:16 +02:00
DEFINE_SIZE_MIN (8);
};
2009-05-20 05:58:54 +02:00
struct ChainRuleSet
{
bool intersects (const hb_set_t *glyphs, ChainContextClosureLookupContext &lookup_context) const
{
2019-03-30 06:40:13 +01:00
return
+ hb_iter (rule)
| hb_map (hb_add (this))
| hb_map ([&] (const ChainRule &_) { return _.intersects (glyphs, lookup_context); })
2019-03-30 06:40:13 +01:00
| hb_any
;
}
void closure (hb_closure_context_t *c, ChainContextClosureLookupContext &lookup_context) const
{
if (unlikely (c->lookup_limit_exceeded ())) return;
2019-03-30 06:40:13 +01:00
return
+ hb_iter (rule)
| hb_map (hb_add (this))
| hb_apply ([&] (const ChainRule &_) { _.closure (c, lookup_context); })
2019-03-30 06:40:13 +01:00
;
}
2019-10-23 01:00:43 +02:00
void closure_lookups (hb_closure_lookups_context_t *c) const
{
if (unlikely (c->lookup_limit_exceeded ())) return;
2019-10-23 01:00:43 +02:00
return
+ hb_iter (rule)
| hb_map (hb_add (this))
| hb_apply ([&] (const ChainRule &_) { _.closure_lookups (c); })
;
}
void collect_glyphs (hb_collect_glyphs_context_t *c, ChainContextCollectGlyphsLookupContext &lookup_context) const
{
2019-03-30 06:40:13 +01:00
return
+ hb_iter (rule)
| hb_map (hb_add (this))
| hb_apply ([&] (const ChainRule &_) { _.collect_glyphs (c, lookup_context); })
2019-03-30 06:40:13 +01:00
;
}
bool would_apply (hb_would_apply_context_t *c, ChainContextApplyLookupContext &lookup_context) const
{
2019-03-30 06:40:13 +01:00
return
+ hb_iter (rule)
| hb_map (hb_add (this))
| hb_map ([&] (const ChainRule &_) { return _.would_apply (c, lookup_context); })
2019-03-30 06:40:13 +01:00
| hb_any
;
}
bool apply (hb_ot_apply_context_t *c, ChainContextApplyLookupContext &lookup_context) const
2009-05-20 05:58:54 +02:00
{
2012-11-23 21:32:14 +01:00
TRACE_APPLY (this);
2019-03-30 06:40:13 +01:00
return_trace (
+ hb_iter (rule)
| hb_map (hb_add (this))
| hb_map ([&] (const ChainRule &_) { return _.apply (c, lookup_context); })
2019-03-30 06:40:13 +01:00
| hb_any
)
;
}
bool subset (hb_subset_context_t *c,
const hb_map_t *lookup_map,
const hb_map_t *backtrack_klass_map = nullptr,
const hb_map_t *input_klass_map = nullptr,
const hb_map_t *lookahead_klass_map = nullptr) const
{
TRACE_SUBSET (this);
auto snap = c->serializer->snapshot ();
auto *out = c->serializer->start_embed (*this);
if (unlikely (!c->serializer->extend_min (out))) return_trace (false);
for (const OffsetTo<ChainRule>& _ : rule)
{
if (!_) continue;
auto *o = out->rule.serialize_append (c->serializer);
if (unlikely (!o)) continue;
auto o_snap = c->serializer->snapshot ();
if (!o->serialize_subset (c, _, this,
lookup_map,
backtrack_klass_map,
input_klass_map,
lookahead_klass_map))
{
out->rule.pop ();
c->serializer->revert (o_snap);
}
}
bool ret = bool (out->rule);
if (!ret) c->serializer->revert (snap);
return_trace (ret);
}
bool sanitize (hb_sanitize_context_t *c) const
{
2012-11-23 21:32:14 +01:00
TRACE_SANITIZE (this);
return_trace (rule.sanitize (c, this));
2009-08-04 06:58:28 +02:00
}
protected:
OffsetArrayOf<ChainRule>
rule; /* Array of ChainRule tables
* ordered by preference */
public:
DEFINE_SIZE_ARRAY (2, rule);
};
2009-05-20 05:58:54 +02:00
struct ChainContextFormat1
{
bool intersects (const hb_set_t *glyphs) const
{
struct ChainContextClosureLookupContext lookup_context = {
{intersects_glyph},
{nullptr, nullptr, nullptr}
};
2019-03-30 06:40:13 +01:00
return
+ hb_zip (this+coverage, ruleSet)
| hb_filter (*glyphs, hb_first)
| hb_map (hb_second)
| hb_map (hb_add (this))
| hb_map ([&] (const ChainRuleSet &_) { return _.intersects (glyphs, lookup_context); })
2019-03-30 06:40:13 +01:00
| hb_any
;
}
void closure (hb_closure_context_t *c) const
{
struct ChainContextClosureLookupContext lookup_context = {
{intersects_glyph},
2017-10-15 12:11:08 +02:00
{nullptr, nullptr, nullptr}
};
2019-03-30 06:40:13 +01:00
+ hb_zip (this+coverage, ruleSet)
| hb_filter (*c->glyphs, hb_first)
| hb_map (hb_second)
| hb_map (hb_add (this))
| hb_apply ([&] (const ChainRuleSet &_) { _.closure (c, lookup_context); })
2019-03-30 06:40:13 +01:00
;
}
2019-10-23 01:00:43 +02:00
void closure_lookups (hb_closure_lookups_context_t *c) const
{
+ hb_iter (ruleSet)
| hb_map (hb_add (this))
| hb_apply ([&] (const ChainRuleSet &_) { _.closure_lookups (c); })
;
}
void collect_variation_indices (hb_collect_variation_indices_context_t *c) const {}
void collect_glyphs (hb_collect_glyphs_context_t *c) const
2012-11-24 00:13:48 +01:00
{
(this+coverage).collect_coverage (c->input);
struct ChainContextCollectGlyphsLookupContext lookup_context = {
{collect_glyph},
2017-10-15 12:11:08 +02:00
{nullptr, nullptr, nullptr}
};
2019-03-30 06:40:13 +01:00
+ hb_iter (ruleSet)
| hb_map (hb_add (this))
| hb_apply ([&] (const ChainRuleSet &_) { _.collect_glyphs (c, lookup_context); })
2019-03-30 06:40:13 +01:00
;
2012-11-24 00:13:48 +01:00
}
bool would_apply (hb_would_apply_context_t *c) const
{
const ChainRuleSet &rule_set = this+ruleSet[(this+coverage).get_coverage (c->glyphs[0])];
struct ChainContextApplyLookupContext lookup_context = {
{match_glyph},
2017-10-15 12:11:08 +02:00
{nullptr, nullptr, nullptr}
};
return rule_set.would_apply (c, lookup_context);
}
const Coverage &get_coverage () const { return this+coverage; }
bool apply (hb_ot_apply_context_t *c) const
2009-05-20 05:58:54 +02:00
{
2012-11-23 21:32:14 +01:00
TRACE_APPLY (this);
unsigned int index = (this+coverage).get_coverage (c->buffer->cur().codepoint);
if (likely (index == NOT_COVERED)) return_trace (false);
const ChainRuleSet &rule_set = this+ruleSet[index];
struct ChainContextApplyLookupContext lookup_context = {
{match_glyph},
2017-10-15 12:11:08 +02:00
{nullptr, nullptr, nullptr}
};
return_trace (rule_set.apply (c, lookup_context));
}
2009-08-04 06:58:28 +02:00
bool subset (hb_subset_context_t *c) const
{
TRACE_SUBSET (this);
const hb_set_t &glyphset = *c->plan->glyphset ();
const hb_map_t &glyph_map = *c->plan->glyph_map;
auto *out = c->serializer->start_embed (*this);
if (unlikely (!c->serializer->extend_min (out))) return_trace (false);
out->format = format;
const hb_map_t *lookup_map = c->table_tag == HB_OT_TAG_GSUB ? c->plan->gsub_lookups : c->plan->gpos_lookups;
hb_sorted_vector_t<hb_codepoint_t> new_coverage;
+ hb_zip (this+coverage, ruleSet)
| hb_filter (glyphset, hb_first)
| hb_filter (subset_offset_array (c, out->ruleSet, this, lookup_map), hb_second)
| hb_map (hb_first)
| hb_map (glyph_map)
| hb_sink (new_coverage)
;
out->coverage.serialize (c->serializer, out)
.serialize (c->serializer, new_coverage.iter ());
return_trace (bool (new_coverage));
}
bool sanitize (hb_sanitize_context_t *c) const
{
2012-11-23 21:32:14 +01:00
TRACE_SANITIZE (this);
return_trace (coverage.sanitize (c, this) && ruleSet.sanitize (c, this));
2009-08-04 06:58:28 +02:00
}
protected:
HBUINT16 format; /* Format identifier--format = 1 */
OffsetTo<Coverage>
coverage; /* Offset to Coverage table--from
* beginning of table */
OffsetArrayOf<ChainRuleSet>
ruleSet; /* Array of ChainRuleSet tables
* ordered by Coverage Index */
public:
DEFINE_SIZE_ARRAY (6, ruleSet);
};
2009-05-20 05:58:54 +02:00
struct ChainContextFormat2
{
bool intersects (const hb_set_t *glyphs) const
{
if (!(this+coverage).intersects (glyphs))
return false;
const ClassDef &backtrack_class_def = this+backtrackClassDef;
const ClassDef &input_class_def = this+inputClassDef;
const ClassDef &lookahead_class_def = this+lookaheadClassDef;
struct ChainContextClosureLookupContext lookup_context = {
{intersects_class},
{&backtrack_class_def,
&input_class_def,
&lookahead_class_def}
};
return
+ hb_enumerate (ruleSet)
2019-05-16 03:15:05 +02:00
| hb_map ([&] (const hb_pair_t<unsigned, const OffsetTo<ChainRuleSet> &> p)
{ return input_class_def.intersects_class (glyphs, p.first) &&
(this+p.second).intersects (glyphs, lookup_context); })
| hb_any
;
}
void closure (hb_closure_context_t *c) const
{
if (!(this+coverage).intersects (c->glyphs))
2012-04-24 05:03:12 +02:00
return;
const ClassDef &backtrack_class_def = this+backtrackClassDef;
const ClassDef &input_class_def = this+inputClassDef;
const ClassDef &lookahead_class_def = this+lookaheadClassDef;
struct ChainContextClosureLookupContext lookup_context = {
{intersects_class},
{&backtrack_class_def,
&input_class_def,
&lookahead_class_def}
};
return
+ hb_enumerate (ruleSet)
2019-05-16 03:15:05 +02:00
| hb_filter ([&] (unsigned _)
{ return input_class_def.intersects_class (c->glyphs, _); },
hb_first)
| hb_map (hb_second)
| hb_map (hb_add (this))
| hb_apply ([&] (const ChainRuleSet &_) { _.closure (c, lookup_context); })
;
}
2019-10-23 01:00:43 +02:00
void closure_lookups (hb_closure_lookups_context_t *c) const
{
+ hb_iter (ruleSet)
| hb_map (hb_add (this))
| hb_apply ([&] (const ChainRuleSet &_) { _.closure_lookups (c); })
;
}
void collect_variation_indices (hb_collect_variation_indices_context_t *c) const {}
void collect_glyphs (hb_collect_glyphs_context_t *c) const
2012-11-24 00:13:48 +01:00
{
(this+coverage).collect_coverage (c->input);
const ClassDef &backtrack_class_def = this+backtrackClassDef;
const ClassDef &input_class_def = this+inputClassDef;
const ClassDef &lookahead_class_def = this+lookaheadClassDef;
struct ChainContextCollectGlyphsLookupContext lookup_context = {
{collect_class},
{&backtrack_class_def,
&input_class_def,
&lookahead_class_def}
};
2019-03-30 06:40:13 +01:00
+ hb_iter (ruleSet)
| hb_map (hb_add (this))
| hb_apply ([&] (const ChainRuleSet &_) { _.collect_glyphs (c, lookup_context); })
2019-03-30 06:40:13 +01:00
;
2012-11-24 00:13:48 +01:00
}
bool would_apply (hb_would_apply_context_t *c) const
{
const ClassDef &backtrack_class_def = this+backtrackClassDef;
const ClassDef &input_class_def = this+inputClassDef;
const ClassDef &lookahead_class_def = this+lookaheadClassDef;
unsigned int index = input_class_def.get_class (c->glyphs[0]);
const ChainRuleSet &rule_set = this+ruleSet[index];
struct ChainContextApplyLookupContext lookup_context = {
{match_class},
{&backtrack_class_def,
&input_class_def,
&lookahead_class_def}
};
return rule_set.would_apply (c, lookup_context);
}
const Coverage &get_coverage () const { return this+coverage; }
bool apply (hb_ot_apply_context_t *c) const
2009-05-20 05:58:54 +02:00
{
2012-11-23 21:32:14 +01:00
TRACE_APPLY (this);
unsigned int index = (this+coverage).get_coverage (c->buffer->cur().codepoint);
if (likely (index == NOT_COVERED)) return_trace (false);
const ClassDef &backtrack_class_def = this+backtrackClassDef;
const ClassDef &input_class_def = this+inputClassDef;
const ClassDef &lookahead_class_def = this+lookaheadClassDef;
index = input_class_def.get_class (c->buffer->cur().codepoint);
const ChainRuleSet &rule_set = this+ruleSet[index];
struct ChainContextApplyLookupContext lookup_context = {
{match_class},
2010-05-10 23:47:22 +02:00
{&backtrack_class_def,
&input_class_def,
&lookahead_class_def}
};
return_trace (rule_set.apply (c, lookup_context));
}
bool subset (hb_subset_context_t *c) const
{
TRACE_SUBSET (this);
auto *out = c->serializer->start_embed (*this);
if (unlikely (!c->serializer->extend_min (out))) return_trace (false);
out->format = format;
out->coverage.serialize_subset (c, coverage, this);
hb_map_t backtrack_klass_map;
out->backtrackClassDef.serialize_subset (c, backtrackClassDef, this, &backtrack_klass_map);
// subset inputClassDef based on glyphs survived in Coverage subsetting
hb_map_t input_klass_map;
out->inputClassDef.serialize_subset (c, inputClassDef, this, &input_klass_map);
hb_map_t lookahead_klass_map;
out->lookaheadClassDef.serialize_subset (c, lookaheadClassDef, this, &lookahead_klass_map);
unsigned non_zero_index = 0, index = 0;
bool ret = true;
const hb_map_t *lookup_map = c->table_tag == HB_OT_TAG_GSUB ? c->plan->gsub_lookups : c->plan->gpos_lookups;
for (const OffsetTo<ChainRuleSet>& _ : + hb_enumerate (ruleSet)
| hb_filter (input_klass_map, hb_first)
| hb_map (hb_second))
{
auto *o = out->ruleSet.serialize_append (c->serializer);
if (unlikely (!o))
{
ret = false;
break;
}
if (o->serialize_subset (c, _, this,
lookup_map,
&backtrack_klass_map,
&input_klass_map,
&lookahead_klass_map))
non_zero_index = index;
index++;
}
if (!ret) return_trace (ret);
//prune empty trailing ruleSets
--index;
while (index > non_zero_index)
{
out->ruleSet.pop ();
index--;
}
return_trace (bool (out->ruleSet));
}
bool sanitize (hb_sanitize_context_t *c) const
{
2012-11-23 21:32:14 +01:00
TRACE_SANITIZE (this);
return_trace (coverage.sanitize (c, this) &&
backtrackClassDef.sanitize (c, this) &&
inputClassDef.sanitize (c, this) &&
lookaheadClassDef.sanitize (c, this) &&
ruleSet.sanitize (c, this));
2009-08-04 06:58:28 +02:00
}
protected:
HBUINT16 format; /* Format identifier--format = 2 */
OffsetTo<Coverage>
coverage; /* Offset to Coverage table--from
* beginning of table */
OffsetTo<ClassDef>
backtrackClassDef; /* Offset to glyph ClassDef table
* containing backtrack sequence
* data--from beginning of table */
OffsetTo<ClassDef>
inputClassDef; /* Offset to glyph ClassDef
* table containing input sequence
* data--from beginning of table */
OffsetTo<ClassDef>
lookaheadClassDef; /* Offset to glyph ClassDef table
* containing lookahead sequence
* data--from beginning of table */
OffsetArrayOf<ChainRuleSet>
ruleSet; /* Array of ChainRuleSet tables
* ordered by class */
public:
DEFINE_SIZE_ARRAY (12, ruleSet);
};
2009-05-20 05:58:54 +02:00
struct ChainContextFormat3
{
bool intersects (const hb_set_t *glyphs) const
{
const OffsetArrayOf<Coverage> &input = StructAfter<OffsetArrayOf<Coverage>> (backtrack);
if (!(this+input[0]).intersects (glyphs))
return false;
const OffsetArrayOf<Coverage> &lookahead = StructAfter<OffsetArrayOf<Coverage>> (input);
struct ChainContextClosureLookupContext lookup_context = {
{intersects_coverage},
{this, this, this}
};
return chain_context_intersects (glyphs,
backtrack.len, (const HBUINT16 *) backtrack.arrayZ,
input.len, (const HBUINT16 *) input.arrayZ + 1,
lookahead.len, (const HBUINT16 *) lookahead.arrayZ,
lookup_context);
}
void closure (hb_closure_context_t *c) const
{
const OffsetArrayOf<Coverage> &input = StructAfter<OffsetArrayOf<Coverage>> (backtrack);
2012-04-24 05:03:12 +02:00
if (!(this+input[0]).intersects (c->glyphs))
return;
const OffsetArrayOf<Coverage> &lookahead = StructAfter<OffsetArrayOf<Coverage>> (input);
const ArrayOf<LookupRecord> &lookup = StructAfter<ArrayOf<LookupRecord>> (lookahead);
2012-04-24 05:03:12 +02:00
struct ChainContextClosureLookupContext lookup_context = {
{intersects_coverage},
2012-04-24 05:03:12 +02:00
{this, this, this}
};
chain_context_closure_lookup (c,
2018-05-09 01:56:11 +02:00
backtrack.len, (const HBUINT16 *) backtrack.arrayZ,
input.len, (const HBUINT16 *) input.arrayZ + 1,
lookahead.len, (const HBUINT16 *) lookahead.arrayZ,
lookup.len, lookup.arrayZ,
2012-04-24 05:03:12 +02:00
lookup_context);
}
2019-10-23 01:00:43 +02:00
void closure_lookups (hb_closure_lookups_context_t *c) const
{
const OffsetArrayOf<Coverage> &input = StructAfter<OffsetArrayOf<Coverage>> (backtrack);
const OffsetArrayOf<Coverage> &lookahead = StructAfter<OffsetArrayOf<Coverage>> (input);
const ArrayOf<LookupRecord> &lookup = StructAfter<ArrayOf<LookupRecord>> (lookahead);
recurse_lookups (c, lookup.len, lookup.arrayZ);
}
void collect_variation_indices (hb_collect_variation_indices_context_t *c) const {}
void collect_glyphs (hb_collect_glyphs_context_t *c) const
2012-11-24 00:13:48 +01:00
{
const OffsetArrayOf<Coverage> &input = StructAfter<OffsetArrayOf<Coverage>> (backtrack);
(this+input[0]).collect_coverage (c->input);
const OffsetArrayOf<Coverage> &lookahead = StructAfter<OffsetArrayOf<Coverage>> (input);
const ArrayOf<LookupRecord> &lookup = StructAfter<ArrayOf<LookupRecord>> (lookahead);
struct ChainContextCollectGlyphsLookupContext lookup_context = {
{collect_coverage},
{this, this, this}
};
chain_context_collect_glyphs_lookup (c,
2018-05-09 01:56:11 +02:00
backtrack.len, (const HBUINT16 *) backtrack.arrayZ,
input.len, (const HBUINT16 *) input.arrayZ + 1,
lookahead.len, (const HBUINT16 *) lookahead.arrayZ,
lookup.len, lookup.arrayZ,
lookup_context);
2012-11-24 00:13:48 +01:00
}
bool would_apply (hb_would_apply_context_t *c) const
{
const OffsetArrayOf<Coverage> &input = StructAfter<OffsetArrayOf<Coverage>> (backtrack);
const OffsetArrayOf<Coverage> &lookahead = StructAfter<OffsetArrayOf<Coverage>> (input);
const ArrayOf<LookupRecord> &lookup = StructAfter<ArrayOf<LookupRecord>> (lookahead);
struct ChainContextApplyLookupContext lookup_context = {
{match_coverage},
{this, this, this}
};
return chain_context_would_apply_lookup (c,
backtrack.len, (const HBUINT16 *) backtrack.arrayZ,
input.len, (const HBUINT16 *) input.arrayZ + 1,
lookahead.len, (const HBUINT16 *) lookahead.arrayZ,
lookup.len, lookup.arrayZ, lookup_context);
}
const Coverage &get_coverage () const
{
const OffsetArrayOf<Coverage> &input = StructAfter<OffsetArrayOf<Coverage>> (backtrack);
return this+input[0];
}
bool apply (hb_ot_apply_context_t *c) const
2009-05-20 05:58:54 +02:00
{
2012-11-23 21:32:14 +01:00
TRACE_APPLY (this);
const OffsetArrayOf<Coverage> &input = StructAfter<OffsetArrayOf<Coverage>> (backtrack);
unsigned int index = (this+input[0]).get_coverage (c->buffer->cur().codepoint);
if (likely (index == NOT_COVERED)) return_trace (false);
const OffsetArrayOf<Coverage> &lookahead = StructAfter<OffsetArrayOf<Coverage>> (input);
const ArrayOf<LookupRecord> &lookup = StructAfter<ArrayOf<LookupRecord>> (lookahead);
struct ChainContextApplyLookupContext lookup_context = {
{match_coverage},
2010-05-10 23:47:22 +02:00
{this, this, this}
};
return_trace (chain_context_apply_lookup (c,
2018-05-09 01:56:11 +02:00
backtrack.len, (const HBUINT16 *) backtrack.arrayZ,
input.len, (const HBUINT16 *) input.arrayZ + 1,
lookahead.len, (const HBUINT16 *) lookahead.arrayZ,
lookup.len, lookup.arrayZ, lookup_context));
}
template<typename Iterator,
hb_requires (hb_is_iterator (Iterator))>
bool serialize_coverage_offsets (hb_subset_context_t *c, Iterator it, const void* base) const
{
TRACE_SERIALIZE (this);
auto *out = c->serializer->start_embed<OffsetArrayOf<Coverage>> ();
if (unlikely (!c->serializer->allocate_size<HBUINT16> (HBUINT16::static_size))) return_trace (false);
+ it
| hb_apply (subset_offset_array (c, *out, base))
;
return_trace (out->len);
}
bool subset (hb_subset_context_t *c) const
{
TRACE_SUBSET (this);
auto *out = c->serializer->start_embed (this);
if (unlikely (!out)) return_trace (false);
if (unlikely (!c->serializer->embed (this->format))) return_trace (false);
if (!serialize_coverage_offsets (c, backtrack.iter (), this))
return_trace (false);
const OffsetArrayOf<Coverage> &input = StructAfter<OffsetArrayOf<Coverage>> (backtrack);
if (!serialize_coverage_offsets (c, input.iter (), this))
return_trace (false);
const OffsetArrayOf<Coverage> &lookahead = StructAfter<OffsetArrayOf<Coverage>> (input);
if (!serialize_coverage_offsets (c, lookahead.iter (), this))
return_trace (false);
const ArrayOf<LookupRecord> &lookupRecord = StructAfter<ArrayOf<LookupRecord>> (lookahead);
HBUINT16 lookupCount;
lookupCount = lookupRecord.len;
if (!c->serializer->copy (lookupCount)) return_trace (false);
const hb_map_t *lookup_map = c->table_tag == HB_OT_TAG_GSUB ? c->plan->gsub_lookups : c->plan->gpos_lookups;
for (unsigned i = 0; i < (unsigned) lookupCount; i++)
if (!c->serializer->copy (lookupRecord[i], lookup_map)) return_trace (false);
return_trace (true);
}
bool sanitize (hb_sanitize_context_t *c) const
{
2012-11-23 21:32:14 +01:00
TRACE_SANITIZE (this);
if (!backtrack.sanitize (c, this)) return_trace (false);
const OffsetArrayOf<Coverage> &input = StructAfter<OffsetArrayOf<Coverage>> (backtrack);
if (!input.sanitize (c, this)) return_trace (false);
if (!input.len) return_trace (false); /* To be consistent with Context. */
const OffsetArrayOf<Coverage> &lookahead = StructAfter<OffsetArrayOf<Coverage>> (input);
if (!lookahead.sanitize (c, this)) return_trace (false);
const ArrayOf<LookupRecord> &lookup = StructAfter<ArrayOf<LookupRecord>> (lookahead);
return_trace (lookup.sanitize (c));
2009-08-04 06:58:28 +02:00
}
protected:
HBUINT16 format; /* Format identifier--format = 3 */
2009-05-18 07:49:57 +02:00
OffsetArrayOf<Coverage>
backtrack; /* Array of coverage tables
* in backtracking sequence, in glyph
* sequence order */
2009-05-18 07:49:57 +02:00
OffsetArrayOf<Coverage>
inputX ; /* Array of coverage
* tables in input sequence, in glyph
* sequence order */
2009-05-18 07:49:57 +02:00
OffsetArrayOf<Coverage>
lookaheadX; /* Array of coverage tables
* in lookahead sequence, in glyph
* sequence order */
2009-05-18 07:49:57 +02:00
ArrayOf<LookupRecord>
lookupX; /* Array of LookupRecords--in
2009-05-18 07:49:57 +02:00
* design order) */
public:
2010-05-10 23:28:16 +02:00
DEFINE_SIZE_MIN (10);
};
2009-05-20 05:58:54 +02:00
struct ChainContext
{
template <typename context_t, typename ...Ts>
2019-05-08 05:58:43 +02:00
typename context_t::return_t dispatch (context_t *c, Ts&&... ds) const
{
2014-12-13 05:36:49 +01:00
TRACE_DISPATCH (this, u.format);
if (unlikely (!c->may_dispatch (this, &u.format))) return_trace (c->no_dispatch_return_value ());
switch (u.format) {
case 1: return_trace (c->dispatch (u.format1, hb_forward<Ts> (ds)...));
case 2: return_trace (c->dispatch (u.format2, hb_forward<Ts> (ds)...));
case 3: return_trace (c->dispatch (u.format3, hb_forward<Ts> (ds)...));
default:return_trace (c->default_return_value ());
}
}
protected:
union {
HBUINT16 format; /* Format identifier */
2010-05-11 01:45:41 +02:00
ChainContextFormat1 format1;
ChainContextFormat2 format2;
ChainContextFormat3 format3;
} u;
};
template <typename T>
struct ExtensionFormat1
{
unsigned int get_type () const { return extensionLookupType; }
template <typename X>
const X& get_subtable () const
2019-12-10 20:18:32 +01:00
{ return this + reinterpret_cast<const LOffsetTo<typename T::SubTable> &> (extensionOffset); }
template <typename context_t, typename ...Ts>
2019-05-08 05:58:43 +02:00
typename context_t::return_t dispatch (context_t *c, Ts&&... ds) const
{
TRACE_DISPATCH (this, format);
if (unlikely (!c->may_dispatch (this, this))) return_trace (c->no_dispatch_return_value ());
return_trace (get_subtable<typename T::SubTable> ().dispatch (c, get_type (), hb_forward<Ts> (ds)...));
}
void collect_variation_indices (hb_collect_variation_indices_context_t *c) const
{ dispatch (c); }
/* This is called from may_dispatch() above with hb_sanitize_context_t. */
bool sanitize (hb_sanitize_context_t *c) const
{
2012-11-23 21:32:14 +01:00
TRACE_SANITIZE (this);
return_trace (c->check_struct (this) &&
2018-09-04 01:53:03 +02:00
extensionLookupType != T::SubTable::Extension);
2009-08-04 06:58:28 +02:00
}
protected:
HBUINT16 format; /* Format identifier. Set to 1. */
HBUINT16 extensionLookupType; /* Lookup type of subtable referenced
* by ExtensionOffset (i.e. the
* extension subtable). */
2019-05-10 22:17:41 +02:00
Offset32 extensionOffset; /* Offset to the extension subtable,
* of lookup type subtable. */
public:
DEFINE_SIZE_STATIC (8);
};
2012-11-23 22:57:36 +01:00
template <typename T>
struct Extension
{
unsigned int get_type () const
{
switch (u.format) {
2010-05-11 01:45:41 +02:00
case 1: return u.format1.get_type ();
default:return 0;
}
}
2012-11-23 23:04:55 +01:00
template <typename X>
const X& get_subtable () const
2012-11-23 23:04:55 +01:00
{
switch (u.format) {
2018-09-04 01:53:03 +02:00
case 1: return u.format1.template get_subtable<typename T::SubTable> ();
default:return Null (typename T::SubTable);
}
2012-11-23 23:04:55 +01:00
}
template <typename context_t, typename ...Ts>
2019-05-08 05:58:43 +02:00
typename context_t::return_t dispatch (context_t *c, Ts&&... ds) const
2012-11-23 22:57:36 +01:00
{
TRACE_DISPATCH (this, u.format);
if (unlikely (!c->may_dispatch (this, &u.format))) return_trace (c->no_dispatch_return_value ());
2009-08-04 06:58:28 +02:00
switch (u.format) {
case 1: return_trace (u.format1.dispatch (c, hb_forward<Ts> (ds)...));
default:return_trace (c->default_return_value ());
2009-08-04 06:58:28 +02:00
}
}
protected:
union {
HBUINT16 format; /* Format identifier */
ExtensionFormat1<T> format1;
} u;
};
2009-05-18 02:13:02 +02:00
/*
* GSUB/GPOS Common
*/
2018-10-10 17:41:05 +02:00
struct hb_ot_layout_lookup_accelerator_t
{
template <typename TLookup>
void init (const TLookup &lookup)
2018-10-10 17:41:05 +02:00
{
digest.init ();
lookup.collect_coverage (&digest);
2018-10-10 17:50:46 +02:00
subtables.init ();
OT::hb_get_subtables_context_t c_get_subtables (subtables);
lookup.dispatch (&c_get_subtables);
}
void fini () { subtables.fini (); }
2018-10-10 17:41:05 +02:00
bool may_have (hb_codepoint_t g) const
2018-10-10 17:41:05 +02:00
{ return digest.may_have (g); }
bool apply (hb_ot_apply_context_t *c) const
2018-10-10 17:54:48 +02:00
{
2018-12-22 00:46:51 +01:00
for (unsigned int i = 0; i < subtables.length; i++)
2018-11-15 20:40:56 +01:00
if (subtables[i].apply (c))
return true;
return false;
2018-10-10 17:54:48 +02:00
}
private:
2018-10-10 17:41:05 +02:00
hb_set_digest_t digest;
2018-10-10 17:50:46 +02:00
hb_get_subtables_context_t::array_t subtables;
2018-10-10 17:41:05 +02:00
};
2009-05-20 05:58:54 +02:00
struct GSUBGPOS
{
bool has_data () const { return version.to_int (); }
unsigned int get_script_count () const
{ return (this+scriptList).len; }
const Tag& get_script_tag (unsigned int i) const
{ return (this+scriptList).get_tag (i); }
unsigned int get_script_tags (unsigned int start_offset,
unsigned int *script_count /* IN/OUT */,
hb_tag_t *script_tags /* OUT */) const
{ return (this+scriptList).get_tags (start_offset, script_count, script_tags); }
const Script& get_script (unsigned int i) const
{ return (this+scriptList)[i]; }
bool find_script_index (hb_tag_t tag, unsigned int *index) const
{ return (this+scriptList).find_index (tag, index); }
unsigned int get_feature_count () const
{ return (this+featureList).len; }
hb_tag_t get_feature_tag (unsigned int i) const
{ return i == Index::NOT_FOUND_INDEX ? HB_TAG_NONE : (this+featureList).get_tag (i); }
unsigned int get_feature_tags (unsigned int start_offset,
unsigned int *feature_count /* IN/OUT */,
hb_tag_t *feature_tags /* OUT */) const
{ return (this+featureList).get_tags (start_offset, feature_count, feature_tags); }
const Feature& get_feature (unsigned int i) const
{ return (this+featureList)[i]; }
bool find_feature_index (hb_tag_t tag, unsigned int *index) const
{ return (this+featureList).find_index (tag, index); }
unsigned int get_lookup_count () const
{ return (this+lookupList).len; }
const Lookup& get_lookup (unsigned int i) const
{ return (this+lookupList)[i]; }
2009-05-18 02:13:02 +02:00
bool find_variations_index (const int *coords, unsigned int num_coords,
unsigned int *index) const
{
2019-10-23 01:00:43 +02:00
#ifdef HB_NO_VAR
*index = FeatureVariations::NOT_FOUND_INDEX;
return false;
#endif
2020-03-26 08:31:53 +01:00
return (version.to_int () >= 0x00010001u ? this+featureVars : Null (FeatureVariations))
.find_index (coords, num_coords, index);
}
const Feature& get_feature_variation (unsigned int feature_index,
2019-01-22 12:02:06 +01:00
unsigned int variations_index) const
{
#ifndef HB_NO_VAR
if (FeatureVariations::NOT_FOUND_INDEX != variations_index &&
version.to_int () >= 0x00010001u)
{
const Feature *feature = (this+featureVars).find_substitute (variations_index,
feature_index);
if (feature)
2018-11-15 20:40:56 +01:00
return *feature;
}
#endif
return get_feature (feature_index);
}
2019-10-23 01:00:43 +02:00
void feature_variation_collect_lookups (const hb_set_t *feature_indexes,
hb_set_t *lookup_indexes /* OUT */) const
{
#ifndef HB_NO_VAR
if (version.to_int () >= 0x00010001u)
(this+featureVars).collect_lookups (feature_indexes, lookup_indexes);
#endif
}
2018-09-04 01:53:03 +02:00
template <typename TLookup>
bool subset (hb_subset_layout_context_t *c) const
{
TRACE_SUBSET (this);
auto *out = c->subset_context->serializer->embed (*this);
if (unlikely (!out)) return_trace (false);
2018-12-14 00:10:48 +01:00
typedef LookupOffsetList<TLookup> TLookupList;
2019-12-10 20:18:32 +01:00
reinterpret_cast<OffsetTo<TLookupList> &> (out->lookupList)
.serialize_subset (c->subset_context,
reinterpret_cast<const OffsetTo<TLookupList> &> (lookupList),
this,
c);
reinterpret_cast<OffsetTo<RecordListOfFeature> &> (out->featureList)
.serialize_subset (c->subset_context,
reinterpret_cast<const OffsetTo<RecordListOfFeature> &> (featureList),
this,
c);
out->scriptList.serialize_subset (c->subset_context,
scriptList,
this,
c);
2018-09-04 01:53:03 +02:00
#ifndef HB_NO_VAR
if (version.to_int () >= 0x00010001u)
{
bool ret = out->featureVars.serialize_subset (c->subset_context, featureVars, this, c);
if (!ret)
{
out->version.major = 1;
out->version.minor = 0;
}
}
#endif
2018-12-14 00:10:48 +01:00
return_trace (true);
}
void closure_features (const hb_map_t *lookup_indexes, /* IN */
hb_set_t *feature_indexes /* OUT */) const
{
unsigned int feature_count = hb_min (get_feature_count (), (unsigned) HB_MAX_FEATURES);
for (unsigned i = 0; i < feature_count; i++)
{
const Feature& f = get_feature (i);
if ((!f.featureParams.is_null ()) || f.intersects_lookup_indexes (lookup_indexes))
feature_indexes->add (i);
}
#ifndef HB_NO_VAR
if (version.to_int () >= 0x00010001u)
(this+featureVars).closure_features (lookup_indexes, feature_indexes);
#endif
}
unsigned int get_size () const
{
return min_size +
(version.to_int () >= 0x00010001u ? featureVars.static_size : 0);
}
2018-09-04 01:41:28 +02:00
template <typename TLookup>
bool sanitize (hb_sanitize_context_t *c) const
{
2012-11-23 21:32:14 +01:00
TRACE_SANITIZE (this);
2018-09-04 01:41:28 +02:00
typedef OffsetListOf<TLookup> TLookupList;
if (unlikely (!(version.sanitize (c) &&
likely (version.major == 1) &&
scriptList.sanitize (c, this) &&
featureList.sanitize (c, this) &&
2019-12-10 20:18:32 +01:00
reinterpret_cast<const OffsetTo<TLookupList> &> (lookupList).sanitize (c, this))))
return_trace (false);
#ifndef HB_NO_VAR
if (unlikely (!(version.to_int () < 0x00010001u || featureVars.sanitize (c, this))))
return_trace (false);
#endif
return_trace (true);
2009-08-04 08:09:34 +02:00
}
2018-08-26 09:47:55 +02:00
template <typename T>
struct accelerator_t
{
void init (hb_face_t *face)
2018-08-26 09:47:55 +02:00
{
this->table = hb_sanitize_context_t ().reference_table<T> (face);
if (unlikely (this->table->is_blacklisted (this->table.get_blob (), face)))
{
hb_blob_destroy (this->table.get_blob ());
this->table = hb_blob_get_empty ();
}
2018-08-26 09:47:55 +02:00
this->lookup_count = table->get_lookup_count ();
2018-08-26 09:47:55 +02:00
this->accels = (hb_ot_layout_lookup_accelerator_t *) calloc (this->lookup_count, sizeof (hb_ot_layout_lookup_accelerator_t));
if (unlikely (!this->accels))
2018-11-15 20:40:56 +01:00
this->lookup_count = 0;
2018-08-26 09:47:55 +02:00
for (unsigned int i = 0; i < this->lookup_count; i++)
this->accels[i].init (table->get_lookup (i));
2018-08-26 09:47:55 +02:00
}
void fini ()
2018-08-26 09:47:55 +02:00
{
for (unsigned int i = 0; i < this->lookup_count; i++)
this->accels[i].fini ();
free (this->accels);
2018-11-11 17:40:57 +01:00
this->table.destroy ();
2018-08-26 09:47:55 +02:00
}
hb_blob_ptr_t<T> table;
unsigned int lookup_count;
hb_ot_layout_lookup_accelerator_t *accels;
2018-08-26 09:47:55 +02:00
};
2009-05-24 06:50:27 +02:00
protected:
2016-02-22 03:44:45 +01:00
FixedVersion<>version; /* Version of the GSUB/GPOS table--initially set
* to 0x00010000u */
2009-05-18 02:13:02 +02:00
OffsetTo<ScriptList>
2019-12-31 13:23:02 +01:00
scriptList; /* ScriptList table */
2009-05-18 02:13:02 +02:00
OffsetTo<FeatureList>
2019-12-31 13:23:02 +01:00
featureList; /* FeatureList table */
2009-05-18 02:13:02 +02:00
OffsetTo<LookupList>
2019-12-31 13:23:02 +01:00
lookupList; /* LookupList table */
2017-01-23 05:28:56 +01:00
LOffsetTo<FeatureVariations>
featureVars; /* Offset to Feature Variations
table--from beginning of table
* (may be NULL). Introduced
* in version 0x00010001. */
public:
DEFINE_SIZE_MIN (10);
2009-05-18 02:13:02 +02:00
};
2012-11-17 03:49:54 +01:00
} /* namespace OT */
2010-07-23 21:11:18 +02:00
#endif /* HB_OT_LAYOUT_GSUBGPOS_HH */