918 lines
26 KiB
C++
918 lines
26 KiB
C++
/*
|
|
* Copyright © 2017 Google, Inc.
|
|
*
|
|
* This is part of HarfBuzz, a text shaping library.
|
|
*
|
|
* Permission is hereby granted, without written agreement and without
|
|
* license or royalty fees, to use, copy, modify, and distribute this
|
|
* software and its documentation for any purpose, provided that the
|
|
* above copyright notice and the following two paragraphs appear in
|
|
* all copies of this software.
|
|
*
|
|
* IN NO EVENT SHALL THE COPYRIGHT HOLDER BE LIABLE TO ANY PARTY FOR
|
|
* DIRECT, INDIRECT, SPECIAL, INCIDENTAL, OR CONSEQUENTIAL DAMAGES
|
|
* ARISING OUT OF THE USE OF THIS SOFTWARE AND ITS DOCUMENTATION, EVEN
|
|
* IF THE COPYRIGHT HOLDER HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH
|
|
* DAMAGE.
|
|
*
|
|
* THE COPYRIGHT HOLDER SPECIFICALLY DISCLAIMS ANY WARRANTIES, INCLUDING,
|
|
* BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND
|
|
* FITNESS FOR A PARTICULAR PURPOSE. THE SOFTWARE PROVIDED HEREUNDER IS
|
|
* ON AN "AS IS" BASIS, AND THE COPYRIGHT HOLDER HAS NO OBLIGATION TO
|
|
* PROVIDE MAINTENANCE, SUPPORT, UPDATES, ENHANCEMENTS, OR MODIFICATIONS.
|
|
*
|
|
* Google Author(s): Behdad Esfahbod
|
|
*/
|
|
|
|
#ifndef HB_AAT_LAYOUT_COMMON_HH
|
|
#define HB_AAT_LAYOUT_COMMON_HH
|
|
|
|
#include "hb-aat-layout.hh"
|
|
#include "hb-aat-map.hh"
|
|
#include "hb-open-type.hh"
|
|
|
|
namespace OT {
|
|
struct GDEF;
|
|
};
|
|
|
|
namespace AAT {
|
|
|
|
using namespace OT;
|
|
|
|
|
|
struct ankr;
|
|
|
|
struct hb_aat_apply_context_t :
|
|
hb_dispatch_context_t<hb_aat_apply_context_t, bool, HB_DEBUG_APPLY>
|
|
{
|
|
const char *get_name () { return "APPLY"; }
|
|
template <typename T>
|
|
return_t dispatch (const T &obj) { return obj.apply (this); }
|
|
static return_t default_return_value () { return false; }
|
|
bool stop_sublookup_iteration (return_t r) const { return r; }
|
|
|
|
const hb_ot_shape_plan_t *plan;
|
|
hb_font_t *font;
|
|
hb_face_t *face;
|
|
hb_buffer_t *buffer;
|
|
hb_sanitize_context_t sanitizer;
|
|
const ankr *ankr_table;
|
|
const OT::GDEF *gdef_table;
|
|
const hb_sorted_vector_t<hb_aat_map_t::range_flags_t> *range_flags = nullptr;
|
|
hb_mask_t subtable_flags = 0;
|
|
|
|
/* Unused. For debug tracing only. */
|
|
unsigned int lookup_index;
|
|
|
|
HB_INTERNAL hb_aat_apply_context_t (const hb_ot_shape_plan_t *plan_,
|
|
hb_font_t *font_,
|
|
hb_buffer_t *buffer_,
|
|
hb_blob_t *blob = const_cast<hb_blob_t *> (&Null (hb_blob_t)));
|
|
|
|
HB_INTERNAL ~hb_aat_apply_context_t ();
|
|
|
|
HB_INTERNAL void set_ankr_table (const AAT::ankr *ankr_table_);
|
|
|
|
void set_lookup_index (unsigned int i) { lookup_index = i; }
|
|
};
|
|
|
|
|
|
/*
|
|
* Lookup Table
|
|
*/
|
|
|
|
template <typename T> struct Lookup;
|
|
|
|
template <typename T>
|
|
struct LookupFormat0
|
|
{
|
|
friend struct Lookup<T>;
|
|
|
|
private:
|
|
const T* get_value (hb_codepoint_t glyph_id, unsigned int num_glyphs) const
|
|
{
|
|
if (unlikely (glyph_id >= num_glyphs)) return nullptr;
|
|
return &arrayZ[glyph_id];
|
|
}
|
|
|
|
bool sanitize (hb_sanitize_context_t *c) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (arrayZ.sanitize (c, c->get_num_glyphs ()));
|
|
}
|
|
bool sanitize (hb_sanitize_context_t *c, const void *base) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (arrayZ.sanitize (c, c->get_num_glyphs (), base));
|
|
}
|
|
|
|
protected:
|
|
HBUINT16 format; /* Format identifier--format = 0 */
|
|
UnsizedArrayOf<T>
|
|
arrayZ; /* Array of lookup values, indexed by glyph index. */
|
|
public:
|
|
DEFINE_SIZE_UNBOUNDED (2);
|
|
};
|
|
|
|
|
|
template <typename T>
|
|
struct LookupSegmentSingle
|
|
{
|
|
static constexpr unsigned TerminationWordCount = 2u;
|
|
|
|
int cmp (hb_codepoint_t g) const
|
|
{ return g < first ? -1 : g <= last ? 0 : +1 ; }
|
|
|
|
bool sanitize (hb_sanitize_context_t *c) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (c->check_struct (this) && value.sanitize (c));
|
|
}
|
|
bool sanitize (hb_sanitize_context_t *c, const void *base) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (c->check_struct (this) && value.sanitize (c, base));
|
|
}
|
|
|
|
HBGlyphID16 last; /* Last GlyphID in this segment */
|
|
HBGlyphID16 first; /* First GlyphID in this segment */
|
|
T value; /* The lookup value (only one) */
|
|
public:
|
|
DEFINE_SIZE_STATIC (4 + T::static_size);
|
|
};
|
|
|
|
template <typename T>
|
|
struct LookupFormat2
|
|
{
|
|
friend struct Lookup<T>;
|
|
|
|
private:
|
|
const T* get_value (hb_codepoint_t glyph_id) const
|
|
{
|
|
const LookupSegmentSingle<T> *v = segments.bsearch (glyph_id);
|
|
return v ? &v->value : nullptr;
|
|
}
|
|
|
|
bool sanitize (hb_sanitize_context_t *c) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (segments.sanitize (c));
|
|
}
|
|
bool sanitize (hb_sanitize_context_t *c, const void *base) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (segments.sanitize (c, base));
|
|
}
|
|
|
|
protected:
|
|
HBUINT16 format; /* Format identifier--format = 2 */
|
|
VarSizedBinSearchArrayOf<LookupSegmentSingle<T>>
|
|
segments; /* The actual segments. These must already be sorted,
|
|
* according to the first word in each one (the last
|
|
* glyph in each segment). */
|
|
public:
|
|
DEFINE_SIZE_ARRAY (8, segments);
|
|
};
|
|
|
|
template <typename T>
|
|
struct LookupSegmentArray
|
|
{
|
|
static constexpr unsigned TerminationWordCount = 2u;
|
|
|
|
const T* get_value (hb_codepoint_t glyph_id, const void *base) const
|
|
{
|
|
return first <= glyph_id && glyph_id <= last ? &(base+valuesZ)[glyph_id - first] : nullptr;
|
|
}
|
|
|
|
int cmp (hb_codepoint_t g) const
|
|
{ return g < first ? -1 : g <= last ? 0 : +1; }
|
|
|
|
bool sanitize (hb_sanitize_context_t *c, const void *base) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (c->check_struct (this) &&
|
|
first <= last &&
|
|
valuesZ.sanitize (c, base, last - first + 1));
|
|
}
|
|
template <typename ...Ts>
|
|
bool sanitize (hb_sanitize_context_t *c, const void *base, Ts&&... ds) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (c->check_struct (this) &&
|
|
first <= last &&
|
|
valuesZ.sanitize (c, base, last - first + 1, std::forward<Ts> (ds)...));
|
|
}
|
|
|
|
HBGlyphID16 last; /* Last GlyphID in this segment */
|
|
HBGlyphID16 first; /* First GlyphID in this segment */
|
|
NNOffset16To<UnsizedArrayOf<T>>
|
|
valuesZ; /* A 16-bit offset from the start of
|
|
* the table to the data. */
|
|
public:
|
|
DEFINE_SIZE_STATIC (6);
|
|
};
|
|
|
|
template <typename T>
|
|
struct LookupFormat4
|
|
{
|
|
friend struct Lookup<T>;
|
|
|
|
private:
|
|
const T* get_value (hb_codepoint_t glyph_id) const
|
|
{
|
|
const LookupSegmentArray<T> *v = segments.bsearch (glyph_id);
|
|
return v ? v->get_value (glyph_id, this) : nullptr;
|
|
}
|
|
|
|
bool sanitize (hb_sanitize_context_t *c) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (segments.sanitize (c, this));
|
|
}
|
|
bool sanitize (hb_sanitize_context_t *c, const void *base) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (segments.sanitize (c, this, base));
|
|
}
|
|
|
|
protected:
|
|
HBUINT16 format; /* Format identifier--format = 4 */
|
|
VarSizedBinSearchArrayOf<LookupSegmentArray<T>>
|
|
segments; /* The actual segments. These must already be sorted,
|
|
* according to the first word in each one (the last
|
|
* glyph in each segment). */
|
|
public:
|
|
DEFINE_SIZE_ARRAY (8, segments);
|
|
};
|
|
|
|
template <typename T>
|
|
struct LookupSingle
|
|
{
|
|
static constexpr unsigned TerminationWordCount = 1u;
|
|
|
|
int cmp (hb_codepoint_t g) const { return glyph.cmp (g); }
|
|
|
|
bool sanitize (hb_sanitize_context_t *c) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (c->check_struct (this) && value.sanitize (c));
|
|
}
|
|
bool sanitize (hb_sanitize_context_t *c, const void *base) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (c->check_struct (this) && value.sanitize (c, base));
|
|
}
|
|
|
|
HBGlyphID16 glyph; /* Last GlyphID */
|
|
T value; /* The lookup value (only one) */
|
|
public:
|
|
DEFINE_SIZE_STATIC (2 + T::static_size);
|
|
};
|
|
|
|
template <typename T>
|
|
struct LookupFormat6
|
|
{
|
|
friend struct Lookup<T>;
|
|
|
|
private:
|
|
const T* get_value (hb_codepoint_t glyph_id) const
|
|
{
|
|
const LookupSingle<T> *v = entries.bsearch (glyph_id);
|
|
return v ? &v->value : nullptr;
|
|
}
|
|
|
|
bool sanitize (hb_sanitize_context_t *c) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (entries.sanitize (c));
|
|
}
|
|
bool sanitize (hb_sanitize_context_t *c, const void *base) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (entries.sanitize (c, base));
|
|
}
|
|
|
|
protected:
|
|
HBUINT16 format; /* Format identifier--format = 6 */
|
|
VarSizedBinSearchArrayOf<LookupSingle<T>>
|
|
entries; /* The actual entries, sorted by glyph index. */
|
|
public:
|
|
DEFINE_SIZE_ARRAY (8, entries);
|
|
};
|
|
|
|
template <typename T>
|
|
struct LookupFormat8
|
|
{
|
|
friend struct Lookup<T>;
|
|
|
|
private:
|
|
const T* get_value (hb_codepoint_t glyph_id) const
|
|
{
|
|
return firstGlyph <= glyph_id && glyph_id - firstGlyph < glyphCount ?
|
|
&valueArrayZ[glyph_id - firstGlyph] : nullptr;
|
|
}
|
|
|
|
bool sanitize (hb_sanitize_context_t *c) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (c->check_struct (this) && valueArrayZ.sanitize (c, glyphCount));
|
|
}
|
|
bool sanitize (hb_sanitize_context_t *c, const void *base) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (c->check_struct (this) && valueArrayZ.sanitize (c, glyphCount, base));
|
|
}
|
|
|
|
protected:
|
|
HBUINT16 format; /* Format identifier--format = 8 */
|
|
HBGlyphID16 firstGlyph; /* First glyph index included in the trimmed array. */
|
|
HBUINT16 glyphCount; /* Total number of glyphs (equivalent to the last
|
|
* glyph minus the value of firstGlyph plus 1). */
|
|
UnsizedArrayOf<T>
|
|
valueArrayZ; /* The lookup values (indexed by the glyph index
|
|
* minus the value of firstGlyph). */
|
|
public:
|
|
DEFINE_SIZE_ARRAY (6, valueArrayZ);
|
|
};
|
|
|
|
template <typename T>
|
|
struct LookupFormat10
|
|
{
|
|
friend struct Lookup<T>;
|
|
|
|
private:
|
|
const typename T::type get_value_or_null (hb_codepoint_t glyph_id) const
|
|
{
|
|
if (!(firstGlyph <= glyph_id && glyph_id - firstGlyph < glyphCount))
|
|
return Null (T);
|
|
|
|
const HBUINT8 *p = &valueArrayZ[(glyph_id - firstGlyph) * valueSize];
|
|
|
|
unsigned int v = 0;
|
|
unsigned int count = valueSize;
|
|
for (unsigned int i = 0; i < count; i++)
|
|
v = (v << 8) | *p++;
|
|
|
|
return v;
|
|
}
|
|
|
|
bool sanitize (hb_sanitize_context_t *c) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (c->check_struct (this) &&
|
|
valueSize <= 4 &&
|
|
valueArrayZ.sanitize (c, glyphCount * valueSize));
|
|
}
|
|
|
|
protected:
|
|
HBUINT16 format; /* Format identifier--format = 8 */
|
|
HBUINT16 valueSize; /* Byte size of each value. */
|
|
HBGlyphID16 firstGlyph; /* First glyph index included in the trimmed array. */
|
|
HBUINT16 glyphCount; /* Total number of glyphs (equivalent to the last
|
|
* glyph minus the value of firstGlyph plus 1). */
|
|
UnsizedArrayOf<HBUINT8>
|
|
valueArrayZ; /* The lookup values (indexed by the glyph index
|
|
* minus the value of firstGlyph). */
|
|
public:
|
|
DEFINE_SIZE_ARRAY (8, valueArrayZ);
|
|
};
|
|
|
|
template <typename T>
|
|
struct Lookup
|
|
{
|
|
const T* get_value (hb_codepoint_t glyph_id, unsigned int num_glyphs) const
|
|
{
|
|
switch (u.format) {
|
|
case 0: return u.format0.get_value (glyph_id, num_glyphs);
|
|
case 2: return u.format2.get_value (glyph_id);
|
|
case 4: return u.format4.get_value (glyph_id);
|
|
case 6: return u.format6.get_value (glyph_id);
|
|
case 8: return u.format8.get_value (glyph_id);
|
|
default:return nullptr;
|
|
}
|
|
}
|
|
|
|
const typename T::type get_value_or_null (hb_codepoint_t glyph_id, unsigned int num_glyphs) const
|
|
{
|
|
switch (u.format) {
|
|
/* Format 10 cannot return a pointer. */
|
|
case 10: return u.format10.get_value_or_null (glyph_id);
|
|
default:
|
|
const T *v = get_value (glyph_id, num_glyphs);
|
|
return v ? *v : Null (T);
|
|
}
|
|
}
|
|
|
|
typename T::type get_class (hb_codepoint_t glyph_id,
|
|
unsigned int num_glyphs,
|
|
unsigned int outOfRange) const
|
|
{
|
|
const T *v = get_value (glyph_id, num_glyphs);
|
|
return v ? *v : outOfRange;
|
|
}
|
|
|
|
bool sanitize (hb_sanitize_context_t *c) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
if (!u.format.sanitize (c)) return_trace (false);
|
|
switch (u.format) {
|
|
case 0: return_trace (u.format0.sanitize (c));
|
|
case 2: return_trace (u.format2.sanitize (c));
|
|
case 4: return_trace (u.format4.sanitize (c));
|
|
case 6: return_trace (u.format6.sanitize (c));
|
|
case 8: return_trace (u.format8.sanitize (c));
|
|
case 10: return_trace (u.format10.sanitize (c));
|
|
default:return_trace (true);
|
|
}
|
|
}
|
|
bool sanitize (hb_sanitize_context_t *c, const void *base) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
if (!u.format.sanitize (c)) return_trace (false);
|
|
switch (u.format) {
|
|
case 0: return_trace (u.format0.sanitize (c, base));
|
|
case 2: return_trace (u.format2.sanitize (c, base));
|
|
case 4: return_trace (u.format4.sanitize (c, base));
|
|
case 6: return_trace (u.format6.sanitize (c, base));
|
|
case 8: return_trace (u.format8.sanitize (c, base));
|
|
case 10: return_trace (false); /* We don't support format10 here currently. */
|
|
default:return_trace (true);
|
|
}
|
|
}
|
|
|
|
protected:
|
|
union {
|
|
HBUINT16 format; /* Format identifier */
|
|
LookupFormat0<T> format0;
|
|
LookupFormat2<T> format2;
|
|
LookupFormat4<T> format4;
|
|
LookupFormat6<T> format6;
|
|
LookupFormat8<T> format8;
|
|
LookupFormat10<T> format10;
|
|
} u;
|
|
public:
|
|
DEFINE_SIZE_UNION (2, format);
|
|
};
|
|
DECLARE_NULL_NAMESPACE_BYTES_TEMPLATE1 (AAT, Lookup, 2);
|
|
|
|
enum { DELETED_GLYPH = 0xFFFF };
|
|
|
|
/*
|
|
* (Extended) State Table
|
|
*/
|
|
|
|
template <typename T>
|
|
struct Entry
|
|
{
|
|
bool sanitize (hb_sanitize_context_t *c, unsigned int count) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
/* Note, we don't recurse-sanitize data because we don't access it.
|
|
* That said, in our DEFINE_SIZE_STATIC we access T::static_size,
|
|
* which ensures that data has a simple sanitize(). To be determined
|
|
* if I need to remove that as well.
|
|
*
|
|
* HOWEVER! Because we are a template, our DEFINE_SIZE_STATIC
|
|
* assertion wouldn't be checked, hence the line below. */
|
|
static_assert (T::static_size, "");
|
|
|
|
return_trace (c->check_struct (this));
|
|
}
|
|
|
|
public:
|
|
HBUINT16 newState; /* Byte offset from beginning of state table
|
|
* to the new state. Really?!?! Or just state
|
|
* number? The latter in morx for sure. */
|
|
HBUINT16 flags; /* Table specific. */
|
|
T data; /* Optional offsets to per-glyph tables. */
|
|
public:
|
|
DEFINE_SIZE_STATIC (4 + T::static_size);
|
|
};
|
|
|
|
template <>
|
|
struct Entry<void>
|
|
{
|
|
bool sanitize (hb_sanitize_context_t *c, unsigned int count /*XXX Unused?*/) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (c->check_struct (this));
|
|
}
|
|
|
|
public:
|
|
HBUINT16 newState; /* Byte offset from beginning of state table to the new state. */
|
|
HBUINT16 flags; /* Table specific. */
|
|
public:
|
|
DEFINE_SIZE_STATIC (4);
|
|
};
|
|
|
|
template <typename Types, typename Extra>
|
|
struct StateTable
|
|
{
|
|
typedef typename Types::HBUINT HBUINT;
|
|
typedef typename Types::HBUSHORT HBUSHORT;
|
|
typedef typename Types::ClassTypeNarrow ClassType;
|
|
|
|
enum State
|
|
{
|
|
STATE_START_OF_TEXT = 0,
|
|
STATE_START_OF_LINE = 1,
|
|
};
|
|
enum Class
|
|
{
|
|
CLASS_END_OF_TEXT = 0,
|
|
CLASS_OUT_OF_BOUNDS = 1,
|
|
CLASS_DELETED_GLYPH = 2,
|
|
CLASS_END_OF_LINE = 3,
|
|
};
|
|
|
|
int new_state (unsigned int newState) const
|
|
{ return Types::extended ? newState : ((int) newState - (int) stateArrayTable) / (int) nClasses; }
|
|
|
|
unsigned int get_class (hb_codepoint_t glyph_id, unsigned int num_glyphs) const
|
|
{
|
|
if (unlikely (glyph_id == DELETED_GLYPH)) return CLASS_DELETED_GLYPH;
|
|
return (this+classTable).get_class (glyph_id, num_glyphs, 1);
|
|
}
|
|
|
|
const Entry<Extra> *get_entries () const
|
|
{ return (this+entryTable).arrayZ; }
|
|
|
|
const Entry<Extra> &get_entry (int state, unsigned int klass) const
|
|
{
|
|
if (unlikely (klass >= nClasses))
|
|
klass = StateTable::CLASS_OUT_OF_BOUNDS;
|
|
|
|
const HBUSHORT *states = (this+stateArrayTable).arrayZ;
|
|
const Entry<Extra> *entries = (this+entryTable).arrayZ;
|
|
|
|
unsigned int entry = states[state * nClasses + klass];
|
|
DEBUG_MSG (APPLY, nullptr, "e%u", entry);
|
|
|
|
return entries[entry];
|
|
}
|
|
|
|
bool sanitize (hb_sanitize_context_t *c,
|
|
unsigned int *num_entries_out = nullptr) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
if (unlikely (!(c->check_struct (this) &&
|
|
nClasses >= 4 /* Ensure pre-defined classes fit. */ &&
|
|
classTable.sanitize (c, this)))) return_trace (false);
|
|
|
|
const HBUSHORT *states = (this+stateArrayTable).arrayZ;
|
|
const Entry<Extra> *entries = (this+entryTable).arrayZ;
|
|
|
|
unsigned int num_classes = nClasses;
|
|
if (unlikely (hb_unsigned_mul_overflows (num_classes, states[0].static_size)))
|
|
return_trace (false);
|
|
unsigned int row_stride = num_classes * states[0].static_size;
|
|
|
|
/* Apple 'kern' table has this peculiarity:
|
|
*
|
|
* "Because the stateTableOffset in the state table header is (strictly
|
|
* speaking) redundant, some 'kern' tables use it to record an initial
|
|
* state where that should not be StartOfText. To determine if this is
|
|
* done, calculate what the stateTableOffset should be. If it's different
|
|
* from the actual stateTableOffset, use it as the initial state."
|
|
*
|
|
* We implement this by calling the initial state zero, but allow *negative*
|
|
* states if the start state indeed was not the first state. Since the code
|
|
* is shared, this will also apply to 'mort' table. The 'kerx' / 'morx'
|
|
* tables are not affected since those address states by index, not offset.
|
|
*/
|
|
|
|
int min_state = 0;
|
|
int max_state = 0;
|
|
unsigned int num_entries = 0;
|
|
|
|
int state_pos = 0;
|
|
int state_neg = 0;
|
|
unsigned int entry = 0;
|
|
while (min_state < state_neg || state_pos <= max_state)
|
|
{
|
|
if (min_state < state_neg)
|
|
{
|
|
/* Negative states. */
|
|
if (unlikely (hb_unsigned_mul_overflows (min_state, num_classes)))
|
|
return_trace (false);
|
|
if (unlikely (!c->check_range (&states[min_state * num_classes],
|
|
-min_state,
|
|
row_stride)))
|
|
return_trace (false);
|
|
if ((c->max_ops -= state_neg - min_state) <= 0)
|
|
return_trace (false);
|
|
{ /* Sweep new states. */
|
|
const HBUSHORT *stop = &states[min_state * num_classes];
|
|
if (unlikely (stop > states))
|
|
return_trace (false);
|
|
for (const HBUSHORT *p = states; stop < p; p--)
|
|
num_entries = hb_max (num_entries, *(p - 1) + 1u);
|
|
state_neg = min_state;
|
|
}
|
|
}
|
|
|
|
if (state_pos <= max_state)
|
|
{
|
|
/* Positive states. */
|
|
if (unlikely (!c->check_range (states,
|
|
max_state + 1,
|
|
row_stride)))
|
|
return_trace (false);
|
|
if ((c->max_ops -= max_state - state_pos + 1) <= 0)
|
|
return_trace (false);
|
|
{ /* Sweep new states. */
|
|
if (unlikely (hb_unsigned_mul_overflows ((max_state + 1), num_classes)))
|
|
return_trace (false);
|
|
const HBUSHORT *stop = &states[(max_state + 1) * num_classes];
|
|
if (unlikely (stop < states))
|
|
return_trace (false);
|
|
for (const HBUSHORT *p = &states[state_pos * num_classes]; p < stop; p++)
|
|
num_entries = hb_max (num_entries, *p + 1u);
|
|
state_pos = max_state + 1;
|
|
}
|
|
}
|
|
|
|
if (unlikely (!c->check_array (entries, num_entries)))
|
|
return_trace (false);
|
|
if ((c->max_ops -= num_entries - entry) <= 0)
|
|
return_trace (false);
|
|
{ /* Sweep new entries. */
|
|
const Entry<Extra> *stop = &entries[num_entries];
|
|
for (const Entry<Extra> *p = &entries[entry]; p < stop; p++)
|
|
{
|
|
int newState = new_state (p->newState);
|
|
min_state = hb_min (min_state, newState);
|
|
max_state = hb_max (max_state, newState);
|
|
}
|
|
entry = num_entries;
|
|
}
|
|
}
|
|
|
|
if (num_entries_out)
|
|
*num_entries_out = num_entries;
|
|
|
|
return_trace (true);
|
|
}
|
|
|
|
protected:
|
|
HBUINT nClasses; /* Number of classes, which is the number of indices
|
|
* in a single line in the state array. */
|
|
NNOffsetTo<ClassType, HBUINT>
|
|
classTable; /* Offset to the class table. */
|
|
NNOffsetTo<UnsizedArrayOf<HBUSHORT>, HBUINT>
|
|
stateArrayTable;/* Offset to the state array. */
|
|
NNOffsetTo<UnsizedArrayOf<Entry<Extra>>, HBUINT>
|
|
entryTable; /* Offset to the entry array. */
|
|
|
|
public:
|
|
DEFINE_SIZE_STATIC (4 * sizeof (HBUINT));
|
|
};
|
|
|
|
template <typename HBUCHAR>
|
|
struct ClassTable
|
|
{
|
|
unsigned int get_class (hb_codepoint_t glyph_id, unsigned int outOfRange) const
|
|
{
|
|
unsigned int i = glyph_id - firstGlyph;
|
|
return i >= classArray.len ? outOfRange : classArray.arrayZ[i];
|
|
}
|
|
unsigned int get_class (hb_codepoint_t glyph_id,
|
|
unsigned int num_glyphs HB_UNUSED,
|
|
unsigned int outOfRange) const
|
|
{
|
|
return get_class (glyph_id, outOfRange);
|
|
}
|
|
bool sanitize (hb_sanitize_context_t *c) const
|
|
{
|
|
TRACE_SANITIZE (this);
|
|
return_trace (c->check_struct (this) && classArray.sanitize (c));
|
|
}
|
|
protected:
|
|
HBGlyphID16 firstGlyph; /* First glyph index included in the trimmed array. */
|
|
Array16Of<HBUCHAR> classArray; /* The class codes (indexed by glyph index minus
|
|
* firstGlyph). */
|
|
public:
|
|
DEFINE_SIZE_ARRAY (4, classArray);
|
|
};
|
|
|
|
struct ObsoleteTypes
|
|
{
|
|
static constexpr bool extended = false;
|
|
typedef HBUINT16 HBUINT;
|
|
typedef HBUINT8 HBUSHORT;
|
|
typedef ClassTable<HBUINT8> ClassTypeNarrow;
|
|
typedef ClassTable<HBUINT16> ClassTypeWide;
|
|
|
|
template <typename T>
|
|
static unsigned int offsetToIndex (unsigned int offset,
|
|
const void *base,
|
|
const T *array)
|
|
{
|
|
/* https://github.com/harfbuzz/harfbuzz/issues/3483 */
|
|
/* If offset is less than base, return an offset that would
|
|
* result in an address half a 32bit address-space away,
|
|
* to make sure sanitize fails even on 32bit builds. */
|
|
if (unlikely (offset < unsigned ((const char *) array - (const char *) base)))
|
|
return INT_MAX / T::static_size;
|
|
|
|
/* https://github.com/harfbuzz/harfbuzz/issues/2816 */
|
|
return (offset - unsigned ((const char *) array - (const char *) base)) / T::static_size;
|
|
}
|
|
template <typename T>
|
|
static unsigned int byteOffsetToIndex (unsigned int offset,
|
|
const void *base,
|
|
const T *array)
|
|
{
|
|
return offsetToIndex (offset, base, array);
|
|
}
|
|
template <typename T>
|
|
static unsigned int wordOffsetToIndex (unsigned int offset,
|
|
const void *base,
|
|
const T *array)
|
|
{
|
|
return offsetToIndex (2 * offset, base, array);
|
|
}
|
|
};
|
|
struct ExtendedTypes
|
|
{
|
|
static constexpr bool extended = true;
|
|
typedef HBUINT32 HBUINT;
|
|
typedef HBUINT16 HBUSHORT;
|
|
typedef Lookup<HBUINT16> ClassTypeNarrow;
|
|
typedef Lookup<HBUINT16> ClassTypeWide;
|
|
|
|
template <typename T>
|
|
static unsigned int offsetToIndex (unsigned int offset,
|
|
const void *base HB_UNUSED,
|
|
const T *array HB_UNUSED)
|
|
{
|
|
return offset;
|
|
}
|
|
template <typename T>
|
|
static unsigned int byteOffsetToIndex (unsigned int offset,
|
|
const void *base HB_UNUSED,
|
|
const T *array HB_UNUSED)
|
|
{
|
|
return offset / 2;
|
|
}
|
|
template <typename T>
|
|
static unsigned int wordOffsetToIndex (unsigned int offset,
|
|
const void *base HB_UNUSED,
|
|
const T *array HB_UNUSED)
|
|
{
|
|
return offset;
|
|
}
|
|
};
|
|
|
|
template <typename Types, typename EntryData>
|
|
struct StateTableDriver
|
|
{
|
|
using StateTableT = StateTable<Types, EntryData>;
|
|
using EntryT = Entry<EntryData>;
|
|
|
|
StateTableDriver (const StateTableT &machine_,
|
|
hb_buffer_t *buffer_,
|
|
hb_face_t *face_) :
|
|
machine (machine_),
|
|
buffer (buffer_),
|
|
num_glyphs (face_->get_num_glyphs ()) {}
|
|
|
|
template <typename context_t>
|
|
void drive (context_t *c, hb_aat_apply_context_t *ac)
|
|
{
|
|
if (!c->in_place)
|
|
buffer->clear_output ();
|
|
|
|
int state = StateTableT::STATE_START_OF_TEXT;
|
|
// If there's only one range, we already checked the flag.
|
|
auto *last_range = ac->range_flags && (ac->range_flags->length > 1) ? &(*ac->range_flags)[0] : nullptr;
|
|
for (buffer->idx = 0; buffer->successful;)
|
|
{
|
|
/* This block is copied in NoncontextualSubtable::apply. Keep in sync. */
|
|
if (last_range)
|
|
{
|
|
auto *range = last_range;
|
|
if (buffer->idx < buffer->len)
|
|
{
|
|
unsigned cluster = buffer->cur().cluster;
|
|
while (cluster < range->cluster_first)
|
|
range--;
|
|
while (cluster > range->cluster_last)
|
|
range++;
|
|
|
|
|
|
last_range = range;
|
|
}
|
|
if (!(range->flags & ac->subtable_flags))
|
|
{
|
|
if (buffer->idx == buffer->len || unlikely (!buffer->successful))
|
|
break;
|
|
|
|
state = StateTableT::STATE_START_OF_TEXT;
|
|
(void) buffer->next_glyph ();
|
|
continue;
|
|
}
|
|
}
|
|
|
|
unsigned int klass = buffer->idx < buffer->len ?
|
|
machine.get_class (buffer->cur().codepoint, num_glyphs) :
|
|
(unsigned) StateTableT::CLASS_END_OF_TEXT;
|
|
DEBUG_MSG (APPLY, nullptr, "c%u at %u", klass, buffer->idx);
|
|
const EntryT &entry = machine.get_entry (state, klass);
|
|
const int next_state = machine.new_state (entry.newState);
|
|
|
|
/* Conditions under which it's guaranteed safe-to-break before current glyph:
|
|
*
|
|
* 1. There was no action in this transition; and
|
|
*
|
|
* 2. If we break before current glyph, the results will be the same. That
|
|
* is guaranteed if:
|
|
*
|
|
* 2a. We were already in start-of-text state; or
|
|
*
|
|
* 2b. We are epsilon-transitioning to start-of-text state; or
|
|
*
|
|
* 2c. Starting from start-of-text state seeing current glyph:
|
|
*
|
|
* 2c'. There won't be any actions; and
|
|
*
|
|
* 2c". We would end up in the same state that we were going to end up
|
|
* in now, including whether epsilon-transitioning.
|
|
*
|
|
* and
|
|
*
|
|
* 3. If we break before current glyph, there won't be any end-of-text action
|
|
* after previous glyph.
|
|
*
|
|
* This triples the transitions we need to look up, but is worth returning
|
|
* granular unsafe-to-break results. See eg.:
|
|
*
|
|
* https://github.com/harfbuzz/harfbuzz/issues/2860
|
|
*/
|
|
const EntryT *wouldbe_entry;
|
|
bool safe_to_break =
|
|
/* 1. */
|
|
!c->is_actionable (this, entry)
|
|
&&
|
|
/* 2. */
|
|
(
|
|
/* 2a. */
|
|
state == StateTableT::STATE_START_OF_TEXT
|
|
||
|
|
/* 2b. */
|
|
(
|
|
(entry.flags & context_t::DontAdvance) &&
|
|
next_state == StateTableT::STATE_START_OF_TEXT
|
|
)
|
|
||
|
|
/* 2c. */
|
|
(
|
|
wouldbe_entry = &machine.get_entry (StateTableT::STATE_START_OF_TEXT, klass)
|
|
,
|
|
/* 2c'. */
|
|
!c->is_actionable (this, *wouldbe_entry)
|
|
&&
|
|
/* 2c". */
|
|
(
|
|
next_state == machine.new_state (wouldbe_entry->newState)
|
|
&&
|
|
(entry.flags & context_t::DontAdvance) == (wouldbe_entry->flags & context_t::DontAdvance)
|
|
)
|
|
)
|
|
)
|
|
&&
|
|
/* 3. */
|
|
!c->is_actionable (this, machine.get_entry (state, StateTableT::CLASS_END_OF_TEXT))
|
|
;
|
|
|
|
if (!safe_to_break && buffer->backtrack_len () && buffer->idx < buffer->len)
|
|
buffer->unsafe_to_break_from_outbuffer (buffer->backtrack_len () - 1, buffer->idx + 1);
|
|
|
|
c->transition (this, entry);
|
|
|
|
state = next_state;
|
|
DEBUG_MSG (APPLY, nullptr, "s%d", state);
|
|
|
|
if (buffer->idx == buffer->len || unlikely (!buffer->successful))
|
|
break;
|
|
|
|
if (!(entry.flags & context_t::DontAdvance) || buffer->max_ops-- <= 0)
|
|
(void) buffer->next_glyph ();
|
|
}
|
|
|
|
if (!c->in_place)
|
|
buffer->sync ();
|
|
}
|
|
|
|
public:
|
|
const StateTableT &machine;
|
|
hb_buffer_t *buffer;
|
|
unsigned int num_glyphs;
|
|
};
|
|
|
|
|
|
} /* namespace AAT */
|
|
|
|
|
|
#endif /* HB_AAT_LAYOUT_COMMON_HH */
|