/* * Copyright © 2007,2008,2009,2010 Red Hat, Inc. * Copyright © 2012 Google, Inc. * * This is part of HarfBuzz, a text shaping library. * * Permission is hereby granted, without written agreement and without * license or royalty fees, to use, copy, modify, and distribute this * software and its documentation for any purpose, provided that the * above copyright notice and the following two paragraphs appear in * all copies of this software. * * IN NO EVENT SHALL THE COPYRIGHT HOLDER BE LIABLE TO ANY PARTY FOR * DIRECT, INDIRECT, SPECIAL, INCIDENTAL, OR CONSEQUENTIAL DAMAGES * ARISING OUT OF THE USE OF THIS SOFTWARE AND ITS DOCUMENTATION, EVEN * IF THE COPYRIGHT HOLDER HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH * DAMAGE. * * THE COPYRIGHT HOLDER SPECIFICALLY DISCLAIMS ANY WARRANTIES, INCLUDING, * BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND * FITNESS FOR A PARTICULAR PURPOSE. THE SOFTWARE PROVIDED HEREUNDER IS * ON AN "AS IS" BASIS, AND THE COPYRIGHT HOLDER HAS NO OBLIGATION TO * PROVIDE MAINTENANCE, SUPPORT, UPDATES, ENHANCEMENTS, OR MODIFICATIONS. * * Red Hat Author(s): Behdad Esfahbod * Google Author(s): Behdad Esfahbod */ #ifndef HB_OPEN_TYPE_HH #define HB_OPEN_TYPE_HH #include "hb.hh" #include "hb-blob.hh" #include "hb-face.hh" #include "hb-machinery.hh" #include "hb-subset.hh" namespace OT { /* * * The OpenType Font File: Data Types */ /* "The following data types are used in the OpenType font file. * All OpenType fonts use Motorola-style byte ordering (Big Endian):" */ /* * Int types */ /* Integer types in big-endian order and no alignment requirement */ template struct IntType { typedef Type type; typedef hb_conditional wide_type; IntType& operator = (wide_type i) { v = i; return *this; } operator wide_type () const { return v; } bool operator == (const IntType &o) const { return (Type) v == (Type) o.v; } bool operator != (const IntType &o) const { return !(*this == o); } IntType& operator += (unsigned count) { *this = *this + count; return *this; } IntType& operator -= (unsigned count) { *this = *this - count; return *this; } IntType& operator ++ () { *this += 1; return *this; } IntType& operator -- () { *this -= 1; return *this; } IntType operator ++ (int) { IntType c (*this); ++*this; return c; } IntType operator -- (int) { IntType c (*this); --*this; return c; } HB_INTERNAL static int cmp (const IntType *a, const IntType *b) { return b->cmp (*a); } HB_INTERNAL static int cmp (const void *a, const void *b) { IntType *pa = (IntType *) a; IntType *pb = (IntType *) b; return pb->cmp (*pa); } template int cmp (Type2 a) const { Type b = v; if (sizeof (Type) < sizeof (int) && sizeof (Type2) < sizeof (int)) return (int) a - (int) b; else return a < b ? -1 : a == b ? 0 : +1; } bool sanitize (hb_sanitize_context_t *c) const { TRACE_SANITIZE (this); return_trace (likely (c->check_struct (this))); } protected: BEInt v; public: DEFINE_SIZE_STATIC (Size); }; typedef IntType HBUINT8; /* 8-bit unsigned integer. */ typedef IntType HBINT8; /* 8-bit signed integer. */ typedef IntType HBUINT16; /* 16-bit unsigned integer. */ typedef IntType HBINT16; /* 16-bit signed integer. */ typedef IntType HBUINT32; /* 32-bit unsigned integer. */ typedef IntType HBINT32; /* 32-bit signed integer. */ /* Note: we cannot defined a signed HBINT24 because there's no corresponding C type. * Works for unsigned, but not signed, since we rely on compiler for sign-extension. */ typedef IntType HBUINT24; /* 24-bit unsigned integer. */ /* 16-bit signed integer (HBINT16) that describes a quantity in FUnits. */ typedef HBINT16 FWORD; /* 32-bit signed integer (HBINT32) that describes a quantity in FUnits. */ typedef HBINT32 FWORD32; /* 16-bit unsigned integer (HBUINT16) that describes a quantity in FUnits. */ typedef HBUINT16 UFWORD; /* 16-bit signed fixed number with the low 14 bits of fraction (2.14). */ struct F2DOT14 : HBINT16 { F2DOT14& operator = (uint16_t i ) { HBINT16::operator= (i); return *this; } // 16384 means 1<<14 float to_float () const { return ((int32_t) v) / 16384.f; } void set_float (float f) { v = roundf (f * 16384.f); } public: DEFINE_SIZE_STATIC (2); }; /* 32-bit signed fixed-point number (16.16). */ struct HBFixed : HBINT32 { HBFixed& operator = (uint32_t i) { HBINT32::operator= (i); return *this; } // 65536 means 1<<16 float to_float () const { return ((int32_t) v) / 65536.f; } void set_float (float f) { v = roundf (f * 65536.f); } public: DEFINE_SIZE_STATIC (4); }; /* Date represented in number of seconds since 12:00 midnight, January 1, * 1904. The value is represented as a signed 64-bit integer. */ struct LONGDATETIME { bool sanitize (hb_sanitize_context_t *c) const { TRACE_SANITIZE (this); return_trace (likely (c->check_struct (this))); } protected: HBINT32 major; HBUINT32 minor; public: DEFINE_SIZE_STATIC (8); }; /* Array of four uint8s (length = 32 bits) used to identify a script, language * system, feature, or baseline */ struct Tag : HBUINT32 { Tag& operator = (hb_tag_t i) { HBUINT32::operator= (i); return *this; } /* What the char* converters return is NOT nul-terminated. Print using "%.4s" */ operator const char* () const { return reinterpret_cast (&this->v); } operator char* () { return reinterpret_cast (&this->v); } public: DEFINE_SIZE_STATIC (4); }; /* Glyph index number, same as uint16 (length = 16 bits) */ struct HBGlyphID : HBUINT16 { HBGlyphID& operator = (uint16_t i) { HBUINT16::operator= (i); return *this; } }; /* Script/language-system/feature index */ struct Index : HBUINT16 { static constexpr unsigned NOT_FOUND_INDEX = 0xFFFFu; Index& operator = (uint16_t i) { HBUINT16::operator= (i); return *this; } }; DECLARE_NULL_NAMESPACE_BYTES (OT, Index); typedef Index NameID; /* Offset, Null offset = 0 */ template struct Offset : Type { Offset& operator = (typename Type::type i) { Type::operator= (i); return *this; } typedef Type type; bool is_null () const { return has_null && 0 == *this; } void *serialize (hb_serialize_context_t *c, const void *base) { void *t = c->start_embed (); c->check_assign (*this, (unsigned) ((char *) t - (char *) base)); return t; } public: DEFINE_SIZE_STATIC (sizeof (Type)); }; typedef Offset Offset16; typedef Offset Offset32; /* CheckSum */ struct CheckSum : HBUINT32 { CheckSum& operator = (uint32_t i) { HBUINT32::operator= (i); return *this; } /* This is reference implementation from the spec. */ static uint32_t CalcTableChecksum (const HBUINT32 *Table, uint32_t Length) { uint32_t Sum = 0L; assert (0 == (Length & 3)); const HBUINT32 *EndPtr = Table + Length / HBUINT32::static_size; while (Table < EndPtr) Sum += *Table++; return Sum; } /* Note: data should be 4byte aligned and have 4byte padding at the end. */ void set_for_data (const void *data, unsigned int length) { *this = CalcTableChecksum ((const HBUINT32 *) data, length); } public: DEFINE_SIZE_STATIC (4); }; /* * Version Numbers */ template struct FixedVersion { uint32_t to_int () const { return (major << (sizeof (FixedType) * 8)) + minor; } bool sanitize (hb_sanitize_context_t *c) const { TRACE_SANITIZE (this); return_trace (c->check_struct (this)); } FixedType major; FixedType minor; public: DEFINE_SIZE_STATIC (2 * sizeof (FixedType)); }; /* * Template subclasses of Offset that do the dereferencing. * Use: (base+offset) */ template struct _hb_has_null { static const Type *get_null () { return nullptr; } static Type *get_crap () { return nullptr; } }; template struct _hb_has_null { static const Type *get_null () { return &Null (Type); } static Type *get_crap () { return &Crap (Type); } }; template struct OffsetTo : Offset { HB_DELETE_COPY_ASSIGN (OffsetTo); OffsetTo () = default; OffsetTo& operator = (typename OffsetType::type i) { OffsetType::operator= (i); return *this; } const Type& operator () (const void *base) const { if (unlikely (this->is_null ())) return *_hb_has_null::get_null (); return StructAtOffset (base, *this); } Type& operator () (void *base) const { if (unlikely (this->is_null ())) return *_hb_has_null::get_crap (); return StructAtOffset (base, *this); } template friend const Type& operator + (const Base &base, const OffsetTo &offset) { return offset ((const void *) base); } template friend const Type& operator + (const OffsetTo &offset, const Base &base) { return offset ((const void *) base); } template friend Type& operator + (Base &&base, OffsetTo &offset) { return offset ((void *) base); } template friend Type& operator + (OffsetTo &offset, Base &&base) { return offset ((void *) base); } Type& serialize (hb_serialize_context_t *c, const void *base) { return * (Type *) Offset::serialize (c, base); } template bool serialize_subset (hb_subset_context_t *c, const OffsetTo& src, const void *src_base, const void *dst_base, Ts&&... ds) { *this = 0; if (src.is_null ()) return false; auto *s = c->serializer; s->push (); bool ret = c->dispatch (src_base+src, hb_forward (ds)...); if (ret || !has_null) s->add_link (*this, s->pop_pack (), dst_base); else s->pop_discard (); return ret; } /* TODO: Somehow merge this with previous function into a serialize_dispatch(). */ template bool serialize_copy (hb_serialize_context_t *c, const OffsetTo& src, const void *src_base, const void *dst_base, Ts&&... ds) { *this = 0; if (src.is_null ()) return false; c->push (); bool ret = c->copy (src_base+src, hb_forward (ds)...); c->add_link (*this, c->pop_pack (), dst_base); return ret; } bool sanitize_shallow (hb_sanitize_context_t *c, const void *base) const { TRACE_SANITIZE (this); if (unlikely (!c->check_struct (this))) return_trace (false); if (unlikely (this->is_null ())) return_trace (true); if (unlikely (!c->check_range (base, *this))) return_trace (false); return_trace (true); } template bool sanitize (hb_sanitize_context_t *c, const void *base, Ts&&... ds) const { TRACE_SANITIZE (this); return_trace (sanitize_shallow (c, base) && (this->is_null () || c->dispatch (StructAtOffset (base, *this), hb_forward (ds)...) || neuter (c))); } /* Set the offset to Null */ bool neuter (hb_sanitize_context_t *c) const { if (!has_null) return false; return c->try_set (this, 0); } DEFINE_SIZE_STATIC (sizeof (OffsetType)); }; /* Partial specializations. */ template using LOffsetTo = OffsetTo; template using NNOffsetTo = OffsetTo; template using LNNOffsetTo = LOffsetTo; /* * Array Types */ template struct UnsizedArrayOf { typedef Type item_t; static constexpr unsigned item_size = hb_static_size (Type); HB_DELETE_CREATE_COPY_ASSIGN (UnsizedArrayOf); const Type& operator [] (int i_) const { unsigned int i = (unsigned int) i_; const Type *p = &arrayZ[i]; if (unlikely (p < arrayZ)) return Null (Type); /* Overflowed. */ return *p; } Type& operator [] (int i_) { unsigned int i = (unsigned int) i_; Type *p = &arrayZ[i]; if (unlikely (p < arrayZ)) return Crap (Type); /* Overflowed. */ return *p; } unsigned int get_size (unsigned int len) const { return len * Type::static_size; } template operator T * () { return arrayZ; } template operator const T * () const { return arrayZ; } hb_array_t as_array (unsigned int len) { return hb_array (arrayZ, len); } hb_array_t as_array (unsigned int len) const { return hb_array (arrayZ, len); } operator hb_array_t< Type> () { return as_array (); } operator hb_array_t () const { return as_array (); } template Type &lsearch (unsigned int len, const T &x, Type ¬_found = Crap (Type)) { return *as_array (len).lsearch (x, ¬_found); } template const Type &lsearch (unsigned int len, const T &x, const Type ¬_found = Null (Type)) const { return *as_array (len).lsearch (x, ¬_found); } void qsort (unsigned int len, unsigned int start = 0, unsigned int end = (unsigned int) -1) { as_array (len).qsort (start, end); } bool serialize (hb_serialize_context_t *c, unsigned int items_len) { TRACE_SERIALIZE (this); if (unlikely (!c->extend (*this, items_len))) return_trace (false); return_trace (true); } template bool serialize (hb_serialize_context_t *c, Iterator items) { TRACE_SERIALIZE (this); unsigned count = items.len (); if (unlikely (!serialize (c, count))) return_trace (false); /* TODO Umm. Just exhaust the iterator instead? Being extra * cautious right now.. */ for (unsigned i = 0; i < count; i++, ++items) arrayZ[i] = *items; return_trace (true); } UnsizedArrayOf* copy (hb_serialize_context_t *c, unsigned count) const { TRACE_SERIALIZE (this); auto *out = c->start_embed (this); if (unlikely (!as_array (count).copy (c))) return_trace (nullptr); return_trace (out); } template bool sanitize (hb_sanitize_context_t *c, unsigned int count, Ts&&... ds) const { TRACE_SANITIZE (this); if (unlikely (!sanitize_shallow (c, count))) return_trace (false); if (!sizeof... (Ts) && hb_is_trivially_copyable (Type)) return_trace (true); for (unsigned int i = 0; i < count; i++) if (unlikely (!c->dispatch (arrayZ[i], hb_forward (ds)...))) return_trace (false); return_trace (true); } bool sanitize_shallow (hb_sanitize_context_t *c, unsigned int count) const { TRACE_SANITIZE (this); return_trace (c->check_array (arrayZ, count)); } public: Type arrayZ[HB_VAR_ARRAY]; public: DEFINE_SIZE_UNBOUNDED (0); }; /* Unsized array of offset's */ template using UnsizedOffsetArrayOf = UnsizedArrayOf>; /* Unsized array of offsets relative to the beginning of the array itself. */ template struct UnsizedOffsetListOf : UnsizedOffsetArrayOf { const Type& operator [] (int i_) const { unsigned int i = (unsigned int) i_; const OffsetTo *p = &this->arrayZ[i]; if (unlikely (p < this->arrayZ)) return Null (Type); /* Overflowed. */ return this+*p; } Type& operator [] (int i_) { unsigned int i = (unsigned int) i_; const OffsetTo *p = &this->arrayZ[i]; if (unlikely (p < this->arrayZ)) return Crap (Type); /* Overflowed. */ return this+*p; } template bool sanitize (hb_sanitize_context_t *c, unsigned int count, Ts&&... ds) const { TRACE_SANITIZE (this); return_trace ((UnsizedOffsetArrayOf ::sanitize (c, count, this, hb_forward (ds)...))); } }; /* An array with sorted elements. Supports binary searching. */ template struct SortedUnsizedArrayOf : UnsizedArrayOf { hb_sorted_array_t as_array (unsigned int len) { return hb_sorted_array (this->arrayZ, len); } hb_sorted_array_t as_array (unsigned int len) const { return hb_sorted_array (this->arrayZ, len); } operator hb_sorted_array_t () { return as_array (); } operator hb_sorted_array_t () const { return as_array (); } template Type &bsearch (unsigned int len, const T &x, Type ¬_found = Crap (Type)) { return *as_array (len).bsearch (x, ¬_found); } template const Type &bsearch (unsigned int len, const T &x, const Type ¬_found = Null (Type)) const { return *as_array (len).bsearch (x, ¬_found); } template bool bfind (unsigned int len, const T &x, unsigned int *i = nullptr, hb_bfind_not_found_t not_found = HB_BFIND_NOT_FOUND_DONT_STORE, unsigned int to_store = (unsigned int) -1) const { return as_array (len).bfind (x, i, not_found, to_store); } }; /* An array with a number of elements. */ template struct ArrayOf { typedef Type item_t; static constexpr unsigned item_size = hb_static_size (Type); HB_DELETE_CREATE_COPY_ASSIGN (ArrayOf); const Type& operator [] (int i_) const { unsigned int i = (unsigned int) i_; if (unlikely (i >= len)) return Null (Type); return arrayZ[i]; } Type& operator [] (int i_) { unsigned int i = (unsigned int) i_; if (unlikely (i >= len)) return Crap (Type); return arrayZ[i]; } unsigned int get_size () const { return len.static_size + len * Type::static_size; } explicit operator bool () const { return len; } void pop () { len--; } hb_array_t< Type> as_array () { return hb_array (arrayZ, len); } hb_array_t as_array () const { return hb_array (arrayZ, len); } /* Iterator. */ typedef hb_array_t iter_t; typedef hb_array_t< Type> writer_t; iter_t iter () const { return as_array (); } writer_t writer () { return as_array (); } operator iter_t () const { return iter (); } operator writer_t () { return writer (); } hb_array_t sub_array (unsigned int start_offset, unsigned int count) const { return as_array ().sub_array (start_offset, count); } hb_array_t sub_array (unsigned int start_offset, unsigned int *count = nullptr /* IN/OUT */) const { return as_array ().sub_array (start_offset, count); } hb_array_t sub_array (unsigned int start_offset, unsigned int count) { return as_array ().sub_array (start_offset, count); } hb_array_t sub_array (unsigned int start_offset, unsigned int *count = nullptr /* IN/OUT */) { return as_array ().sub_array (start_offset, count); } bool serialize (hb_serialize_context_t *c, unsigned int items_len) { TRACE_SERIALIZE (this); if (unlikely (!c->extend_min (*this))) return_trace (false); c->check_assign (len, items_len); if (unlikely (!c->extend (*this))) return_trace (false); return_trace (true); } template bool serialize (hb_serialize_context_t *c, Iterator items) { TRACE_SERIALIZE (this); unsigned count = items.len (); if (unlikely (!serialize (c, count))) return_trace (false); /* TODO Umm. Just exhaust the iterator instead? Being extra * cautious right now.. */ for (unsigned i = 0; i < count; i++, ++items) arrayZ[i] = *items; return_trace (true); } Type* serialize_append (hb_serialize_context_t *c) { TRACE_SERIALIZE (this); len++; if (unlikely (!len || !c->extend (*this))) { len--; return_trace (nullptr); } return_trace (&arrayZ[len - 1]); } ArrayOf* copy (hb_serialize_context_t *c) const { TRACE_SERIALIZE (this); auto *out = c->start_embed (this); if (unlikely (!c->extend_min (out))) return_trace (nullptr); c->check_assign (out->len, len); if (unlikely (!as_array ().copy (c))) return_trace (nullptr); return_trace (out); } template bool sanitize (hb_sanitize_context_t *c, Ts&&... ds) const { TRACE_SANITIZE (this); if (unlikely (!sanitize_shallow (c))) return_trace (false); if (!sizeof... (Ts) && hb_is_trivially_copyable (Type)) return_trace (true); unsigned int count = len; for (unsigned int i = 0; i < count; i++) if (unlikely (!c->dispatch (arrayZ[i], hb_forward (ds)...))) return_trace (false); return_trace (true); } template Type &lsearch (const T &x, Type ¬_found = Crap (Type)) { return *as_array ().lsearch (x, ¬_found); } template const Type &lsearch (const T &x, const Type ¬_found = Null (Type)) const { return *as_array ().lsearch (x, ¬_found); } void qsort (unsigned int start = 0, unsigned int end = (unsigned int) -1) { as_array ().qsort (start, end); } bool sanitize_shallow (hb_sanitize_context_t *c) const { TRACE_SANITIZE (this); return_trace (len.sanitize (c) && c->check_array (arrayZ, len)); } public: LenType len; Type arrayZ[HB_VAR_ARRAY]; public: DEFINE_SIZE_ARRAY (sizeof (LenType), arrayZ); }; template using LArrayOf = ArrayOf; using PString = ArrayOf; /* Array of Offset's */ template using OffsetArrayOf = ArrayOf>; template using LOffsetArrayOf = ArrayOf>; template using LOffsetLArrayOf = ArrayOf, HBUINT32>; /* Array of offsets relative to the beginning of the array itself. */ template struct OffsetListOf : OffsetArrayOf { const Type& operator [] (int i_) const { unsigned int i = (unsigned int) i_; if (unlikely (i >= this->len)) return Null (Type); return this+this->arrayZ[i]; } const Type& operator [] (int i_) { unsigned int i = (unsigned int) i_; if (unlikely (i >= this->len)) return Crap (Type); return this+this->arrayZ[i]; } bool subset (hb_subset_context_t *c) const { TRACE_SUBSET (this); struct OffsetListOf *out = c->serializer->embed (*this); if (unlikely (!out)) return_trace (false); unsigned int count = this->len; for (unsigned int i = 0; i < count; i++) out->arrayZ[i].serialize_subset (c, this->arrayZ[i], this, out); return_trace (true); } template bool sanitize (hb_sanitize_context_t *c, Ts&&... ds) const { TRACE_SANITIZE (this); return_trace (OffsetArrayOf::sanitize (c, this, hb_forward (ds)...)); } }; /* An array starting at second element. */ template struct HeadlessArrayOf { static constexpr unsigned item_size = Type::static_size; HB_DELETE_CREATE_COPY_ASSIGN (HeadlessArrayOf); const Type& operator [] (int i_) const { unsigned int i = (unsigned int) i_; if (unlikely (i >= lenP1 || !i)) return Null (Type); return arrayZ[i-1]; } Type& operator [] (int i_) { unsigned int i = (unsigned int) i_; if (unlikely (i >= lenP1 || !i)) return Crap (Type); return arrayZ[i-1]; } unsigned int get_size () const { return lenP1.static_size + get_length () * Type::static_size; } unsigned get_length () const { return lenP1 ? lenP1 - 1 : 0; } hb_array_t< Type> as_array () { return hb_array (arrayZ, get_length ()); } hb_array_t as_array () const { return hb_array (arrayZ, get_length ()); } /* Iterator. */ typedef hb_array_t iter_t; typedef hb_array_t< Type> writer_t; iter_t iter () const { return as_array (); } writer_t writer () { return as_array (); } operator iter_t () const { return iter (); } operator writer_t () { return writer (); } bool serialize (hb_serialize_context_t *c, unsigned int items_len) { TRACE_SERIALIZE (this); if (unlikely (!c->extend_min (*this))) return_trace (false); c->check_assign (lenP1, items_len + 1); if (unlikely (!c->extend (*this))) return_trace (false); return_trace (true); } template bool serialize (hb_serialize_context_t *c, Iterator items) { TRACE_SERIALIZE (this); unsigned count = items.len (); if (unlikely (!serialize (c, count))) return_trace (false); /* TODO Umm. Just exhaust the iterator instead? Being extra * cautious right now.. */ for (unsigned i = 0; i < count; i++, ++items) arrayZ[i] = *items; return_trace (true); } template bool sanitize (hb_sanitize_context_t *c, Ts&&... ds) const { TRACE_SANITIZE (this); if (unlikely (!sanitize_shallow (c))) return_trace (false); if (!sizeof... (Ts) && hb_is_trivially_copyable (Type)) return_trace (true); unsigned int count = get_length (); for (unsigned int i = 0; i < count; i++) if (unlikely (!c->dispatch (arrayZ[i], hb_forward (ds)...))) return_trace (false); return_trace (true); } private: bool sanitize_shallow (hb_sanitize_context_t *c) const { TRACE_SANITIZE (this); return_trace (lenP1.sanitize (c) && (!lenP1 || c->check_array (arrayZ, lenP1 - 1))); } public: LenType lenP1; Type arrayZ[HB_VAR_ARRAY]; public: DEFINE_SIZE_ARRAY (sizeof (LenType), arrayZ); }; /* An array storing length-1. */ template struct ArrayOfM1 { HB_DELETE_CREATE_COPY_ASSIGN (ArrayOfM1); const Type& operator [] (int i_) const { unsigned int i = (unsigned int) i_; if (unlikely (i > lenM1)) return Null (Type); return arrayZ[i]; } Type& operator [] (int i_) { unsigned int i = (unsigned int) i_; if (unlikely (i > lenM1)) return Crap (Type); return arrayZ[i]; } unsigned int get_size () const { return lenM1.static_size + (lenM1 + 1) * Type::static_size; } template bool sanitize (hb_sanitize_context_t *c, Ts&&... ds) const { TRACE_SANITIZE (this); if (unlikely (!sanitize_shallow (c))) return_trace (false); unsigned int count = lenM1 + 1; for (unsigned int i = 0; i < count; i++) if (unlikely (!c->dispatch (arrayZ[i], hb_forward (ds)...))) return_trace (false); return_trace (true); } private: bool sanitize_shallow (hb_sanitize_context_t *c) const { TRACE_SANITIZE (this); return_trace (lenM1.sanitize (c) && (c->check_array (arrayZ, lenM1 + 1))); } public: LenType lenM1; Type arrayZ[HB_VAR_ARRAY]; public: DEFINE_SIZE_ARRAY (sizeof (LenType), arrayZ); }; /* An array with sorted elements. Supports binary searching. */ template struct SortedArrayOf : ArrayOf { hb_sorted_array_t< Type> as_array () { return hb_sorted_array (this->arrayZ, this->len); } hb_sorted_array_t as_array () const { return hb_sorted_array (this->arrayZ, this->len); } /* Iterator. */ typedef hb_sorted_array_t iter_t; typedef hb_sorted_array_t< Type> writer_t; iter_t iter () const { return as_array (); } writer_t writer () { return as_array (); } operator iter_t () const { return iter (); } operator writer_t () { return writer (); } hb_sorted_array_t sub_array (unsigned int start_offset, unsigned int count) const { return as_array ().sub_array (start_offset, count); } hb_sorted_array_t sub_array (unsigned int start_offset, unsigned int *count = nullptr /* IN/OUT */) const { return as_array ().sub_array (start_offset, count); } hb_sorted_array_t sub_array (unsigned int start_offset, unsigned int count) { return as_array ().sub_array (start_offset, count); } hb_sorted_array_t sub_array (unsigned int start_offset, unsigned int *count = nullptr /* IN/OUT */) { return as_array ().sub_array (start_offset, count); } bool serialize (hb_serialize_context_t *c, unsigned int items_len) { TRACE_SERIALIZE (this); bool ret = ArrayOf::serialize (c, items_len); return_trace (ret); } template bool serialize (hb_serialize_context_t *c, Iterator items) { TRACE_SERIALIZE (this); bool ret = ArrayOf::serialize (c, items); return_trace (ret); } template Type &bsearch (const T &x, Type ¬_found = Crap (Type)) { return *as_array ().bsearch (x, ¬_found); } template const Type &bsearch (const T &x, const Type ¬_found = Null (Type)) const { return *as_array ().bsearch (x, ¬_found); } template bool bfind (const T &x, unsigned int *i = nullptr, hb_bfind_not_found_t not_found = HB_BFIND_NOT_FOUND_DONT_STORE, unsigned int to_store = (unsigned int) -1) const { return as_array ().bfind (x, i, not_found, to_store); } }; /* * Binary-search arrays */ template struct BinSearchHeader { operator uint32_t () const { return len; } bool sanitize (hb_sanitize_context_t *c) const { TRACE_SANITIZE (this); return_trace (c->check_struct (this)); } BinSearchHeader& operator = (unsigned int v) { len = v; assert (len == v); entrySelector = hb_max (1u, hb_bit_storage (v)) - 1; searchRange = 16 * (1u << entrySelector); rangeShift = v * 16 > searchRange ? 16 * v - searchRange : 0; return *this; } protected: LenType len; LenType searchRange; LenType entrySelector; LenType rangeShift; public: DEFINE_SIZE_STATIC (8); }; template using BinSearchArrayOf = SortedArrayOf>; struct VarSizedBinSearchHeader { bool sanitize (hb_sanitize_context_t *c) const { TRACE_SANITIZE (this); return_trace (c->check_struct (this)); } HBUINT16 unitSize; /* Size of a lookup unit for this search in bytes. */ HBUINT16 nUnits; /* Number of units of the preceding size to be searched. */ HBUINT16 searchRange; /* The value of unitSize times the largest power of 2 * that is less than or equal to the value of nUnits. */ HBUINT16 entrySelector; /* The log base 2 of the largest power of 2 less than * or equal to the value of nUnits. */ HBUINT16 rangeShift; /* The value of unitSize times the difference of the * value of nUnits minus the largest power of 2 less * than or equal to the value of nUnits. */ public: DEFINE_SIZE_STATIC (10); }; template struct VarSizedBinSearchArrayOf { static constexpr unsigned item_size = Type::static_size; HB_DELETE_CREATE_COPY_ASSIGN (VarSizedBinSearchArrayOf); bool last_is_terminator () const { if (unlikely (!header.nUnits)) return false; /* Gah. * * "The number of termination values that need to be included is table-specific. * The value that indicates binary search termination is 0xFFFF." */ const HBUINT16 *words = &StructAtOffset (&bytesZ, (header.nUnits - 1) * header.unitSize); unsigned int count = Type::TerminationWordCount; for (unsigned int i = 0; i < count; i++) if (words[i] != 0xFFFFu) return false; return true; } const Type& operator [] (int i_) const { unsigned int i = (unsigned int) i_; if (unlikely (i >= get_length ())) return Null (Type); return StructAtOffset (&bytesZ, i * header.unitSize); } Type& operator [] (int i_) { unsigned int i = (unsigned int) i_; if (unlikely (i >= get_length ())) return Crap (Type); return StructAtOffset (&bytesZ, i * header.unitSize); } unsigned int get_length () const { return header.nUnits - last_is_terminator (); } unsigned int get_size () const { return header.static_size + header.nUnits * header.unitSize; } template bool sanitize (hb_sanitize_context_t *c, Ts&&... ds) const { TRACE_SANITIZE (this); if (unlikely (!sanitize_shallow (c))) return_trace (false); if (!sizeof... (Ts) && hb_is_trivially_copyable (Type)) return_trace (true); unsigned int count = get_length (); for (unsigned int i = 0; i < count; i++) if (unlikely (!(*this)[i].sanitize (c, hb_forward (ds)...))) return_trace (false); return_trace (true); } template const Type *bsearch (const T &key) const { unsigned pos; return hb_bsearch_impl (&pos, key, (const void *) bytesZ, get_length (), header.unitSize, _hb_cmp_method) ? (const Type *) (((const char *) &bytesZ) + (pos * header.unitSize)) : nullptr; } private: bool sanitize_shallow (hb_sanitize_context_t *c) const { TRACE_SANITIZE (this); return_trace (header.sanitize (c) && Type::static_size <= header.unitSize && c->check_range (bytesZ.arrayZ, header.nUnits, header.unitSize)); } protected: VarSizedBinSearchHeader header; UnsizedArrayOf bytesZ; public: DEFINE_SIZE_ARRAY (10, bytesZ); }; } /* namespace OT */ #endif /* HB_OPEN_TYPE_HH */