/* * Copyright © 1998-2004 David Turner and Werner Lemberg * Copyright © 2004,2007,2009,2010 Red Hat, Inc. * Copyright © 2011,2012 Google, Inc. * * This is part of HarfBuzz, a text shaping library. * * Permission is hereby granted, without written agreement and without * license or royalty fees, to use, copy, modify, and distribute this * software and its documentation for any purpose, provided that the * above copyright notice and the following two paragraphs appear in * all copies of this software. * * IN NO EVENT SHALL THE COPYRIGHT HOLDER BE LIABLE TO ANY PARTY FOR * DIRECT, INDIRECT, SPECIAL, INCIDENTAL, OR CONSEQUENTIAL DAMAGES * ARISING OUT OF THE USE OF THIS SOFTWARE AND ITS DOCUMENTATION, EVEN * IF THE COPYRIGHT HOLDER HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH * DAMAGE. * * THE COPYRIGHT HOLDER SPECIFICALLY DISCLAIMS ANY WARRANTIES, INCLUDING, * BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND * FITNESS FOR A PARTICULAR PURPOSE. THE SOFTWARE PROVIDED HEREUNDER IS * ON AN "AS IS" BASIS, AND THE COPYRIGHT HOLDER HAS NO OBLIGATION TO * PROVIDE MAINTENANCE, SUPPORT, UPDATES, ENHANCEMENTS, OR MODIFICATIONS. * * Red Hat Author(s): Owen Taylor, Behdad Esfahbod * Google Author(s): Behdad Esfahbod */ #ifndef HB_BUFFER_HH #define HB_BUFFER_HH #include "hb.hh" #include "hb-unicode.hh" #ifndef HB_BUFFER_MAX_LEN_FACTOR #define HB_BUFFER_MAX_LEN_FACTOR 64 #endif #ifndef HB_BUFFER_MAX_LEN_MIN #define HB_BUFFER_MAX_LEN_MIN 16384 #endif #ifndef HB_BUFFER_MAX_LEN_DEFAULT #define HB_BUFFER_MAX_LEN_DEFAULT 0x3FFFFFFF /* Shaping more than a billion chars? Let us know! */ #endif #ifndef HB_BUFFER_MAX_OPS_FACTOR #define HB_BUFFER_MAX_OPS_FACTOR 1024 #endif #ifndef HB_BUFFER_MAX_OPS_MIN #define HB_BUFFER_MAX_OPS_MIN 16384 #endif #ifndef HB_BUFFER_MAX_OPS_DEFAULT #define HB_BUFFER_MAX_OPS_DEFAULT 0x1FFFFFFF /* Shaping more than a billion operations? Let us know! */ #endif static_assert ((sizeof (hb_glyph_info_t) == 20), ""); static_assert ((sizeof (hb_glyph_info_t) == sizeof (hb_glyph_position_t)), ""); HB_MARK_AS_FLAG_T (hb_buffer_flags_t); HB_MARK_AS_FLAG_T (hb_buffer_serialize_flags_t); HB_MARK_AS_FLAG_T (hb_buffer_diff_flags_t); enum hb_buffer_scratch_flags_t { HB_BUFFER_SCRATCH_FLAG_DEFAULT = 0x00000000u, HB_BUFFER_SCRATCH_FLAG_HAS_NON_ASCII = 0x00000001u, HB_BUFFER_SCRATCH_FLAG_HAS_DEFAULT_IGNORABLES = 0x00000002u, HB_BUFFER_SCRATCH_FLAG_HAS_SPACE_FALLBACK = 0x00000004u, HB_BUFFER_SCRATCH_FLAG_HAS_GPOS_ATTACHMENT = 0x00000008u, HB_BUFFER_SCRATCH_FLAG_HAS_UNSAFE_TO_BREAK = 0x00000010u, HB_BUFFER_SCRATCH_FLAG_HAS_CGJ = 0x00000020u, /* Reserved for complex shapers' internal use. */ HB_BUFFER_SCRATCH_FLAG_COMPLEX0 = 0x01000000u, HB_BUFFER_SCRATCH_FLAG_COMPLEX1 = 0x02000000u, HB_BUFFER_SCRATCH_FLAG_COMPLEX2 = 0x04000000u, HB_BUFFER_SCRATCH_FLAG_COMPLEX3 = 0x08000000u, }; HB_MARK_AS_FLAG_T (hb_buffer_scratch_flags_t); /* * hb_buffer_t */ struct hb_buffer_t { hb_object_header_t header; /* Information about how the text in the buffer should be treated */ hb_unicode_funcs_t *unicode; /* Unicode functions */ hb_buffer_flags_t flags; /* BOT / EOT / etc. */ hb_buffer_cluster_level_t cluster_level; hb_codepoint_t replacement; /* U+FFFD or something else. */ hb_codepoint_t invisible; /* 0 or something else. */ hb_codepoint_t not_found; /* 0 or something else. */ hb_buffer_scratch_flags_t scratch_flags; /* Have space-fallback, etc. */ unsigned int max_len; /* Maximum allowed len. */ int max_ops; /* Maximum allowed operations. */ /* Buffer contents */ hb_buffer_content_type_t content_type; hb_segment_properties_t props; /* Script, language, direction */ bool successful; /* Allocations successful */ bool have_output; /* Whether we have an output buffer going on */ bool have_positions; /* Whether we have positions */ unsigned int idx; /* Cursor into ->info and ->pos arrays */ unsigned int len; /* Length of ->info and ->pos arrays */ unsigned int out_len; /* Length of ->out_info array if have_output */ unsigned int allocated; /* Length of allocated arrays */ hb_glyph_info_t *info; hb_glyph_info_t *out_info; hb_glyph_position_t *pos; unsigned int serial; /* Text before / after the main buffer contents. * Always in Unicode, and ordered outward. * Index 0 is for "pre-context", 1 for "post-context". */ static constexpr unsigned CONTEXT_LENGTH = 5u; hb_codepoint_t context[2][CONTEXT_LENGTH]; unsigned int context_len[2]; /* Debugging API */ #ifndef HB_NO_BUFFER_MESSAGE hb_buffer_message_func_t message_func; void *message_data; hb_destroy_func_t message_destroy; unsigned message_depth; /* How deeply are we inside a message callback? */ #else static constexpr unsigned message_depth = 0u; #endif /* Internal debugging. */ /* The bits here reflect current allocations of the bytes in glyph_info_t's var1 and var2. */ #ifndef HB_NDEBUG uint8_t allocated_var_bits; #endif /* Methods */ HB_NODISCARD bool in_error () const { return !successful; } void allocate_var (unsigned int start, unsigned int count) { #ifndef HB_NDEBUG unsigned int end = start + count; assert (end <= 8); unsigned int bits = (1u< (info, len).reverse (start, end); if (have_positions) hb_array_t (pos, len).reverse (start, end); } void reverse () { reverse_range (0, len); } template void reverse_groups (const FuncType& group, bool merge_clusters = false) { if (unlikely (!len)) return; unsigned start = 0; unsigned i; for (i = 1; i < len; i++) { if (!group (info[i - 1], info[i])) { if (merge_clusters) this->merge_clusters (start, i); reverse_range (start, i); start = i; } } if (merge_clusters) this->merge_clusters (start, i); reverse_range (start, i); reverse (); } template unsigned group_end (unsigned start, const FuncType& group) const { while (++start < len && group (info[start - 1], info[start])) ; return start; } static bool _cluster_group_func (const hb_glyph_info_t& a, const hb_glyph_info_t& b) { return a.cluster == b.cluster; } void reverse_clusters () { reverse_groups (_cluster_group_func); } HB_INTERNAL void guess_segment_properties (); HB_INTERNAL void swap_buffers (); HB_INTERNAL void clear_output (); HB_INTERNAL void clear_positions (); template HB_NODISCARD bool replace_glyphs (unsigned int num_in, unsigned int num_out, const T *glyph_data) { if (unlikely (!make_room_for (num_in, num_out))) return false; assert (idx + num_in <= len); merge_clusters (idx, idx + num_in); hb_glyph_info_t &orig_info = idx < len ? cur() : prev(); hb_glyph_info_t *pinfo = &out_info[out_len]; for (unsigned int i = 0; i < num_out; i++) { *pinfo = orig_info; pinfo->codepoint = glyph_data[i]; pinfo++; } idx += num_in; out_len += num_out; return true; } HB_NODISCARD bool replace_glyph (hb_codepoint_t glyph_index) { return replace_glyphs (1, 1, &glyph_index); } /* Makes a copy of the glyph at idx to output and replace glyph_index */ HB_NODISCARD bool output_glyph (hb_codepoint_t glyph_index) { return replace_glyphs (0, 1, &glyph_index); } HB_NODISCARD bool output_info (const hb_glyph_info_t &glyph_info) { if (unlikely (!make_room_for (0, 1))) return false; out_info[out_len] = glyph_info; out_len++; return true; } /* Copies glyph at idx to output but doesn't advance idx */ HB_NODISCARD bool copy_glyph () { /* Extra copy because cur()'s return can be freed within * output_info() call if buffer reallocates. */ return output_info (hb_glyph_info_t (cur())); } /* Copies glyph at idx to output and advance idx. * If there's no output, just advance idx. */ HB_NODISCARD bool next_glyph () { if (have_output) { if (out_info != info || out_len != idx) { if (unlikely (!make_room_for (1, 1))) return false; out_info[out_len] = info[idx]; } out_len++; } idx++; return true; } /* Copies n glyphs at idx to output and advance idx. * If there's no output, just advance idx. */ HB_NODISCARD bool next_glyphs (unsigned int n) { if (have_output) { if (out_info != info || out_len != idx) { if (unlikely (!make_room_for (n, n))) return false; memmove (out_info + out_len, info + idx, n * sizeof (out_info[0])); } out_len += n; } idx += n; return true; } /* Advance idx without copying to output. */ void skip_glyph () { idx++; } void reset_masks (hb_mask_t mask) { for (unsigned int j = 0; j < len; j++) info[j].mask = mask; } void add_masks (hb_mask_t mask) { for (unsigned int j = 0; j < len; j++) info[j].mask |= mask; } HB_INTERNAL void set_masks (hb_mask_t value, hb_mask_t mask, unsigned int cluster_start, unsigned int cluster_end); void merge_clusters (unsigned int start, unsigned int end) { if (end - start < 2) return; merge_clusters_impl (start, end); } HB_INTERNAL void merge_clusters_impl (unsigned int start, unsigned int end); HB_INTERNAL void merge_out_clusters (unsigned int start, unsigned int end); /* Merge clusters for deleting current glyph, and skip it. */ HB_INTERNAL void delete_glyph (); void unsafe_to_break (unsigned int start, unsigned int end) { if (end - start < 2) return; unsafe_to_break_impl (start, end); } HB_INTERNAL void unsafe_to_break_impl (unsigned int start, unsigned int end); HB_INTERNAL void unsafe_to_break_from_outbuffer (unsigned int start, unsigned int end); /* Internal methods */ HB_NODISCARD HB_INTERNAL bool move_to (unsigned int i); /* i is output-buffer index. */ HB_NODISCARD HB_INTERNAL bool enlarge (unsigned int size); HB_NODISCARD bool ensure (unsigned int size) { return likely (!size || size < allocated) ? true : enlarge (size); } HB_NODISCARD bool ensure_inplace (unsigned int size) { return likely (!size || size < allocated); } void assert_glyphs () { assert ((content_type == HB_BUFFER_CONTENT_TYPE_GLYPHS) || (!len && (content_type == HB_BUFFER_CONTENT_TYPE_INVALID))); } void assert_unicode () { assert ((content_type == HB_BUFFER_CONTENT_TYPE_UNICODE) || (!len && (content_type == HB_BUFFER_CONTENT_TYPE_INVALID))); } HB_NODISCARD bool ensure_glyphs () { if (unlikely (content_type != HB_BUFFER_CONTENT_TYPE_GLYPHS)) { if (content_type != HB_BUFFER_CONTENT_TYPE_INVALID) return false; assert (len == 0); content_type = HB_BUFFER_CONTENT_TYPE_GLYPHS; } return true; } HB_NODISCARD bool ensure_unicode () { if (unlikely (content_type != HB_BUFFER_CONTENT_TYPE_UNICODE)) { if (content_type != HB_BUFFER_CONTENT_TYPE_INVALID) return false; assert (len == 0); content_type = HB_BUFFER_CONTENT_TYPE_UNICODE; } return true; } HB_NODISCARD HB_INTERNAL bool make_room_for (unsigned int num_in, unsigned int num_out); HB_NODISCARD HB_INTERNAL bool shift_forward (unsigned int count); typedef long scratch_buffer_t; HB_INTERNAL scratch_buffer_t *get_scratch_buffer (unsigned int *size); void clear_context (unsigned int side) { context_len[side] = 0; } HB_INTERNAL void sort (unsigned int start, unsigned int end, int(*compar)(const hb_glyph_info_t *, const hb_glyph_info_t *)); bool messaging () { #ifdef HB_NO_BUFFER_MESSAGE return false; #else return unlikely (message_func); #endif } bool message (hb_font_t *font, const char *fmt, ...) HB_PRINTF_FUNC(3, 4) { #ifdef HB_NO_BUFFER_MESSAGE return true; #else if (!messaging ()) return true; message_depth++; va_list ap; va_start (ap, fmt); bool ret = message_impl (font, fmt, ap); va_end (ap); message_depth--; return ret; #endif } HB_INTERNAL bool message_impl (hb_font_t *font, const char *fmt, va_list ap) HB_PRINTF_FUNC(3, 0); static void set_cluster (hb_glyph_info_t &inf, unsigned int cluster, unsigned int mask = 0) { if (inf.cluster != cluster) { if (mask & HB_GLYPH_FLAG_UNSAFE_TO_BREAK) inf.mask |= HB_GLYPH_FLAG_UNSAFE_TO_BREAK; else inf.mask &= ~HB_GLYPH_FLAG_UNSAFE_TO_BREAK; } inf.cluster = cluster; } static unsigned _infos_find_min_cluster (const hb_glyph_info_t *infos, unsigned start, unsigned end, unsigned cluster) { for (unsigned int i = start; i < end; i++) cluster = hb_min (cluster, infos[i].cluster); return cluster; } void _unsafe_to_break_set_mask (hb_glyph_info_t *infos, unsigned int start, unsigned int end, unsigned int cluster) { for (unsigned int i = start; i < end; i++) if (cluster != infos[i].cluster) { scratch_flags |= HB_BUFFER_SCRATCH_FLAG_HAS_UNSAFE_TO_BREAK; infos[i].mask |= HB_GLYPH_FLAG_UNSAFE_TO_BREAK; } } void clear_glyph_flags (hb_mask_t mask = 0) { for (unsigned int i = 0; i < len; i++) info[i].mask = (info[i].mask & ~HB_GLYPH_FLAG_DEFINED) | (mask & HB_GLYPH_FLAG_DEFINED); } }; DECLARE_NULL_INSTANCE (hb_buffer_t); #define foreach_group(buffer, start, end, group_func) \ for (unsigned int \ _count = buffer->len, \ start = 0, end = _count ? buffer->group_end (0, group_func) : 0; \ start < _count; \ start = end, end = buffer->group_end (start, group_func)) #define foreach_cluster(buffer, start, end) \ foreach_group (buffer, start, end, hb_buffer_t::_cluster_group_func) #define HB_BUFFER_XALLOCATE_VAR(b, func, var) \ b->func (offsetof (hb_glyph_info_t, var) - offsetof(hb_glyph_info_t, var1), \ sizeof (b->info[0].var)) #define HB_BUFFER_ALLOCATE_VAR(b, var) HB_BUFFER_XALLOCATE_VAR (b, allocate_var, var ()) #define HB_BUFFER_DEALLOCATE_VAR(b, var) HB_BUFFER_XALLOCATE_VAR (b, deallocate_var, var ()) #define HB_BUFFER_ASSERT_VAR(b, var) HB_BUFFER_XALLOCATE_VAR (b, assert_var, var ()) #endif /* HB_BUFFER_HH */