Spaces:
Paused
Paused
| # Copyright 2013 Google, Inc. All Rights Reserved. | |
| # | |
| # Google Author(s): Behdad Esfahbod | |
| from fontTools import config | |
| from fontTools.misc.roundTools import otRound | |
| from fontTools import ttLib | |
| from fontTools.ttLib.tables import otTables | |
| from fontTools.ttLib.tables.otBase import USE_HARFBUZZ_REPACKER | |
| from fontTools.otlLib.maxContextCalc import maxCtxFont | |
| from fontTools.pens.basePen import NullPen | |
| from fontTools.misc.loggingTools import Timer | |
| from fontTools.misc.cliTools import makeOutputFileName | |
| from fontTools.subset.util import _add_method, _uniq_sort | |
| from fontTools.subset.cff import * | |
| from fontTools.subset.svg import * | |
| from fontTools.varLib import varStore, multiVarStore # For monkey-patching | |
| from fontTools.ttLib.tables._n_a_m_e import NameRecordVisitor | |
| import sys | |
| import struct | |
| import array | |
| import logging | |
| from collections import Counter, defaultdict | |
| from functools import reduce | |
| from types import MethodType | |
| __usage__ = "pyftsubset font-file [glyph...] [--option=value]..." | |
| __doc__ = ( | |
| """\ | |
| pyftsubset -- OpenType font subsetter and optimizer | |
| pyftsubset is an OpenType font subsetter and optimizer, based on fontTools. | |
| It accepts any TT- or CFF-flavored OpenType (.otf or .ttf) or WOFF (.woff) | |
| font file. The subsetted glyph set is based on the specified glyphs | |
| or characters, and specified OpenType layout features. | |
| The tool also performs some size-reducing optimizations, aimed for using | |
| subset fonts as webfonts. Individual optimizations can be enabled or | |
| disabled, and are enabled by default when they are safe. | |
| Usage: """ | |
| + __usage__ | |
| + """ | |
| At least one glyph or one of --gids, --gids-file, --glyphs, --glyphs-file, | |
| --text, --text-file, --unicodes, or --unicodes-file, must be specified. | |
| Args: | |
| font-file | |
| The input font file. | |
| glyph | |
| Specify one or more glyph identifiers to include in the subset. Must be | |
| PS glyph names, or the special string '*' to keep the entire glyph set. | |
| Initial glyph set specification | |
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ | |
| These options populate the initial glyph set. Same option can appear | |
| multiple times, and the results are accummulated. | |
| --gids=<NNN>[,<NNN>...] | |
| Specify comma/whitespace-separated list of glyph IDs or ranges as decimal | |
| numbers. For example, --gids=10-12,14 adds glyphs with numbers 10, 11, | |
| 12, and 14. | |
| --gids-file=<path> | |
| Like --gids but reads from a file. Anything after a '#' on any line is | |
| ignored as comments. | |
| --glyphs=<glyphname>[,<glyphname>...] | |
| Specify comma/whitespace-separated PS glyph names to add to the subset. | |
| Note that only PS glyph names are accepted, not gidNNN, U+XXXX, etc | |
| that are accepted on the command line. The special string '*' will keep | |
| the entire glyph set. | |
| --glyphs-file=<path> | |
| Like --glyphs but reads from a file. Anything after a '#' on any line | |
| is ignored as comments. | |
| --text=<text> | |
| Specify characters to include in the subset, as UTF-8 string. | |
| --text-file=<path> | |
| Like --text but reads from a file. Newline character are not added to | |
| the subset. | |
| --unicodes=<XXXX>[,<XXXX>...] | |
| Specify comma/whitespace-separated list of Unicode codepoints or | |
| ranges as hex numbers, optionally prefixed with 'U+', 'u', etc. | |
| For example, --unicodes=41-5a,61-7a adds ASCII letters, so does | |
| the more verbose --unicodes=U+0041-005A,U+0061-007A. | |
| The special strings '*' will choose all Unicode characters mapped | |
| by the font. | |
| --unicodes-file=<path> | |
| Like --unicodes, but reads from a file. Anything after a '#' on any | |
| line in the file is ignored as comments. | |
| --ignore-missing-glyphs | |
| Do not fail if some requested glyphs or gids are not available in | |
| the font. | |
| --no-ignore-missing-glyphs | |
| Stop and fail if some requested glyphs or gids are not available | |
| in the font. [default] | |
| --ignore-missing-unicodes [default] | |
| Do not fail if some requested Unicode characters (including those | |
| indirectly specified using --text or --text-file) are not available | |
| in the font. | |
| --no-ignore-missing-unicodes | |
| Stop and fail if some requested Unicode characters are not available | |
| in the font. | |
| Note the default discrepancy between ignoring missing glyphs versus | |
| unicodes. This is for historical reasons and in the future | |
| --no-ignore-missing-unicodes might become default. | |
| Other options | |
| ^^^^^^^^^^^^^ | |
| For the other options listed below, to see the current value of the option, | |
| pass a value of '?' to it, with or without a '='. In some environments, | |
| you might need to escape the question mark, like this: '--glyph-names\\?'. | |
| Examples:: | |
| $ pyftsubset --glyph-names? | |
| Current setting for 'glyph-names' is: False | |
| $ pyftsubset --name-IDs=? | |
| Current setting for 'name-IDs' is: [0, 1, 2, 3, 4, 5, 6] | |
| $ pyftsubset --hinting? --no-hinting --hinting? | |
| Current setting for 'hinting' is: True | |
| Current setting for 'hinting' is: False | |
| Output options | |
| ^^^^^^^^^^^^^^ | |
| --output-file=<path> | |
| The output font file. If not specified, the subsetted font | |
| will be saved in as font-file.subset. | |
| --flavor=<type> | |
| Specify flavor of output font file. May be 'woff' or 'woff2'. | |
| Note that WOFF2 requires the Brotli Python extension, available | |
| at https://github.com/google/brotli | |
| --with-zopfli | |
| Use the Google Zopfli algorithm to compress WOFF. The output is 3-8 % | |
| smaller than pure zlib, but the compression speed is much slower. | |
| The Zopfli Python bindings are available at: | |
| https://pypi.python.org/pypi/zopfli | |
| --harfbuzz-repacker | |
| By default, we serialize GPOS/GSUB using the HarfBuzz Repacker when | |
| uharfbuzz can be imported and is successful, otherwise fall back to | |
| the pure-python serializer. Set the option to force using the HarfBuzz | |
| Repacker (raises an error if uharfbuzz can't be found or fails). | |
| --no-harfbuzz-repacker | |
| Always use the pure-python serializer even if uharfbuzz is available. | |
| Glyph set expansion | |
| ^^^^^^^^^^^^^^^^^^^ | |
| These options control how additional glyphs are added to the subset. | |
| --retain-gids | |
| Retain glyph indices; just empty glyphs not needed in-place. | |
| --notdef-glyph | |
| Add the '.notdef' glyph to the subset (ie, keep it). [default] | |
| --no-notdef-glyph | |
| Drop the '.notdef' glyph unless specified in the glyph set. This | |
| saves a few bytes, but is not possible for Postscript-flavored | |
| fonts, as those require '.notdef'. For TrueType-flavored fonts, | |
| this works fine as long as no unsupported glyphs are requested | |
| from the font. | |
| --notdef-outline | |
| Keep the outline of '.notdef' glyph. The '.notdef' glyph outline is | |
| used when glyphs not supported by the font are to be shown. It is not | |
| needed otherwise. | |
| --no-notdef-outline | |
| When including a '.notdef' glyph, remove its outline. This saves | |
| a few bytes. [default] | |
| --recommended-glyphs | |
| Add glyphs 0, 1, 2, and 3 to the subset, as recommended for | |
| TrueType-flavored fonts: '.notdef', 'NULL' or '.null', 'CR', 'space'. | |
| Some legacy software might require this, but no modern system does. | |
| --no-recommended-glyphs | |
| Do not add glyphs 0, 1, 2, and 3 to the subset, unless specified in | |
| glyph set. [default] | |
| --no-layout-closure | |
| Do not expand glyph set to add glyphs produced by OpenType layout | |
| features. Instead, OpenType layout features will be subset to only | |
| rules that are relevant to the otherwise-specified glyph set. | |
| --layout-features[+|-]=<feature>[,<feature>...] | |
| Specify (=), add to (+=) or exclude from (-=) the comma-separated | |
| set of OpenType layout feature tags that will be preserved. | |
| Glyph variants used by the preserved features are added to the | |
| specified subset glyph set. By default, 'calt', 'ccmp', 'clig', 'curs', | |
| 'dnom', 'frac', 'kern', 'liga', 'locl', 'mark', 'mkmk', 'numr', 'rclt', | |
| 'rlig', 'rvrn', and all features required for script shaping are | |
| preserved. To see the full list, try '--layout-features=?'. | |
| Use '*' to keep all features. | |
| Multiple --layout-features options can be provided if necessary. | |
| Examples: | |
| --layout-features+=onum,pnum,ss01 | |
| * Keep the default set of features and 'onum', 'pnum', 'ss01'. | |
| --layout-features-='mark','mkmk' | |
| * Keep the default set of features but drop 'mark' and 'mkmk'. | |
| --layout-features='kern' | |
| * Only keep the 'kern' feature, drop all others. | |
| --layout-features='' | |
| * Drop all features. | |
| --layout-features='*' | |
| * Keep all features. | |
| --layout-features+=aalt --layout-features-=vrt2 | |
| * Keep default set of features plus 'aalt', but drop 'vrt2'. | |
| --layout-scripts[+|-]=<script>[,<script>...] | |
| Specify (=), add to (+=) or exclude from (-=) the comma-separated | |
| set of OpenType layout script tags that will be preserved. LangSys tags | |
| can be appended to script tag, separated by '.', for example: | |
| 'arab.dflt,arab.URD,latn.TRK'. By default all scripts are retained ('*'). | |
| Hinting options | |
| ^^^^^^^^^^^^^^^ | |
| --hinting | |
| Keep hinting [default] | |
| --no-hinting | |
| Drop glyph-specific hinting and font-wide hinting tables, as well | |
| as remove hinting-related bits and pieces from other tables (eg. GPOS). | |
| See --hinting-tables for list of tables that are dropped by default. | |
| Instructions and hints are stripped from 'glyf' and 'CFF ' tables | |
| respectively. This produces (sometimes up to 30%) smaller fonts that | |
| are suitable for extremely high-resolution systems, like high-end | |
| mobile devices and retina displays. | |
| Optimization options | |
| ^^^^^^^^^^^^^^^^^^^^ | |
| --desubroutinize | |
| Remove CFF use of subroutinizes. Subroutinization is a way to make CFF | |
| fonts smaller. For small subsets however, desubroutinizing might make | |
| the font smaller. It has even been reported that desubroutinized CFF | |
| fonts compress better (produce smaller output) WOFF and WOFF2 fonts. | |
| Also see note under --no-hinting. | |
| --no-desubroutinize [default] | |
| Leave CFF subroutinizes as is, only throw away unused subroutinizes. | |
| Font table options | |
| ^^^^^^^^^^^^^^^^^^ | |
| --drop-tables[+|-]=<table>[,<table>...] | |
| Specify (=), add to (+=) or exclude from (-=) the comma-separated | |
| set of tables that will be be dropped. | |
| By default, the following tables are dropped: | |
| 'BASE', 'JSTF', 'DSIG', 'EBDT', 'EBLC', 'EBSC', 'PCLT', 'LTSH' | |
| and Graphite tables: 'Feat', 'Glat', 'Gloc', 'Silf', 'Sill'. | |
| The tool will attempt to subset the remaining tables. | |
| Examples: | |
| --drop-tables-=BASE | |
| * Drop the default set of tables but keep 'BASE'. | |
| --drop-tables+=GSUB | |
| * Drop the default set of tables and 'GSUB'. | |
| --drop-tables=DSIG | |
| * Only drop the 'DSIG' table, keep all others. | |
| --drop-tables= | |
| * Keep all tables. | |
| --no-subset-tables+=<table>[,<table>...] | |
| Add to the set of tables that will not be subsetted. | |
| By default, the following tables are included in this list, as | |
| they do not need subsetting (ignore the fact that 'loca' is listed | |
| here): 'gasp', 'head', 'hhea', 'maxp', 'vhea', 'OS/2', 'loca', 'name', | |
| 'cvt ', 'fpgm', 'prep', 'VMDX', 'DSIG', 'CPAL', 'MVAR', 'cvar', 'STAT'. | |
| By default, tables that the tool does not know how to subset and are not | |
| specified here will be dropped from the font, unless --passthrough-tables | |
| option is passed. | |
| Example: | |
| --no-subset-tables+=FFTM | |
| * Keep 'FFTM' table in the font by preventing subsetting. | |
| --passthrough-tables | |
| Do not drop tables that the tool does not know how to subset. | |
| --no-passthrough-tables | |
| Tables that the tool does not know how to subset and are not specified | |
| in --no-subset-tables will be dropped from the font. [default] | |
| --hinting-tables[-]=<table>[,<table>...] | |
| Specify (=), add to (+=) or exclude from (-=) the list of font-wide | |
| hinting tables that will be dropped if --no-hinting is specified. | |
| Examples: | |
| --hinting-tables-=VDMX | |
| * Drop font-wide hinting tables except 'VDMX'. | |
| --hinting-tables= | |
| * Keep all font-wide hinting tables (but strip hints from glyphs). | |
| --legacy-kern | |
| Keep TrueType 'kern' table even when OpenType 'GPOS' is available. | |
| --no-legacy-kern | |
| Drop TrueType 'kern' table if OpenType 'GPOS' is available. [default] | |
| Font naming options | |
| ^^^^^^^^^^^^^^^^^^^ | |
| These options control what is retained in the 'name' table. For numerical | |
| codes, see: http://www.microsoft.com/typography/otspec/name.htm | |
| --name-IDs[+|-]=<nameID>[,<nameID>...] | |
| Specify (=), add to (+=) or exclude from (-=) the set of 'name' table | |
| entry nameIDs that will be preserved. By default, only nameIDs between 0 | |
| and 6 are preserved, the rest are dropped. Use '*' to keep all entries. | |
| Examples: | |
| --name-IDs+=7,8,9 | |
| * Also keep Trademark, Manufacturer and Designer name entries. | |
| --name-IDs= | |
| * Drop all 'name' table entries. | |
| --name-IDs=* | |
| * keep all 'name' table entries | |
| --name-legacy | |
| Keep legacy (non-Unicode) 'name' table entries (0.x, 1.x etc.). | |
| XXX Note: This might be needed for some fonts that have no Unicode name | |
| entires for English. See: https://github.com/fonttools/fonttools/issues/146 | |
| --no-name-legacy | |
| Drop legacy (non-Unicode) 'name' table entries [default] | |
| --name-languages[+|-]=<langID>[,<langID>] | |
| Specify (=), add to (+=) or exclude from (-=) the set of 'name' table | |
| langIDs that will be preserved. By default only records with langID | |
| 0x0409 (English) are preserved. Use '*' to keep all langIDs. | |
| --obfuscate-names | |
| Make the font unusable as a system font by replacing name IDs 1, 2, 3, 4, | |
| and 6 with dummy strings (it is still fully functional as webfont). | |
| Glyph naming and encoding options | |
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ | |
| --glyph-names | |
| Keep PS glyph names in TT-flavored fonts. In general glyph names are | |
| not needed for correct use of the font. However, some PDF generators | |
| and PDF viewers might rely on glyph names to extract Unicode text | |
| from PDF documents. | |
| --no-glyph-names | |
| Drop PS glyph names in TT-flavored fonts, by using 'post' table | |
| version 3.0. [default] | |
| --legacy-cmap | |
| Keep the legacy 'cmap' subtables (0.x, 1.x, 4.x etc.). | |
| --no-legacy-cmap | |
| Drop the legacy 'cmap' subtables. [default] | |
| --symbol-cmap | |
| Keep the 3.0 symbol 'cmap'. | |
| --no-symbol-cmap | |
| Drop the 3.0 symbol 'cmap'. [default] | |
| Other font-specific options | |
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^ | |
| --recalc-bounds | |
| Recalculate font bounding boxes. | |
| --no-recalc-bounds | |
| Keep original font bounding boxes. This is faster and still safe | |
| for all practical purposes. [default] | |
| --recalc-timestamp | |
| Set font 'modified' timestamp to current time. | |
| --no-recalc-timestamp | |
| Do not modify font 'modified' timestamp. [default] | |
| --canonical-order | |
| Order tables as recommended in the OpenType standard. This is not | |
| required by the standard, nor by any known implementation. | |
| --no-canonical-order | |
| Keep original order of font tables. This is faster. [default] | |
| --prune-unicode-ranges | |
| Update the 'OS/2 ulUnicodeRange*' bits after subsetting. The Unicode | |
| ranges defined in the OpenType specification v1.7 are intersected with | |
| the Unicode codepoints specified in the font's Unicode 'cmap' subtables: | |
| when no overlap is found, the bit will be switched off. However, it will | |
| *not* be switched on if an intersection is found. [default] | |
| --no-prune-unicode-ranges | |
| Don't change the 'OS/2 ulUnicodeRange*' bits. | |
| --prune-codepage-ranges | |
| Update the 'OS/2 ulCodePageRange*' bits after subsetting. [default] | |
| --no-prune-codepage-ranges | |
| Don't change the 'OS/2 ulCodePageRange*' bits. | |
| --recalc-average-width | |
| Update the 'OS/2 xAvgCharWidth' field after subsetting. | |
| --no-recalc-average-width | |
| Don't change the 'OS/2 xAvgCharWidth' field. [default] | |
| --recalc-max-context | |
| Update the 'OS/2 usMaxContext' field after subsetting. | |
| --no-recalc-max-context | |
| Don't change the 'OS/2 usMaxContext' field. [default] | |
| --font-number=<number> | |
| Select font number for TrueType Collection (.ttc/.otc), starting from 0. | |
| --pretty-svg | |
| When subsetting SVG table, use lxml pretty_print=True option to indent | |
| the XML output (only recommended for debugging purposes). | |
| Application options | |
| ^^^^^^^^^^^^^^^^^^^ | |
| --verbose | |
| Display verbose information of the subsetting process. | |
| --timing | |
| Display detailed timing information of the subsetting process. | |
| --xml | |
| Display the TTX XML representation of subsetted font. | |
| Example | |
| ^^^^^^^ | |
| Produce a subset containing the characters ' !"#$%' without performing | |
| size-reducing optimizations:: | |
| $ pyftsubset font.ttf --unicodes="U+0020-0025" \\ | |
| --layout-features=* --glyph-names --symbol-cmap --legacy-cmap \\ | |
| --notdef-glyph --notdef-outline --recommended-glyphs \\ | |
| --name-IDs=* --name-legacy --name-languages=* | |
| """ | |
| ) | |
| log = logging.getLogger("fontTools.subset") | |
| def _log_glyphs(self, glyphs, font=None): | |
| self.info("Glyph names: %s", sorted(glyphs)) | |
| if font: | |
| reverseGlyphMap = font.getReverseGlyphMap() | |
| self.info("Glyph IDs: %s", sorted(reverseGlyphMap[g] for g in glyphs)) | |
| # bind "glyphs" function to 'log' object | |
| log.glyphs = MethodType(_log_glyphs, log) | |
| # I use a different timing channel so I can configure it separately from the | |
| # main module's logger | |
| timer = Timer(logger=logging.getLogger("fontTools.subset.timer")) | |
| def _dict_subset(d, glyphs): | |
| return {g: d[g] for g in glyphs} | |
| def _list_subset(l, indices): | |
| count = len(l) | |
| return [l[i] for i in indices if i < count] | |
| def intersect(self, glyphs): | |
| """Returns ascending list of matching coverage values.""" | |
| return [i for i, g in enumerate(self.glyphs) if g in glyphs] | |
| def intersect_glyphs(self, glyphs): | |
| """Returns set of intersecting glyphs.""" | |
| return set(g for g in self.glyphs if g in glyphs) | |
| def subset(self, glyphs): | |
| """Returns ascending list of remaining coverage values.""" | |
| indices = self.intersect(glyphs) | |
| self.glyphs = [g for g in self.glyphs if g in glyphs] | |
| return indices | |
| def remap(self, coverage_map): | |
| """Remaps coverage.""" | |
| self.glyphs = [self.glyphs[i] for i in coverage_map] | |
| def intersect(self, glyphs): | |
| """Returns ascending list of matching class values.""" | |
| return _uniq_sort( | |
| ([0] if any(g not in self.classDefs for g in glyphs) else []) | |
| + [v for g, v in self.classDefs.items() if g in glyphs] | |
| ) | |
| def intersect_class(self, glyphs, klass): | |
| """Returns set of glyphs matching class.""" | |
| if klass == 0: | |
| return set(g for g in glyphs if g not in self.classDefs) | |
| return set(g for g, v in self.classDefs.items() if v == klass and g in glyphs) | |
| def subset(self, glyphs, remap=False, useClass0=True): | |
| """Returns ascending list of remaining classes.""" | |
| self.classDefs = {g: v for g, v in self.classDefs.items() if g in glyphs} | |
| # Note: while class 0 has the special meaning of "not matched", | |
| # if no glyph will ever /not match/, we can optimize class 0 out too. | |
| # Only do this if allowed. | |
| indices = _uniq_sort( | |
| ( | |
| [0] | |
| if ((not useClass0) or any(g not in self.classDefs for g in glyphs)) | |
| else [] | |
| ) | |
| + list(self.classDefs.values()) | |
| ) | |
| if remap: | |
| self.remap(indices) | |
| return indices | |
| def remap(self, class_map): | |
| """Remaps classes.""" | |
| self.classDefs = {g: class_map.index(v) for g, v in self.classDefs.items()} | |
| def closure_glyphs(self, s, cur_glyphs): | |
| s.glyphs.update(v for g, v in self.mapping.items() if g in cur_glyphs) | |
| def subset_glyphs(self, s): | |
| self.mapping = { | |
| g: v for g, v in self.mapping.items() if g in s.glyphs and v in s.glyphs | |
| } | |
| return bool(self.mapping) | |
| def closure_glyphs(self, s, cur_glyphs): | |
| for glyph, subst in self.mapping.items(): | |
| if glyph in cur_glyphs: | |
| s.glyphs.update(subst) | |
| def subset_glyphs(self, s): | |
| self.mapping = { | |
| g: v | |
| for g, v in self.mapping.items() | |
| if g in s.glyphs and all(sub in s.glyphs for sub in v) | |
| } | |
| return bool(self.mapping) | |
| def closure_glyphs(self, s, cur_glyphs): | |
| s.glyphs.update(*(vlist for g, vlist in self.alternates.items() if g in cur_glyphs)) | |
| def subset_glyphs(self, s): | |
| self.alternates = { | |
| g: [v for v in vlist if v in s.glyphs] | |
| for g, vlist in self.alternates.items() | |
| if g in s.glyphs and any(v in s.glyphs for v in vlist) | |
| } | |
| return bool(self.alternates) | |
| def closure_glyphs(self, s, cur_glyphs): | |
| s.glyphs.update( | |
| *( | |
| [seq.LigGlyph for seq in seqs if all(c in s.glyphs for c in seq.Component)] | |
| for g, seqs in self.ligatures.items() | |
| if g in cur_glyphs | |
| ) | |
| ) | |
| def subset_glyphs(self, s): | |
| self.ligatures = {g: v for g, v in self.ligatures.items() if g in s.glyphs} | |
| self.ligatures = { | |
| g: [ | |
| seq | |
| for seq in seqs | |
| if seq.LigGlyph in s.glyphs and all(c in s.glyphs for c in seq.Component) | |
| ] | |
| for g, seqs in self.ligatures.items() | |
| } | |
| self.ligatures = {g: v for g, v in self.ligatures.items() if v} | |
| return bool(self.ligatures) | |
| def closure_glyphs(self, s, cur_glyphs): | |
| if self.Format == 1: | |
| indices = self.Coverage.intersect(cur_glyphs) | |
| if not indices or not all( | |
| c.intersect(s.glyphs) | |
| for c in self.LookAheadCoverage + self.BacktrackCoverage | |
| ): | |
| return | |
| s.glyphs.update(self.Substitute[i] for i in indices) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def subset_glyphs(self, s): | |
| if self.Format == 1: | |
| indices = self.Coverage.subset(s.glyphs) | |
| self.Substitute = _list_subset(self.Substitute, indices) | |
| # Now drop rules generating glyphs we don't want | |
| indices = [i for i, sub in enumerate(self.Substitute) if sub in s.glyphs] | |
| self.Substitute = _list_subset(self.Substitute, indices) | |
| self.Coverage.remap(indices) | |
| self.GlyphCount = len(self.Substitute) | |
| return bool( | |
| self.GlyphCount | |
| and all( | |
| c.subset(s.glyphs) | |
| for c in self.LookAheadCoverage + self.BacktrackCoverage | |
| ) | |
| ) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def is_hinting(self): | |
| return self.DeltaFormat in (1, 2, 3) | |
| def prune_hints(self): | |
| for name in ["XPlaDevice", "YPlaDevice", "XAdvDevice", "YAdvDevice"]: | |
| v = getattr(self, name, None) | |
| if v is not None and v.is_hinting(): | |
| delattr(self, name) | |
| def subset_glyphs(self, s): | |
| if self.Format == 1: | |
| return len(self.Coverage.subset(s.glyphs)) | |
| elif self.Format == 2: | |
| indices = self.Coverage.subset(s.glyphs) | |
| values = self.Value | |
| count = len(values) | |
| self.Value = [values[i] for i in indices if i < count] | |
| self.ValueCount = len(self.Value) | |
| return bool(self.ValueCount) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def prune_post_subset(self, font, options): | |
| if self.Value is None: | |
| assert self.ValueFormat == 0 | |
| return True | |
| # Shrink ValueFormat | |
| if self.Format == 1: | |
| if not options.hinting: | |
| self.Value.prune_hints() | |
| self.ValueFormat = self.Value.getEffectiveFormat() | |
| elif self.Format == 2: | |
| if None in self.Value: | |
| assert self.ValueFormat == 0 | |
| assert all(v is None for v in self.Value) | |
| else: | |
| if not options.hinting: | |
| for v in self.Value: | |
| v.prune_hints() | |
| self.ValueFormat = reduce( | |
| int.__or__, [v.getEffectiveFormat() for v in self.Value], 0 | |
| ) | |
| # Downgrade to Format 1 if all ValueRecords are the same | |
| if self.Format == 2 and all(v == self.Value[0] for v in self.Value): | |
| self.Format = 1 | |
| self.Value = self.Value[0] if self.ValueFormat != 0 else None | |
| del self.ValueCount | |
| return True | |
| def subset_glyphs(self, s): | |
| if self.Format == 1: | |
| indices = self.Coverage.subset(s.glyphs) | |
| pairs = self.PairSet | |
| count = len(pairs) | |
| self.PairSet = [pairs[i] for i in indices if i < count] | |
| for p in self.PairSet: | |
| p.PairValueRecord = [ | |
| r for r in p.PairValueRecord if r.SecondGlyph in s.glyphs | |
| ] | |
| p.PairValueCount = len(p.PairValueRecord) | |
| # Remove empty pairsets | |
| indices = [i for i, p in enumerate(self.PairSet) if p.PairValueCount] | |
| self.Coverage.remap(indices) | |
| self.PairSet = _list_subset(self.PairSet, indices) | |
| self.PairSetCount = len(self.PairSet) | |
| return bool(self.PairSetCount) | |
| elif self.Format == 2: | |
| class1_map = [ | |
| c | |
| for c in self.ClassDef1.subset( | |
| s.glyphs.intersection(self.Coverage.glyphs), remap=True | |
| ) | |
| if c < self.Class1Count | |
| ] | |
| class2_map = [ | |
| c | |
| for c in self.ClassDef2.subset(s.glyphs, remap=True, useClass0=False) | |
| if c < self.Class2Count | |
| ] | |
| self.Class1Record = [self.Class1Record[i] for i in class1_map] | |
| for c in self.Class1Record: | |
| c.Class2Record = [c.Class2Record[i] for i in class2_map] | |
| self.Class1Count = len(class1_map) | |
| self.Class2Count = len(class2_map) | |
| # If only Class2 0 left, no need to keep anything. | |
| return bool( | |
| self.Class1Count | |
| and (self.Class2Count > 1) | |
| and self.Coverage.subset(s.glyphs) | |
| ) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def prune_post_subset(self, font, options): | |
| if not options.hinting: | |
| attr1, attr2 = { | |
| 1: ("PairSet", "PairValueRecord"), | |
| 2: ("Class1Record", "Class2Record"), | |
| }[self.Format] | |
| self.ValueFormat1 = self.ValueFormat2 = 0 | |
| for row in getattr(self, attr1): | |
| for r in getattr(row, attr2): | |
| if r.Value1: | |
| r.Value1.prune_hints() | |
| self.ValueFormat1 |= r.Value1.getEffectiveFormat() | |
| if r.Value2: | |
| r.Value2.prune_hints() | |
| self.ValueFormat2 |= r.Value2.getEffectiveFormat() | |
| return bool(self.ValueFormat1 | self.ValueFormat2) | |
| def subset_glyphs(self, s): | |
| if self.Format == 1: | |
| indices = self.Coverage.subset(s.glyphs) | |
| records = self.EntryExitRecord | |
| count = len(records) | |
| self.EntryExitRecord = [records[i] for i in indices if i < count] | |
| self.EntryExitCount = len(self.EntryExitRecord) | |
| return bool(self.EntryExitCount) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def prune_hints(self): | |
| if self.Format == 2: | |
| self.Format = 1 | |
| elif self.Format == 3: | |
| for name in ("XDeviceTable", "YDeviceTable"): | |
| v = getattr(self, name, None) | |
| if v is not None and v.is_hinting(): | |
| setattr(self, name, None) | |
| if self.XDeviceTable is None and self.YDeviceTable is None: | |
| self.Format = 1 | |
| def prune_post_subset(self, font, options): | |
| if not options.hinting: | |
| for rec in self.EntryExitRecord: | |
| if rec.EntryAnchor: | |
| rec.EntryAnchor.prune_hints() | |
| if rec.ExitAnchor: | |
| rec.ExitAnchor.prune_hints() | |
| return True | |
| def subset_glyphs(self, s): | |
| if self.Format == 1: | |
| mark_indices = self.MarkCoverage.subset(s.glyphs) | |
| self.MarkArray.MarkRecord = _list_subset( | |
| self.MarkArray.MarkRecord, mark_indices | |
| ) | |
| self.MarkArray.MarkCount = len(self.MarkArray.MarkRecord) | |
| base_indices = self.BaseCoverage.subset(s.glyphs) | |
| self.BaseArray.BaseRecord = _list_subset( | |
| self.BaseArray.BaseRecord, base_indices | |
| ) | |
| self.BaseArray.BaseCount = len(self.BaseArray.BaseRecord) | |
| # Prune empty classes | |
| class_indices = _uniq_sort(v.Class for v in self.MarkArray.MarkRecord) | |
| self.ClassCount = len(class_indices) | |
| for m in self.MarkArray.MarkRecord: | |
| m.Class = class_indices.index(m.Class) | |
| for b in self.BaseArray.BaseRecord: | |
| b.BaseAnchor = _list_subset(b.BaseAnchor, class_indices) | |
| return bool( | |
| self.ClassCount and self.MarkArray.MarkCount and self.BaseArray.BaseCount | |
| ) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def prune_post_subset(self, font, options): | |
| if not options.hinting: | |
| for m in self.MarkArray.MarkRecord: | |
| if m.MarkAnchor: | |
| m.MarkAnchor.prune_hints() | |
| for b in self.BaseArray.BaseRecord: | |
| for a in b.BaseAnchor: | |
| if a: | |
| a.prune_hints() | |
| return True | |
| def subset_glyphs(self, s): | |
| if self.Format == 1: | |
| mark_indices = self.MarkCoverage.subset(s.glyphs) | |
| self.MarkArray.MarkRecord = _list_subset( | |
| self.MarkArray.MarkRecord, mark_indices | |
| ) | |
| self.MarkArray.MarkCount = len(self.MarkArray.MarkRecord) | |
| ligature_indices = self.LigatureCoverage.subset(s.glyphs) | |
| self.LigatureArray.LigatureAttach = _list_subset( | |
| self.LigatureArray.LigatureAttach, ligature_indices | |
| ) | |
| self.LigatureArray.LigatureCount = len(self.LigatureArray.LigatureAttach) | |
| # Prune empty classes | |
| class_indices = _uniq_sort(v.Class for v in self.MarkArray.MarkRecord) | |
| self.ClassCount = len(class_indices) | |
| for m in self.MarkArray.MarkRecord: | |
| m.Class = class_indices.index(m.Class) | |
| for l in self.LigatureArray.LigatureAttach: | |
| if l is None: | |
| continue | |
| for c in l.ComponentRecord: | |
| c.LigatureAnchor = _list_subset(c.LigatureAnchor, class_indices) | |
| return bool( | |
| self.ClassCount | |
| and self.MarkArray.MarkCount | |
| and self.LigatureArray.LigatureCount | |
| ) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def prune_post_subset(self, font, options): | |
| if not options.hinting: | |
| for m in self.MarkArray.MarkRecord: | |
| if m.MarkAnchor: | |
| m.MarkAnchor.prune_hints() | |
| for l in self.LigatureArray.LigatureAttach: | |
| if l is None: | |
| continue | |
| for c in l.ComponentRecord: | |
| for a in c.LigatureAnchor: | |
| if a: | |
| a.prune_hints() | |
| return True | |
| def subset_glyphs(self, s): | |
| if self.Format == 1: | |
| mark1_indices = self.Mark1Coverage.subset(s.glyphs) | |
| self.Mark1Array.MarkRecord = _list_subset( | |
| self.Mark1Array.MarkRecord, mark1_indices | |
| ) | |
| self.Mark1Array.MarkCount = len(self.Mark1Array.MarkRecord) | |
| mark2_indices = self.Mark2Coverage.subset(s.glyphs) | |
| self.Mark2Array.Mark2Record = _list_subset( | |
| self.Mark2Array.Mark2Record, mark2_indices | |
| ) | |
| self.Mark2Array.MarkCount = len(self.Mark2Array.Mark2Record) | |
| # Prune empty classes | |
| class_indices = _uniq_sort(v.Class for v in self.Mark1Array.MarkRecord) | |
| self.ClassCount = len(class_indices) | |
| for m in self.Mark1Array.MarkRecord: | |
| m.Class = class_indices.index(m.Class) | |
| for b in self.Mark2Array.Mark2Record: | |
| b.Mark2Anchor = _list_subset(b.Mark2Anchor, class_indices) | |
| return bool( | |
| self.ClassCount and self.Mark1Array.MarkCount and self.Mark2Array.MarkCount | |
| ) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def prune_post_subset(self, font, options): | |
| if not options.hinting: | |
| for m in self.Mark1Array.MarkRecord: | |
| if m.MarkAnchor: | |
| m.MarkAnchor.prune_hints() | |
| for b in self.Mark2Array.Mark2Record: | |
| for m in b.Mark2Anchor: | |
| if m: | |
| m.prune_hints() | |
| return True | |
| def subset_lookups(self, lookup_indices): | |
| pass | |
| def collect_lookups(self): | |
| return [] | |
| def prune_post_subset(self, font, options): | |
| return True | |
| def may_have_non_1to1(self): | |
| return False | |
| def may_have_non_1to1(self): | |
| return True | |
| def __subset_classify_context(self): | |
| class ContextHelper(object): | |
| def __init__(self, klass, Format): | |
| if klass.__name__.endswith("Subst"): | |
| Typ = "Sub" | |
| Type = "Subst" | |
| else: | |
| Typ = "Pos" | |
| Type = "Pos" | |
| if klass.__name__.startswith("Chain"): | |
| Chain = "Chain" | |
| InputIdx = 1 | |
| DataLen = 3 | |
| else: | |
| Chain = "" | |
| InputIdx = 0 | |
| DataLen = 1 | |
| ChainTyp = Chain + Typ | |
| self.Typ = Typ | |
| self.Type = Type | |
| self.Chain = Chain | |
| self.ChainTyp = ChainTyp | |
| self.InputIdx = InputIdx | |
| self.DataLen = DataLen | |
| self.LookupRecord = Type + "LookupRecord" | |
| if Format == 1: | |
| Coverage = lambda r: r.Coverage | |
| ChainCoverage = lambda r: r.Coverage | |
| ContextData = lambda r: (None,) | |
| ChainContextData = lambda r: (None, None, None) | |
| SetContextData = None | |
| SetChainContextData = None | |
| RuleData = lambda r: (r.Input,) | |
| ChainRuleData = lambda r: (r.Backtrack, r.Input, r.LookAhead) | |
| def SetRuleData(r, d): | |
| (r.Input,) = d | |
| (r.GlyphCount,) = (len(x) + 1 for x in d) | |
| def ChainSetRuleData(r, d): | |
| (r.Backtrack, r.Input, r.LookAhead) = d | |
| ( | |
| r.BacktrackGlyphCount, | |
| r.InputGlyphCount, | |
| r.LookAheadGlyphCount, | |
| ) = (len(d[0]), len(d[1]) + 1, len(d[2])) | |
| elif Format == 2: | |
| Coverage = lambda r: r.Coverage | |
| ChainCoverage = lambda r: r.Coverage | |
| ContextData = lambda r: (r.ClassDef,) | |
| ChainContextData = lambda r: ( | |
| r.BacktrackClassDef, | |
| r.InputClassDef, | |
| r.LookAheadClassDef, | |
| ) | |
| def SetContextData(r, d): | |
| (r.ClassDef,) = d | |
| def SetChainContextData(r, d): | |
| (r.BacktrackClassDef, r.InputClassDef, r.LookAheadClassDef) = d | |
| RuleData = lambda r: (r.Class,) | |
| ChainRuleData = lambda r: (r.Backtrack, r.Input, r.LookAhead) | |
| def SetRuleData(r, d): | |
| (r.Class,) = d | |
| (r.GlyphCount,) = (len(x) + 1 for x in d) | |
| def ChainSetRuleData(r, d): | |
| (r.Backtrack, r.Input, r.LookAhead) = d | |
| ( | |
| r.BacktrackGlyphCount, | |
| r.InputGlyphCount, | |
| r.LookAheadGlyphCount, | |
| ) = (len(d[0]), len(d[1]) + 1, len(d[2])) | |
| elif Format == 3: | |
| Coverage = lambda r: r.Coverage[0] | |
| ChainCoverage = lambda r: r.InputCoverage[0] | |
| ContextData = None | |
| ChainContextData = None | |
| SetContextData = None | |
| SetChainContextData = None | |
| RuleData = lambda r: r.Coverage | |
| ChainRuleData = lambda r: ( | |
| r.BacktrackCoverage + r.InputCoverage + r.LookAheadCoverage | |
| ) | |
| def SetRuleData(r, d): | |
| (r.Coverage,) = d | |
| (r.GlyphCount,) = (len(x) for x in d) | |
| def ChainSetRuleData(r, d): | |
| (r.BacktrackCoverage, r.InputCoverage, r.LookAheadCoverage) = d | |
| ( | |
| r.BacktrackGlyphCount, | |
| r.InputGlyphCount, | |
| r.LookAheadGlyphCount, | |
| ) = (len(x) for x in d) | |
| else: | |
| assert 0, "unknown format: %s" % Format | |
| if Chain: | |
| self.Coverage = ChainCoverage | |
| self.ContextData = ChainContextData | |
| self.SetContextData = SetChainContextData | |
| self.RuleData = ChainRuleData | |
| self.SetRuleData = ChainSetRuleData | |
| else: | |
| self.Coverage = Coverage | |
| self.ContextData = ContextData | |
| self.SetContextData = SetContextData | |
| self.RuleData = RuleData | |
| self.SetRuleData = SetRuleData | |
| if Format == 1: | |
| self.Rule = ChainTyp + "Rule" | |
| self.RuleCount = ChainTyp + "RuleCount" | |
| self.RuleSet = ChainTyp + "RuleSet" | |
| self.RuleSetCount = ChainTyp + "RuleSetCount" | |
| self.Intersect = lambda glyphs, c, r: [r] if r in glyphs else [] | |
| elif Format == 2: | |
| self.Rule = ChainTyp + "ClassRule" | |
| self.RuleCount = ChainTyp + "ClassRuleCount" | |
| self.RuleSet = ChainTyp + "ClassSet" | |
| self.RuleSetCount = ChainTyp + "ClassSetCount" | |
| self.Intersect = lambda glyphs, c, r: ( | |
| c.intersect_class(glyphs, r) | |
| if c | |
| else (set(glyphs) if r == 0 else set()) | |
| ) | |
| self.ClassDef = "InputClassDef" if Chain else "ClassDef" | |
| self.ClassDefIndex = 1 if Chain else 0 | |
| self.Input = "Input" if Chain else "Class" | |
| elif Format == 3: | |
| self.Input = "InputCoverage" if Chain else "Coverage" | |
| if self.Format not in [1, 2, 3]: | |
| return None # Don't shoot the messenger; let it go | |
| if not hasattr(self.__class__, "_subset__ContextHelpers"): | |
| self.__class__._subset__ContextHelpers = {} | |
| if self.Format not in self.__class__._subset__ContextHelpers: | |
| helper = ContextHelper(self.__class__, self.Format) | |
| self.__class__._subset__ContextHelpers[self.Format] = helper | |
| return self.__class__._subset__ContextHelpers[self.Format] | |
| def closure_glyphs(self, s, cur_glyphs): | |
| c = self.__subset_classify_context() | |
| indices = c.Coverage(self).intersect(cur_glyphs) | |
| if not indices: | |
| return [] | |
| cur_glyphs = c.Coverage(self).intersect_glyphs(cur_glyphs) | |
| if self.Format == 1: | |
| ContextData = c.ContextData(self) | |
| rss = getattr(self, c.RuleSet) | |
| rssCount = getattr(self, c.RuleSetCount) | |
| for i in indices: | |
| if i >= rssCount or not rss[i]: | |
| continue | |
| for r in getattr(rss[i], c.Rule): | |
| if not r: | |
| continue | |
| if not all( | |
| all(c.Intersect(s.glyphs, cd, k) for k in klist) | |
| for cd, klist in zip(ContextData, c.RuleData(r)) | |
| ): | |
| continue | |
| chaos = set() | |
| for ll in getattr(r, c.LookupRecord): | |
| if not ll: | |
| continue | |
| seqi = ll.SequenceIndex | |
| if seqi in chaos: | |
| # TODO Can we improve this? | |
| pos_glyphs = None | |
| else: | |
| if seqi == 0: | |
| pos_glyphs = frozenset([c.Coverage(self).glyphs[i]]) | |
| else: | |
| pos_glyphs = frozenset([r.Input[seqi - 1]]) | |
| lookup = s.table.LookupList.Lookup[ll.LookupListIndex] | |
| chaos.add(seqi) | |
| if lookup.may_have_non_1to1(): | |
| chaos.update(range(seqi, len(r.Input) + 2)) | |
| lookup.closure_glyphs(s, cur_glyphs=pos_glyphs) | |
| elif self.Format == 2: | |
| ClassDef = getattr(self, c.ClassDef) | |
| indices = ClassDef.intersect(cur_glyphs) | |
| ContextData = c.ContextData(self) | |
| rss = getattr(self, c.RuleSet) | |
| rssCount = getattr(self, c.RuleSetCount) | |
| for i in indices: | |
| if i >= rssCount or not rss[i]: | |
| continue | |
| for r in getattr(rss[i], c.Rule): | |
| if not r: | |
| continue | |
| if not all( | |
| all(c.Intersect(s.glyphs, cd, k) for k in klist) | |
| for cd, klist in zip(ContextData, c.RuleData(r)) | |
| ): | |
| continue | |
| chaos = set() | |
| for ll in getattr(r, c.LookupRecord): | |
| if not ll: | |
| continue | |
| seqi = ll.SequenceIndex | |
| if seqi in chaos: | |
| # TODO Can we improve this? | |
| pos_glyphs = None | |
| else: | |
| if seqi == 0: | |
| pos_glyphs = frozenset( | |
| ClassDef.intersect_class(cur_glyphs, i) | |
| ) | |
| else: | |
| pos_glyphs = frozenset( | |
| ClassDef.intersect_class( | |
| s.glyphs, getattr(r, c.Input)[seqi - 1] | |
| ) | |
| ) | |
| lookup = s.table.LookupList.Lookup[ll.LookupListIndex] | |
| chaos.add(seqi) | |
| if lookup.may_have_non_1to1(): | |
| chaos.update(range(seqi, len(getattr(r, c.Input)) + 2)) | |
| lookup.closure_glyphs(s, cur_glyphs=pos_glyphs) | |
| elif self.Format == 3: | |
| if not all(x is not None and x.intersect(s.glyphs) for x in c.RuleData(self)): | |
| return [] | |
| r = self | |
| input_coverages = getattr(r, c.Input) | |
| chaos = set() | |
| for ll in getattr(r, c.LookupRecord): | |
| if not ll: | |
| continue | |
| seqi = ll.SequenceIndex | |
| if seqi in chaos: | |
| # TODO Can we improve this? | |
| pos_glyphs = None | |
| else: | |
| if seqi == 0: | |
| pos_glyphs = frozenset(cur_glyphs) | |
| else: | |
| pos_glyphs = frozenset( | |
| input_coverages[seqi].intersect_glyphs(s.glyphs) | |
| ) | |
| lookup = s.table.LookupList.Lookup[ll.LookupListIndex] | |
| chaos.add(seqi) | |
| if lookup.may_have_non_1to1(): | |
| chaos.update(range(seqi, len(input_coverages) + 1)) | |
| lookup.closure_glyphs(s, cur_glyphs=pos_glyphs) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def subset_glyphs(self, s): | |
| c = self.__subset_classify_context() | |
| if self.Format == 1: | |
| indices = self.Coverage.subset(s.glyphs) | |
| rss = getattr(self, c.RuleSet) | |
| rssCount = getattr(self, c.RuleSetCount) | |
| rss = [rss[i] for i in indices if i < rssCount] | |
| for rs in rss: | |
| if not rs: | |
| continue | |
| ss = getattr(rs, c.Rule) | |
| ss = [ | |
| r | |
| for r in ss | |
| if r | |
| and all(all(g in s.glyphs for g in glist) for glist in c.RuleData(r)) | |
| ] | |
| setattr(rs, c.Rule, ss) | |
| setattr(rs, c.RuleCount, len(ss)) | |
| # Prune empty rulesets | |
| indices = [i for i, rs in enumerate(rss) if rs and getattr(rs, c.Rule)] | |
| self.Coverage.remap(indices) | |
| rss = _list_subset(rss, indices) | |
| setattr(self, c.RuleSet, rss) | |
| setattr(self, c.RuleSetCount, len(rss)) | |
| return bool(rss) | |
| elif self.Format == 2: | |
| if not self.Coverage.subset(s.glyphs): | |
| return False | |
| ContextData = c.ContextData(self) | |
| klass_maps = [ | |
| x.subset(s.glyphs, remap=True) if x else None for x in ContextData | |
| ] | |
| # Keep rulesets for class numbers that survived. | |
| indices = klass_maps[c.ClassDefIndex] | |
| rss = getattr(self, c.RuleSet) | |
| rssCount = getattr(self, c.RuleSetCount) | |
| rss = [rss[i] for i in indices if i < rssCount] | |
| del rssCount | |
| # Delete, but not renumber, unreachable rulesets. | |
| indices = getattr(self, c.ClassDef).intersect(self.Coverage.glyphs) | |
| rss = [rss if i in indices else None for i, rss in enumerate(rss)] | |
| for rs in rss: | |
| if not rs: | |
| continue | |
| ss = getattr(rs, c.Rule) | |
| ss = [ | |
| r | |
| for r in ss | |
| if r | |
| and all( | |
| all(k in klass_map for k in klist) | |
| for klass_map, klist in zip(klass_maps, c.RuleData(r)) | |
| ) | |
| ] | |
| setattr(rs, c.Rule, ss) | |
| setattr(rs, c.RuleCount, len(ss)) | |
| # Remap rule classes | |
| for r in ss: | |
| c.SetRuleData( | |
| r, | |
| [ | |
| [klass_map.index(k) for k in klist] | |
| for klass_map, klist in zip(klass_maps, c.RuleData(r)) | |
| ], | |
| ) | |
| # Prune empty rulesets | |
| rss = [rs if rs and getattr(rs, c.Rule) else None for rs in rss] | |
| while rss and rss[-1] is None: | |
| del rss[-1] | |
| setattr(self, c.RuleSet, rss) | |
| setattr(self, c.RuleSetCount, len(rss)) | |
| # TODO: We can do a second round of remapping class values based | |
| # on classes that are actually used in at least one rule. Right | |
| # now we subset classes to c.glyphs only. Or better, rewrite | |
| # the above to do that. | |
| return bool(rss) | |
| elif self.Format == 3: | |
| return all(x is not None and x.subset(s.glyphs) for x in c.RuleData(self)) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def subset_lookups(self, lookup_indices): | |
| c = self.__subset_classify_context() | |
| if self.Format in [1, 2]: | |
| for rs in getattr(self, c.RuleSet): | |
| if not rs: | |
| continue | |
| for r in getattr(rs, c.Rule): | |
| if not r: | |
| continue | |
| setattr( | |
| r, | |
| c.LookupRecord, | |
| [ | |
| ll | |
| for ll in getattr(r, c.LookupRecord) | |
| if ll and ll.LookupListIndex in lookup_indices | |
| ], | |
| ) | |
| for ll in getattr(r, c.LookupRecord): | |
| if not ll: | |
| continue | |
| ll.LookupListIndex = lookup_indices.index(ll.LookupListIndex) | |
| elif self.Format == 3: | |
| setattr( | |
| self, | |
| c.LookupRecord, | |
| [ | |
| ll | |
| for ll in getattr(self, c.LookupRecord) | |
| if ll and ll.LookupListIndex in lookup_indices | |
| ], | |
| ) | |
| for ll in getattr(self, c.LookupRecord): | |
| if not ll: | |
| continue | |
| ll.LookupListIndex = lookup_indices.index(ll.LookupListIndex) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def collect_lookups(self): | |
| c = self.__subset_classify_context() | |
| if self.Format in [1, 2]: | |
| return [ | |
| ll.LookupListIndex | |
| for rs in getattr(self, c.RuleSet) | |
| if rs | |
| for r in getattr(rs, c.Rule) | |
| if r | |
| for ll in getattr(r, c.LookupRecord) | |
| if ll | |
| ] | |
| elif self.Format == 3: | |
| return [ll.LookupListIndex for ll in getattr(self, c.LookupRecord) if ll] | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def closure_glyphs(self, s, cur_glyphs): | |
| if self.Format == 1: | |
| self.ExtSubTable.closure_glyphs(s, cur_glyphs) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def may_have_non_1to1(self): | |
| if self.Format == 1: | |
| return self.ExtSubTable.may_have_non_1to1() | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def subset_glyphs(self, s): | |
| if self.Format == 1: | |
| return self.ExtSubTable.subset_glyphs(s) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def prune_post_subset(self, font, options): | |
| if self.Format == 1: | |
| return self.ExtSubTable.prune_post_subset(font, options) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def subset_lookups(self, lookup_indices): | |
| if self.Format == 1: | |
| return self.ExtSubTable.subset_lookups(lookup_indices) | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def collect_lookups(self): | |
| if self.Format == 1: | |
| return self.ExtSubTable.collect_lookups() | |
| else: | |
| assert 0, "unknown format: %s" % self.Format | |
| def closure_glyphs(self, s, cur_glyphs=None): | |
| if cur_glyphs is None: | |
| cur_glyphs = frozenset(s.glyphs) | |
| # Memoize | |
| key = id(self) | |
| doneLookups = s._doneLookups | |
| count, covered = doneLookups.get(key, (0, None)) | |
| if count != len(s.glyphs): | |
| count, covered = doneLookups[key] = (len(s.glyphs), set()) | |
| if cur_glyphs.issubset(covered): | |
| return | |
| covered.update(cur_glyphs) | |
| for st in self.SubTable: | |
| if not st: | |
| continue | |
| st.closure_glyphs(s, cur_glyphs) | |
| def subset_glyphs(self, s): | |
| self.SubTable = [st for st in self.SubTable if st and st.subset_glyphs(s)] | |
| self.SubTableCount = len(self.SubTable) | |
| if hasattr(self, "MarkFilteringSet") and self.MarkFilteringSet is not None: | |
| if self.MarkFilteringSet not in s.used_mark_sets: | |
| self.MarkFilteringSet = None | |
| self.LookupFlag &= ~0x10 | |
| else: | |
| self.MarkFilteringSet = s.used_mark_sets.index(self.MarkFilteringSet) | |
| return bool(self.SubTableCount) | |
| def prune_post_subset(self, font, options): | |
| ret = False | |
| for st in self.SubTable: | |
| if not st: | |
| continue | |
| if st.prune_post_subset(font, options): | |
| ret = True | |
| return ret | |
| def subset_lookups(self, lookup_indices): | |
| for s in self.SubTable: | |
| s.subset_lookups(lookup_indices) | |
| def collect_lookups(self): | |
| return sum((st.collect_lookups() for st in self.SubTable if st), []) | |
| def may_have_non_1to1(self): | |
| return any(st.may_have_non_1to1() for st in self.SubTable if st) | |
| def subset_glyphs(self, s): | |
| """Returns the indices of nonempty lookups.""" | |
| return [i for i, l in enumerate(self.Lookup) if l and l.subset_glyphs(s)] | |
| def prune_post_subset(self, font, options): | |
| ret = False | |
| for l in self.Lookup: | |
| if not l: | |
| continue | |
| if l.prune_post_subset(font, options): | |
| ret = True | |
| return ret | |
| def subset_lookups(self, lookup_indices): | |
| self.ensureDecompiled() | |
| self.Lookup = [self.Lookup[i] for i in lookup_indices if i < self.LookupCount] | |
| self.LookupCount = len(self.Lookup) | |
| for l in self.Lookup: | |
| l.subset_lookups(lookup_indices) | |
| def neuter_lookups(self, lookup_indices): | |
| """Sets lookups not in lookup_indices to None.""" | |
| self.ensureDecompiled() | |
| self.Lookup = [ | |
| l if i in lookup_indices else None for i, l in enumerate(self.Lookup) | |
| ] | |
| def closure_lookups(self, lookup_indices): | |
| """Returns sorted index of all lookups reachable from lookup_indices.""" | |
| lookup_indices = _uniq_sort(lookup_indices) | |
| recurse = lookup_indices | |
| while True: | |
| recurse_lookups = sum( | |
| (self.Lookup[i].collect_lookups() for i in recurse if i < self.LookupCount), | |
| [], | |
| ) | |
| recurse_lookups = [ | |
| l | |
| for l in recurse_lookups | |
| if l not in lookup_indices and l < self.LookupCount | |
| ] | |
| if not recurse_lookups: | |
| return _uniq_sort(lookup_indices) | |
| recurse_lookups = _uniq_sort(recurse_lookups) | |
| lookup_indices.extend(recurse_lookups) | |
| recurse = recurse_lookups | |
| def subset_lookups(self, lookup_indices): | |
| """ "Returns True if feature is non-empty afterwards.""" | |
| self.LookupListIndex = [l for l in self.LookupListIndex if l in lookup_indices] | |
| # Now map them. | |
| self.LookupListIndex = [lookup_indices.index(l) for l in self.LookupListIndex] | |
| self.LookupCount = len(self.LookupListIndex) | |
| # keep 'size' feature even if it contains no lookups; but drop any other | |
| # empty feature (e.g. FeatureParams for stylistic set names) | |
| # https://github.com/fonttools/fonttools/issues/2324 | |
| return self.LookupCount or isinstance( | |
| self.FeatureParams, otTables.FeatureParamsSize | |
| ) | |
| def subset_lookups(self, lookup_indices): | |
| """Returns the indices of nonempty features.""" | |
| # Note: Never ever drop feature 'pref', even if it's empty. | |
| # HarfBuzz chooses shaper for Khmer based on presence of this | |
| # feature. See thread at: | |
| # http://lists.freedesktop.org/archives/harfbuzz/2012-November/002660.html | |
| return [ | |
| i | |
| for i, f in enumerate(self.FeatureRecord) | |
| if (f.Feature.subset_lookups(lookup_indices) or f.FeatureTag == "pref") | |
| ] | |
| def collect_lookups(self, feature_indices): | |
| return sum( | |
| ( | |
| self.FeatureRecord[i].Feature.LookupListIndex | |
| for i in feature_indices | |
| if i < self.FeatureCount | |
| ), | |
| [], | |
| ) | |
| def subset_features(self, feature_indices): | |
| self.ensureDecompiled() | |
| self.FeatureRecord = _list_subset(self.FeatureRecord, feature_indices) | |
| self.FeatureCount = len(self.FeatureRecord) | |
| return bool(self.FeatureCount) | |
| def subset_lookups(self, lookup_indices): | |
| """Returns the indices of nonempty features.""" | |
| return [ | |
| r.FeatureIndex | |
| for r in self.SubstitutionRecord | |
| if r.Feature.subset_lookups(lookup_indices) | |
| ] | |
| def subset_lookups(self, lookup_indices): | |
| """Returns the indices of nonempty features.""" | |
| return sum( | |
| ( | |
| f.FeatureTableSubstitution.subset_lookups(lookup_indices) | |
| for f in self.FeatureVariationRecord | |
| ), | |
| [], | |
| ) | |
| def collect_lookups(self, feature_indices): | |
| return sum( | |
| ( | |
| r.Feature.LookupListIndex | |
| for vr in self.FeatureVariationRecord | |
| for r in vr.FeatureTableSubstitution.SubstitutionRecord | |
| if r.FeatureIndex in feature_indices | |
| ), | |
| [], | |
| ) | |
| def subset_features(self, feature_indices): | |
| self.ensureDecompiled() | |
| self.SubstitutionRecord = [ | |
| r for r in self.SubstitutionRecord if r.FeatureIndex in feature_indices | |
| ] | |
| # remap feature indices | |
| for r in self.SubstitutionRecord: | |
| r.FeatureIndex = feature_indices.index(r.FeatureIndex) | |
| self.SubstitutionCount = len(self.SubstitutionRecord) | |
| return bool(self.SubstitutionCount) | |
| def subset_features(self, feature_indices): | |
| self.ensureDecompiled() | |
| for r in self.FeatureVariationRecord: | |
| r.FeatureTableSubstitution.subset_features(feature_indices) | |
| # Prune empty records at the end only | |
| # https://github.com/fonttools/fonttools/issues/1881 | |
| while ( | |
| self.FeatureVariationRecord | |
| and not self.FeatureVariationRecord[ | |
| -1 | |
| ].FeatureTableSubstitution.SubstitutionCount | |
| ): | |
| self.FeatureVariationRecord.pop() | |
| self.FeatureVariationCount = len(self.FeatureVariationRecord) | |
| return bool(self.FeatureVariationCount) | |
| def subset_features(self, feature_indices): | |
| if self.ReqFeatureIndex in feature_indices: | |
| self.ReqFeatureIndex = feature_indices.index(self.ReqFeatureIndex) | |
| else: | |
| self.ReqFeatureIndex = 65535 | |
| self.FeatureIndex = [f for f in self.FeatureIndex if f in feature_indices] | |
| # Now map them. | |
| self.FeatureIndex = [ | |
| feature_indices.index(f) for f in self.FeatureIndex if f in feature_indices | |
| ] | |
| self.FeatureCount = len(self.FeatureIndex) | |
| return bool(self.FeatureCount or self.ReqFeatureIndex != 65535) | |
| def collect_features(self): | |
| feature_indices = self.FeatureIndex[:] | |
| if self.ReqFeatureIndex != 65535: | |
| feature_indices.append(self.ReqFeatureIndex) | |
| return _uniq_sort(feature_indices) | |
| def subset_features(self, feature_indices, keepEmptyDefaultLangSys=False): | |
| if ( | |
| self.DefaultLangSys | |
| and not self.DefaultLangSys.subset_features(feature_indices) | |
| and not keepEmptyDefaultLangSys | |
| ): | |
| self.DefaultLangSys = None | |
| self.LangSysRecord = [ | |
| l for l in self.LangSysRecord if l.LangSys.subset_features(feature_indices) | |
| ] | |
| self.LangSysCount = len(self.LangSysRecord) | |
| return bool(self.LangSysCount or self.DefaultLangSys) | |
| def collect_features(self): | |
| feature_indices = [l.LangSys.collect_features() for l in self.LangSysRecord] | |
| if self.DefaultLangSys: | |
| feature_indices.append(self.DefaultLangSys.collect_features()) | |
| return _uniq_sort(sum(feature_indices, [])) | |
| def subset_features(self, feature_indices, retain_empty): | |
| # https://bugzilla.mozilla.org/show_bug.cgi?id=1331737#c32 | |
| self.ScriptRecord = [ | |
| s | |
| for s in self.ScriptRecord | |
| if s.Script.subset_features(feature_indices, s.ScriptTag == "DFLT") | |
| or retain_empty | |
| ] | |
| self.ScriptCount = len(self.ScriptRecord) | |
| return bool(self.ScriptCount) | |
| def collect_features(self): | |
| return _uniq_sort(sum((s.Script.collect_features() for s in self.ScriptRecord), [])) | |
| # CBLC will inherit it | |
| def subset_glyphs(self, s): | |
| for strike in self.strikes: | |
| for indexSubTable in strike.indexSubTables: | |
| indexSubTable.names = [n for n in indexSubTable.names if n in s.glyphs] | |
| strike.indexSubTables = [i for i in strike.indexSubTables if i.names] | |
| self.strikes = [s for s in self.strikes if s.indexSubTables] | |
| return True | |
| # CBDT will inherit it | |
| def subset_glyphs(self, s): | |
| strikeData = [ | |
| {g: strike[g] for g in s.glyphs if g in strike} for strike in self.strikeData | |
| ] | |
| # Prune empty strikes | |
| # https://github.com/fonttools/fonttools/issues/1633 | |
| self.strikeData = [strike for strike in strikeData if strike] | |
| return True | |
| def subset_glyphs(self, s): | |
| for strike in self.strikes.values(): | |
| strike.glyphs = {g: strike.glyphs[g] for g in s.glyphs if g in strike.glyphs} | |
| return True | |
| def closure_glyphs(self, s): | |
| s.table = self.table | |
| if self.table.ScriptList: | |
| feature_indices = self.table.ScriptList.collect_features() | |
| else: | |
| feature_indices = [] | |
| if self.table.FeatureList: | |
| lookup_indices = self.table.FeatureList.collect_lookups(feature_indices) | |
| else: | |
| lookup_indices = [] | |
| if getattr(self.table, "FeatureVariations", None): | |
| lookup_indices += self.table.FeatureVariations.collect_lookups(feature_indices) | |
| lookup_indices = _uniq_sort(lookup_indices) | |
| if self.table.LookupList: | |
| s._doneLookups = {} | |
| while True: | |
| orig_glyphs = frozenset(s.glyphs) | |
| for i in lookup_indices: | |
| if i >= self.table.LookupList.LookupCount: | |
| continue | |
| if not self.table.LookupList.Lookup[i]: | |
| continue | |
| self.table.LookupList.Lookup[i].closure_glyphs(s) | |
| if orig_glyphs == s.glyphs: | |
| break | |
| del s._doneLookups | |
| del s.table | |
| def subset_glyphs(self, s): | |
| s.glyphs = s.glyphs_gsubed | |
| if self.table.LookupList: | |
| lookup_indices = self.table.LookupList.subset_glyphs(s) | |
| else: | |
| lookup_indices = [] | |
| self.subset_lookups(lookup_indices) | |
| return True | |
| def retain_empty_scripts(self): | |
| # https://github.com/fonttools/fonttools/issues/518 | |
| # https://bugzilla.mozilla.org/show_bug.cgi?id=1080739#c15 | |
| return self.__class__ == ttLib.getTableClass("GSUB") | |
| def subset_lookups(self, lookup_indices): | |
| """Retains specified lookups, then removes empty features, language | |
| systems, and scripts.""" | |
| if self.table.LookupList: | |
| self.table.LookupList.subset_lookups(lookup_indices) | |
| if self.table.FeatureList: | |
| feature_indices = self.table.FeatureList.subset_lookups(lookup_indices) | |
| else: | |
| feature_indices = [] | |
| if getattr(self.table, "FeatureVariations", None): | |
| feature_indices += self.table.FeatureVariations.subset_lookups(lookup_indices) | |
| feature_indices = _uniq_sort(feature_indices) | |
| if self.table.FeatureList: | |
| self.table.FeatureList.subset_features(feature_indices) | |
| if getattr(self.table, "FeatureVariations", None): | |
| self.table.FeatureVariations.subset_features(feature_indices) | |
| if self.table.ScriptList: | |
| self.table.ScriptList.subset_features( | |
| feature_indices, self.retain_empty_scripts() | |
| ) | |
| def neuter_lookups(self, lookup_indices): | |
| """Sets lookups not in lookup_indices to None.""" | |
| if self.table.LookupList: | |
| self.table.LookupList.neuter_lookups(lookup_indices) | |
| def prune_lookups(self, remap=True): | |
| """Remove (default) or neuter unreferenced lookups""" | |
| if self.table.ScriptList: | |
| feature_indices = self.table.ScriptList.collect_features() | |
| else: | |
| feature_indices = [] | |
| if self.table.FeatureList: | |
| lookup_indices = self.table.FeatureList.collect_lookups(feature_indices) | |
| else: | |
| lookup_indices = [] | |
| if getattr(self.table, "FeatureVariations", None): | |
| lookup_indices += self.table.FeatureVariations.collect_lookups(feature_indices) | |
| lookup_indices = _uniq_sort(lookup_indices) | |
| if self.table.LookupList: | |
| lookup_indices = self.table.LookupList.closure_lookups(lookup_indices) | |
| else: | |
| lookup_indices = [] | |
| if remap: | |
| self.subset_lookups(lookup_indices) | |
| else: | |
| self.neuter_lookups(lookup_indices) | |
| def subset_feature_tags(self, feature_tags): | |
| if self.table.FeatureList: | |
| feature_indices = [ | |
| i | |
| for i, f in enumerate(self.table.FeatureList.FeatureRecord) | |
| if f.FeatureTag in feature_tags | |
| ] | |
| self.table.FeatureList.subset_features(feature_indices) | |
| if getattr(self.table, "FeatureVariations", None): | |
| self.table.FeatureVariations.subset_features(feature_indices) | |
| else: | |
| feature_indices = [] | |
| if self.table.ScriptList: | |
| self.table.ScriptList.subset_features( | |
| feature_indices, self.retain_empty_scripts() | |
| ) | |
| def subset_script_tags(self, tags): | |
| langsys = {} | |
| script_tags = set() | |
| for tag in tags: | |
| script_tag, lang_tag = tag.split(".") if "." in tag else (tag, "*") | |
| script_tags.add(script_tag.ljust(4)) | |
| langsys.setdefault(script_tag, set()).add(lang_tag.ljust(4)) | |
| if self.table.ScriptList: | |
| self.table.ScriptList.ScriptRecord = [ | |
| s for s in self.table.ScriptList.ScriptRecord if s.ScriptTag in script_tags | |
| ] | |
| self.table.ScriptList.ScriptCount = len(self.table.ScriptList.ScriptRecord) | |
| for record in self.table.ScriptList.ScriptRecord: | |
| if record.ScriptTag in langsys and "* " not in langsys[record.ScriptTag]: | |
| record.Script.LangSysRecord = [ | |
| l | |
| for l in record.Script.LangSysRecord | |
| if l.LangSysTag in langsys[record.ScriptTag] | |
| ] | |
| record.Script.LangSysCount = len(record.Script.LangSysRecord) | |
| if "dflt" not in langsys[record.ScriptTag]: | |
| record.Script.DefaultLangSys = None | |
| def prune_features(self): | |
| """Remove unreferenced features""" | |
| if self.table.ScriptList: | |
| feature_indices = self.table.ScriptList.collect_features() | |
| else: | |
| feature_indices = [] | |
| if self.table.FeatureList: | |
| self.table.FeatureList.subset_features(feature_indices) | |
| if getattr(self.table, "FeatureVariations", None): | |
| self.table.FeatureVariations.subset_features(feature_indices) | |
| if self.table.ScriptList: | |
| self.table.ScriptList.subset_features( | |
| feature_indices, self.retain_empty_scripts() | |
| ) | |
| def prune_pre_subset(self, font, options): | |
| # Drop undesired features | |
| if "*" not in options.layout_scripts: | |
| self.subset_script_tags(options.layout_scripts) | |
| if "*" not in options.layout_features: | |
| self.subset_feature_tags(options.layout_features) | |
| # Neuter unreferenced lookups | |
| self.prune_lookups(remap=False) | |
| return True | |
| def remove_redundant_langsys(self): | |
| table = self.table | |
| if not table.ScriptList or not table.FeatureList: | |
| return | |
| features = table.FeatureList.FeatureRecord | |
| for s in table.ScriptList.ScriptRecord: | |
| d = s.Script.DefaultLangSys | |
| if not d: | |
| continue | |
| for lr in s.Script.LangSysRecord[:]: | |
| l = lr.LangSys | |
| # Compare d and l | |
| if len(d.FeatureIndex) != len(l.FeatureIndex): | |
| continue | |
| if (d.ReqFeatureIndex == 65535) != (l.ReqFeatureIndex == 65535): | |
| continue | |
| if d.ReqFeatureIndex != 65535: | |
| if features[d.ReqFeatureIndex] != features[l.ReqFeatureIndex]: | |
| continue | |
| for i in range(len(d.FeatureIndex)): | |
| if features[d.FeatureIndex[i]] != features[l.FeatureIndex[i]]: | |
| break | |
| else: | |
| # LangSys and default are equal; delete LangSys | |
| s.Script.LangSysRecord.remove(lr) | |
| def prune_post_subset(self, font, options): | |
| table = self.table | |
| self.prune_lookups() # XXX Is this actually needed?! | |
| if table.LookupList: | |
| table.LookupList.prune_post_subset(font, options) | |
| # XXX Next two lines disabled because OTS is stupid and | |
| # doesn't like NULL offsets here. | |
| # if not table.LookupList.Lookup: | |
| # table.LookupList = None | |
| if not table.LookupList: | |
| table.FeatureList = None | |
| if table.FeatureList: | |
| self.remove_redundant_langsys() | |
| # Remove unreferenced features | |
| self.prune_features() | |
| # XXX Next two lines disabled because OTS is stupid and | |
| # doesn't like NULL offsets here. | |
| # if table.FeatureList and not table.FeatureList.FeatureRecord: | |
| # table.FeatureList = None | |
| # Never drop scripts themselves as them just being available | |
| # holds semantic significance. | |
| # XXX Next two lines disabled because OTS is stupid and | |
| # doesn't like NULL offsets here. | |
| # if table.ScriptList and not table.ScriptList.ScriptRecord: | |
| # table.ScriptList = None | |
| if hasattr(table, "FeatureVariations"): | |
| # drop FeatureVariations if there are no features to substitute | |
| if table.FeatureVariations and not ( | |
| table.FeatureList and table.FeatureVariations.FeatureVariationRecord | |
| ): | |
| table.FeatureVariations = None | |
| # downgrade table version if there are no FeatureVariations | |
| if not table.FeatureVariations and table.Version == 0x00010001: | |
| table.Version = 0x00010000 | |
| return True | |
| def subset_glyphs(self, s): | |
| glyphs = s.glyphs_gsubed | |
| table = self.table | |
| if table.LigCaretList: | |
| indices = table.LigCaretList.Coverage.subset(glyphs) | |
| table.LigCaretList.LigGlyph = _list_subset(table.LigCaretList.LigGlyph, indices) | |
| table.LigCaretList.LigGlyphCount = len(table.LigCaretList.LigGlyph) | |
| if table.MarkAttachClassDef: | |
| table.MarkAttachClassDef.classDefs = { | |
| g: v for g, v in table.MarkAttachClassDef.classDefs.items() if g in glyphs | |
| } | |
| if table.GlyphClassDef: | |
| table.GlyphClassDef.classDefs = { | |
| g: v for g, v in table.GlyphClassDef.classDefs.items() if g in glyphs | |
| } | |
| if table.AttachList: | |
| indices = table.AttachList.Coverage.subset(glyphs) | |
| GlyphCount = table.AttachList.GlyphCount | |
| table.AttachList.AttachPoint = [ | |
| table.AttachList.AttachPoint[i] for i in indices if i < GlyphCount | |
| ] | |
| table.AttachList.GlyphCount = len(table.AttachList.AttachPoint) | |
| if hasattr(table, "MarkGlyphSetsDef") and table.MarkGlyphSetsDef: | |
| markGlyphSets = table.MarkGlyphSetsDef | |
| for coverage in markGlyphSets.Coverage: | |
| if coverage: | |
| coverage.subset(glyphs) | |
| s.used_mark_sets = [i for i, c in enumerate(markGlyphSets.Coverage) if c.glyphs] | |
| markGlyphSets.Coverage = [c for c in markGlyphSets.Coverage if c.glyphs] | |
| return True | |
| def _pruneGDEF(font): | |
| if "GDEF" not in font: | |
| return | |
| gdef = font["GDEF"] | |
| table = gdef.table | |
| if not hasattr(table, "VarStore"): | |
| return | |
| store = table.VarStore | |
| usedVarIdxes = set() | |
| # Collect. | |
| table.collect_device_varidxes(usedVarIdxes) | |
| if "GPOS" in font: | |
| font["GPOS"].table.collect_device_varidxes(usedVarIdxes) | |
| # Subset. | |
| varidx_map = store.subset_varidxes(usedVarIdxes) | |
| # Map. | |
| table.remap_device_varidxes(varidx_map) | |
| if "GPOS" in font: | |
| font["GPOS"].table.remap_device_varidxes(varidx_map) | |
| def prune_post_subset(self, font, options): | |
| table = self.table | |
| # XXX check these against OTS | |
| if table.LigCaretList and not table.LigCaretList.LigGlyphCount: | |
| table.LigCaretList = None | |
| if table.MarkAttachClassDef and not table.MarkAttachClassDef.classDefs: | |
| table.MarkAttachClassDef = None | |
| if table.GlyphClassDef and not table.GlyphClassDef.classDefs: | |
| table.GlyphClassDef = None | |
| if table.AttachList and not table.AttachList.GlyphCount: | |
| table.AttachList = None | |
| if hasattr(table, "VarStore"): | |
| _pruneGDEF(font) | |
| if table.VarStore.VarDataCount == 0: | |
| if table.Version == 0x00010003: | |
| table.Version = 0x00010002 | |
| if ( | |
| not hasattr(table, "MarkGlyphSetsDef") | |
| or not table.MarkGlyphSetsDef | |
| or not table.MarkGlyphSetsDef.Coverage | |
| ): | |
| table.MarkGlyphSetsDef = None | |
| if table.Version == 0x00010002: | |
| table.Version = 0x00010000 | |
| return bool( | |
| table.LigCaretList | |
| or table.MarkAttachClassDef | |
| or table.GlyphClassDef | |
| or table.AttachList | |
| or (table.Version >= 0x00010002 and table.MarkGlyphSetsDef) | |
| or (table.Version >= 0x00010003 and table.VarStore) | |
| ) | |
| def prune_pre_subset(self, font, options): | |
| # Prune unknown kern table types | |
| self.kernTables = [t for t in self.kernTables if hasattr(t, "kernTable")] | |
| return bool(self.kernTables) | |
| def subset_glyphs(self, s): | |
| glyphs = s.glyphs_gsubed | |
| for t in self.kernTables: | |
| t.kernTable = { | |
| (a, b): v | |
| for (a, b), v in t.kernTable.items() | |
| if a in glyphs and b in glyphs | |
| } | |
| self.kernTables = [t for t in self.kernTables if t.kernTable] | |
| return bool(self.kernTables) | |
| def subset_glyphs(self, s): | |
| self.metrics = _dict_subset(self.metrics, s.glyphs) | |
| for g in s.glyphs_emptied: | |
| self.metrics[g] = (0, 0) | |
| return bool(self.metrics) | |
| def subset_glyphs(self, s): | |
| self.metrics = _dict_subset(self.metrics, s.glyphs) | |
| for g in s.glyphs_emptied: | |
| self.metrics[g] = (0, 0) | |
| return True # Required table | |
| def subset_glyphs(self, s): | |
| self.hdmx = {sz: _dict_subset(l, s.glyphs) for sz, l in self.hdmx.items()} | |
| for sz in self.hdmx: | |
| for g in s.glyphs_emptied: | |
| self.hdmx[sz][g] = 0 | |
| return bool(self.hdmx) | |
| def subset_glyphs(self, s): | |
| table = self.table.AnchorPoints | |
| assert table.Format == 0, "unknown 'ankr' format %s" % table.Format | |
| table.Anchors = { | |
| glyph: table.Anchors[glyph] for glyph in s.glyphs if glyph in table.Anchors | |
| } | |
| return len(table.Anchors) > 0 | |
| def closure_glyphs(self, s): | |
| table = self.table.Baseline | |
| if table.Format in (2, 3): | |
| s.glyphs.add(table.StandardGlyph) | |
| def subset_glyphs(self, s): | |
| table = self.table.Baseline | |
| if table.Format in (1, 3): | |
| baselines = { | |
| glyph: table.BaselineValues.get(glyph, table.DefaultBaseline) | |
| for glyph in s.glyphs | |
| } | |
| if len(baselines) > 0: | |
| mostCommon, _cnt = Counter(baselines.values()).most_common(1)[0] | |
| table.DefaultBaseline = mostCommon | |
| baselines = {glyph: b for glyph, b in baselines.items() if b != mostCommon} | |
| if len(baselines) > 0: | |
| table.BaselineValues = baselines | |
| else: | |
| table.Format = {1: 0, 3: 2}[table.Format] | |
| del table.BaselineValues | |
| return True | |
| def subset_glyphs(self, s): | |
| table = self.table.LigatureCarets | |
| if table.Format in (0, 1): | |
| table.Carets = { | |
| glyph: table.Carets[glyph] for glyph in s.glyphs if glyph in table.Carets | |
| } | |
| return len(table.Carets) > 0 | |
| else: | |
| assert False, "unknown 'lcar' format %s" % table.Format | |
| def prune_pre_subset(self, font, options): | |
| if options.notdef_glyph and not options.notdef_outline: | |
| self.variations[font.glyphOrder[0]] = [] | |
| return True | |
| def subset_glyphs(self, s): | |
| self.variations = _dict_subset(self.variations, s.glyphs) | |
| self.glyphCount = len(self.variations) | |
| return bool(self.variations) | |
| def _remap_index_map(s, varidx_map, table_map): | |
| map_ = {k: varidx_map[v] for k, v in table_map.mapping.items()} | |
| # Emptied glyphs are remapped to: | |
| # if GID <= last retained GID, 0/0: delta set for 0/0 is expected to exist & zeros compress well | |
| # if GID > last retained GID, major/minor of the last retained glyph: will be optimized out by table compiler | |
| last_idx = varidx_map[table_map.mapping[s.last_retained_glyph]] | |
| for g, i in s.reverseEmptiedGlyphMap.items(): | |
| map_[g] = last_idx if i > s.last_retained_order else 0 | |
| return map_ | |
| def subset_glyphs(self, s): | |
| table = self.table | |
| used = set() | |
| advIdxes_ = set() | |
| retainAdvMap = False | |
| if table.AdvWidthMap: | |
| table.AdvWidthMap.mapping = _dict_subset(table.AdvWidthMap.mapping, s.glyphs) | |
| used.update(table.AdvWidthMap.mapping.values()) | |
| else: | |
| used.update(s.reverseOrigGlyphMap.values()) | |
| advIdxes_ = used.copy() | |
| retainAdvMap = s.options.retain_gids | |
| if table.LsbMap: | |
| table.LsbMap.mapping = _dict_subset(table.LsbMap.mapping, s.glyphs) | |
| used.update(table.LsbMap.mapping.values()) | |
| if table.RsbMap: | |
| table.RsbMap.mapping = _dict_subset(table.RsbMap.mapping, s.glyphs) | |
| used.update(table.RsbMap.mapping.values()) | |
| varidx_map = table.VarStore.subset_varidxes( | |
| used, retainFirstMap=retainAdvMap, advIdxes=advIdxes_ | |
| ) | |
| if table.AdvWidthMap: | |
| table.AdvWidthMap.mapping = _remap_index_map(s, varidx_map, table.AdvWidthMap) | |
| if table.LsbMap: | |
| table.LsbMap.mapping = _remap_index_map(s, varidx_map, table.LsbMap) | |
| if table.RsbMap: | |
| table.RsbMap.mapping = _remap_index_map(s, varidx_map, table.RsbMap) | |
| # TODO Return emptiness... | |
| return True | |
| def subset_glyphs(self, s): | |
| table = self.table | |
| used = set() | |
| advIdxes_ = set() | |
| retainAdvMap = False | |
| if table.AdvHeightMap: | |
| table.AdvHeightMap.mapping = _dict_subset(table.AdvHeightMap.mapping, s.glyphs) | |
| used.update(table.AdvHeightMap.mapping.values()) | |
| else: | |
| used.update(s.reverseOrigGlyphMap.values()) | |
| advIdxes_ = used.copy() | |
| retainAdvMap = s.options.retain_gids | |
| if table.TsbMap: | |
| table.TsbMap.mapping = _dict_subset(table.TsbMap.mapping, s.glyphs) | |
| used.update(table.TsbMap.mapping.values()) | |
| if table.BsbMap: | |
| table.BsbMap.mapping = _dict_subset(table.BsbMap.mapping, s.glyphs) | |
| used.update(table.BsbMap.mapping.values()) | |
| if table.VOrgMap: | |
| table.VOrgMap.mapping = _dict_subset(table.VOrgMap.mapping, s.glyphs) | |
| used.update(table.VOrgMap.mapping.values()) | |
| varidx_map = table.VarStore.subset_varidxes( | |
| used, retainFirstMap=retainAdvMap, advIdxes=advIdxes_ | |
| ) | |
| if table.AdvHeightMap: | |
| table.AdvHeightMap.mapping = _remap_index_map(s, varidx_map, table.AdvHeightMap) | |
| if table.TsbMap: | |
| table.TsbMap.mapping = _remap_index_map(s, varidx_map, table.TsbMap) | |
| if table.BsbMap: | |
| table.BsbMap.mapping = _remap_index_map(s, varidx_map, table.BsbMap) | |
| if table.VOrgMap: | |
| table.VOrgMap.mapping = _remap_index_map(s, varidx_map, table.VOrgMap) | |
| # TODO Return emptiness... | |
| return True | |
| def subset_glyphs(self, s): | |
| self.VOriginRecords = { | |
| g: v for g, v in self.VOriginRecords.items() if g in s.glyphs | |
| } | |
| self.numVertOriginYMetrics = len(self.VOriginRecords) | |
| return True # Never drop; has default metrics | |
| def subset_glyphs(self, s): | |
| table = self.table.OpticalBounds | |
| if table.Format == 0: | |
| table.OpticalBoundsDeltas = { | |
| glyph: table.OpticalBoundsDeltas[glyph] | |
| for glyph in s.glyphs | |
| if glyph in table.OpticalBoundsDeltas | |
| } | |
| return len(table.OpticalBoundsDeltas) > 0 | |
| elif table.Format == 1: | |
| table.OpticalBoundsPoints = { | |
| glyph: table.OpticalBoundsPoints[glyph] | |
| for glyph in s.glyphs | |
| if glyph in table.OpticalBoundsPoints | |
| } | |
| return len(table.OpticalBoundsPoints) > 0 | |
| else: | |
| assert False, "unknown 'opbd' format %s" % table.Format | |
| def prune_pre_subset(self, font, options): | |
| if not options.glyph_names: | |
| self.formatType = 3.0 | |
| return True # Required table | |
| def subset_glyphs(self, s): | |
| self.extraNames = [] # This seems to do it | |
| return True # Required table | |
| def subset_glyphs(self, s): | |
| prop = self.table.GlyphProperties | |
| if prop.Format == 0: | |
| return prop.DefaultProperties != 0 | |
| elif prop.Format == 1: | |
| prop.Properties = { | |
| g: prop.Properties.get(g, prop.DefaultProperties) for g in s.glyphs | |
| } | |
| mostCommon, _cnt = Counter(prop.Properties.values()).most_common(1)[0] | |
| prop.DefaultProperties = mostCommon | |
| prop.Properties = { | |
| g: prop for g, prop in prop.Properties.items() if prop != mostCommon | |
| } | |
| if len(prop.Properties) == 0: | |
| del prop.Properties | |
| prop.Format = 0 | |
| return prop.DefaultProperties != 0 | |
| return True | |
| else: | |
| assert False, "unknown 'prop' format %s" % prop.Format | |
| def _paint_glyph_names(paint, colr): | |
| result = set() | |
| def callback(paint): | |
| if paint.Format in { | |
| otTables.PaintFormat.PaintGlyph, | |
| otTables.PaintFormat.PaintColrGlyph, | |
| }: | |
| result.add(paint.Glyph) | |
| paint.traverse(colr, callback) | |
| return result | |
| def closure_glyphs(self, s): | |
| if self.version > 0: | |
| # on decompiling COLRv1, we only keep around the raw otTables | |
| # but for subsetting we need dicts with fully decompiled layers; | |
| # we store them temporarily in the C_O_L_R_ instance and delete | |
| # them after we have finished subsetting. | |
| self.ColorLayers = self._decompileColorLayersV0(self.table) | |
| self.ColorLayersV1 = { | |
| rec.BaseGlyph: rec.Paint | |
| for rec in self.table.BaseGlyphList.BaseGlyphPaintRecord | |
| } | |
| decompose = s.glyphs | |
| while decompose: | |
| layers = set() | |
| for g in decompose: | |
| for layer in self.ColorLayers.get(g, []): | |
| layers.add(layer.name) | |
| if self.version > 0: | |
| paint = self.ColorLayersV1.get(g) | |
| if paint is not None: | |
| layers.update(_paint_glyph_names(paint, self.table)) | |
| layers -= s.glyphs | |
| s.glyphs.update(layers) | |
| decompose = layers | |
| def subset_glyphs(self, s): | |
| from fontTools.colorLib.unbuilder import unbuildColrV1 | |
| from fontTools.colorLib.builder import buildColrV1, populateCOLRv0 | |
| # only include glyphs after COLR closure, which in turn comes after cmap and GSUB | |
| # closure, but importantly before glyf/CFF closures. COLR layers can refer to | |
| # composite glyphs, and that's ok, since glyf/CFF closures happen after COLR closure | |
| # and take care of those. If we also included glyphs resulting from glyf/CFF closures | |
| # when deciding which COLR base glyphs to retain, then we may end up with a situation | |
| # whereby a COLR base glyph is kept, not because directly requested (cmap) | |
| # or substituted (GSUB) or referenced by another COLRv1 PaintColrGlyph, but because | |
| # it corresponds to (has same GID as) a non-COLR glyph that happens to be used as a | |
| # component in glyf or CFF table. Best case scenario we retain more glyphs than | |
| # required; worst case we retain incomplete COLR records that try to reference | |
| # glyphs that are no longer in the final subset font. | |
| # https://github.com/fonttools/fonttools/issues/2461 | |
| s.glyphs = s.glyphs_colred | |
| self.ColorLayers = { | |
| g: self.ColorLayers[g] for g in s.glyphs if g in self.ColorLayers | |
| } | |
| if self.version == 0: | |
| return bool(self.ColorLayers) | |
| colorGlyphsV1 = unbuildColrV1(self.table.LayerList, self.table.BaseGlyphList) | |
| self.table.LayerList, self.table.BaseGlyphList = buildColrV1( | |
| {g: colorGlyphsV1[g] for g in colorGlyphsV1 if g in s.glyphs} | |
| ) | |
| del self.ColorLayersV1 | |
| if self.table.ClipList is not None: | |
| clips = self.table.ClipList.clips | |
| self.table.ClipList.clips = {g: clips[g] for g in clips if g in s.glyphs} | |
| layersV0 = self.ColorLayers | |
| if not self.table.BaseGlyphList.BaseGlyphPaintRecord: | |
| # no more COLRv1 glyphs: downgrade to version 0 | |
| self.version = 0 | |
| del self.table | |
| return bool(layersV0) | |
| populateCOLRv0( | |
| self.table, | |
| {g: [(layer.name, layer.colorID) for layer in layersV0[g]] for g in layersV0}, | |
| ) | |
| del self.ColorLayers | |
| # TODO: also prune ununsed varIndices in COLR.VarStore | |
| return True | |
| def prune_post_subset(self, font, options): | |
| # Keep whole "CPAL" if "SVG " is present as it may be referenced by the latter | |
| # via 'var(--color{palette_entry_index}, ...)' CSS color variables. | |
| # For now we just assume this is the case by the mere presence of "SVG " table, | |
| # for parsing SVG to collect all the used indices is too much work... | |
| # TODO(anthrotype): Do The Right Thing (TM). | |
| if "SVG " in font: | |
| return True | |
| colr = font.get("COLR") | |
| if not colr: # drop CPAL if COLR was subsetted to empty | |
| return False | |
| colors_by_index = defaultdict(list) | |
| def collect_colors_by_index(paint): | |
| if hasattr(paint, "PaletteIndex"): # either solid colors... | |
| colors_by_index[paint.PaletteIndex].append(paint) | |
| elif hasattr(paint, "ColorLine"): # ... or gradient color stops | |
| for stop in paint.ColorLine.ColorStop: | |
| colors_by_index[stop.PaletteIndex].append(stop) | |
| if colr.version == 0: | |
| for layers in colr.ColorLayers.values(): | |
| for layer in layers: | |
| colors_by_index[layer.colorID].append(layer) | |
| else: | |
| if colr.table.LayerRecordArray: | |
| for layer in colr.table.LayerRecordArray.LayerRecord: | |
| colors_by_index[layer.PaletteIndex].append(layer) | |
| for record in colr.table.BaseGlyphList.BaseGlyphPaintRecord: | |
| record.Paint.traverse(colr.table, collect_colors_by_index) | |
| # don't remap palette entry index 0xFFFF, this is always the foreground color | |
| # https://github.com/fonttools/fonttools/issues/2257 | |
| retained_palette_indices = set(colors_by_index.keys()) - {0xFFFF} | |
| for palette in self.palettes: | |
| palette[:] = [c for i, c in enumerate(palette) if i in retained_palette_indices] | |
| assert len(palette) == len(retained_palette_indices) | |
| for new_index, old_index in enumerate(sorted(retained_palette_indices)): | |
| for record in colors_by_index[old_index]: | |
| if hasattr(record, "colorID"): # v0 | |
| record.colorID = new_index | |
| elif hasattr(record, "PaletteIndex"): # v1 | |
| record.PaletteIndex = new_index | |
| else: | |
| raise AssertionError(record) | |
| self.numPaletteEntries = len(self.palettes[0]) | |
| if self.version == 1: | |
| kept_labels = [] | |
| for i, label in enumerate(self.paletteEntryLabels): | |
| if i in retained_palette_indices: | |
| kept_labels.append(label) | |
| self.paletteEntryLabels = kept_labels | |
| return bool(self.numPaletteEntries) | |
| def closure_glyphs(self, glyphs): | |
| variants = set() | |
| for v in self.MathGlyphVariantRecord: | |
| variants.add(v.VariantGlyph) | |
| if self.GlyphAssembly: | |
| for p in self.GlyphAssembly.PartRecords: | |
| variants.add(p.glyph) | |
| return variants | |
| def closure_glyphs(self, s): | |
| glyphs = frozenset(s.glyphs) | |
| variants = set() | |
| if self.VertGlyphCoverage: | |
| indices = self.VertGlyphCoverage.intersect(glyphs) | |
| for i in indices: | |
| variants.update(self.VertGlyphConstruction[i].closure_glyphs(glyphs)) | |
| if self.HorizGlyphCoverage: | |
| indices = self.HorizGlyphCoverage.intersect(glyphs) | |
| for i in indices: | |
| variants.update(self.HorizGlyphConstruction[i].closure_glyphs(glyphs)) | |
| s.glyphs.update(variants) | |
| def subset_glyphs(self, s): | |
| indices = self.table.Coverage.subset(s.glyphs) | |
| self.table.VarCompositeGlyphs.VarCompositeGlyph = _list_subset( | |
| self.table.VarCompositeGlyphs.VarCompositeGlyph, indices | |
| ) | |
| return bool(self.table.VarCompositeGlyphs.VarCompositeGlyph) | |
| def closure_glyphs(self, s): | |
| if self.table.VarCompositeGlyphs is None: | |
| return | |
| glyphMap = {glyphName: i for i, glyphName in enumerate(self.table.Coverage.glyphs)} | |
| glyphRecords = self.table.VarCompositeGlyphs.VarCompositeGlyph | |
| glyphs = s.glyphs | |
| covered = set() | |
| new = set(glyphs) | |
| while new: | |
| oldNew = new | |
| new = set() | |
| for glyphName in oldNew: | |
| if glyphName in covered: | |
| continue | |
| idx = glyphMap.get(glyphName) | |
| if idx is None: | |
| continue | |
| glyph = glyphRecords[idx] | |
| for comp in glyph.components: | |
| name = comp.glyphName | |
| glyphs.add(name) | |
| if name not in covered: | |
| new.add(name) | |
| def prune_post_subset(self, font, options): | |
| table = self.table | |
| store = table.MultiVarStore | |
| if store is not None: | |
| usedVarIdxes = set() | |
| table.collect_varidxes(usedVarIdxes) | |
| varidx_map = store.subset_varidxes(usedVarIdxes) | |
| table.remap_varidxes(varidx_map) | |
| axisIndicesList = table.AxisIndicesList.Item | |
| if axisIndicesList is not None: | |
| usedIndices = set() | |
| for glyph in table.VarCompositeGlyphs.VarCompositeGlyph: | |
| for comp in glyph.components: | |
| if comp.axisIndicesIndex is not None: | |
| usedIndices.add(comp.axisIndicesIndex) | |
| usedIndices = sorted(usedIndices) | |
| table.AxisIndicesList.Item = _list_subset(axisIndicesList, usedIndices) | |
| mapping = {old: new for new, old in enumerate(usedIndices)} | |
| for glyph in table.VarCompositeGlyphs.VarCompositeGlyph: | |
| for comp in glyph.components: | |
| if comp.axisIndicesIndex is not None: | |
| comp.axisIndicesIndex = mapping[comp.axisIndicesIndex] | |
| conditionList = table.ConditionList | |
| if conditionList is not None: | |
| conditionTables = conditionList.ConditionTable | |
| usedIndices = set() | |
| for glyph in table.VarCompositeGlyphs.VarCompositeGlyph: | |
| for comp in glyph.components: | |
| if comp.conditionIndex is not None: | |
| usedIndices.add(comp.conditionIndex) | |
| usedIndices = sorted(usedIndices) | |
| conditionList.ConditionTable = _list_subset(conditionTables, usedIndices) | |
| mapping = {old: new for new, old in enumerate(usedIndices)} | |
| for glyph in table.VarCompositeGlyphs.VarCompositeGlyph: | |
| for comp in glyph.components: | |
| if comp.conditionIndex is not None: | |
| comp.conditionIndex = mapping[comp.conditionIndex] | |
| return True | |
| def closure_glyphs(self, s): | |
| if self.table.MathVariants: | |
| self.table.MathVariants.closure_glyphs(s) | |
| def subset_glyphs(self, s): | |
| indices = self.Coverage.subset(s.glyphs) | |
| self.ItalicsCorrection = _list_subset(self.ItalicsCorrection, indices) | |
| self.ItalicsCorrectionCount = len(self.ItalicsCorrection) | |
| return bool(self.ItalicsCorrectionCount) | |
| def subset_glyphs(self, s): | |
| indices = self.TopAccentCoverage.subset(s.glyphs) | |
| self.TopAccentAttachment = _list_subset(self.TopAccentAttachment, indices) | |
| self.TopAccentAttachmentCount = len(self.TopAccentAttachment) | |
| return bool(self.TopAccentAttachmentCount) | |
| def subset_glyphs(self, s): | |
| indices = self.MathKernCoverage.subset(s.glyphs) | |
| self.MathKernInfoRecords = _list_subset(self.MathKernInfoRecords, indices) | |
| self.MathKernCount = len(self.MathKernInfoRecords) | |
| return bool(self.MathKernCount) | |
| def subset_glyphs(self, s): | |
| if self.MathItalicsCorrectionInfo: | |
| self.MathItalicsCorrectionInfo.subset_glyphs(s) | |
| if self.MathTopAccentAttachment: | |
| self.MathTopAccentAttachment.subset_glyphs(s) | |
| if self.MathKernInfo: | |
| self.MathKernInfo.subset_glyphs(s) | |
| if self.ExtendedShapeCoverage: | |
| self.ExtendedShapeCoverage.subset(s.glyphs) | |
| return True | |
| def subset_glyphs(self, s): | |
| if self.VertGlyphCoverage: | |
| indices = self.VertGlyphCoverage.subset(s.glyphs) | |
| self.VertGlyphConstruction = _list_subset(self.VertGlyphConstruction, indices) | |
| self.VertGlyphCount = len(self.VertGlyphConstruction) | |
| if self.HorizGlyphCoverage: | |
| indices = self.HorizGlyphCoverage.subset(s.glyphs) | |
| self.HorizGlyphConstruction = _list_subset(self.HorizGlyphConstruction, indices) | |
| self.HorizGlyphCount = len(self.HorizGlyphConstruction) | |
| return True | |
| def subset_glyphs(self, s): | |
| s.glyphs = s.glyphs_mathed | |
| if self.table.MathGlyphInfo: | |
| self.table.MathGlyphInfo.subset_glyphs(s) | |
| if self.table.MathVariants: | |
| self.table.MathVariants.subset_glyphs(s) | |
| return True | |
| def remapComponentsFast(self, glyphidmap): | |
| if not self.data or struct.unpack(">h", self.data[:2])[0] >= 0: | |
| return # Not composite | |
| data = self.data = bytearray(self.data) | |
| i = 10 | |
| more = 1 | |
| while more: | |
| flags = (data[i] << 8) | data[i + 1] | |
| glyphID = (data[i + 2] << 8) | data[i + 3] | |
| # Remap | |
| glyphID = glyphidmap[glyphID] | |
| data[i + 2] = glyphID >> 8 | |
| data[i + 3] = glyphID & 0xFF | |
| i += 4 | |
| flags = int(flags) | |
| if flags & 0x0001: | |
| i += 4 # ARG_1_AND_2_ARE_WORDS | |
| else: | |
| i += 2 | |
| if flags & 0x0008: | |
| i += 2 # WE_HAVE_A_SCALE | |
| elif flags & 0x0040: | |
| i += 4 # WE_HAVE_AN_X_AND_Y_SCALE | |
| elif flags & 0x0080: | |
| i += 8 # WE_HAVE_A_TWO_BY_TWO | |
| more = flags & 0x0020 # MORE_COMPONENTS | |
| def closure_glyphs(self, s): | |
| glyphSet = self.glyphs | |
| decompose = s.glyphs | |
| while decompose: | |
| components = set() | |
| for g in decompose: | |
| if g not in glyphSet: | |
| continue | |
| gl = glyphSet[g] | |
| for c in gl.getComponentNames(self): | |
| components.add(c) | |
| components -= s.glyphs | |
| s.glyphs.update(components) | |
| decompose = components | |
| def prune_pre_subset(self, font, options): | |
| if options.notdef_glyph and not options.notdef_outline: | |
| g = self[self.glyphOrder[0]] | |
| # Yay, easy! | |
| g.__dict__.clear() | |
| g.data = b"" | |
| return True | |
| def subset_glyphs(self, s): | |
| self.glyphs = _dict_subset(self.glyphs, s.glyphs) | |
| if not s.options.retain_gids: | |
| indices = [i for i, g in enumerate(self.glyphOrder) if g in s.glyphs] | |
| glyphmap = {o: n for n, o in enumerate(indices)} | |
| for v in self.glyphs.values(): | |
| if hasattr(v, "data"): | |
| v.remapComponentsFast(glyphmap) | |
| Glyph = ttLib.getTableModule("glyf").Glyph | |
| for g in s.glyphs_emptied: | |
| self.glyphs[g] = Glyph() | |
| self.glyphs[g].data = b"" | |
| self.glyphOrder = [ | |
| g for g in self.glyphOrder if g in s.glyphs or g in s.glyphs_emptied | |
| ] | |
| # Don't drop empty 'glyf' tables, otherwise 'loca' doesn't get subset. | |
| return True | |
| def prune_post_subset(self, font, options): | |
| remove_hinting = not options.hinting | |
| for v in self.glyphs.values(): | |
| v.trim(remove_hinting=remove_hinting) | |
| return True | |
| def closure_glyphs(self, s): | |
| tables = [t for t in self.tables if t.isUnicode()] | |
| # Close glyphs | |
| for table in tables: | |
| if table.format == 14: | |
| for cmap in table.uvsDict.values(): | |
| glyphs = {g for u, g in cmap if u in s.unicodes_requested} | |
| if None in glyphs: | |
| glyphs.remove(None) | |
| s.glyphs.update(glyphs) | |
| else: | |
| cmap = table.cmap | |
| intersection = s.unicodes_requested.intersection(cmap.keys()) | |
| s.glyphs.update(cmap[u] for u in intersection) | |
| # Calculate unicodes_missing | |
| s.unicodes_missing = s.unicodes_requested.copy() | |
| for table in tables: | |
| s.unicodes_missing.difference_update(table.cmap) | |
| def prune_pre_subset(self, font, options): | |
| if not options.legacy_cmap: | |
| # Drop non-Unicode / non-Symbol cmaps | |
| self.tables = [t for t in self.tables if t.isUnicode() or t.isSymbol()] | |
| if not options.symbol_cmap: | |
| self.tables = [t for t in self.tables if not t.isSymbol()] | |
| # TODO(behdad) Only keep one subtable? | |
| # For now, drop format=0 which can't be subset_glyphs easily? | |
| self.tables = [t for t in self.tables if t.format != 0] | |
| self.numSubTables = len(self.tables) | |
| return True # Required table | |
| def subset_glyphs(self, s): | |
| s.glyphs = None # We use s.glyphs_requested and s.unicodes_requested only | |
| tables_format12_bmp = [] | |
| table_plat0_enc3 = {} # Unicode platform, Unicode BMP only, keyed by language | |
| table_plat3_enc1 = {} # Windows platform, Unicode BMP, keyed by language | |
| for t in self.tables: | |
| if t.platformID == 0 and t.platEncID == 3: | |
| table_plat0_enc3[t.language] = t | |
| if t.platformID == 3 and t.platEncID == 1: | |
| table_plat3_enc1[t.language] = t | |
| if t.format == 14: | |
| # TODO(behdad) We drop all the default-UVS mappings | |
| # for glyphs_requested. So it's the caller's responsibility to make | |
| # sure those are included. | |
| t.uvsDict = { | |
| v: [ | |
| (u, g) | |
| for u, g in l | |
| if g in s.glyphs_requested or u in s.unicodes_requested | |
| ] | |
| for v, l in t.uvsDict.items() | |
| } | |
| t.uvsDict = {v: l for v, l in t.uvsDict.items() if l} | |
| elif t.isUnicode(): | |
| t.cmap = { | |
| u: g | |
| for u, g in t.cmap.items() | |
| if g in s.glyphs_requested or u in s.unicodes_requested | |
| } | |
| # Collect format 12 tables that hold only basic multilingual plane | |
| # codepoints. | |
| if t.format == 12 and t.cmap and max(t.cmap.keys()) < 0x10000: | |
| tables_format12_bmp.append(t) | |
| else: | |
| t.cmap = {u: g for u, g in t.cmap.items() if g in s.glyphs_requested} | |
| # Fomat 12 tables are redundant if they contain just the same BMP codepoints | |
| # their little BMP-only encoding siblings contain. | |
| for t in tables_format12_bmp: | |
| if ( | |
| t.platformID == 0 # Unicode platform | |
| and t.platEncID == 4 # Unicode full repertoire | |
| and t.language in table_plat0_enc3 # Have a BMP-only sibling? | |
| and table_plat0_enc3[t.language].cmap == t.cmap | |
| ): | |
| t.cmap.clear() | |
| elif ( | |
| t.platformID == 3 # Windows platform | |
| and t.platEncID == 10 # Unicode full repertoire | |
| and t.language in table_plat3_enc1 # Have a BMP-only sibling? | |
| and table_plat3_enc1[t.language].cmap == t.cmap | |
| ): | |
| t.cmap.clear() | |
| self.tables = [t for t in self.tables if (t.cmap if t.format != 14 else t.uvsDict)] | |
| self.numSubTables = len(self.tables) | |
| # TODO(behdad) Convert formats when needed. | |
| # In particular, if we have a format=12 without non-BMP | |
| # characters, convert it to format=4 if there's not one. | |
| return True # Required table | |
| def prune_pre_subset(self, font, options): | |
| # Drop all signatures since they will be invalid | |
| self.usNumSigs = 0 | |
| self.signatureRecords = [] | |
| return True | |
| def prune_pre_subset(self, font, options): | |
| if not options.hinting: | |
| if self.tableVersion == 0x00010000: | |
| self.maxZones = 1 | |
| self.maxTwilightPoints = 0 | |
| self.maxStorage = 0 | |
| self.maxFunctionDefs = 0 | |
| self.maxInstructionDefs = 0 | |
| self.maxStackElements = 0 | |
| self.maxSizeOfInstructions = 0 | |
| return True | |
| def prune_post_subset(self, font, options): | |
| visitor = NameRecordVisitor() | |
| visitor.visit(font) | |
| nameIDs = set(options.name_IDs) | visitor.seen | |
| if "*" in options.name_IDs: | |
| nameIDs |= {n.nameID for n in self.names if n.nameID < 256} | |
| self.names = [n for n in self.names if n.nameID in nameIDs] | |
| if not options.name_legacy: | |
| # TODO(behdad) Sometimes (eg Apple Color Emoji) there's only a macroman | |
| # entry for Latin and no Unicode names. | |
| self.names = [n for n in self.names if n.isUnicode()] | |
| # TODO(behdad) Option to keep only one platform's | |
| if "*" not in options.name_languages: | |
| # TODO(behdad) This is Windows-platform specific! | |
| self.names = [n for n in self.names if n.langID in options.name_languages] | |
| if options.obfuscate_names: | |
| namerecs = [] | |
| for n in self.names: | |
| if n.nameID in [1, 4]: | |
| n.string = ".\x7f".encode("utf_16_be") if n.isUnicode() else ".\x7f" | |
| elif n.nameID in [2, 6]: | |
| n.string = "\x7f".encode("utf_16_be") if n.isUnicode() else "\x7f" | |
| elif n.nameID == 3: | |
| n.string = "" | |
| elif n.nameID in [16, 17, 18]: | |
| continue | |
| namerecs.append(n) | |
| self.names = namerecs | |
| return True # Required table | |
| def prune_post_subset(self, font, options): | |
| # Force re-compiling head table, to update any recalculated values. | |
| return True | |
| # TODO(behdad) OS/2 ulCodePageRange? | |
| # TODO(behdad) Drop AAT tables. | |
| # TODO(behdad) Drop unneeded GSUB/GPOS Script/LangSys entries. | |
| # TODO(behdad) Drop empty GSUB/GPOS, and GDEF if no GSUB/GPOS left | |
| # TODO(behdad) Drop GDEF subitems if unused by lookups | |
| # TODO(behdad) Avoid recursing too much (in GSUB/GPOS and in CFF) | |
| # TODO(behdad) Text direction considerations. | |
| # TODO(behdad) Text script / language considerations. | |
| # TODO(behdad) Optionally drop 'kern' table if GPOS available | |
| # TODO(behdad) Implement --unicode='*' to choose all cmap'ed | |
| # TODO(behdad) Drop old-spec Indic scripts | |
| class Options(object): | |
| class OptionError(Exception): | |
| pass | |
| class UnknownOptionError(OptionError): | |
| pass | |
| # spaces in tag names (e.g. "SVG ", "cvt ") are stripped by the argument parser | |
| _drop_tables_default = [ | |
| "BASE", | |
| "JSTF", | |
| "DSIG", | |
| "EBDT", | |
| "EBLC", | |
| "EBSC", | |
| "PCLT", | |
| "LTSH", | |
| ] | |
| _drop_tables_default += ["Feat", "Glat", "Gloc", "Silf", "Sill"] # Graphite | |
| _no_subset_tables_default = [ | |
| "avar", | |
| "fvar", | |
| "gasp", | |
| "head", | |
| "hhea", | |
| "maxp", | |
| "vhea", | |
| "OS/2", | |
| "loca", | |
| "name", | |
| "cvt", | |
| "fpgm", | |
| "prep", | |
| "VDMX", | |
| "DSIG", | |
| "CPAL", | |
| "MVAR", | |
| "cvar", | |
| "STAT", | |
| ] | |
| _hinting_tables_default = ["cvt", "cvar", "fpgm", "prep", "hdmx", "VDMX"] | |
| # Based on HarfBuzz shapers | |
| _layout_features_groups = { | |
| # Default shaper | |
| "common": ["rvrn", "ccmp", "liga", "locl", "mark", "mkmk", "rlig"], | |
| "fractions": ["frac", "numr", "dnom"], | |
| "horizontal": ["calt", "clig", "curs", "kern", "rclt"], | |
| "vertical": ["valt", "vert", "vkrn", "vpal", "vrt2"], | |
| "ltr": ["ltra", "ltrm"], | |
| "rtl": ["rtla", "rtlm"], | |
| "rand": ["rand"], | |
| "justify": ["jalt"], | |
| "private": ["Harf", "HARF", "Buzz", "BUZZ"], | |
| "east_asian_spacing": ["chws", "vchw", "halt", "vhal"], | |
| # Complex shapers | |
| "arabic": [ | |
| "init", | |
| "medi", | |
| "fina", | |
| "isol", | |
| "med2", | |
| "fin2", | |
| "fin3", | |
| "cswh", | |
| "mset", | |
| "stch", | |
| ], | |
| "hangul": ["ljmo", "vjmo", "tjmo"], | |
| "tibetan": ["abvs", "blws", "abvm", "blwm"], | |
| "indic": [ | |
| "nukt", | |
| "akhn", | |
| "rphf", | |
| "rkrf", | |
| "pref", | |
| "blwf", | |
| "half", | |
| "abvf", | |
| "pstf", | |
| "cfar", | |
| "vatu", | |
| "cjct", | |
| "init", | |
| "pres", | |
| "abvs", | |
| "blws", | |
| "psts", | |
| "haln", | |
| "dist", | |
| "abvm", | |
| "blwm", | |
| ], | |
| } | |
| _layout_features_default = _uniq_sort( | |
| sum(iter(_layout_features_groups.values()), []) | |
| ) | |
| def __init__(self, **kwargs): | |
| self.drop_tables = self._drop_tables_default[:] | |
| self.no_subset_tables = self._no_subset_tables_default[:] | |
| self.passthrough_tables = False # keep/drop tables we can't subset | |
| self.hinting_tables = self._hinting_tables_default[:] | |
| self.legacy_kern = False # drop 'kern' table if GPOS available | |
| self.layout_closure = True | |
| self.layout_features = self._layout_features_default[:] | |
| self.layout_scripts = ["*"] | |
| self.ignore_missing_glyphs = False | |
| self.ignore_missing_unicodes = True | |
| self.hinting = True | |
| self.glyph_names = False | |
| self.legacy_cmap = False | |
| self.symbol_cmap = False | |
| self.name_IDs = [ | |
| 0, | |
| 1, | |
| 2, | |
| 3, | |
| 4, | |
| 5, | |
| 6, | |
| ] # https://github.com/fonttools/fonttools/issues/1170#issuecomment-364631225 | |
| self.name_legacy = False | |
| self.name_languages = [0x0409] # English | |
| self.obfuscate_names = False # to make webfont unusable as a system font | |
| self.retain_gids = False | |
| self.notdef_glyph = True # gid0 for TrueType / .notdef for CFF | |
| self.notdef_outline = False # No need for notdef to have an outline really | |
| self.recommended_glyphs = False # gid1, gid2, gid3 for TrueType | |
| self.recalc_bounds = False # Recalculate font bounding boxes | |
| self.recalc_timestamp = False # Recalculate font modified timestamp | |
| self.prune_unicode_ranges = True # Clear unused 'ulUnicodeRange' bits | |
| self.prune_codepage_ranges = True # Clear unused 'ulCodePageRange' bits | |
| self.recalc_average_width = False # update 'xAvgCharWidth' | |
| self.recalc_max_context = False # update 'usMaxContext' | |
| self.canonical_order = None # Order tables as recommended | |
| self.flavor = None # May be 'woff' or 'woff2' | |
| self.with_zopfli = False # use zopfli instead of zlib for WOFF 1.0 | |
| self.desubroutinize = False # Desubroutinize CFF CharStrings | |
| self.harfbuzz_repacker = USE_HARFBUZZ_REPACKER.default | |
| self.verbose = False | |
| self.timing = False | |
| self.xml = False | |
| self.font_number = -1 | |
| self.pretty_svg = False | |
| self.lazy = True | |
| self.set(**kwargs) | |
| def set(self, **kwargs): | |
| for k, v in kwargs.items(): | |
| if not hasattr(self, k): | |
| raise self.UnknownOptionError("Unknown option '%s'" % k) | |
| setattr(self, k, v) | |
| def parse_opts(self, argv, ignore_unknown=[]): | |
| posargs = [] | |
| passthru_options = [] | |
| for a in argv: | |
| orig_a = a | |
| if not a.startswith("--"): | |
| posargs.append(a) | |
| continue | |
| a = a[2:] | |
| i = a.find("=") | |
| op = "=" | |
| if i == -1: | |
| if a.startswith("no-"): | |
| k = a[3:] | |
| if k == "canonical-order": | |
| # reorderTables=None is faster than False (the latter | |
| # still reorders to "keep" the original table order) | |
| v = None | |
| else: | |
| v = False | |
| else: | |
| k = a | |
| v = True | |
| if k.endswith("?"): | |
| k = k[:-1] | |
| v = "?" | |
| else: | |
| k = a[:i] | |
| if k[-1] in "-+": | |
| op = k[-1] + "=" # Op is '-=' or '+=' now. | |
| k = k[:-1] | |
| v = a[i + 1 :] | |
| ok = k | |
| k = k.replace("-", "_") | |
| if not hasattr(self, k): | |
| if ignore_unknown is True or ok in ignore_unknown: | |
| passthru_options.append(orig_a) | |
| continue | |
| else: | |
| raise self.UnknownOptionError("Unknown option '%s'" % a) | |
| ov = getattr(self, k) | |
| if v == "?": | |
| print("Current setting for '%s' is: %s" % (ok, ov)) | |
| continue | |
| if isinstance(ov, bool): | |
| v = bool(v) | |
| elif isinstance(ov, int): | |
| v = int(v) | |
| elif isinstance(ov, str): | |
| v = str(v) # redundant | |
| elif isinstance(ov, list): | |
| if isinstance(v, bool): | |
| raise self.OptionError( | |
| "Option '%s' requires values to be specified using '='" % a | |
| ) | |
| vv = v.replace(",", " ").split() | |
| if vv == [""]: | |
| vv = [] | |
| vv = [int(x, 0) if len(x) and x[0] in "0123456789" else x for x in vv] | |
| if op == "=": | |
| v = vv | |
| elif op == "+=": | |
| v = ov | |
| v.extend(vv) | |
| elif op == "-=": | |
| v = ov | |
| for x in vv: | |
| if x in v: | |
| v.remove(x) | |
| else: | |
| assert False | |
| setattr(self, k, v) | |
| return posargs + passthru_options | |
| class Subsetter(object): | |
| class SubsettingError(Exception): | |
| pass | |
| class MissingGlyphsSubsettingError(SubsettingError): | |
| pass | |
| class MissingUnicodesSubsettingError(SubsettingError): | |
| pass | |
| def __init__(self, options=None): | |
| if not options: | |
| options = Options() | |
| self.options = options | |
| self.unicodes_requested = set() | |
| self.glyph_names_requested = set() | |
| self.glyph_ids_requested = set() | |
| def populate(self, glyphs=[], gids=[], unicodes=[], text=""): | |
| self.unicodes_requested.update(unicodes) | |
| if isinstance(text, bytes): | |
| text = text.decode("utf_8") | |
| text_utf32 = text.encode("utf-32-be") | |
| nchars = len(text_utf32) // 4 | |
| for u in struct.unpack(">%dL" % nchars, text_utf32): | |
| self.unicodes_requested.add(u) | |
| self.glyph_names_requested.update(glyphs) | |
| self.glyph_ids_requested.update(gids) | |
| def _prune_pre_subset(self, font): | |
| for tag in self._sort_tables(font): | |
| if ( | |
| tag.strip() in self.options.drop_tables | |
| or ( | |
| tag.strip() in self.options.hinting_tables | |
| and not self.options.hinting | |
| ) | |
| or (tag == "kern" and (not self.options.legacy_kern and "GPOS" in font)) | |
| ): | |
| log.info("%s dropped", tag) | |
| del font[tag] | |
| continue | |
| clazz = ttLib.getTableClass(tag) | |
| if hasattr(clazz, "prune_pre_subset"): | |
| with timer("load '%s'" % tag): | |
| table = font[tag] | |
| with timer("prune '%s'" % tag): | |
| retain = table.prune_pre_subset(font, self.options) | |
| if not retain: | |
| log.info("%s pruned to empty; dropped", tag) | |
| del font[tag] | |
| continue | |
| else: | |
| log.info("%s pruned", tag) | |
| def _closure_glyphs(self, font): | |
| realGlyphs = set(font.getGlyphOrder()) | |
| self.orig_glyph_order = glyph_order = font.getGlyphOrder() | |
| self.glyphs_requested = set() | |
| self.glyphs_requested.update(self.glyph_names_requested) | |
| self.glyphs_requested.update( | |
| glyph_order[i] for i in self.glyph_ids_requested if i < len(glyph_order) | |
| ) | |
| self.glyphs_missing = set() | |
| self.glyphs_missing.update(self.glyphs_requested.difference(realGlyphs)) | |
| self.glyphs_missing.update( | |
| i for i in self.glyph_ids_requested if i >= len(glyph_order) | |
| ) | |
| if self.glyphs_missing: | |
| log.info("Missing requested glyphs: %s", self.glyphs_missing) | |
| if not self.options.ignore_missing_glyphs: | |
| raise self.MissingGlyphsSubsettingError(self.glyphs_missing) | |
| self.glyphs = self.glyphs_requested.copy() | |
| self.unicodes_missing = set() | |
| if "cmap" in font: | |
| with timer("close glyph list over 'cmap'"): | |
| font["cmap"].closure_glyphs(self) | |
| self.glyphs.intersection_update(realGlyphs) | |
| self.glyphs_cmaped = frozenset(self.glyphs) | |
| if self.unicodes_missing: | |
| missing = ["U+%04X" % u for u in self.unicodes_missing] | |
| log.info("Missing glyphs for requested Unicodes: %s", missing) | |
| if not self.options.ignore_missing_unicodes: | |
| raise self.MissingUnicodesSubsettingError(missing) | |
| del missing | |
| if self.options.notdef_glyph: | |
| if "glyf" in font: | |
| self.glyphs.add(font.getGlyphName(0)) | |
| log.info("Added gid0 to subset") | |
| else: | |
| self.glyphs.add(".notdef") | |
| log.info("Added .notdef to subset") | |
| if self.options.recommended_glyphs: | |
| if "glyf" in font: | |
| for i in range(min(4, len(font.getGlyphOrder()))): | |
| self.glyphs.add(font.getGlyphName(i)) | |
| log.info("Added first four glyphs to subset") | |
| if "MATH" in font: | |
| with timer("close glyph list over 'MATH'"): | |
| log.info( | |
| "Closing glyph list over 'MATH': %d glyphs before", len(self.glyphs) | |
| ) | |
| log.glyphs(self.glyphs, font=font) | |
| font["MATH"].closure_glyphs(self) | |
| self.glyphs.intersection_update(realGlyphs) | |
| log.info( | |
| "Closed glyph list over 'MATH': %d glyphs after", len(self.glyphs) | |
| ) | |
| log.glyphs(self.glyphs, font=font) | |
| self.glyphs_mathed = frozenset(self.glyphs) | |
| if self.options.layout_closure and "GSUB" in font: | |
| with timer("close glyph list over 'GSUB'"): | |
| log.info( | |
| "Closing glyph list over 'GSUB': %d glyphs before", len(self.glyphs) | |
| ) | |
| log.glyphs(self.glyphs, font=font) | |
| font["GSUB"].closure_glyphs(self) | |
| self.glyphs.intersection_update(realGlyphs) | |
| log.info( | |
| "Closed glyph list over 'GSUB': %d glyphs after", len(self.glyphs) | |
| ) | |
| log.glyphs(self.glyphs, font=font) | |
| self.glyphs_gsubed = frozenset(self.glyphs) | |
| for table in ("COLR", "bsln"): | |
| if table in font: | |
| with timer("close glyph list over '%s'" % table): | |
| log.info( | |
| "Closing glyph list over '%s': %d glyphs before", | |
| table, | |
| len(self.glyphs), | |
| ) | |
| log.glyphs(self.glyphs, font=font) | |
| font[table].closure_glyphs(self) | |
| self.glyphs.intersection_update(realGlyphs) | |
| log.info( | |
| "Closed glyph list over '%s': %d glyphs after", | |
| table, | |
| len(self.glyphs), | |
| ) | |
| log.glyphs(self.glyphs, font=font) | |
| setattr(self, f"glyphs_{table.lower()}ed", frozenset(self.glyphs)) | |
| if "VARC" in font: | |
| with timer("close glyph list over 'VARC'"): | |
| log.info( | |
| "Closing glyph list over 'VARC': %d glyphs before", len(self.glyphs) | |
| ) | |
| log.glyphs(self.glyphs, font=font) | |
| font["VARC"].closure_glyphs(self) | |
| self.glyphs.intersection_update(realGlyphs) | |
| log.info( | |
| "Closed glyph list over 'VARC': %d glyphs after", len(self.glyphs) | |
| ) | |
| log.glyphs(self.glyphs, font=font) | |
| self.glyphs_glyfed = frozenset(self.glyphs) | |
| if "glyf" in font: | |
| with timer("close glyph list over 'glyf'"): | |
| log.info( | |
| "Closing glyph list over 'glyf': %d glyphs before", len(self.glyphs) | |
| ) | |
| log.glyphs(self.glyphs, font=font) | |
| font["glyf"].closure_glyphs(self) | |
| self.glyphs.intersection_update(realGlyphs) | |
| log.info( | |
| "Closed glyph list over 'glyf': %d glyphs after", len(self.glyphs) | |
| ) | |
| log.glyphs(self.glyphs, font=font) | |
| self.glyphs_glyfed = frozenset(self.glyphs) | |
| if "CFF " in font: | |
| with timer("close glyph list over 'CFF '"): | |
| log.info( | |
| "Closing glyph list over 'CFF ': %d glyphs before", len(self.glyphs) | |
| ) | |
| log.glyphs(self.glyphs, font=font) | |
| font["CFF "].closure_glyphs(self) | |
| self.glyphs.intersection_update(realGlyphs) | |
| log.info( | |
| "Closed glyph list over 'CFF ': %d glyphs after", len(self.glyphs) | |
| ) | |
| log.glyphs(self.glyphs, font=font) | |
| self.glyphs_cffed = frozenset(self.glyphs) | |
| self.glyphs_retained = frozenset(self.glyphs) | |
| order = font.getReverseGlyphMap() | |
| self.reverseOrigGlyphMap = {g: order[g] for g in self.glyphs_retained} | |
| self.last_retained_order = max(self.reverseOrigGlyphMap.values()) | |
| self.last_retained_glyph = font.getGlyphOrder()[self.last_retained_order] | |
| self.glyphs_emptied = frozenset() | |
| if self.options.retain_gids: | |
| self.glyphs_emptied = { | |
| g | |
| for g in realGlyphs - self.glyphs_retained | |
| if order[g] <= self.last_retained_order | |
| } | |
| self.reverseEmptiedGlyphMap = {g: order[g] for g in self.glyphs_emptied} | |
| if not self.options.retain_gids: | |
| new_glyph_order = [g for g in glyph_order if g in self.glyphs_retained] | |
| else: | |
| new_glyph_order = [ | |
| g for g in glyph_order if font.getGlyphID(g) <= self.last_retained_order | |
| ] | |
| # We'll call font.setGlyphOrder() at the end of _subset_glyphs when all | |
| # tables have been subsetted. Below, we use the new glyph order to get | |
| # a map from old to new glyph indices, which can be useful when | |
| # subsetting individual tables (e.g. SVG) that refer to GIDs. | |
| self.new_glyph_order = new_glyph_order | |
| self.glyph_index_map = { | |
| order[new_glyph_order[i]]: i for i in range(len(new_glyph_order)) | |
| } | |
| log.info("Retaining %d glyphs", len(self.glyphs_retained)) | |
| del self.glyphs | |
| def _subset_glyphs(self, font): | |
| self.used_mark_sets = [] | |
| for tag in self._sort_tables(font): | |
| clazz = ttLib.getTableClass(tag) | |
| if tag.strip() in self.options.no_subset_tables: | |
| log.info("%s subsetting not needed", tag) | |
| elif hasattr(clazz, "subset_glyphs"): | |
| with timer("subset '%s'" % tag): | |
| table = font[tag] | |
| self.glyphs = self.glyphs_retained | |
| retain = table.subset_glyphs(self) | |
| del self.glyphs | |
| if not retain: | |
| log.info("%s subsetted to empty; dropped", tag) | |
| del font[tag] | |
| else: | |
| log.info("%s subsetted", tag) | |
| elif self.options.passthrough_tables: | |
| log.info("%s NOT subset; don't know how to subset", tag) | |
| else: | |
| log.warning("%s NOT subset; don't know how to subset; dropped", tag) | |
| del font[tag] | |
| with timer("subset GlyphOrder"): | |
| font.setGlyphOrder(self.new_glyph_order) | |
| def _prune_post_subset(self, font): | |
| tableTags = font.keys() | |
| # Prune the name table last because when we're pruning the name table, | |
| # we visit each table in the font to see what name table records are | |
| # still in use. | |
| if "name" in tableTags: | |
| tableTags.remove("name") | |
| tableTags.append("name") | |
| for tag in tableTags: | |
| if tag == "GlyphOrder": | |
| continue | |
| if tag == "OS/2": | |
| if self.options.prune_unicode_ranges: | |
| old_uniranges = font[tag].getUnicodeRanges() | |
| new_uniranges = font[tag].recalcUnicodeRanges(font, pruneOnly=True) | |
| if old_uniranges != new_uniranges: | |
| log.info( | |
| "%s Unicode ranges pruned: %s", tag, sorted(new_uniranges) | |
| ) | |
| if self.options.prune_codepage_ranges and font[tag].version >= 1: | |
| # codepage range fields were added with OS/2 format 1 | |
| # https://learn.microsoft.com/en-us/typography/opentype/spec/os2#version-1 | |
| old_codepages = font[tag].getCodePageRanges() | |
| new_codepages = font[tag].recalcCodePageRanges(font, pruneOnly=True) | |
| if old_codepages != new_codepages: | |
| log.info( | |
| "%s CodePage ranges pruned: %s", | |
| tag, | |
| sorted(new_codepages), | |
| ) | |
| if self.options.recalc_average_width: | |
| old_avg_width = font[tag].xAvgCharWidth | |
| new_avg_width = font[tag].recalcAvgCharWidth(font) | |
| if old_avg_width != new_avg_width: | |
| log.info("%s xAvgCharWidth updated: %d", tag, new_avg_width) | |
| if self.options.recalc_max_context: | |
| max_context = maxCtxFont(font) | |
| if max_context != font[tag].usMaxContext: | |
| font[tag].usMaxContext = max_context | |
| log.info("%s usMaxContext updated: %d", tag, max_context) | |
| clazz = ttLib.getTableClass(tag) | |
| if hasattr(clazz, "prune_post_subset"): | |
| with timer("prune '%s'" % tag): | |
| table = font[tag] | |
| retain = table.prune_post_subset(font, self.options) | |
| if not retain: | |
| log.info("%s pruned to empty; dropped", tag) | |
| del font[tag] | |
| else: | |
| log.info("%s pruned", tag) | |
| def _sort_tables(self, font): | |
| tagOrder = ["GDEF", "GPOS", "GSUB", "fvar", "avar", "gvar", "name", "glyf"] | |
| tagOrder = {t: i + 1 for i, t in enumerate(tagOrder)} | |
| tags = sorted(font.keys(), key=lambda tag: tagOrder.get(tag, 0)) | |
| return [t for t in tags if t != "GlyphOrder"] | |
| def subset(self, font): | |
| self._prune_pre_subset(font) | |
| self._closure_glyphs(font) | |
| self._subset_glyphs(font) | |
| self._prune_post_subset(font) | |
| def load_font(fontFile, options, checkChecksums=0, dontLoadGlyphNames=False, lazy=True): | |
| font = ttLib.TTFont( | |
| fontFile, | |
| checkChecksums=checkChecksums, | |
| recalcBBoxes=options.recalc_bounds, | |
| recalcTimestamp=options.recalc_timestamp, | |
| lazy=lazy, | |
| fontNumber=options.font_number, | |
| ) | |
| # Hack: | |
| # | |
| # If we don't need glyph names, change 'post' class to not try to | |
| # load them. It avoid lots of headache with broken fonts as well | |
| # as loading time. | |
| # | |
| # Ideally ttLib should provide a way to ask it to skip loading | |
| # glyph names. But it currently doesn't provide such a thing. | |
| # | |
| if dontLoadGlyphNames: | |
| post = ttLib.getTableClass("post") | |
| saved = post.decode_format_2_0 | |
| post.decode_format_2_0 = post.decode_format_3_0 | |
| f = font["post"] | |
| if f.formatType == 2.0: | |
| f.formatType = 3.0 | |
| post.decode_format_2_0 = saved | |
| return font | |
| def save_font(font, outfile, options): | |
| if options.with_zopfli and options.flavor == "woff": | |
| from fontTools.ttLib import sfnt | |
| sfnt.USE_ZOPFLI = True | |
| font.flavor = options.flavor | |
| font.cfg[USE_HARFBUZZ_REPACKER] = options.harfbuzz_repacker | |
| font.save(outfile, reorderTables=options.canonical_order) | |
| def parse_unicodes(s): | |
| import re | |
| s = re.sub(r"0[xX]", " ", s) | |
| s = re.sub(r"[<+>,;&#\\xXuU\n ]", " ", s) | |
| l = [] | |
| for item in s.split(): | |
| fields = item.split("-") | |
| if len(fields) == 1: | |
| l.append(int(item, 16)) | |
| else: | |
| start, end = fields | |
| l.extend(range(int(start, 16), int(end, 16) + 1)) | |
| return l | |
| def parse_gids(s): | |
| l = [] | |
| for item in s.replace(",", " ").split(): | |
| fields = item.split("-") | |
| if len(fields) == 1: | |
| l.append(int(fields[0])) | |
| else: | |
| l.extend(range(int(fields[0]), int(fields[1]) + 1)) | |
| return l | |
| def parse_glyphs(s): | |
| return s.replace(",", " ").split() | |
| def usage(): | |
| print("usage:", __usage__, file=sys.stderr) | |
| print("Try pyftsubset --help for more information.\n", file=sys.stderr) | |
| def main(args=None): | |
| """OpenType font subsetter and optimizer""" | |
| from os.path import splitext | |
| from fontTools import configLogger | |
| if args is None: | |
| args = sys.argv[1:] | |
| if "--help" in args: | |
| print(__doc__) | |
| return 0 | |
| options = Options() | |
| try: | |
| args = options.parse_opts( | |
| args, | |
| ignore_unknown=[ | |
| "gids", | |
| "gids-file", | |
| "glyphs", | |
| "glyphs-file", | |
| "text", | |
| "text-file", | |
| "unicodes", | |
| "unicodes-file", | |
| "output-file", | |
| ], | |
| ) | |
| except options.OptionError as e: | |
| usage() | |
| print("ERROR:", e, file=sys.stderr) | |
| return 2 | |
| if len(args) < 2: | |
| usage() | |
| return 1 | |
| configLogger(level=logging.INFO if options.verbose else logging.WARNING) | |
| if options.timing: | |
| timer.logger.setLevel(logging.DEBUG) | |
| else: | |
| timer.logger.disabled = True | |
| fontfile = args[0] | |
| args = args[1:] | |
| subsetter = Subsetter(options=options) | |
| outfile = None | |
| glyphs = [] | |
| gids = [] | |
| unicodes = [] | |
| wildcard_glyphs = False | |
| wildcard_unicodes = False | |
| text = "" | |
| for g in args: | |
| if g == "*": | |
| wildcard_glyphs = True | |
| continue | |
| if g.startswith("--output-file="): | |
| outfile = g[14:] | |
| continue | |
| if g.startswith("--text="): | |
| text += g[7:] | |
| continue | |
| if g.startswith("--text-file="): | |
| with open(g[12:], encoding="utf-8") as f: | |
| text += f.read().replace("\n", "") | |
| continue | |
| if g.startswith("--unicodes="): | |
| if g[11:] == "*": | |
| wildcard_unicodes = True | |
| else: | |
| unicodes.extend(parse_unicodes(g[11:])) | |
| continue | |
| if g.startswith("--unicodes-file="): | |
| with open(g[16:]) as f: | |
| for line in f.readlines(): | |
| unicodes.extend(parse_unicodes(line.split("#")[0])) | |
| continue | |
| if g.startswith("--gids="): | |
| gids.extend(parse_gids(g[7:])) | |
| continue | |
| if g.startswith("--gids-file="): | |
| with open(g[12:]) as f: | |
| for line in f.readlines(): | |
| gids.extend(parse_gids(line.split("#")[0])) | |
| continue | |
| if g.startswith("--glyphs="): | |
| if g[9:] == "*": | |
| wildcard_glyphs = True | |
| else: | |
| glyphs.extend(parse_glyphs(g[9:])) | |
| continue | |
| if g.startswith("--glyphs-file="): | |
| with open(g[14:]) as f: | |
| for line in f.readlines(): | |
| glyphs.extend(parse_glyphs(line.split("#")[0])) | |
| continue | |
| glyphs.append(g) | |
| dontLoadGlyphNames = not options.glyph_names and not glyphs | |
| lazy = options.lazy | |
| font = load_font( | |
| fontfile, options, dontLoadGlyphNames=dontLoadGlyphNames, lazy=lazy | |
| ) | |
| if outfile is None: | |
| ext = "." + options.flavor.lower() if options.flavor is not None else None | |
| outfile = makeOutputFileName( | |
| fontfile, extension=ext, overWrite=True, suffix=".subset" | |
| ) | |
| with timer("compile glyph list"): | |
| if wildcard_glyphs: | |
| glyphs.extend(font.getGlyphOrder()) | |
| if wildcard_unicodes: | |
| for t in font["cmap"].tables: | |
| if t.isUnicode(): | |
| unicodes.extend(t.cmap.keys()) | |
| assert "" not in glyphs | |
| log.info("Text: '%s'" % text) | |
| log.info("Unicodes: %s", unicodes) | |
| log.info("Glyphs: %s", glyphs) | |
| log.info("Gids: %s", gids) | |
| subsetter.populate(glyphs=glyphs, gids=gids, unicodes=unicodes, text=text) | |
| subsetter.subset(font) | |
| save_font(font, outfile, options) | |
| if options.verbose: | |
| import os | |
| log.info("Input font:% 7d bytes: %s" % (os.path.getsize(fontfile), fontfile)) | |
| log.info("Subset font:% 7d bytes: %s" % (os.path.getsize(outfile), outfile)) | |
| if options.xml: | |
| font.saveXML(sys.stdout) | |
| font.close() | |
| __all__ = [ | |
| "Options", | |
| "Subsetter", | |
| "load_font", | |
| "save_font", | |
| "parse_gids", | |
| "parse_glyphs", | |
| "parse_unicodes", | |
| "main", | |
| ] | |