2012-09-01 07:48:21 +09:00
|
|
|
|
/* Copyright 2012 Mozilla Foundation
|
|
|
|
|
*
|
|
|
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
|
* you may not use this file except in compliance with the License.
|
|
|
|
|
* You may obtain a copy of the License at
|
|
|
|
|
*
|
|
|
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
|
*
|
|
|
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
|
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
|
* See the License for the specific language governing permissions and
|
|
|
|
|
* limitations under the License.
|
|
|
|
|
*/
|
2020-10-03 19:26:25 +09:00
|
|
|
|
/* eslint-disable no-var */
|
2011-10-26 10:18:22 +09:00
|
|
|
|
|
2017-04-02 23:14:30 +09:00
|
|
|
|
import {
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
AbortException,
|
|
|
|
|
assert,
|
|
|
|
|
CMapCompressionType,
|
|
|
|
|
createPromiseCapability,
|
|
|
|
|
FONT_IDENTITY_MATRIX,
|
|
|
|
|
FormatError,
|
|
|
|
|
IDENTITY_MATRIX,
|
|
|
|
|
info,
|
|
|
|
|
isArrayEqual,
|
|
|
|
|
isNum,
|
|
|
|
|
isString,
|
|
|
|
|
OPS,
|
Add local caching of `Function`s, by reference, in the `PDFFunctionFactory` (issue 2541)
Note that compared other structures, such as e.g. Images and ColorSpaces, `Function`s are not referred to by name, which however does bring the advantage of being able to share the cache for an *entire* page.
Furthermore, similar to ColorSpaces, the parsing of individual `Function`s are generally fast enough to not really warrant trying to cache them in any "smarter" way than by reference. (Hence trying to do caching similar to e.g. Fonts would most likely be a losing proposition, given the amount of data lookup/parsing that'd be required.)
Originally I tried implementing this similar to e.g. the recently added ColorSpace caching (and in a couple of different ways), however it unfortunately turned out to be quite ugly/unwieldy given the sheer number of functions/methods where you'd thus need to pass in a `LocalFunctionCache` instance. (Also, the affected functions/methods didn't exactly have short signatures as-is.)
After going back and forth on this for a while it seemed to me that the simplest, or least "invasive" if you will, solution would be if each `PartialEvaluator` instance had its *own* `PDFFunctionFactory` instance (since the latter is already passed to all of the required code). This way each `PDFFunctionFactory` instances could have a local `Function` cache, without it being necessary to provide a `LocalFunctionCache` instance manually at every `PDFFunctionFactory.{create, createFromArray}` call-site.
Obviously, with this patch, there's now (potentially) more `PDFFunctionFactory` instances than before when the entire document shared just one. However, each such instance is really quite small and it's also tied to a `PartialEvaluator` instance and those are *not* kept alive and/or cached. To reduce the impact of these changes, I've tried to make as many of these structures as possible *lazily initialized*, specifically:
- The `PDFFunctionFactory`, on `PartialEvaluator` instances, since not all kinds of general parsing actually requires it. For example: `getTextContent` calls won't cause any `Function` to be parsed, and even some `getOperatorList` calls won't trigger `Function` parsing (if a page contains e.g. no Patterns or "complex" ColorSpaces).
- The `LocalFunctionCache`, on `PDFFunctionFactory` instances, since only certain parsing requires it. Generally speaking, only e.g. Patterns, "complex" ColorSpaces, and/or (some) SoftMasks will trigger any `Function` parsing.
To put these changes into perspective, when loading/rendering all (14) pages of the default `tracemonkey.pdf` file there's now a total of 6 `PDFFunctionFactory` and 1 `LocalFunctionCache` instances created thanks to the lazy initialization.
(If you instead would keep the document-"global" `PDFFunctionFactory` instance and pass around `LocalFunctionCache` instances everywhere, the numbers for the `tracemonkey.pdf` file would be instead be something like 1 `PDFFunctionFactory` and 6 `LocalFunctionCache` instances.)
All-in-all, I thus don't think that the `PDFFunctionFactory` changes should be generally problematic.
With these changes, we can also modify (some) call-sites to pass in a `Reference` rather than the actual `Function` data. This is nice since `Function`s can also be `Streams`, which are not cached on the `XRef` instance (given their potential size), and this way we can avoid unnecessary lookups and thus save some additional time/resources.
Obviously I had intended to include (standard) benchmark results with these changes, but for reasons I don't really understand the test run-time (even with `master`) of the document in issue 2541 is quite a bit slower than in the development viewer.
However, logging the time it takes for the relevant `PDFFunctionFactory`/`PDFFunction ` parsing shows that it takes *approximately* `0.5 ms` for the `Function` in question. Looking up a cached `Function`, on the other hand, is *one order of magnitude faster* which does add up when the same `Function` is invoked close to 2000 times.
2020-06-28 20:12:24 +09:00
|
|
|
|
shadow,
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
stringToPDFString,
|
|
|
|
|
TextRenderingMode,
|
|
|
|
|
UNSUPPORTED_FEATURES,
|
|
|
|
|
Util,
|
|
|
|
|
warn,
|
2020-01-02 20:00:16 +09:00
|
|
|
|
} from "../shared/util.js";
|
|
|
|
|
import { CMapFactory, IdentityCMap } from "./cmap.js";
|
2017-04-02 23:14:30 +09:00
|
|
|
|
import {
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
Cmd,
|
|
|
|
|
Dict,
|
|
|
|
|
EOF,
|
|
|
|
|
isDict,
|
|
|
|
|
isName,
|
|
|
|
|
isRef,
|
|
|
|
|
isStream,
|
|
|
|
|
Name,
|
|
|
|
|
Ref,
|
2020-07-15 19:05:05 +09:00
|
|
|
|
RefSet,
|
2020-01-02 20:00:16 +09:00
|
|
|
|
} from "./primitives.js";
|
2017-04-02 23:14:30 +09:00
|
|
|
|
import {
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
ErrorFont,
|
|
|
|
|
Font,
|
|
|
|
|
FontFlags,
|
|
|
|
|
getFontType,
|
|
|
|
|
IdentityToUnicodeMap,
|
|
|
|
|
ToUnicodeMap,
|
2020-01-02 20:00:16 +09:00
|
|
|
|
} from "./fonts.js";
|
2017-04-02 23:14:30 +09:00
|
|
|
|
import {
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
getEncoding,
|
|
|
|
|
MacRomanEncoding,
|
|
|
|
|
StandardEncoding,
|
|
|
|
|
SymbolSetEncoding,
|
|
|
|
|
WinAnsiEncoding,
|
|
|
|
|
ZapfDingbatsEncoding,
|
2020-01-02 20:00:16 +09:00
|
|
|
|
} from "./encodings.js";
|
2020-03-09 19:37:33 +09:00
|
|
|
|
import { getLookupTableFactory, MissingDataException } from "./core_utils.js";
|
2017-04-02 23:14:30 +09:00
|
|
|
|
import {
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
getNormalizedUnicodes,
|
|
|
|
|
getUnicodeForGlyph,
|
|
|
|
|
reverseIfRtl,
|
2020-01-02 20:00:16 +09:00
|
|
|
|
} from "./unicode.js";
|
2017-04-02 23:14:30 +09:00
|
|
|
|
import {
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
getSerifFonts,
|
|
|
|
|
getStdFontMap,
|
|
|
|
|
getSymbolsFonts,
|
2020-01-02 20:00:16 +09:00
|
|
|
|
} from "./standard_fonts.js";
|
|
|
|
|
import { getTilingPatternIR, Pattern } from "./pattern.js";
|
Add local caching of `Function`s, by reference, in the `PDFFunctionFactory` (issue 2541)
Note that compared other structures, such as e.g. Images and ColorSpaces, `Function`s are not referred to by name, which however does bring the advantage of being able to share the cache for an *entire* page.
Furthermore, similar to ColorSpaces, the parsing of individual `Function`s are generally fast enough to not really warrant trying to cache them in any "smarter" way than by reference. (Hence trying to do caching similar to e.g. Fonts would most likely be a losing proposition, given the amount of data lookup/parsing that'd be required.)
Originally I tried implementing this similar to e.g. the recently added ColorSpace caching (and in a couple of different ways), however it unfortunately turned out to be quite ugly/unwieldy given the sheer number of functions/methods where you'd thus need to pass in a `LocalFunctionCache` instance. (Also, the affected functions/methods didn't exactly have short signatures as-is.)
After going back and forth on this for a while it seemed to me that the simplest, or least "invasive" if you will, solution would be if each `PartialEvaluator` instance had its *own* `PDFFunctionFactory` instance (since the latter is already passed to all of the required code). This way each `PDFFunctionFactory` instances could have a local `Function` cache, without it being necessary to provide a `LocalFunctionCache` instance manually at every `PDFFunctionFactory.{create, createFromArray}` call-site.
Obviously, with this patch, there's now (potentially) more `PDFFunctionFactory` instances than before when the entire document shared just one. However, each such instance is really quite small and it's also tied to a `PartialEvaluator` instance and those are *not* kept alive and/or cached. To reduce the impact of these changes, I've tried to make as many of these structures as possible *lazily initialized*, specifically:
- The `PDFFunctionFactory`, on `PartialEvaluator` instances, since not all kinds of general parsing actually requires it. For example: `getTextContent` calls won't cause any `Function` to be parsed, and even some `getOperatorList` calls won't trigger `Function` parsing (if a page contains e.g. no Patterns or "complex" ColorSpaces).
- The `LocalFunctionCache`, on `PDFFunctionFactory` instances, since only certain parsing requires it. Generally speaking, only e.g. Patterns, "complex" ColorSpaces, and/or (some) SoftMasks will trigger any `Function` parsing.
To put these changes into perspective, when loading/rendering all (14) pages of the default `tracemonkey.pdf` file there's now a total of 6 `PDFFunctionFactory` and 1 `LocalFunctionCache` instances created thanks to the lazy initialization.
(If you instead would keep the document-"global" `PDFFunctionFactory` instance and pass around `LocalFunctionCache` instances everywhere, the numbers for the `tracemonkey.pdf` file would be instead be something like 1 `PDFFunctionFactory` and 6 `LocalFunctionCache` instances.)
All-in-all, I thus don't think that the `PDFFunctionFactory` changes should be generally problematic.
With these changes, we can also modify (some) call-sites to pass in a `Reference` rather than the actual `Function` data. This is nice since `Function`s can also be `Streams`, which are not cached on the `XRef` instance (given their potential size), and this way we can avoid unnecessary lookups and thus save some additional time/resources.
Obviously I had intended to include (standard) benchmark results with these changes, but for reasons I don't really understand the test run-time (even with `master`) of the document in issue 2541 is quite a bit slower than in the development viewer.
However, logging the time it takes for the relevant `PDFFunctionFactory`/`PDFFunction ` parsing shows that it takes *approximately* `0.5 ms` for the `Function` in question. Looking up a cached `Function`, on the other hand, is *one order of magnitude faster* which does add up when the same `Function` is invoked close to 2000 times.
2020-06-28 20:12:24 +09:00
|
|
|
|
import { isPDFFunction, PDFFunctionFactory } from "./function.js";
|
2020-01-02 20:00:16 +09:00
|
|
|
|
import { Lexer, Parser } from "./parser.js";
|
2020-07-11 20:52:11 +09:00
|
|
|
|
import {
|
|
|
|
|
LocalColorSpaceCache,
|
|
|
|
|
LocalGStateCache,
|
|
|
|
|
LocalImageCache,
|
2020-10-09 00:33:23 +09:00
|
|
|
|
LocalTilingPatternCache,
|
2020-07-11 20:52:11 +09:00
|
|
|
|
} from "./image_utils.js";
|
2020-01-02 20:00:16 +09:00
|
|
|
|
import { bidi } from "./bidi.js";
|
|
|
|
|
import { ColorSpace } from "./colorspace.js";
|
|
|
|
|
import { DecodeStream } from "./stream.js";
|
|
|
|
|
import { getGlyphsUnicode } from "./glyphlist.js";
|
|
|
|
|
import { getMetrics } from "./metrics.js";
|
|
|
|
|
import { MurmurHash3_64 } from "./murmurhash3.js";
|
|
|
|
|
import { OperatorList } from "./operator_list.js";
|
|
|
|
|
import { PDFImage } from "./image.js";
|
2015-11-22 01:32:47 +09:00
|
|
|
|
|
2020-07-05 19:06:56 +09:00
|
|
|
|
const DefaultPartialEvaluatorOptions = Object.freeze({
|
|
|
|
|
maxImageSize: -1,
|
|
|
|
|
disableFontFace: false,
|
|
|
|
|
ignoreErrors: false,
|
|
|
|
|
isEvalSupported: true,
|
|
|
|
|
fontExtraProperties: false,
|
|
|
|
|
});
|
|
|
|
|
|
|
|
|
|
const PatternType = {
|
|
|
|
|
TILING: 1,
|
|
|
|
|
SHADING: 2,
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
const deferred = Promise.resolve();
|
|
|
|
|
|
|
|
|
|
// Convert PDF blend mode names to HTML5 blend mode names.
|
|
|
|
|
function normalizeBlendMode(value, parsingArray = false) {
|
|
|
|
|
if (Array.isArray(value)) {
|
|
|
|
|
// Use the first *supported* BM value in the Array (fixes issue11279.pdf).
|
|
|
|
|
for (let i = 0, ii = value.length; i < ii; i++) {
|
|
|
|
|
const maybeBM = normalizeBlendMode(value[i], /* parsingArray = */ true);
|
|
|
|
|
if (maybeBM) {
|
|
|
|
|
return maybeBM;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
warn(`Unsupported blend mode Array: ${value}`);
|
|
|
|
|
return "source-over";
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (!isName(value)) {
|
|
|
|
|
if (parsingArray) {
|
|
|
|
|
return null;
|
|
|
|
|
}
|
|
|
|
|
return "source-over";
|
|
|
|
|
}
|
|
|
|
|
switch (value.name) {
|
|
|
|
|
case "Normal":
|
|
|
|
|
case "Compatible":
|
|
|
|
|
return "source-over";
|
|
|
|
|
case "Multiply":
|
|
|
|
|
return "multiply";
|
|
|
|
|
case "Screen":
|
|
|
|
|
return "screen";
|
|
|
|
|
case "Overlay":
|
|
|
|
|
return "overlay";
|
|
|
|
|
case "Darken":
|
|
|
|
|
return "darken";
|
|
|
|
|
case "Lighten":
|
|
|
|
|
return "lighten";
|
|
|
|
|
case "ColorDodge":
|
|
|
|
|
return "color-dodge";
|
|
|
|
|
case "ColorBurn":
|
|
|
|
|
return "color-burn";
|
|
|
|
|
case "HardLight":
|
|
|
|
|
return "hard-light";
|
|
|
|
|
case "SoftLight":
|
|
|
|
|
return "soft-light";
|
|
|
|
|
case "Difference":
|
|
|
|
|
return "difference";
|
|
|
|
|
case "Exclusion":
|
|
|
|
|
return "exclusion";
|
|
|
|
|
case "Hue":
|
|
|
|
|
return "hue";
|
|
|
|
|
case "Saturation":
|
|
|
|
|
return "saturation";
|
|
|
|
|
case "Color":
|
|
|
|
|
return "color";
|
|
|
|
|
case "Luminosity":
|
|
|
|
|
return "luminosity";
|
|
|
|
|
}
|
|
|
|
|
if (parsingArray) {
|
|
|
|
|
return null;
|
|
|
|
|
}
|
|
|
|
|
warn(`Unsupported blend mode: ${value.name}`);
|
|
|
|
|
return "source-over";
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Trying to minimize Date.now() usage and check every 100 time.
|
2020-07-05 19:20:10 +09:00
|
|
|
|
class TimeSlotManager {
|
|
|
|
|
static get TIME_SLOT_DURATION_MS() {
|
|
|
|
|
return shadow(this, "TIME_SLOT_DURATION_MS", 20);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static get CHECK_TIME_EVERY() {
|
|
|
|
|
return shadow(this, "CHECK_TIME_EVERY", 100);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
constructor() {
|
|
|
|
|
this.reset();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
check() {
|
|
|
|
|
if (++this.checked < TimeSlotManager.CHECK_TIME_EVERY) {
|
2020-07-05 19:06:56 +09:00
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
this.checked = 0;
|
|
|
|
|
return this.endTime <= Date.now();
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
reset() {
|
|
|
|
|
this.endTime = Date.now() + TimeSlotManager.TIME_SLOT_DURATION_MS;
|
2020-07-05 19:06:56 +09:00
|
|
|
|
this.checked = 0;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
}
|
2016-03-03 09:48:21 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
class PartialEvaluator {
|
|
|
|
|
constructor({
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
xref,
|
|
|
|
|
handler,
|
|
|
|
|
pageIndex,
|
|
|
|
|
idFactory,
|
|
|
|
|
fontCache,
|
|
|
|
|
builtInCMapCache,
|
Attempt to cache repeated images at the document, rather than the page, level (issue 11878)
Currently image resources, as opposed to e.g. font resources, are handled exclusively on a page-specific basis. Generally speaking this makes sense, since pages are separate from each other, however there's PDF documents where many (or even all) pages actually references exactly the same image resources (through the XRef table). Hence, in some cases, we're decoding the *same* images over and over for every page which is obviously slow and wasting both CPU and memory resources better used elsewhere.[1]
Obviously we cannot simply treat all image resources as-if they're used throughout the entire PDF document, since that would end up increasing memory usage too much.[2]
However, by introducing a `GlobalImageCache` in the worker we can track image resources that appear on more than one page. Hence we can switch image resources from being page-specific to being document-specific, once the image resource has been seen on more than a certain number of pages.
In many cases, such as e.g. the referenced issue, this patch will thus lead to reduced memory usage for image resources. Scrolling through all pages of the document, there's now only a few main-thread copies of the same image data, as opposed to one for each rendered page (i.e. there could theoretically be *twenty* copies of the image data).
While this obviously benefit both CPU and memory usage in this case, for *very* large image data this patch *may* possibly increase persistent main-thread memory usage a tiny bit. Thus to avoid negatively affecting memory usage too much in general, particularly on the main-thread, the `GlobalImageCache` will *only* cache a certain number of image resources at the document level and simply fallback to the default behaviour.
Unfortunately the asynchronous nature of the code, with ranged/streamed loading of data, actually makes all of this much more complicated than if all data could be assumed to be immediately available.[3]
*Please note:* The patch will lead to *small* movement in some existing test-cases, since we're now using the built-in PDF.js JPEG decoder more. This was done in order to simplify the overall implementation, especially on the main-thread, by limiting it to only the `OPS.paintImageXObject` operator.
---
[1] There's e.g. PDF documents that use the same image as background on all pages.
[2] Given that data stored in the `commonObjs`, on the main-thread, are only cleared manually through `PDFDocumentProxy.cleanup`. This as opposed to data stored in the `objs` of each page, which is automatically removed when the page is cleaned-up e.g. by being evicted from the cache in the default viewer.
[3] If the latter case were true, we could simply check for repeat images *before* parsing started and thus avoid handling *any* duplicate image resources.
2020-05-18 21:17:56 +09:00
|
|
|
|
globalImageCache,
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
options = null,
|
|
|
|
|
}) {
|
2011-10-25 08:55:23 +09:00
|
|
|
|
this.xref = xref;
|
|
|
|
|
this.handler = handler;
|
2012-10-29 05:10:34 +09:00
|
|
|
|
this.pageIndex = pageIndex;
|
2017-01-09 00:51:30 +09:00
|
|
|
|
this.idFactory = idFactory;
|
2013-11-15 06:43:38 +09:00
|
|
|
|
this.fontCache = fontCache;
|
2017-02-14 22:28:31 +09:00
|
|
|
|
this.builtInCMapCache = builtInCMapCache;
|
Attempt to cache repeated images at the document, rather than the page, level (issue 11878)
Currently image resources, as opposed to e.g. font resources, are handled exclusively on a page-specific basis. Generally speaking this makes sense, since pages are separate from each other, however there's PDF documents where many (or even all) pages actually references exactly the same image resources (through the XRef table). Hence, in some cases, we're decoding the *same* images over and over for every page which is obviously slow and wasting both CPU and memory resources better used elsewhere.[1]
Obviously we cannot simply treat all image resources as-if they're used throughout the entire PDF document, since that would end up increasing memory usage too much.[2]
However, by introducing a `GlobalImageCache` in the worker we can track image resources that appear on more than one page. Hence we can switch image resources from being page-specific to being document-specific, once the image resource has been seen on more than a certain number of pages.
In many cases, such as e.g. the referenced issue, this patch will thus lead to reduced memory usage for image resources. Scrolling through all pages of the document, there's now only a few main-thread copies of the same image data, as opposed to one for each rendered page (i.e. there could theoretically be *twenty* copies of the image data).
While this obviously benefit both CPU and memory usage in this case, for *very* large image data this patch *may* possibly increase persistent main-thread memory usage a tiny bit. Thus to avoid negatively affecting memory usage too much in general, particularly on the main-thread, the `GlobalImageCache` will *only* cache a certain number of image resources at the document level and simply fallback to the default behaviour.
Unfortunately the asynchronous nature of the code, with ranged/streamed loading of data, actually makes all of this much more complicated than if all data could be assumed to be immediately available.[3]
*Please note:* The patch will lead to *small* movement in some existing test-cases, since we're now using the built-in PDF.js JPEG decoder more. This was done in order to simplify the overall implementation, especially on the main-thread, by limiting it to only the `OPS.paintImageXObject` operator.
---
[1] There's e.g. PDF documents that use the same image as background on all pages.
[2] Given that data stored in the `commonObjs`, on the main-thread, are only cleared manually through `PDFDocumentProxy.cleanup`. This as opposed to data stored in the `objs` of each page, which is automatically removed when the page is cleaned-up e.g. by being evicted from the cache in the default viewer.
[3] If the latter case were true, we could simply check for repeat images *before* parsing started and thus avoid handling *any* duplicate image resources.
2020-05-18 21:17:56 +09:00
|
|
|
|
this.globalImageCache = globalImageCache;
|
2016-03-03 09:48:21 +09:00
|
|
|
|
this.options = options || DefaultPartialEvaluatorOptions;
|
2019-04-11 19:26:15 +09:00
|
|
|
|
this.parsingType3Font = false;
|
2017-02-12 23:54:41 +09:00
|
|
|
|
|
2020-06-25 00:23:41 +09:00
|
|
|
|
this._fetchBuiltInCMapBound = this.fetchBuiltInCMap.bind(this);
|
2011-10-25 08:55:23 +09:00
|
|
|
|
}
|
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
/**
|
|
|
|
|
* Since Functions are only cached (locally) by reference, we can share one
|
|
|
|
|
* `PDFFunctionFactory` instance within this `PartialEvaluator` instance.
|
|
|
|
|
*/
|
|
|
|
|
get _pdfFunctionFactory() {
|
|
|
|
|
const pdfFunctionFactory = new PDFFunctionFactory({
|
|
|
|
|
xref: this.xref,
|
|
|
|
|
isEvalSupported: this.options.isEvalSupported,
|
|
|
|
|
});
|
|
|
|
|
return shadow(this, "_pdfFunctionFactory", pdfFunctionFactory);
|
|
|
|
|
}
|
2019-10-28 19:28:13 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
clone(newOptions = DefaultPartialEvaluatorOptions) {
|
|
|
|
|
var newEvaluator = Object.create(this);
|
|
|
|
|
newEvaluator.options = newOptions;
|
|
|
|
|
return newEvaluator;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
hasBlendModes(resources) {
|
|
|
|
|
if (!(resources instanceof Dict)) {
|
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
|
2020-07-15 19:05:05 +09:00
|
|
|
|
const processed = new RefSet();
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (resources.objId) {
|
2020-07-15 19:05:05 +09:00
|
|
|
|
processed.put(resources.objId);
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
var nodes = [resources],
|
|
|
|
|
xref = this.xref;
|
|
|
|
|
while (nodes.length) {
|
|
|
|
|
var node = nodes.shift();
|
|
|
|
|
// First check the current resources for blend modes.
|
|
|
|
|
var graphicStates = node.get("ExtGState");
|
|
|
|
|
if (graphicStates instanceof Dict) {
|
2020-07-17 19:57:34 +09:00
|
|
|
|
for (let graphicState of graphicStates.getRawValues()) {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (graphicState instanceof Ref) {
|
2020-07-15 19:05:05 +09:00
|
|
|
|
if (processed.has(graphicState)) {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
continue; // The ExtGState has already been processed.
|
2016-02-10 01:09:17 +09:00
|
|
|
|
}
|
2020-03-09 19:37:33 +09:00
|
|
|
|
try {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
graphicState = xref.fetch(graphicState);
|
2020-03-09 19:37:33 +09:00
|
|
|
|
} catch (ex) {
|
|
|
|
|
if (ex instanceof MissingDataException) {
|
|
|
|
|
throw ex;
|
|
|
|
|
}
|
2020-07-14 20:00:35 +09:00
|
|
|
|
// Avoid parsing a corrupt ExtGState more than once.
|
2020-07-15 19:05:05 +09:00
|
|
|
|
processed.put(graphicState);
|
2020-07-14 20:00:35 +09:00
|
|
|
|
|
|
|
|
|
info(`hasBlendModes - ignoring ExtGState: "${ex}".`);
|
|
|
|
|
continue;
|
2020-03-09 19:37:33 +09:00
|
|
|
|
}
|
2016-02-10 01:09:17 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (!(graphicState instanceof Dict)) {
|
2013-08-01 03:17:36 +09:00
|
|
|
|
continue;
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (graphicState.objId) {
|
2020-07-15 19:05:05 +09:00
|
|
|
|
processed.put(graphicState.objId);
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
const bm = graphicState.get("BM");
|
|
|
|
|
if (bm instanceof Name) {
|
|
|
|
|
if (bm.name !== "Normal") {
|
|
|
|
|
return true;
|
2014-06-10 18:29:25 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
continue;
|
2014-06-10 18:29:25 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (bm !== undefined && Array.isArray(bm)) {
|
2020-07-15 18:51:45 +09:00
|
|
|
|
for (const element of bm) {
|
|
|
|
|
if (element instanceof Name && element.name !== "Normal") {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return true;
|
|
|
|
|
}
|
2014-03-26 23:07:38 +09:00
|
|
|
|
}
|
2013-08-01 03:17:36 +09:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Descend into the XObjects to look for more resources and blend modes.
|
|
|
|
|
var xObjects = node.get("XObject");
|
|
|
|
|
if (!(xObjects instanceof Dict)) {
|
|
|
|
|
continue;
|
2020-06-25 00:23:41 +09:00
|
|
|
|
}
|
2020-07-17 19:57:34 +09:00
|
|
|
|
for (let xObject of xObjects.getRawValues()) {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (xObject instanceof Ref) {
|
2020-07-15 19:05:05 +09:00
|
|
|
|
if (processed.has(xObject)) {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// The XObject has already been processed, and by avoiding a
|
|
|
|
|
// redundant `xref.fetch` we can *significantly* reduce the load
|
|
|
|
|
// time for badly generated PDF files (fixes issue6961.pdf).
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
try {
|
|
|
|
|
xObject = xref.fetch(xObject);
|
|
|
|
|
} catch (ex) {
|
|
|
|
|
if (ex instanceof MissingDataException) {
|
|
|
|
|
throw ex;
|
|
|
|
|
}
|
2020-07-14 20:00:35 +09:00
|
|
|
|
// Avoid parsing a corrupt XObject more than once.
|
2020-07-15 19:05:05 +09:00
|
|
|
|
processed.put(xObject);
|
2020-07-14 20:00:35 +09:00
|
|
|
|
|
|
|
|
|
info(`hasBlendModes - ignoring XObject: "${ex}".`);
|
|
|
|
|
continue;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
if (!isStream(xObject)) {
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
if (xObject.dict.objId) {
|
2020-07-15 19:05:05 +09:00
|
|
|
|
processed.put(xObject.dict.objId);
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
var xResources = xObject.dict.get("Resources");
|
2020-07-15 18:51:45 +09:00
|
|
|
|
if (!(xResources instanceof Dict)) {
|
|
|
|
|
continue;
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Checking objId to detect an infinite loop.
|
2020-07-15 19:05:05 +09:00
|
|
|
|
if (xResources.objId && processed.has(xResources.objId)) {
|
2020-07-15 18:51:45 +09:00
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
nodes.push(xResources);
|
|
|
|
|
if (xResources.objId) {
|
2020-07-15 19:05:05 +09:00
|
|
|
|
processed.put(xResources.objId);
|
2020-06-25 00:23:41 +09:00
|
|
|
|
}
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
return false;
|
|
|
|
|
}
|
2020-06-25 00:23:41 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
async fetchBuiltInCMap(name) {
|
|
|
|
|
const cachedData = this.builtInCMapCache.get(name);
|
|
|
|
|
if (cachedData) {
|
|
|
|
|
return cachedData;
|
|
|
|
|
}
|
|
|
|
|
const readableStream = this.handler.sendWithStream("FetchBuiltInCMap", {
|
|
|
|
|
name,
|
|
|
|
|
});
|
|
|
|
|
const reader = readableStream.getReader();
|
|
|
|
|
|
|
|
|
|
const data = await new Promise(function (resolve, reject) {
|
|
|
|
|
function pump() {
|
|
|
|
|
reader.read().then(function ({ value, done }) {
|
|
|
|
|
if (done) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
resolve(value);
|
|
|
|
|
pump();
|
|
|
|
|
}, reject);
|
2018-10-16 22:23:14 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
pump();
|
|
|
|
|
});
|
2013-04-09 07:14:56 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (data.compressionType !== CMapCompressionType.NONE) {
|
|
|
|
|
// Given the size of uncompressed CMaps, only cache compressed ones.
|
|
|
|
|
this.builtInCMapCache.set(name, data);
|
|
|
|
|
}
|
|
|
|
|
return data;
|
|
|
|
|
}
|
Add local caching of `ColorSpace`s, by name, in `PartialEvaluator.getOperatorList` (issue 2504)
By caching parsed `ColorSpace`s, we thus don't need to re-parse the same data over and over which saves CPU cycles *and* reduces peak memory usage. (Obviously persistent memory usage *may* increase a tiny bit, but since the caching is done per `PartialEvaluator.getOperatorList` invocation and given that `ColorSpace` instances generally hold very little data this shouldn't be much of an issue.)
Furthermore, by caching `ColorSpace`s we can also lookup the already parsed ones *synchronously* during the `OperatorList` building, instead of having to defer to the event loop/microtask queue since the parsing is done asynchronously (such that error handling is easier).
Possible future improvements:
- Cache/lookup parsed `ColorSpaces` used in `Pattern`s and `Image`s.
- Attempt to cache *local* `ColorSpace`s by reference as well, in addition to only by name, assuming that there's documents where that would be beneficial and that it's not too difficult to implement.
- Assuming there's documents that would benefit from it, also cache repeated `ColorSpace`s *globally* as well.
Given that we've never, until now, been doing *any* caching of parsed `ColorSpace`s and that even using a simple name-only *local* cache helps tremendously in pathological cases, I purposely decided against complicating the implementation too much initially.
Also, compared to parsing of `Image`s, simply creating a `ColorSpace` instance isn't that expensive (hence I'd be somewhat surprised if adding a *global* cache would help much).
---
This patch was tested using:
- The default `tracemonkey` PDF file, which was included mostly to show that "normal" documents aren't negatively affected by these changes.
- The PDF file from issue 2504, i.e. https://dl-ctlg.panasonic.com/jp/manual/sd/sd_rbm1000_0.pdf, where most pages will switch *thousands* of times between a handful of `ColorSpace`s.
with the following manifest file:
```
[
{ "id": "tracemonkey",
"file": "pdfs/tracemonkey.pdf",
"md5": "9a192d8b1a7dc652a19835f6f08098bd",
"rounds": 100,
"type": "eq"
},
{ "id": "issue2504",
"file": "../web/pdfs/issue2504.pdf",
"md5": "",
"rounds": 20,
"type": "eq"
}
]
```
which gave the following results when comparing this patch against the `master` branch:
- Overall
```
-- Grouped By browser, pdf, stat --
browser | pdf | stat | Count | Baseline(ms) | Current(ms) | +/- | % | Result(P<.05)
------- | ----------- | ------------ | ----- | ------------ | ----------- | ---- | ------ | -------------
firefox | issue2504 | Overall | 640 | 977 | 497 | -479 | -49.08 | faster
firefox | issue2504 | Page Request | 640 | 3 | 4 | 1 | 59.18 |
firefox | issue2504 | Rendering | 640 | 974 | 493 | -481 | -49.37 | faster
firefox | tracemonkey | Overall | 1400 | 116 | 111 | -5 | -4.43 |
firefox | tracemonkey | Page Request | 1400 | 2 | 2 | 0 | -2.86 |
firefox | tracemonkey | Rendering | 1400 | 114 | 109 | -5 | -4.47 |
```
- Page-specific
```
-- Grouped By browser, pdf, page, stat --
browser | pdf | page | stat | Count | Baseline(ms) | Current(ms) | +/- | % | Result(P<.05)
------- | ----------- | ---- | ------------ | ----- | ------------ | ----------- | ----- | ------- | -------------
firefox | issue2504 | 0 | Overall | 20 | 2295 | 1268 | -1027 | -44.76 | faster
firefox | issue2504 | 0 | Page Request | 20 | 6 | 7 | 1 | 15.32 |
firefox | issue2504 | 0 | Rendering | 20 | 2288 | 1260 | -1028 | -44.93 | faster
firefox | issue2504 | 1 | Overall | 20 | 3059 | 2806 | -252 | -8.25 | faster
firefox | issue2504 | 1 | Page Request | 20 | 11 | 14 | 3 | 23.25 | slower
firefox | issue2504 | 1 | Rendering | 20 | 3047 | 2792 | -255 | -8.37 | faster
firefox | issue2504 | 2 | Overall | 20 | 411 | 295 | -116 | -28.20 | faster
firefox | issue2504 | 2 | Page Request | 20 | 2 | 42 | 40 | 1897.62 |
firefox | issue2504 | 2 | Rendering | 20 | 409 | 253 | -156 | -38.09 | faster
firefox | issue2504 | 3 | Overall | 20 | 736 | 299 | -437 | -59.34 | faster
firefox | issue2504 | 3 | Page Request | 20 | 2 | 2 | 0 | 0.00 |
firefox | issue2504 | 3 | Rendering | 20 | 734 | 297 | -437 | -59.49 | faster
firefox | issue2504 | 4 | Overall | 20 | 356 | 458 | 102 | 28.63 |
firefox | issue2504 | 4 | Page Request | 20 | 1 | 2 | 1 | 57.14 | slower
firefox | issue2504 | 4 | Rendering | 20 | 354 | 455 | 101 | 28.53 |
firefox | issue2504 | 5 | Overall | 20 | 1381 | 765 | -616 | -44.59 | faster
firefox | issue2504 | 5 | Page Request | 20 | 3 | 5 | 2 | 50.00 | slower
firefox | issue2504 | 5 | Rendering | 20 | 1378 | 760 | -617 | -44.81 | faster
firefox | issue2504 | 6 | Overall | 20 | 757 | 299 | -459 | -60.57 | faster
firefox | issue2504 | 6 | Page Request | 20 | 2 | 5 | 3 | 150.00 | slower
firefox | issue2504 | 6 | Rendering | 20 | 755 | 294 | -462 | -61.11 | faster
firefox | issue2504 | 7 | Overall | 20 | 394 | 302 | -92 | -23.39 | faster
firefox | issue2504 | 7 | Page Request | 20 | 2 | 1 | -1 | -34.88 | faster
firefox | issue2504 | 7 | Rendering | 20 | 392 | 301 | -91 | -23.32 | faster
firefox | issue2504 | 8 | Overall | 20 | 2875 | 979 | -1896 | -65.95 | faster
firefox | issue2504 | 8 | Page Request | 20 | 1 | 2 | 0 | 11.11 |
firefox | issue2504 | 8 | Rendering | 20 | 2874 | 978 | -1896 | -65.99 | faster
firefox | issue2504 | 9 | Overall | 20 | 700 | 332 | -368 | -52.60 | faster
firefox | issue2504 | 9 | Page Request | 20 | 3 | 2 | 0 | -4.00 |
firefox | issue2504 | 9 | Rendering | 20 | 698 | 329 | -368 | -52.78 | faster
firefox | issue2504 | 10 | Overall | 20 | 3296 | 926 | -2370 | -71.91 | faster
firefox | issue2504 | 10 | Page Request | 20 | 2 | 2 | 0 | -18.75 |
firefox | issue2504 | 10 | Rendering | 20 | 3293 | 924 | -2370 | -71.96 | faster
firefox | issue2504 | 11 | Overall | 20 | 524 | 197 | -327 | -62.34 | faster
firefox | issue2504 | 11 | Page Request | 20 | 2 | 3 | 1 | 58.54 |
firefox | issue2504 | 11 | Rendering | 20 | 522 | 194 | -328 | -62.81 | faster
firefox | issue2504 | 12 | Overall | 20 | 752 | 369 | -384 | -50.98 | faster
firefox | issue2504 | 12 | Page Request | 20 | 3 | 2 | -1 | -36.51 | faster
firefox | issue2504 | 12 | Rendering | 20 | 749 | 367 | -382 | -51.05 | faster
firefox | issue2504 | 13 | Overall | 20 | 679 | 487 | -193 | -28.38 | faster
firefox | issue2504 | 13 | Page Request | 20 | 4 | 2 | -2 | -48.68 | faster
firefox | issue2504 | 13 | Rendering | 20 | 676 | 485 | -191 | -28.28 | faster
firefox | issue2504 | 14 | Overall | 20 | 474 | 283 | -191 | -40.26 | faster
firefox | issue2504 | 14 | Page Request | 20 | 2 | 4 | 2 | 78.57 |
firefox | issue2504 | 14 | Rendering | 20 | 471 | 279 | -192 | -40.79 | faster
firefox | issue2504 | 15 | Overall | 20 | 860 | 618 | -241 | -28.05 | faster
firefox | issue2504 | 15 | Page Request | 20 | 2 | 3 | 0 | 10.87 |
firefox | issue2504 | 15 | Rendering | 20 | 857 | 616 | -241 | -28.15 | faster
firefox | issue2504 | 16 | Overall | 20 | 389 | 243 | -147 | -37.71 | faster
firefox | issue2504 | 16 | Page Request | 20 | 2 | 2 | 0 | 2.33 |
firefox | issue2504 | 16 | Rendering | 20 | 387 | 240 | -147 | -37.94 | faster
firefox | issue2504 | 17 | Overall | 20 | 1484 | 672 | -812 | -54.70 | faster
firefox | issue2504 | 17 | Page Request | 20 | 2 | 3 | 1 | 37.21 |
firefox | issue2504 | 17 | Rendering | 20 | 1482 | 669 | -812 | -54.84 | faster
firefox | issue2504 | 18 | Overall | 20 | 575 | 252 | -323 | -56.12 | faster
firefox | issue2504 | 18 | Page Request | 20 | 2 | 2 | 0 | -16.22 |
firefox | issue2504 | 18 | Rendering | 20 | 573 | 251 | -322 | -56.24 | faster
firefox | issue2504 | 19 | Overall | 20 | 517 | 227 | -290 | -56.08 | faster
firefox | issue2504 | 19 | Page Request | 20 | 2 | 2 | 0 | 21.62 |
firefox | issue2504 | 19 | Rendering | 20 | 515 | 225 | -290 | -56.37 | faster
firefox | issue2504 | 20 | Overall | 20 | 668 | 670 | 2 | 0.31 |
firefox | issue2504 | 20 | Page Request | 20 | 4 | 2 | -1 | -34.29 |
firefox | issue2504 | 20 | Rendering | 20 | 664 | 667 | 3 | 0.49 |
firefox | issue2504 | 21 | Overall | 20 | 486 | 309 | -177 | -36.44 | faster
firefox | issue2504 | 21 | Page Request | 20 | 2 | 2 | 0 | 16.13 |
firefox | issue2504 | 21 | Rendering | 20 | 484 | 307 | -177 | -36.60 | faster
firefox | issue2504 | 22 | Overall | 20 | 543 | 267 | -276 | -50.85 | faster
firefox | issue2504 | 22 | Page Request | 20 | 2 | 2 | 0 | 10.26 |
firefox | issue2504 | 22 | Rendering | 20 | 541 | 265 | -276 | -51.07 | faster
firefox | issue2504 | 23 | Overall | 20 | 3246 | 871 | -2375 | -73.17 | faster
firefox | issue2504 | 23 | Page Request | 20 | 2 | 3 | 1 | 37.21 |
firefox | issue2504 | 23 | Rendering | 20 | 3243 | 868 | -2376 | -73.25 | faster
firefox | issue2504 | 24 | Overall | 20 | 379 | 156 | -223 | -58.83 | faster
firefox | issue2504 | 24 | Page Request | 20 | 2 | 2 | 0 | -2.86 |
firefox | issue2504 | 24 | Rendering | 20 | 378 | 154 | -223 | -59.10 | faster
firefox | issue2504 | 25 | Overall | 20 | 176 | 127 | -50 | -28.19 | faster
firefox | issue2504 | 25 | Page Request | 20 | 2 | 1 | 0 | -15.63 |
firefox | issue2504 | 25 | Rendering | 20 | 175 | 125 | -49 | -28.31 | faster
firefox | issue2504 | 26 | Overall | 20 | 181 | 108 | -74 | -40.67 | faster
firefox | issue2504 | 26 | Page Request | 20 | 3 | 2 | -1 | -39.13 | faster
firefox | issue2504 | 26 | Rendering | 20 | 178 | 105 | -72 | -40.69 | faster
firefox | issue2504 | 27 | Overall | 20 | 208 | 104 | -104 | -49.92 | faster
firefox | issue2504 | 27 | Page Request | 20 | 2 | 2 | 1 | 48.39 |
firefox | issue2504 | 27 | Rendering | 20 | 206 | 102 | -104 | -50.64 | faster
firefox | issue2504 | 28 | Overall | 20 | 241 | 111 | -131 | -54.16 | faster
firefox | issue2504 | 28 | Page Request | 20 | 2 | 2 | -1 | -33.33 |
firefox | issue2504 | 28 | Rendering | 20 | 239 | 109 | -130 | -54.39 | faster
firefox | issue2504 | 29 | Overall | 20 | 321 | 196 | -125 | -39.05 | faster
firefox | issue2504 | 29 | Page Request | 20 | 1 | 2 | 0 | 17.86 |
firefox | issue2504 | 29 | Rendering | 20 | 319 | 194 | -126 | -39.35 | faster
firefox | issue2504 | 30 | Overall | 20 | 651 | 271 | -380 | -58.41 | faster
firefox | issue2504 | 30 | Page Request | 20 | 1 | 2 | 1 | 50.00 |
firefox | issue2504 | 30 | Rendering | 20 | 649 | 269 | -381 | -58.60 | faster
firefox | issue2504 | 31 | Overall | 20 | 1635 | 647 | -988 | -60.42 | faster
firefox | issue2504 | 31 | Page Request | 20 | 1 | 2 | 0 | 30.43 |
firefox | issue2504 | 31 | Rendering | 20 | 1634 | 645 | -988 | -60.49 | faster
firefox | tracemonkey | 0 | Overall | 100 | 51 | 51 | 0 | 0.02 |
firefox | tracemonkey | 0 | Page Request | 100 | 1 | 1 | 0 | -4.76 |
firefox | tracemonkey | 0 | Rendering | 100 | 50 | 50 | 0 | 0.12 |
firefox | tracemonkey | 1 | Overall | 100 | 97 | 91 | -5 | -5.52 | faster
firefox | tracemonkey | 1 | Page Request | 100 | 3 | 3 | 0 | -1.32 |
firefox | tracemonkey | 1 | Rendering | 100 | 94 | 88 | -5 | -5.73 | faster
firefox | tracemonkey | 2 | Overall | 100 | 40 | 40 | 0 | 0.50 |
firefox | tracemonkey | 2 | Page Request | 100 | 1 | 1 | 0 | 3.16 |
firefox | tracemonkey | 2 | Rendering | 100 | 39 | 39 | 0 | 0.54 |
firefox | tracemonkey | 3 | Overall | 100 | 62 | 62 | -1 | -0.94 |
firefox | tracemonkey | 3 | Page Request | 100 | 1 | 1 | 0 | 17.05 |
firefox | tracemonkey | 3 | Rendering | 100 | 61 | 61 | -1 | -1.11 |
firefox | tracemonkey | 4 | Overall | 100 | 56 | 58 | 2 | 3.41 |
firefox | tracemonkey | 4 | Page Request | 100 | 1 | 1 | 0 | 15.31 |
firefox | tracemonkey | 4 | Rendering | 100 | 55 | 57 | 2 | 3.23 |
firefox | tracemonkey | 5 | Overall | 100 | 73 | 71 | -2 | -2.28 |
firefox | tracemonkey | 5 | Page Request | 100 | 2 | 2 | 0 | 12.20 |
firefox | tracemonkey | 5 | Rendering | 100 | 71 | 69 | -2 | -2.69 |
firefox | tracemonkey | 6 | Overall | 100 | 85 | 69 | -16 | -18.73 | faster
firefox | tracemonkey | 6 | Page Request | 100 | 2 | 2 | 0 | -9.90 |
firefox | tracemonkey | 6 | Rendering | 100 | 83 | 67 | -16 | -18.97 | faster
firefox | tracemonkey | 7 | Overall | 100 | 65 | 64 | 0 | -0.37 |
firefox | tracemonkey | 7 | Page Request | 100 | 1 | 1 | 0 | -11.94 |
firefox | tracemonkey | 7 | Rendering | 100 | 63 | 63 | 0 | -0.05 |
firefox | tracemonkey | 8 | Overall | 100 | 53 | 54 | 1 | 2.04 |
firefox | tracemonkey | 8 | Page Request | 100 | 1 | 1 | 0 | 17.02 |
firefox | tracemonkey | 8 | Rendering | 100 | 52 | 53 | 1 | 1.82 |
firefox | tracemonkey | 9 | Overall | 100 | 79 | 73 | -6 | -7.86 | faster
firefox | tracemonkey | 9 | Page Request | 100 | 2 | 2 | 0 | -15.14 |
firefox | tracemonkey | 9 | Rendering | 100 | 77 | 71 | -6 | -7.86 | faster
firefox | tracemonkey | 10 | Overall | 100 | 545 | 519 | -27 | -4.86 | faster
firefox | tracemonkey | 10 | Page Request | 100 | 14 | 13 | 0 | -3.56 |
firefox | tracemonkey | 10 | Rendering | 100 | 532 | 506 | -26 | -4.90 | faster
firefox | tracemonkey | 11 | Overall | 100 | 42 | 41 | -1 | -2.50 |
firefox | tracemonkey | 11 | Page Request | 100 | 1 | 1 | 0 | -27.42 | faster
firefox | tracemonkey | 11 | Rendering | 100 | 41 | 40 | -1 | -1.75 |
firefox | tracemonkey | 12 | Overall | 100 | 350 | 332 | -18 | -5.16 | faster
firefox | tracemonkey | 12 | Page Request | 100 | 3 | 3 | 0 | -5.17 |
firefox | tracemonkey | 12 | Rendering | 100 | 347 | 329 | -18 | -5.15 | faster
firefox | tracemonkey | 13 | Overall | 100 | 31 | 31 | 0 | 0.52 |
firefox | tracemonkey | 13 | Page Request | 100 | 1 | 1 | 0 | 4.95 |
firefox | tracemonkey | 13 | Rendering | 100 | 30 | 30 | 0 | 0.20 |
```
2020-06-13 21:12:40 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
async buildFormXObject(
|
|
|
|
|
resources,
|
|
|
|
|
xobj,
|
|
|
|
|
smask,
|
|
|
|
|
operatorList,
|
|
|
|
|
task,
|
|
|
|
|
initialState,
|
|
|
|
|
localColorSpaceCache
|
|
|
|
|
) {
|
|
|
|
|
var dict = xobj.dict;
|
|
|
|
|
var matrix = dict.getArray("Matrix");
|
|
|
|
|
var bbox = dict.getArray("BBox");
|
|
|
|
|
if (Array.isArray(bbox) && bbox.length === 4) {
|
|
|
|
|
bbox = Util.normalizeRect(bbox);
|
|
|
|
|
} else {
|
|
|
|
|
bbox = null;
|
|
|
|
|
}
|
2020-07-15 07:17:27 +09:00
|
|
|
|
let optionalContent = null;
|
|
|
|
|
if (dict.has("OC")) {
|
|
|
|
|
optionalContent = await this.parseMarkedContentProps(
|
|
|
|
|
dict.get("OC"),
|
|
|
|
|
resources
|
|
|
|
|
);
|
|
|
|
|
operatorList.addOp(OPS.beginMarkedContentProps, ["OC", optionalContent]);
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var group = dict.get("Group");
|
|
|
|
|
if (group) {
|
|
|
|
|
var groupOptions = {
|
|
|
|
|
matrix,
|
|
|
|
|
bbox,
|
|
|
|
|
smask,
|
|
|
|
|
isolated: false,
|
|
|
|
|
knockout: false,
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
var groupSubtype = group.get("S");
|
|
|
|
|
var colorSpace = null;
|
|
|
|
|
if (isName(groupSubtype, "Transparency")) {
|
|
|
|
|
groupOptions.isolated = group.get("I") || false;
|
|
|
|
|
groupOptions.knockout = group.get("K") || false;
|
|
|
|
|
if (group.has("CS")) {
|
|
|
|
|
const cs = group.getRaw("CS");
|
|
|
|
|
|
|
|
|
|
const cachedColorSpace = ColorSpace.getCached(
|
|
|
|
|
cs,
|
|
|
|
|
this.xref,
|
|
|
|
|
localColorSpaceCache
|
|
|
|
|
);
|
|
|
|
|
if (cachedColorSpace) {
|
|
|
|
|
colorSpace = cachedColorSpace;
|
|
|
|
|
} else {
|
|
|
|
|
colorSpace = await this.parseColorSpace({
|
2020-06-18 01:45:11 +09:00
|
|
|
|
cs,
|
2020-07-05 19:20:10 +09:00
|
|
|
|
resources,
|
|
|
|
|
localColorSpaceCache,
|
|
|
|
|
});
|
2017-09-19 20:49:30 +09:00
|
|
|
|
}
|
2013-04-09 07:14:56 +09:00
|
|
|
|
}
|
2012-10-16 01:48:45 +09:00
|
|
|
|
}
|
2012-09-14 00:09:46 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (smask && smask.backdrop) {
|
|
|
|
|
colorSpace = colorSpace || ColorSpace.singletons.rgb;
|
|
|
|
|
smask.backdrop = colorSpace.getRgb(smask.backdrop, 0);
|
|
|
|
|
}
|
2012-09-14 00:09:46 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
operatorList.addOp(OPS.beginGroup, [groupOptions]);
|
|
|
|
|
}
|
Change the signatures of the `PartialEvaluator` "constructor" and its `getOperatorList`/`getTextContent` methods to take parameter objects
Currently these methods accept a large number of parameters, which creates quite unwieldy call-sites. When invoking them, you have to remember not only what arguments to supply, but also the correct order, to avoid runtime errors.
Furthermore, since some of the parameters are optional, you also have to remember to pass e.g. `null` or `undefined` for those ones.
Also, adding new parameters to these methods (which happens occasionally), often becomes unnecessarily tedious (based on personal experience).
Please note that I do *not* think that we need/should convert *every* single method in `evaluator.js` (or elsewhere in `/core` files) to take parameter objects. However, in my opinion, once a method starts relying on approximately five parameter (or even more), passing them in individually becomes quite cumbersome.
With these changes, I obviously needed to update the `evaluator_spec.js` unit-tests. The main change there, except the new method signatures[1], is that it's now re-using *one* `PartialEvalutor` instance, since I couldn't see any compelling reason for creating a new one in every single test.
*Note:* If this patch is accepted, my intention is to (time permitting) see if it makes sense to convert additional methods in `evaluator.js` (and other `/core` files) in a similar fashion, but I figured that it'd be a good idea to limit the initial scope somewhat.
---
[1] A fun fact here, note how the `PartialEvaluator` signature used in `evaluator_spec.js` wasn't even correct in the current `master`.
2017-04-30 06:13:51 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
operatorList.addOp(OPS.paintFormXObjectBegin, [matrix, bbox]);
|
2011-10-25 08:55:23 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return this.getOperatorList({
|
|
|
|
|
stream: xobj,
|
|
|
|
|
task,
|
|
|
|
|
resources: dict.get("Resources") || resources,
|
|
|
|
|
operatorList,
|
|
|
|
|
initialState,
|
|
|
|
|
}).then(function () {
|
|
|
|
|
operatorList.addOp(OPS.paintFormXObjectEnd, []);
|
2020-06-07 19:01:51 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (group) {
|
|
|
|
|
operatorList.addOp(OPS.endGroup, [groupOptions]);
|
2020-06-07 19:01:51 +09:00
|
|
|
|
}
|
2020-07-15 07:17:27 +09:00
|
|
|
|
|
|
|
|
|
if (optionalContent) {
|
|
|
|
|
operatorList.addOp(OPS.endMarkedContent, []);
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
});
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
_sendImgData(objId, imgData, cacheGlobally = false) {
|
|
|
|
|
const transfers = imgData ? [imgData.data.buffer] : null;
|
|
|
|
|
|
2020-07-27 01:05:38 +09:00
|
|
|
|
if (this.parsingType3Font || cacheGlobally) {
|
2020-06-07 19:01:51 +09:00
|
|
|
|
return this.handler.send(
|
2020-07-05 19:20:10 +09:00
|
|
|
|
"commonobj",
|
|
|
|
|
[objId, "Image", imgData],
|
2020-06-07 19:01:51 +09:00
|
|
|
|
transfers
|
|
|
|
|
);
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
return this.handler.send(
|
|
|
|
|
"obj",
|
|
|
|
|
[objId, this.pageIndex, "Image", imgData],
|
|
|
|
|
transfers
|
|
|
|
|
);
|
|
|
|
|
}
|
2013-07-11 01:52:37 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
async buildPaintImageXObject({
|
|
|
|
|
resources,
|
|
|
|
|
image,
|
|
|
|
|
isInline = false,
|
|
|
|
|
operatorList,
|
|
|
|
|
cacheKey,
|
|
|
|
|
localImageCache,
|
|
|
|
|
localColorSpaceCache,
|
|
|
|
|
}) {
|
|
|
|
|
var dict = image.dict;
|
|
|
|
|
const imageRef = dict.objId;
|
|
|
|
|
var w = dict.get("Width", "W");
|
|
|
|
|
var h = dict.get("Height", "H");
|
2019-03-16 20:09:14 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (!(w && isNum(w)) || !(h && isNum(h))) {
|
|
|
|
|
warn("Image dimensions are missing, or not numbers.");
|
|
|
|
|
return undefined;
|
|
|
|
|
}
|
|
|
|
|
var maxImageSize = this.options.maxImageSize;
|
|
|
|
|
if (maxImageSize !== -1 && w * h > maxImageSize) {
|
|
|
|
|
warn("Image exceeded maximum allowed size and was removed.");
|
|
|
|
|
return undefined;
|
|
|
|
|
}
|
2013-04-09 07:14:56 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var imageMask = dict.get("ImageMask", "IM") || false;
|
|
|
|
|
var imgData, args;
|
|
|
|
|
if (imageMask) {
|
|
|
|
|
// This depends on a tmpCanvas being filled with the
|
|
|
|
|
// current fillStyle, such that processing the pixel
|
|
|
|
|
// data can't be done here. Instead of creating a
|
|
|
|
|
// complete PDFImage, only read the information needed
|
|
|
|
|
// for later.
|
|
|
|
|
|
|
|
|
|
var width = dict.get("Width", "W");
|
|
|
|
|
var height = dict.get("Height", "H");
|
|
|
|
|
var bitStrideLength = (width + 7) >> 3;
|
|
|
|
|
var imgArray = image.getBytes(
|
|
|
|
|
bitStrideLength * height,
|
|
|
|
|
/* forceClamped = */ true
|
|
|
|
|
);
|
|
|
|
|
var decode = dict.getArray("Decode", "D");
|
|
|
|
|
|
|
|
|
|
imgData = PDFImage.createMask({
|
|
|
|
|
imgArray,
|
|
|
|
|
width,
|
|
|
|
|
height,
|
|
|
|
|
imageIsFromDecodeStream: image instanceof DecodeStream,
|
|
|
|
|
inverseDecode: !!decode && decode[0] > 0,
|
|
|
|
|
});
|
|
|
|
|
imgData.cached = !!cacheKey;
|
|
|
|
|
args = [imgData];
|
2013-04-09 07:14:56 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
operatorList.addOp(OPS.paintImageMaskXObject, args);
|
|
|
|
|
if (cacheKey) {
|
|
|
|
|
localImageCache.set(cacheKey, imageRef, {
|
|
|
|
|
fn: OPS.paintImageMaskXObject,
|
|
|
|
|
args,
|
2017-09-21 19:14:05 +09:00
|
|
|
|
});
|
2019-04-11 19:26:15 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return undefined;
|
|
|
|
|
}
|
2013-04-09 07:14:56 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var softMask = dict.get("SMask", "SM") || false;
|
|
|
|
|
var mask = dict.get("Mask") || false;
|
2018-02-02 00:43:10 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var SMALL_IMAGE_DIMENSIONS = 200;
|
|
|
|
|
// Inlining small images into the queue as RGB data
|
|
|
|
|
if (isInline && !softMask && !mask && w + h < SMALL_IMAGE_DIMENSIONS) {
|
|
|
|
|
const imageObj = new PDFImage({
|
2017-09-21 19:14:05 +09:00
|
|
|
|
xref: this.xref,
|
|
|
|
|
res: resources,
|
|
|
|
|
image,
|
2018-02-01 23:44:49 +09:00
|
|
|
|
isInline,
|
Add local caching of `Function`s, by reference, in the `PDFFunctionFactory` (issue 2541)
Note that compared other structures, such as e.g. Images and ColorSpaces, `Function`s are not referred to by name, which however does bring the advantage of being able to share the cache for an *entire* page.
Furthermore, similar to ColorSpaces, the parsing of individual `Function`s are generally fast enough to not really warrant trying to cache them in any "smarter" way than by reference. (Hence trying to do caching similar to e.g. Fonts would most likely be a losing proposition, given the amount of data lookup/parsing that'd be required.)
Originally I tried implementing this similar to e.g. the recently added ColorSpace caching (and in a couple of different ways), however it unfortunately turned out to be quite ugly/unwieldy given the sheer number of functions/methods where you'd thus need to pass in a `LocalFunctionCache` instance. (Also, the affected functions/methods didn't exactly have short signatures as-is.)
After going back and forth on this for a while it seemed to me that the simplest, or least "invasive" if you will, solution would be if each `PartialEvaluator` instance had its *own* `PDFFunctionFactory` instance (since the latter is already passed to all of the required code). This way each `PDFFunctionFactory` instances could have a local `Function` cache, without it being necessary to provide a `LocalFunctionCache` instance manually at every `PDFFunctionFactory.{create, createFromArray}` call-site.
Obviously, with this patch, there's now (potentially) more `PDFFunctionFactory` instances than before when the entire document shared just one. However, each such instance is really quite small and it's also tied to a `PartialEvaluator` instance and those are *not* kept alive and/or cached. To reduce the impact of these changes, I've tried to make as many of these structures as possible *lazily initialized*, specifically:
- The `PDFFunctionFactory`, on `PartialEvaluator` instances, since not all kinds of general parsing actually requires it. For example: `getTextContent` calls won't cause any `Function` to be parsed, and even some `getOperatorList` calls won't trigger `Function` parsing (if a page contains e.g. no Patterns or "complex" ColorSpaces).
- The `LocalFunctionCache`, on `PDFFunctionFactory` instances, since only certain parsing requires it. Generally speaking, only e.g. Patterns, "complex" ColorSpaces, and/or (some) SoftMasks will trigger any `Function` parsing.
To put these changes into perspective, when loading/rendering all (14) pages of the default `tracemonkey.pdf` file there's now a total of 6 `PDFFunctionFactory` and 1 `LocalFunctionCache` instances created thanks to the lazy initialization.
(If you instead would keep the document-"global" `PDFFunctionFactory` instance and pass around `LocalFunctionCache` instances everywhere, the numbers for the `tracemonkey.pdf` file would be instead be something like 1 `PDFFunctionFactory` and 6 `LocalFunctionCache` instances.)
All-in-all, I thus don't think that the `PDFFunctionFactory` changes should be generally problematic.
With these changes, we can also modify (some) call-sites to pass in a `Reference` rather than the actual `Function` data. This is nice since `Function`s can also be `Streams`, which are not cached on the `XRef` instance (given their potential size), and this way we can avoid unnecessary lookups and thus save some additional time/resources.
Obviously I had intended to include (standard) benchmark results with these changes, but for reasons I don't really understand the test run-time (even with `master`) of the document in issue 2541 is quite a bit slower than in the development viewer.
However, logging the time it takes for the relevant `PDFFunctionFactory`/`PDFFunction ` parsing shows that it takes *approximately* `0.5 ms` for the `Function` in question. Looking up a cached `Function`, on the other hand, is *one order of magnitude faster* which does add up when the same `Function` is invoked close to 2000 times.
2020-06-28 20:12:24 +09:00
|
|
|
|
pdfFunctionFactory: this._pdfFunctionFactory,
|
2020-06-18 01:45:11 +09:00
|
|
|
|
localColorSpaceCache,
|
2020-07-05 19:20:10 +09:00
|
|
|
|
});
|
|
|
|
|
// We force the use of RGBA_32BPP images here, because we can't handle
|
|
|
|
|
// any other kind.
|
|
|
|
|
imgData = imageObj.createImageData(/* forceRGBA = */ true);
|
|
|
|
|
operatorList.addOp(OPS.paintInlineImageXObject, [imgData]);
|
|
|
|
|
return undefined;
|
|
|
|
|
}
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// If there is no imageMask, create the PDFImage and a lot
|
|
|
|
|
// of image processing can be done here.
|
|
|
|
|
let objId = `img_${this.idFactory.createObjId()}`,
|
|
|
|
|
cacheGlobally = false;
|
|
|
|
|
|
|
|
|
|
if (this.parsingType3Font) {
|
Re-factor the `idFactory` functionality, used in the `core/`-code, and move the `fontID` generation into it
Note how the `getFontID`-method in `src/core/fonts.js` is *completely* global, rather than properly tied to the current document. This means that if you repeatedly open and parse/render, and then close, even the *same* PDF document the `fontID`s will still be incremented continuously.
For comparison the `createObjId` method, on `idFactory`, will always create a *consistent* id, assuming of course that the document and its pages are parsed/rendered in the same order.
In order to address this inconsistency, it thus seems reasonable to add a new `createFontId` method on the `idFactory` and use that when obtaining `fontID`s. (When the current `getFontID` method was added the `idFactory` didn't actually exist yet, which explains why the code looks the way it does.)
*Please note:* Since the document id is (still) part of the `loadedName`, it's thus not possible for different documents to have identical font names.
2020-07-07 23:00:05 +09:00
|
|
|
|
objId = `${this.idFactory.getDocId()}_type3_${objId}`;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
} else if (imageRef) {
|
|
|
|
|
cacheGlobally = this.globalImageCache.shouldCache(
|
|
|
|
|
imageRef,
|
|
|
|
|
this.pageIndex
|
|
|
|
|
);
|
2014-03-23 03:15:51 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (cacheGlobally) {
|
|
|
|
|
objId = `${this.idFactory.getDocId()}_${objId}`;
|
2019-04-11 19:26:15 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
2019-04-11 19:26:15 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Ensure that the dependency is added before the image is decoded.
|
|
|
|
|
operatorList.addDependency(objId);
|
|
|
|
|
args = [objId, w, h];
|
Attempt to cache repeated images at the document, rather than the page, level (issue 11878)
Currently image resources, as opposed to e.g. font resources, are handled exclusively on a page-specific basis. Generally speaking this makes sense, since pages are separate from each other, however there's PDF documents where many (or even all) pages actually references exactly the same image resources (through the XRef table). Hence, in some cases, we're decoding the *same* images over and over for every page which is obviously slow and wasting both CPU and memory resources better used elsewhere.[1]
Obviously we cannot simply treat all image resources as-if they're used throughout the entire PDF document, since that would end up increasing memory usage too much.[2]
However, by introducing a `GlobalImageCache` in the worker we can track image resources that appear on more than one page. Hence we can switch image resources from being page-specific to being document-specific, once the image resource has been seen on more than a certain number of pages.
In many cases, such as e.g. the referenced issue, this patch will thus lead to reduced memory usage for image resources. Scrolling through all pages of the document, there's now only a few main-thread copies of the same image data, as opposed to one for each rendered page (i.e. there could theoretically be *twenty* copies of the image data).
While this obviously benefit both CPU and memory usage in this case, for *very* large image data this patch *may* possibly increase persistent main-thread memory usage a tiny bit. Thus to avoid negatively affecting memory usage too much in general, particularly on the main-thread, the `GlobalImageCache` will *only* cache a certain number of image resources at the document level and simply fallback to the default behaviour.
Unfortunately the asynchronous nature of the code, with ranged/streamed loading of data, actually makes all of this much more complicated than if all data could be assumed to be immediately available.[3]
*Please note:* The patch will lead to *small* movement in some existing test-cases, since we're now using the built-in PDF.js JPEG decoder more. This was done in order to simplify the overall implementation, especially on the main-thread, by limiting it to only the `OPS.paintImageXObject` operator.
---
[1] There's e.g. PDF documents that use the same image as background on all pages.
[2] Given that data stored in the `commonObjs`, on the main-thread, are only cleared manually through `PDFDocumentProxy.cleanup`. This as opposed to data stored in the `objs` of each page, which is automatically removed when the page is cleaned-up e.g. by being evicted from the cache in the default viewer.
[3] If the latter case were true, we could simply check for repeat images *before* parsing started and thus avoid handling *any* duplicate image resources.
2020-05-18 21:17:56 +09:00
|
|
|
|
|
Improve how Type3-fonts with dependencies are handled
While the `CharProcs` streams of Type3-fonts *usually* don't rely on dependencies, such as e.g. images, it does happen in some cases.
Currently any dependencies are simply appended to the parent operatorList, which in practice means *only* the operatorList of the *first* page where the Type3-font is being used.
However, there's one thing that's slightly unfortunate with that approach: Since fonts are global to the PDF document, we really ought to ensure that any Type3 dependencies are appended to the operatorList of *all* pages where the Type3-font is being used. Otherwise there's a theoretical risk that, if one page has its rendering paused, another page may try to use a Type3-font whose dependencies are not yet fully resolved. In that case there would be errors, since Type3 operatorLists are executed synchronously.
Hence this patch, which ensures that all relevant pages will have Type3 dependencies appended to the main operatorList. (Note here that the `OperatorList.addDependencies` method, via `OperatorList.addDependency`, ensures that a dependency is only added *once* to any operatorList.)
Finally, these changes also remove the need for the "waiting for the main-thread"-hack that was added to `PartialEvaluator.buildPaintImageXObject` as part of fixing issue 10717.
2020-07-26 19:23:28 +09:00
|
|
|
|
PDFImage.buildImage({
|
2020-07-05 19:20:10 +09:00
|
|
|
|
xref: this.xref,
|
|
|
|
|
res: resources,
|
|
|
|
|
image,
|
|
|
|
|
isInline,
|
|
|
|
|
pdfFunctionFactory: this._pdfFunctionFactory,
|
|
|
|
|
localColorSpaceCache,
|
|
|
|
|
})
|
|
|
|
|
.then(imageObj => {
|
|
|
|
|
imgData = imageObj.createImageData(/* forceRGBA = */ false);
|
Attempt to cache repeated images at the document, rather than the page, level (issue 11878)
Currently image resources, as opposed to e.g. font resources, are handled exclusively on a page-specific basis. Generally speaking this makes sense, since pages are separate from each other, however there's PDF documents where many (or even all) pages actually references exactly the same image resources (through the XRef table). Hence, in some cases, we're decoding the *same* images over and over for every page which is obviously slow and wasting both CPU and memory resources better used elsewhere.[1]
Obviously we cannot simply treat all image resources as-if they're used throughout the entire PDF document, since that would end up increasing memory usage too much.[2]
However, by introducing a `GlobalImageCache` in the worker we can track image resources that appear on more than one page. Hence we can switch image resources from being page-specific to being document-specific, once the image resource has been seen on more than a certain number of pages.
In many cases, such as e.g. the referenced issue, this patch will thus lead to reduced memory usage for image resources. Scrolling through all pages of the document, there's now only a few main-thread copies of the same image data, as opposed to one for each rendered page (i.e. there could theoretically be *twenty* copies of the image data).
While this obviously benefit both CPU and memory usage in this case, for *very* large image data this patch *may* possibly increase persistent main-thread memory usage a tiny bit. Thus to avoid negatively affecting memory usage too much in general, particularly on the main-thread, the `GlobalImageCache` will *only* cache a certain number of image resources at the document level and simply fallback to the default behaviour.
Unfortunately the asynchronous nature of the code, with ranged/streamed loading of data, actually makes all of this much more complicated than if all data could be assumed to be immediately available.[3]
*Please note:* The patch will lead to *small* movement in some existing test-cases, since we're now using the built-in PDF.js JPEG decoder more. This was done in order to simplify the overall implementation, especially on the main-thread, by limiting it to only the `OPS.paintImageXObject` operator.
---
[1] There's e.g. PDF documents that use the same image as background on all pages.
[2] Given that data stored in the `commonObjs`, on the main-thread, are only cleared manually through `PDFDocumentProxy.cleanup`. This as opposed to data stored in the `objs` of each page, which is automatically removed when the page is cleaned-up e.g. by being evicted from the cache in the default viewer.
[3] If the latter case were true, we could simply check for repeat images *before* parsing started and thus avoid handling *any* duplicate image resources.
2020-05-18 21:17:56 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return this._sendImgData(objId, imgData, cacheGlobally);
|
|
|
|
|
})
|
|
|
|
|
.catch(reason => {
|
|
|
|
|
warn(`Unable to decode image "${objId}": "${reason}".`);
|
2013-04-09 07:14:56 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return this._sendImgData(objId, /* imgData = */ null, cacheGlobally);
|
|
|
|
|
});
|
|
|
|
|
|
|
|
|
|
operatorList.addOp(OPS.paintImageXObject, args);
|
|
|
|
|
if (cacheKey) {
|
|
|
|
|
localImageCache.set(cacheKey, imageRef, {
|
|
|
|
|
fn: OPS.paintImageXObject,
|
|
|
|
|
args,
|
|
|
|
|
});
|
|
|
|
|
|
|
|
|
|
if (imageRef) {
|
|
|
|
|
assert(!isInline, "Cannot cache an inline image globally.");
|
|
|
|
|
this.globalImageCache.addPageIndex(imageRef, this.pageIndex);
|
|
|
|
|
|
|
|
|
|
if (cacheGlobally) {
|
|
|
|
|
this.globalImageCache.setData(imageRef, {
|
|
|
|
|
objId,
|
|
|
|
|
fn: OPS.paintImageXObject,
|
|
|
|
|
args,
|
|
|
|
|
});
|
2015-12-05 03:52:45 +09:00
|
|
|
|
}
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
return undefined;
|
|
|
|
|
}
|
2015-12-05 03:52:45 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
handleSMask(
|
|
|
|
|
smask,
|
|
|
|
|
resources,
|
|
|
|
|
operatorList,
|
|
|
|
|
task,
|
|
|
|
|
stateManager,
|
|
|
|
|
localColorSpaceCache
|
|
|
|
|
) {
|
|
|
|
|
var smaskContent = smask.get("G");
|
|
|
|
|
var smaskOptions = {
|
|
|
|
|
subtype: smask.get("S").name,
|
|
|
|
|
backdrop: smask.get("BC"),
|
|
|
|
|
};
|
2014-01-24 02:13:32 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// The SMask might have a alpha/luminosity value transfer function --
|
|
|
|
|
// we will build a map of integer values in range 0..255 to be fast.
|
|
|
|
|
var transferObj = smask.get("TR");
|
|
|
|
|
if (isPDFFunction(transferObj)) {
|
|
|
|
|
const transferFn = this._pdfFunctionFactory.create(transferObj);
|
|
|
|
|
var transferMap = new Uint8Array(256);
|
|
|
|
|
var tmp = new Float32Array(1);
|
|
|
|
|
for (var i = 0; i < 256; i++) {
|
|
|
|
|
tmp[0] = i / 255;
|
|
|
|
|
transferFn(tmp, 0, tmp, 0);
|
|
|
|
|
transferMap[i] = (tmp[0] * 255) | 0;
|
|
|
|
|
}
|
|
|
|
|
smaskOptions.transferMap = transferMap;
|
|
|
|
|
}
|
2012-09-14 00:09:46 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return this.buildFormXObject(
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
resources,
|
2020-07-05 19:20:10 +09:00
|
|
|
|
smaskContent,
|
|
|
|
|
smaskOptions,
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
operatorList,
|
|
|
|
|
task,
|
2020-07-05 19:20:10 +09:00
|
|
|
|
stateManager.state.clone(),
|
|
|
|
|
localColorSpaceCache
|
|
|
|
|
);
|
|
|
|
|
}
|
2013-08-01 03:17:36 +09:00
|
|
|
|
|
2020-08-17 15:49:19 +09:00
|
|
|
|
handleTransferFunction(tr) {
|
|
|
|
|
let transferArray;
|
|
|
|
|
if (Array.isArray(tr)) {
|
|
|
|
|
transferArray = tr;
|
|
|
|
|
} else if (isPDFFunction(tr)) {
|
|
|
|
|
transferArray = [tr];
|
|
|
|
|
} else {
|
|
|
|
|
return null; // Not a valid transfer function entry.
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
const transferMaps = [];
|
|
|
|
|
let numFns = 0,
|
|
|
|
|
numEffectfulFns = 0;
|
|
|
|
|
for (const entry of transferArray) {
|
|
|
|
|
const transferObj = this.xref.fetchIfRef(entry);
|
|
|
|
|
numFns++;
|
|
|
|
|
|
|
|
|
|
if (isName(transferObj, "Identity")) {
|
|
|
|
|
transferMaps.push(null);
|
|
|
|
|
continue;
|
|
|
|
|
} else if (!isPDFFunction(transferObj)) {
|
|
|
|
|
return null; // Not a valid transfer function object.
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
const transferFn = this._pdfFunctionFactory.create(transferObj);
|
|
|
|
|
const transferMap = new Uint8Array(256),
|
|
|
|
|
tmp = new Float32Array(1);
|
|
|
|
|
for (let j = 0; j < 256; j++) {
|
|
|
|
|
tmp[0] = j / 255;
|
|
|
|
|
transferFn(tmp, 0, tmp, 0);
|
|
|
|
|
transferMap[j] = (tmp[0] * 255) | 0;
|
|
|
|
|
}
|
|
|
|
|
transferMaps.push(transferMap);
|
|
|
|
|
numEffectfulFns++;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (!(numFns === 1 || numFns === 4)) {
|
|
|
|
|
return null; // Only 1 or 4 functions are supported, by the specification.
|
|
|
|
|
}
|
|
|
|
|
if (numEffectfulFns === 0) {
|
|
|
|
|
return null; // Only /Identity transfer functions found, which are no-ops.
|
|
|
|
|
}
|
|
|
|
|
return transferMaps;
|
|
|
|
|
}
|
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
handleTilingType(
|
|
|
|
|
fn,
|
2020-10-09 00:33:23 +09:00
|
|
|
|
color,
|
2020-07-05 19:20:10 +09:00
|
|
|
|
resources,
|
|
|
|
|
pattern,
|
|
|
|
|
patternDict,
|
|
|
|
|
operatorList,
|
2020-10-09 00:33:23 +09:00
|
|
|
|
task,
|
|
|
|
|
cacheKey,
|
|
|
|
|
localTilingPatternCache
|
2020-07-05 19:20:10 +09:00
|
|
|
|
) {
|
|
|
|
|
// Create an IR of the pattern code.
|
|
|
|
|
const tilingOpList = new OperatorList();
|
|
|
|
|
// Merge the available resources, to prevent issues when the patternDict
|
|
|
|
|
// is missing some /Resources entries (fixes issue6541.pdf).
|
2020-08-28 08:05:33 +09:00
|
|
|
|
const patternResources = Dict.merge({
|
|
|
|
|
xref: this.xref,
|
|
|
|
|
dictArray: [patternDict.get("Resources"), resources],
|
|
|
|
|
});
|
2020-07-05 19:20:10 +09:00
|
|
|
|
|
|
|
|
|
return this.getOperatorList({
|
|
|
|
|
stream: pattern,
|
|
|
|
|
task,
|
|
|
|
|
resources: patternResources,
|
|
|
|
|
operatorList: tilingOpList,
|
|
|
|
|
})
|
|
|
|
|
.then(function () {
|
2020-10-09 00:33:23 +09:00
|
|
|
|
const operatorListIR = tilingOpList.getIR();
|
|
|
|
|
const tilingPatternIR = getTilingPatternIR(
|
|
|
|
|
operatorListIR,
|
2020-07-05 19:20:10 +09:00
|
|
|
|
patternDict,
|
2020-10-09 00:33:23 +09:00
|
|
|
|
color
|
2020-07-05 19:20:10 +09:00
|
|
|
|
);
|
2020-10-09 00:33:23 +09:00
|
|
|
|
// Add the dependencies to the parent operator list so they are
|
|
|
|
|
// resolved before the sub operator list is executed synchronously.
|
|
|
|
|
operatorList.addDependencies(tilingOpList.dependencies);
|
|
|
|
|
operatorList.addOp(fn, tilingPatternIR);
|
|
|
|
|
|
|
|
|
|
if (cacheKey) {
|
|
|
|
|
localTilingPatternCache.set(cacheKey, patternDict.objId, {
|
|
|
|
|
operatorListIR,
|
|
|
|
|
dict: patternDict,
|
|
|
|
|
});
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
})
|
2020-10-09 00:33:23 +09:00
|
|
|
|
.catch(reason => {
|
|
|
|
|
if (reason instanceof AbortException) {
|
|
|
|
|
return;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
2020-10-09 00:33:23 +09:00
|
|
|
|
if (this.options.ignoreErrors) {
|
|
|
|
|
// Error(s) in the TilingPattern -- sending unsupported feature
|
|
|
|
|
// notification and allow rendering to continue.
|
|
|
|
|
this.handler.send("UnsupportedFeature", {
|
|
|
|
|
featureId: UNSUPPORTED_FEATURES.errorTilingPattern,
|
|
|
|
|
});
|
|
|
|
|
warn(`handleTilingType - ignoring pattern: "${reason}".`);
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
throw reason;
|
|
|
|
|
});
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
2013-03-13 09:20:38 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
handleSetFont(resources, fontArgs, fontRef, operatorList, task, state) {
|
|
|
|
|
// TODO(mack): Not needed?
|
2020-08-05 21:40:31 +09:00
|
|
|
|
var fontName,
|
|
|
|
|
fontSize = 0;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (fontArgs) {
|
|
|
|
|
fontArgs = fontArgs.slice();
|
|
|
|
|
fontName = fontArgs[0].name;
|
2020-08-05 21:40:31 +09:00
|
|
|
|
fontSize = fontArgs[1];
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
2014-05-04 00:28:30 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return this.loadFont(fontName, fontRef, resources)
|
|
|
|
|
.then(translated => {
|
|
|
|
|
if (!translated.font.isType3Font) {
|
|
|
|
|
return translated;
|
2013-08-20 08:33:20 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return translated
|
Improve how Type3-fonts with dependencies are handled
While the `CharProcs` streams of Type3-fonts *usually* don't rely on dependencies, such as e.g. images, it does happen in some cases.
Currently any dependencies are simply appended to the parent operatorList, which in practice means *only* the operatorList of the *first* page where the Type3-font is being used.
However, there's one thing that's slightly unfortunate with that approach: Since fonts are global to the PDF document, we really ought to ensure that any Type3 dependencies are appended to the operatorList of *all* pages where the Type3-font is being used. Otherwise there's a theoretical risk that, if one page has its rendering paused, another page may try to use a Type3-font whose dependencies are not yet fully resolved. In that case there would be errors, since Type3 operatorLists are executed synchronously.
Hence this patch, which ensures that all relevant pages will have Type3 dependencies appended to the main operatorList. (Note here that the `OperatorList.addDependencies` method, via `OperatorList.addDependency`, ensures that a dependency is only added *once* to any operatorList.)
Finally, these changes also remove the need for the "waiting for the main-thread"-hack that was added to `PartialEvaluator.buildPaintImageXObject` as part of fixing issue 10717.
2020-07-26 19:23:28 +09:00
|
|
|
|
.loadType3Data(this, resources, task)
|
2020-07-05 19:20:10 +09:00
|
|
|
|
.then(function () {
|
Improve how Type3-fonts with dependencies are handled
While the `CharProcs` streams of Type3-fonts *usually* don't rely on dependencies, such as e.g. images, it does happen in some cases.
Currently any dependencies are simply appended to the parent operatorList, which in practice means *only* the operatorList of the *first* page where the Type3-font is being used.
However, there's one thing that's slightly unfortunate with that approach: Since fonts are global to the PDF document, we really ought to ensure that any Type3 dependencies are appended to the operatorList of *all* pages where the Type3-font is being used. Otherwise there's a theoretical risk that, if one page has its rendering paused, another page may try to use a Type3-font whose dependencies are not yet fully resolved. In that case there would be errors, since Type3 operatorLists are executed synchronously.
Hence this patch, which ensures that all relevant pages will have Type3 dependencies appended to the main operatorList. (Note here that the `OperatorList.addDependencies` method, via `OperatorList.addDependency`, ensures that a dependency is only added *once* to any operatorList.)
Finally, these changes also remove the need for the "waiting for the main-thread"-hack that was added to `PartialEvaluator.buildPaintImageXObject` as part of fixing issue 10717.
2020-07-26 19:23:28 +09:00
|
|
|
|
// Add the dependencies to the parent operatorList so they are
|
|
|
|
|
// resolved before Type3 operatorLists are executed synchronously.
|
|
|
|
|
operatorList.addDependencies(translated.type3Dependencies);
|
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return translated;
|
|
|
|
|
})
|
|
|
|
|
.catch(reason => {
|
|
|
|
|
// Error in the font data -- sending unsupported feature
|
|
|
|
|
// notification.
|
|
|
|
|
this.handler.send("UnsupportedFeature", {
|
|
|
|
|
featureId: UNSUPPORTED_FEATURES.errorFontLoadType3,
|
|
|
|
|
});
|
|
|
|
|
return new TranslatedFont({
|
|
|
|
|
loadedName: "g_font_error",
|
|
|
|
|
font: new ErrorFont(`Type3 font load error: ${reason}`),
|
|
|
|
|
dict: translated.font,
|
|
|
|
|
extraProperties: this.options.fontExtraProperties,
|
|
|
|
|
});
|
|
|
|
|
});
|
|
|
|
|
})
|
|
|
|
|
.then(translated => {
|
|
|
|
|
state.font = translated.font;
|
2020-08-05 21:40:31 +09:00
|
|
|
|
state.fontSize = fontSize;
|
|
|
|
|
state.fontName = fontName;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
translated.send(this.handler);
|
|
|
|
|
return translated.loadedName;
|
|
|
|
|
});
|
|
|
|
|
}
|
2020-03-02 23:34:00 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
handleText(chars, state) {
|
|
|
|
|
const font = state.font;
|
|
|
|
|
const glyphs = font.charsToGlyphs(chars);
|
|
|
|
|
|
|
|
|
|
if (font.data) {
|
|
|
|
|
const isAddToPathSet = !!(
|
|
|
|
|
state.textRenderingMode & TextRenderingMode.ADD_TO_PATH_FLAG
|
|
|
|
|
);
|
|
|
|
|
if (
|
|
|
|
|
isAddToPathSet ||
|
|
|
|
|
state.fillColorSpace.name === "Pattern" ||
|
|
|
|
|
font.disableFontFace ||
|
|
|
|
|
this.options.disableFontFace
|
|
|
|
|
) {
|
|
|
|
|
PartialEvaluator.buildFontPaths(font, glyphs, this.handler);
|
2020-03-02 23:34:00 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
return glyphs;
|
|
|
|
|
}
|
2020-03-02 23:34:00 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
ensureStateFont(state) {
|
|
|
|
|
if (state.font) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
const reason = new FormatError(
|
|
|
|
|
"Missing setFont (Tf) operator before text rendering operator."
|
|
|
|
|
);
|
|
|
|
|
|
|
|
|
|
if (this.options.ignoreErrors) {
|
|
|
|
|
// Missing setFont operator before text rendering operator -- sending
|
|
|
|
|
// unsupported feature notification and allow rendering to continue.
|
|
|
|
|
this.handler.send("UnsupportedFeature", {
|
|
|
|
|
featureId: UNSUPPORTED_FEATURES.errorFontState,
|
|
|
|
|
});
|
|
|
|
|
warn(`ensureStateFont: "${reason}".`);
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
throw reason;
|
|
|
|
|
}
|
|
|
|
|
|
2020-07-15 21:25:24 +09:00
|
|
|
|
async setGState({
|
2020-07-05 19:20:10 +09:00
|
|
|
|
resources,
|
|
|
|
|
gState,
|
|
|
|
|
operatorList,
|
2020-07-11 20:52:11 +09:00
|
|
|
|
cacheKey,
|
2020-07-05 19:20:10 +09:00
|
|
|
|
task,
|
|
|
|
|
stateManager,
|
2020-07-11 20:52:11 +09:00
|
|
|
|
localGStateCache,
|
|
|
|
|
localColorSpaceCache,
|
|
|
|
|
}) {
|
|
|
|
|
const gStateRef = gState.objId;
|
|
|
|
|
let isSimpleGState = true;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// This array holds the converted/processed state data.
|
|
|
|
|
var gStateObj = [];
|
|
|
|
|
var gStateKeys = gState.getKeys();
|
|
|
|
|
var promise = Promise.resolve();
|
|
|
|
|
for (var i = 0, ii = gStateKeys.length; i < ii; i++) {
|
|
|
|
|
const key = gStateKeys[i];
|
|
|
|
|
const value = gState.get(key);
|
|
|
|
|
switch (key) {
|
|
|
|
|
case "Type":
|
|
|
|
|
break;
|
|
|
|
|
case "LW":
|
|
|
|
|
case "LC":
|
|
|
|
|
case "LJ":
|
|
|
|
|
case "ML":
|
|
|
|
|
case "D":
|
|
|
|
|
case "RI":
|
|
|
|
|
case "FL":
|
|
|
|
|
case "CA":
|
|
|
|
|
case "ca":
|
|
|
|
|
gStateObj.push([key, value]);
|
|
|
|
|
break;
|
|
|
|
|
case "Font":
|
2020-08-17 00:32:15 +09:00
|
|
|
|
isSimpleGState = false;
|
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
promise = promise.then(() => {
|
|
|
|
|
return this.handleSetFont(
|
|
|
|
|
resources,
|
|
|
|
|
null,
|
|
|
|
|
value[0],
|
|
|
|
|
operatorList,
|
|
|
|
|
task,
|
|
|
|
|
stateManager.state
|
|
|
|
|
).then(function (loadedName) {
|
|
|
|
|
operatorList.addDependency(loadedName);
|
|
|
|
|
gStateObj.push([key, [loadedName, value[1]]]);
|
|
|
|
|
});
|
|
|
|
|
});
|
|
|
|
|
break;
|
|
|
|
|
case "BM":
|
|
|
|
|
gStateObj.push([key, normalizeBlendMode(value)]);
|
|
|
|
|
break;
|
|
|
|
|
case "SMask":
|
|
|
|
|
if (isName(value, "None")) {
|
|
|
|
|
gStateObj.push([key, false]);
|
2013-04-09 07:14:56 +09:00
|
|
|
|
break;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
if (isDict(value)) {
|
2020-07-11 20:52:11 +09:00
|
|
|
|
isSimpleGState = false;
|
|
|
|
|
|
2017-04-30 06:36:43 +09:00
|
|
|
|
promise = promise.then(() => {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return this.handleSMask(
|
|
|
|
|
value,
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
resources,
|
|
|
|
|
operatorList,
|
|
|
|
|
task,
|
2020-07-05 19:20:10 +09:00
|
|
|
|
stateManager,
|
|
|
|
|
localColorSpaceCache
|
|
|
|
|
);
|
2014-05-10 10:21:15 +09:00
|
|
|
|
});
|
2020-07-05 19:20:10 +09:00
|
|
|
|
gStateObj.push([key, true]);
|
|
|
|
|
} else {
|
|
|
|
|
warn("Unsupported SMask type");
|
|
|
|
|
}
|
2020-08-17 15:49:19 +09:00
|
|
|
|
break;
|
|
|
|
|
case "TR":
|
|
|
|
|
const transferMaps = this.handleTransferFunction(value);
|
|
|
|
|
gStateObj.push([key, transferMaps]);
|
2020-07-05 19:20:10 +09:00
|
|
|
|
break;
|
|
|
|
|
// Only generate info log messages for the following since
|
|
|
|
|
// they are unlikely to have a big impact on the rendering.
|
|
|
|
|
case "OP":
|
|
|
|
|
case "op":
|
|
|
|
|
case "OPM":
|
|
|
|
|
case "BG":
|
|
|
|
|
case "BG2":
|
|
|
|
|
case "UCR":
|
|
|
|
|
case "UCR2":
|
|
|
|
|
case "TR2":
|
|
|
|
|
case "HT":
|
|
|
|
|
case "SM":
|
|
|
|
|
case "SA":
|
|
|
|
|
case "AIS":
|
|
|
|
|
case "TK":
|
|
|
|
|
// TODO implement these operators.
|
|
|
|
|
info("graphic state operator " + key);
|
|
|
|
|
break;
|
|
|
|
|
default:
|
|
|
|
|
info("Unknown graphic state operator " + key);
|
|
|
|
|
break;
|
2013-04-09 07:14:56 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
return promise.then(function () {
|
|
|
|
|
if (gStateObj.length > 0) {
|
|
|
|
|
operatorList.addOp(OPS.setGState, [gStateObj]);
|
2013-06-26 02:33:53 +09:00
|
|
|
|
}
|
2020-07-11 20:52:11 +09:00
|
|
|
|
|
|
|
|
|
if (isSimpleGState) {
|
|
|
|
|
localGStateCache.set(cacheKey, gStateRef, gStateObj);
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
});
|
|
|
|
|
}
|
2019-10-02 21:13:49 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
loadFont(fontName, font, resources) {
|
|
|
|
|
const errorFont = () => {
|
|
|
|
|
return Promise.resolve(
|
|
|
|
|
new TranslatedFont({
|
|
|
|
|
loadedName: "g_font_error",
|
|
|
|
|
font: new ErrorFont(`Font "${fontName}" is not available.`),
|
|
|
|
|
dict: font,
|
|
|
|
|
extraProperties: this.options.fontExtraProperties,
|
|
|
|
|
})
|
|
|
|
|
);
|
|
|
|
|
};
|
2019-10-02 21:13:49 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var fontRef,
|
|
|
|
|
xref = this.xref;
|
|
|
|
|
if (font) {
|
|
|
|
|
// Loading by ref.
|
|
|
|
|
if (!isRef(font)) {
|
|
|
|
|
throw new FormatError('The "font" object should be a reference.');
|
2014-06-12 19:46:39 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
fontRef = font;
|
|
|
|
|
} else {
|
|
|
|
|
// Loading by name.
|
|
|
|
|
var fontRes = resources.get("Font");
|
|
|
|
|
if (fontRes) {
|
|
|
|
|
fontRef = fontRes.getRaw(fontName);
|
2013-06-26 02:33:53 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
if (!fontRef) {
|
|
|
|
|
const partialMsg = `Font "${
|
|
|
|
|
fontName || (font && font.toString())
|
|
|
|
|
}" is not available`;
|
2013-06-26 02:33:53 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (!this.options.ignoreErrors && !this.parsingType3Font) {
|
|
|
|
|
warn(`${partialMsg}.`);
|
2013-08-01 03:17:36 +09:00
|
|
|
|
return errorFont();
|
2013-05-04 03:13:45 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Font not found -- sending unsupported feature notification.
|
|
|
|
|
this.handler.send("UnsupportedFeature", {
|
|
|
|
|
featureId: UNSUPPORTED_FEATURES.errorFontMissing,
|
|
|
|
|
});
|
|
|
|
|
warn(`${partialMsg} -- attempting to fallback to a default font.`);
|
2014-03-04 02:44:45 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Falling back to a default font to avoid completely broken rendering,
|
|
|
|
|
// but note that there're no guarantees that things will look "correct".
|
|
|
|
|
fontRef = PartialEvaluator.fallbackFontDict;
|
|
|
|
|
}
|
2014-05-20 06:27:54 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (this.fontCache.has(fontRef)) {
|
|
|
|
|
return this.fontCache.get(fontRef);
|
|
|
|
|
}
|
2014-05-10 10:21:15 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
font = xref.fetchIfRef(fontRef);
|
|
|
|
|
if (!isDict(font)) {
|
|
|
|
|
return errorFont();
|
|
|
|
|
}
|
Slightly refactor the `fontRef` handling in `PartialEvaluator_loadFont` (issue 7403 and issue 7402)
Originally, I was just going to change this code to use `Ref_toString` in a couple more places. When I started reading the code, I figured that it wouldn't hurt to clean up a couple of comments. While doing this, I noticed that the logic for the (rare) `isDict(fontRef)` case could do with a few improvements.
There should be no functional changes with this patch, but given the added reference checks, we will now avoid bogus `Ref`s when resolving font aliases. In practice, as issue 7403 shows, the current code can break certain PDF files even if it's very rare.
Note that the only thing that this patch will change, is the `font.loadedName` in the case where a `fontRef` is a reference *and* the font doesn't have a descriptor. Previously for `fontRef = Ref(4, 0)` we'd get `font.loadedName = 'g_d0_f4_0'`, and with this patch `font.loadedName = g_d0_f4R`, which is actually one character shorted in most cases. (Given that `Ref_toString` contains an optimization for the `gen === 0` case, which is by far the most common `gen` value.)
In the already existing fallback case, where the `fontName` is used to when creating the `font.loadedName`, we allow any alphanumeric character. Hence I don't see how (as mentioned above) e.g. `font.loadedName = g_d0_f4R` would be an issue here.
2016-05-23 22:32:04 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// We are holding `font.translated` references just for `fontRef`s that
|
|
|
|
|
// are not actually `Ref`s, but rather `Dict`s. See explanation below.
|
|
|
|
|
if (font.translated) {
|
|
|
|
|
return font.translated;
|
|
|
|
|
}
|
Slightly refactor the `fontRef` handling in `PartialEvaluator_loadFont` (issue 7403 and issue 7402)
Originally, I was just going to change this code to use `Ref_toString` in a couple more places. When I started reading the code, I figured that it wouldn't hurt to clean up a couple of comments. While doing this, I noticed that the logic for the (rare) `isDict(fontRef)` case could do with a few improvements.
There should be no functional changes with this patch, but given the added reference checks, we will now avoid bogus `Ref`s when resolving font aliases. In practice, as issue 7403 shows, the current code can break certain PDF files even if it's very rare.
Note that the only thing that this patch will change, is the `font.loadedName` in the case where a `fontRef` is a reference *and* the font doesn't have a descriptor. Previously for `fontRef = Ref(4, 0)` we'd get `font.loadedName = 'g_d0_f4_0'`, and with this patch `font.loadedName = g_d0_f4R`, which is actually one character shorted in most cases. (Given that `Ref_toString` contains an optimization for the `gen === 0` case, which is by far the most common `gen` value.)
In the already existing fallback case, where the `fontName` is used to when creating the `font.loadedName`, we allow any alphanumeric character. Hence I don't see how (as mentioned above) e.g. `font.loadedName = g_d0_f4R` would be an issue here.
2016-05-23 22:32:04 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var fontCapability = createPromiseCapability();
|
|
|
|
|
|
|
|
|
|
var preEvaluatedFont = this.preEvaluateFont(font);
|
|
|
|
|
const { descriptor, hash } = preEvaluatedFont;
|
|
|
|
|
|
|
|
|
|
var fontRefIsRef = isRef(fontRef),
|
|
|
|
|
fontID;
|
|
|
|
|
if (fontRefIsRef) {
|
Re-factor the `idFactory` functionality, used in the `core/`-code, and move the `fontID` generation into it
Note how the `getFontID`-method in `src/core/fonts.js` is *completely* global, rather than properly tied to the current document. This means that if you repeatedly open and parse/render, and then close, even the *same* PDF document the `fontID`s will still be incremented continuously.
For comparison the `createObjId` method, on `idFactory`, will always create a *consistent* id, assuming of course that the document and its pages are parsed/rendered in the same order.
In order to address this inconsistency, it thus seems reasonable to add a new `createFontId` method on the `idFactory` and use that when obtaining `fontID`s. (When the current `getFontID` method was added the `idFactory` didn't actually exist yet, which explains why the code looks the way it does.)
*Please note:* Since the document id is (still) part of the `loadedName`, it's thus not possible for different documents to have identical font names.
2020-07-07 23:00:05 +09:00
|
|
|
|
fontID = `f${fontRef.toString()}`;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (hash && isDict(descriptor)) {
|
|
|
|
|
if (!descriptor.fontAliases) {
|
|
|
|
|
descriptor.fontAliases = Object.create(null);
|
|
|
|
|
}
|
|
|
|
|
var fontAliases = descriptor.fontAliases;
|
2014-03-04 02:44:45 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (fontAliases[hash]) {
|
|
|
|
|
var aliasFontRef = fontAliases[hash].aliasRef;
|
|
|
|
|
if (fontRefIsRef && aliasFontRef && this.fontCache.has(aliasFontRef)) {
|
|
|
|
|
this.fontCache.putAlias(fontRef, aliasFontRef);
|
|
|
|
|
return this.fontCache.get(fontRef);
|
Slightly refactor the `fontRef` handling in `PartialEvaluator_loadFont` (issue 7403 and issue 7402)
Originally, I was just going to change this code to use `Ref_toString` in a couple more places. When I started reading the code, I figured that it wouldn't hurt to clean up a couple of comments. While doing this, I noticed that the logic for the (rare) `isDict(fontRef)` case could do with a few improvements.
There should be no functional changes with this patch, but given the added reference checks, we will now avoid bogus `Ref`s when resolving font aliases. In practice, as issue 7403 shows, the current code can break certain PDF files even if it's very rare.
Note that the only thing that this patch will change, is the `font.loadedName` in the case where a `fontRef` is a reference *and* the font doesn't have a descriptor. Previously for `fontRef = Ref(4, 0)` we'd get `font.loadedName = 'g_d0_f4_0'`, and with this patch `font.loadedName = g_d0_f4R`, which is actually one character shorted in most cases. (Given that `Ref_toString` contains an optimization for the `gen === 0` case, which is by far the most common `gen` value.)
In the already existing fallback case, where the `fontName` is used to when creating the `font.loadedName`, we allow any alphanumeric character. Hence I don't see how (as mentioned above) e.g. `font.loadedName = g_d0_f4R` would be an issue here.
2016-05-23 22:32:04 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
} else {
|
|
|
|
|
fontAliases[hash] = {
|
Re-factor the `idFactory` functionality, used in the `core/`-code, and move the `fontID` generation into it
Note how the `getFontID`-method in `src/core/fonts.js` is *completely* global, rather than properly tied to the current document. This means that if you repeatedly open and parse/render, and then close, even the *same* PDF document the `fontID`s will still be incremented continuously.
For comparison the `createObjId` method, on `idFactory`, will always create a *consistent* id, assuming of course that the document and its pages are parsed/rendered in the same order.
In order to address this inconsistency, it thus seems reasonable to add a new `createFontId` method on the `idFactory` and use that when obtaining `fontID`s. (When the current `getFontID` method was added the `idFactory` didn't actually exist yet, which explains why the code looks the way it does.)
*Please note:* Since the document id is (still) part of the `loadedName`, it's thus not possible for different documents to have identical font names.
2020-07-07 23:00:05 +09:00
|
|
|
|
fontID: this.idFactory.createFontId(),
|
2020-07-05 19:20:10 +09:00
|
|
|
|
};
|
2014-03-04 02:44:45 +09:00
|
|
|
|
}
|
|
|
|
|
|
Slightly refactor the `fontRef` handling in `PartialEvaluator_loadFont` (issue 7403 and issue 7402)
Originally, I was just going to change this code to use `Ref_toString` in a couple more places. When I started reading the code, I figured that it wouldn't hurt to clean up a couple of comments. While doing this, I noticed that the logic for the (rare) `isDict(fontRef)` case could do with a few improvements.
There should be no functional changes with this patch, but given the added reference checks, we will now avoid bogus `Ref`s when resolving font aliases. In practice, as issue 7403 shows, the current code can break certain PDF files even if it's very rare.
Note that the only thing that this patch will change, is the `font.loadedName` in the case where a `fontRef` is a reference *and* the font doesn't have a descriptor. Previously for `fontRef = Ref(4, 0)` we'd get `font.loadedName = 'g_d0_f4_0'`, and with this patch `font.loadedName = g_d0_f4R`, which is actually one character shorted in most cases. (Given that `Ref_toString` contains an optimization for the `gen === 0` case, which is by far the most common `gen` value.)
In the already existing fallback case, where the `fontName` is used to when creating the `font.loadedName`, we allow any alphanumeric character. Hence I don't see how (as mentioned above) e.g. `font.loadedName = g_d0_f4R` would be an issue here.
2016-05-23 22:32:04 +09:00
|
|
|
|
if (fontRefIsRef) {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
fontAliases[hash].aliasRef = fontRef;
|
2013-12-17 08:19:31 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
fontID = fontAliases[hash].fontID;
|
|
|
|
|
}
|
2013-05-04 03:13:45 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Workaround for bad PDF generators that reference fonts incorrectly,
|
|
|
|
|
// where `fontRef` is a `Dict` rather than a `Ref` (fixes bug946506.pdf).
|
|
|
|
|
// In this case we should not put the font into `this.fontCache` (which is
|
|
|
|
|
// a `RefSetCache`), since it's not meaningful to use a `Dict` as a key.
|
|
|
|
|
//
|
|
|
|
|
// However, if we don't cache the font it's not possible to remove it
|
|
|
|
|
// when `cleanup` is triggered from the API, which causes issues on
|
|
|
|
|
// subsequent rendering operations (see issue7403.pdf).
|
|
|
|
|
// A simple workaround would be to just not hold `font.translated`
|
|
|
|
|
// references in this case, but this would force us to unnecessarily load
|
|
|
|
|
// the same fonts over and over.
|
|
|
|
|
//
|
|
|
|
|
// Instead, we cheat a bit by attempting to use a modified `fontID` as a
|
|
|
|
|
// key in `this.fontCache`, to allow the font to be cached.
|
|
|
|
|
// NOTE: This works because `RefSetCache` calls `toString()` on provided
|
|
|
|
|
// keys. Also, since `fontRef` is used when getting cached fonts,
|
|
|
|
|
// we'll not accidentally match fonts cached with the `fontID`.
|
|
|
|
|
if (fontRefIsRef) {
|
|
|
|
|
this.fontCache.put(fontRef, fontCapability.promise);
|
|
|
|
|
} else {
|
|
|
|
|
if (!fontID) {
|
Re-factor the `idFactory` functionality, used in the `core/`-code, and move the `fontID` generation into it
Note how the `getFontID`-method in `src/core/fonts.js` is *completely* global, rather than properly tied to the current document. This means that if you repeatedly open and parse/render, and then close, even the *same* PDF document the `fontID`s will still be incremented continuously.
For comparison the `createObjId` method, on `idFactory`, will always create a *consistent* id, assuming of course that the document and its pages are parsed/rendered in the same order.
In order to address this inconsistency, it thus seems reasonable to add a new `createFontId` method on the `idFactory` and use that when obtaining `fontID`s. (When the current `getFontID` method was added the `idFactory` didn't actually exist yet, which explains why the code looks the way it does.)
*Please note:* Since the document id is (still) part of the `loadedName`, it's thus not possible for different documents to have identical font names.
2020-07-07 23:00:05 +09:00
|
|
|
|
fontID = this.idFactory.createFontId();
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
this.fontCache.put(`id_${fontID}`, fontCapability.promise);
|
|
|
|
|
}
|
Re-factor the `idFactory` functionality, used in the `core/`-code, and move the `fontID` generation into it
Note how the `getFontID`-method in `src/core/fonts.js` is *completely* global, rather than properly tied to the current document. This means that if you repeatedly open and parse/render, and then close, even the *same* PDF document the `fontID`s will still be incremented continuously.
For comparison the `createObjId` method, on `idFactory`, will always create a *consistent* id, assuming of course that the document and its pages are parsed/rendered in the same order.
In order to address this inconsistency, it thus seems reasonable to add a new `createFontId` method on the `idFactory` and use that when obtaining `fontID`s. (When the current `getFontID` method was added the `idFactory` didn't actually exist yet, which explains why the code looks the way it does.)
*Please note:* Since the document id is (still) part of the `loadedName`, it's thus not possible for different documents to have identical font names.
2020-07-07 23:00:05 +09:00
|
|
|
|
assert(
|
|
|
|
|
fontID && fontID.startsWith("f"),
|
|
|
|
|
'The "fontID" must be (correctly) defined.'
|
|
|
|
|
);
|
2012-02-21 22:28:42 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Keep track of each font we translated so the caller can
|
|
|
|
|
// load them asynchronously before calling display on a page.
|
Re-factor the `idFactory` functionality, used in the `core/`-code, and move the `fontID` generation into it
Note how the `getFontID`-method in `src/core/fonts.js` is *completely* global, rather than properly tied to the current document. This means that if you repeatedly open and parse/render, and then close, even the *same* PDF document the `fontID`s will still be incremented continuously.
For comparison the `createObjId` method, on `idFactory`, will always create a *consistent* id, assuming of course that the document and its pages are parsed/rendered in the same order.
In order to address this inconsistency, it thus seems reasonable to add a new `createFontId` method on the `idFactory` and use that when obtaining `fontID`s. (When the current `getFontID` method was added the `idFactory` didn't actually exist yet, which explains why the code looks the way it does.)
*Please note:* Since the document id is (still) part of the `loadedName`, it's thus not possible for different documents to have identical font names.
2020-07-07 23:00:05 +09:00
|
|
|
|
font.loadedName = `${this.idFactory.getDocId()}_${fontID}`;
|
2013-04-09 07:14:56 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
font.translated = fontCapability.promise;
|
2014-05-20 06:27:54 +09:00
|
|
|
|
|
2020-10-15 16:30:54 +09:00
|
|
|
|
this.translateFont(preEvaluatedFont)
|
2020-07-05 19:20:10 +09:00
|
|
|
|
.then(translatedFont => {
|
|
|
|
|
if (translatedFont.fontType !== undefined) {
|
|
|
|
|
var xrefFontStats = xref.stats.fontTypes;
|
|
|
|
|
xrefFontStats[translatedFont.fontType] = true;
|
|
|
|
|
}
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
fontCapability.resolve(
|
|
|
|
|
new TranslatedFont({
|
|
|
|
|
loadedName: font.loadedName,
|
|
|
|
|
font: translatedFont,
|
|
|
|
|
dict: font,
|
|
|
|
|
extraProperties: this.options.fontExtraProperties,
|
|
|
|
|
})
|
|
|
|
|
);
|
|
|
|
|
})
|
|
|
|
|
.catch(reason => {
|
|
|
|
|
// TODO fontCapability.reject?
|
|
|
|
|
// Error in the font data -- sending unsupported feature notification.
|
|
|
|
|
this.handler.send("UnsupportedFeature", {
|
|
|
|
|
featureId: UNSUPPORTED_FEATURES.errorFontTranslate,
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
});
|
2013-04-09 07:14:56 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
try {
|
|
|
|
|
// error, but it's still nice to have font type reported
|
|
|
|
|
var fontFile3 = descriptor && descriptor.get("FontFile3");
|
|
|
|
|
var subtype = fontFile3 && fontFile3.get("Subtype");
|
|
|
|
|
var fontType = getFontType(
|
|
|
|
|
preEvaluatedFont.type,
|
|
|
|
|
subtype && subtype.name
|
|
|
|
|
);
|
|
|
|
|
var xrefFontStats = xref.stats.fontTypes;
|
|
|
|
|
xrefFontStats[fontType] = true;
|
|
|
|
|
} catch (ex) {}
|
2019-04-22 00:03:38 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
fontCapability.resolve(
|
|
|
|
|
new TranslatedFont({
|
|
|
|
|
loadedName: font.loadedName,
|
|
|
|
|
font: new ErrorFont(
|
|
|
|
|
reason instanceof Error ? reason.message : reason
|
|
|
|
|
),
|
|
|
|
|
dict: font,
|
|
|
|
|
extraProperties: this.options.fontExtraProperties,
|
|
|
|
|
})
|
|
|
|
|
);
|
|
|
|
|
});
|
|
|
|
|
return fontCapability.promise;
|
|
|
|
|
}
|
2019-04-22 00:03:38 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
buildPath(operatorList, fn, args, parsingText = false) {
|
|
|
|
|
var lastIndex = operatorList.length - 1;
|
|
|
|
|
if (!args) {
|
|
|
|
|
args = [];
|
|
|
|
|
}
|
|
|
|
|
if (
|
|
|
|
|
lastIndex < 0 ||
|
|
|
|
|
operatorList.fnArray[lastIndex] !== OPS.constructPath
|
|
|
|
|
) {
|
|
|
|
|
// Handle corrupt PDF documents that contains path operators inside of
|
|
|
|
|
// text objects, which may shift subsequent text, by enclosing the path
|
|
|
|
|
// operator in save/restore operators (fixes issue10542_reduced.pdf).
|
|
|
|
|
//
|
|
|
|
|
// Note that this will effectively disable the optimization in the
|
|
|
|
|
// `else` branch below, but given that this type of corruption is
|
|
|
|
|
// *extremely* rare that shouldn't really matter much in practice.
|
|
|
|
|
if (parsingText) {
|
|
|
|
|
warn(`Encountered path operator "${fn}" inside of a text object.`);
|
|
|
|
|
operatorList.addOp(OPS.save, null);
|
2014-04-30 23:09:04 +09:00
|
|
|
|
}
|
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
operatorList.addOp(OPS.constructPath, [[fn], args]);
|
2019-10-31 23:53:51 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (parsingText) {
|
|
|
|
|
operatorList.addOp(OPS.restore, null);
|
|
|
|
|
}
|
|
|
|
|
} else {
|
|
|
|
|
var opArgs = operatorList.argsArray[lastIndex];
|
|
|
|
|
opArgs[0].push(fn);
|
|
|
|
|
Array.prototype.push.apply(opArgs[1], args);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
parseColorSpace({ cs, resources, localColorSpaceCache }) {
|
|
|
|
|
return ColorSpace.parseAsync({
|
2020-06-18 01:45:11 +09:00
|
|
|
|
cs,
|
2020-07-05 19:20:10 +09:00
|
|
|
|
xref: this.xref,
|
2020-06-18 01:45:11 +09:00
|
|
|
|
resources,
|
2020-07-05 19:20:10 +09:00
|
|
|
|
pdfFunctionFactory: this._pdfFunctionFactory,
|
|
|
|
|
localColorSpaceCache,
|
|
|
|
|
}).catch(reason => {
|
|
|
|
|
if (reason instanceof AbortException) {
|
|
|
|
|
return null;
|
|
|
|
|
}
|
|
|
|
|
if (this.options.ignoreErrors) {
|
|
|
|
|
// Error(s) in the ColorSpace -- sending unsupported feature
|
|
|
|
|
// notification and allow rendering to continue.
|
|
|
|
|
this.handler.send("UnsupportedFeature", {
|
|
|
|
|
featureId: UNSUPPORTED_FEATURES.errorColorSpace,
|
|
|
|
|
});
|
|
|
|
|
warn(`parseColorSpace - ignoring ColorSpace: "${reason}".`);
|
|
|
|
|
return null;
|
2014-05-22 02:47:42 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
throw reason;
|
|
|
|
|
});
|
|
|
|
|
}
|
2014-05-22 02:47:42 +09:00
|
|
|
|
|
2020-10-09 00:33:23 +09:00
|
|
|
|
handleColorN(
|
2020-07-05 19:20:10 +09:00
|
|
|
|
operatorList,
|
|
|
|
|
fn,
|
|
|
|
|
args,
|
|
|
|
|
cs,
|
|
|
|
|
patterns,
|
|
|
|
|
resources,
|
|
|
|
|
task,
|
2020-10-09 00:33:23 +09:00
|
|
|
|
localColorSpaceCache,
|
|
|
|
|
localTilingPatternCache
|
2020-07-05 19:20:10 +09:00
|
|
|
|
) {
|
|
|
|
|
// compile tiling patterns
|
2020-10-09 00:33:23 +09:00
|
|
|
|
const patternName = args[args.length - 1];
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// SCN/scn applies patterns along with normal colors
|
2020-10-09 00:33:23 +09:00
|
|
|
|
if (patternName instanceof Name) {
|
|
|
|
|
const localTilingPattern = localTilingPatternCache.getByName(patternName);
|
|
|
|
|
if (localTilingPattern) {
|
|
|
|
|
try {
|
|
|
|
|
const color = cs.base ? cs.base.getRgb(args, 0) : null;
|
|
|
|
|
const tilingPatternIR = getTilingPatternIR(
|
|
|
|
|
localTilingPattern.operatorListIR,
|
|
|
|
|
localTilingPattern.dict,
|
|
|
|
|
color
|
|
|
|
|
);
|
|
|
|
|
operatorList.addOp(fn, tilingPatternIR);
|
|
|
|
|
return undefined;
|
|
|
|
|
} catch (ex) {
|
|
|
|
|
if (ex instanceof MissingDataException) {
|
|
|
|
|
throw ex;
|
|
|
|
|
}
|
|
|
|
|
// Handle any errors during normal TilingPattern parsing.
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
// TODO: Attempt to lookup cached TilingPatterns by reference as well,
|
|
|
|
|
// if and only if there are PDF documents where doing so would
|
|
|
|
|
// significantly improve performance.
|
|
|
|
|
|
|
|
|
|
let pattern = patterns.get(patternName.name);
|
|
|
|
|
if (pattern) {
|
|
|
|
|
var dict = isStream(pattern) ? pattern.dict : pattern;
|
|
|
|
|
var typeNum = dict.get("PatternType");
|
|
|
|
|
|
|
|
|
|
if (typeNum === PatternType.TILING) {
|
|
|
|
|
const color = cs.base ? cs.base.getRgb(args, 0) : null;
|
|
|
|
|
return this.handleTilingType(
|
|
|
|
|
fn,
|
|
|
|
|
color,
|
|
|
|
|
resources,
|
|
|
|
|
pattern,
|
|
|
|
|
dict,
|
|
|
|
|
operatorList,
|
|
|
|
|
task,
|
|
|
|
|
patternName,
|
|
|
|
|
localTilingPatternCache
|
|
|
|
|
);
|
|
|
|
|
} else if (typeNum === PatternType.SHADING) {
|
|
|
|
|
var shading = dict.get("Shading");
|
|
|
|
|
var matrix = dict.getArray("Matrix");
|
|
|
|
|
pattern = Pattern.parseShading(
|
|
|
|
|
shading,
|
|
|
|
|
matrix,
|
|
|
|
|
this.xref,
|
|
|
|
|
resources,
|
|
|
|
|
this.handler,
|
|
|
|
|
this._pdfFunctionFactory,
|
|
|
|
|
localColorSpaceCache
|
|
|
|
|
);
|
|
|
|
|
operatorList.addOp(fn, pattern.getIR());
|
|
|
|
|
return undefined;
|
|
|
|
|
}
|
|
|
|
|
throw new FormatError(`Unknown PatternType: ${typeNum}`);
|
[api-minor] Always allow e.g. rendering to continue even if there are errors, and add a `stopAtErrors` parameter to `getDocument` to opt-out of this behaviour (issue 6342, issue 3795, bug 1130815)
Other PDF readers, e.g. Adobe Reader and PDFium (in Chrome), will attempt to render as much of a page as possible even if there are errors present.
Currently we just bail as soon the first error is hit, which means that we'll usually not render anything in these cases and just display a blank page instead.
NOTE: This patch changes the default behaviour of the PDF.js API to always attempt to recover as much data as possible, even when encountering errors during e.g. `getOperatorList`/`getTextContent`, which thus improve our handling of corrupt PDF files and allow the default viewer to handle errors slightly more gracefully.
In the event that an API consumer wishes to use the old behaviour, where we stop parsing as soon as an error is encountered, the `stopAtErrors` parameter can be set at `getDocument`.
Fixes, inasmuch it's possible since the PDF files are corrupt, e.g. issue 6342, issue 3795, and [bug 1130815](https://bugzilla.mozilla.org/show_bug.cgi?id=1130815) (and probably others too).
2017-02-19 22:03:08 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
throw new FormatError(`Unknown PatternName: ${patternName}`);
|
|
|
|
|
}
|
[api-minor] Always allow e.g. rendering to continue even if there are errors, and add a `stopAtErrors` parameter to `getDocument` to opt-out of this behaviour (issue 6342, issue 3795, bug 1130815)
Other PDF readers, e.g. Adobe Reader and PDFium (in Chrome), will attempt to render as much of a page as possible even if there are errors present.
Currently we just bail as soon the first error is hit, which means that we'll usually not render anything in these cases and just display a blank page instead.
NOTE: This patch changes the default behaviour of the PDF.js API to always attempt to recover as much data as possible, even when encountering errors during e.g. `getOperatorList`/`getTextContent`, which thus improve our handling of corrupt PDF files and allow the default viewer to handle errors slightly more gracefully.
In the event that an API consumer wishes to use the old behaviour, where we stop parsing as soon as an error is encountered, the `stopAtErrors` parameter can be set at `getDocument`.
Fixes, inasmuch it's possible since the PDF files are corrupt, e.g. issue 6342, issue 3795, and [bug 1130815](https://bugzilla.mozilla.org/show_bug.cgi?id=1130815) (and probably others too).
2017-02-19 22:03:08 +09:00
|
|
|
|
|
2020-07-15 07:17:27 +09:00
|
|
|
|
async parseMarkedContentProps(contentProperties, resources) {
|
|
|
|
|
let optionalContent;
|
|
|
|
|
if (isName(contentProperties)) {
|
|
|
|
|
const properties = resources.get("Properties");
|
|
|
|
|
optionalContent = properties.get(contentProperties.name);
|
|
|
|
|
} else if (isDict(contentProperties)) {
|
|
|
|
|
optionalContent = contentProperties;
|
|
|
|
|
} else {
|
|
|
|
|
throw new FormatError("Optional content properties malformed.");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
const optionalContentType = optionalContent.get("Type").name;
|
|
|
|
|
if (optionalContentType === "OCG") {
|
|
|
|
|
return {
|
|
|
|
|
type: optionalContentType,
|
|
|
|
|
id: optionalContent.objId,
|
|
|
|
|
};
|
|
|
|
|
} else if (optionalContentType === "OCMD") {
|
|
|
|
|
const optionalContentGroups = optionalContent.get("OCGs");
|
|
|
|
|
if (
|
|
|
|
|
Array.isArray(optionalContentGroups) ||
|
|
|
|
|
isDict(optionalContentGroups)
|
|
|
|
|
) {
|
|
|
|
|
const groupIds = [];
|
|
|
|
|
if (Array.isArray(optionalContentGroups)) {
|
|
|
|
|
optionalContent.get("OCGs").forEach(ocg => {
|
|
|
|
|
groupIds.push(ocg.toString());
|
|
|
|
|
});
|
|
|
|
|
} else {
|
|
|
|
|
// Dictionary, just use the obj id.
|
|
|
|
|
groupIds.push(optionalContentGroups.objId);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
let expression = null;
|
|
|
|
|
if (optionalContent.get("VE")) {
|
|
|
|
|
// TODO support visibility expression.
|
|
|
|
|
expression = true;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return {
|
|
|
|
|
type: optionalContentType,
|
|
|
|
|
ids: groupIds,
|
|
|
|
|
policy: isName(optionalContent.get("P"))
|
|
|
|
|
? optionalContent.get("P").name
|
|
|
|
|
: null,
|
|
|
|
|
expression,
|
|
|
|
|
};
|
|
|
|
|
} else if (isRef(optionalContentGroups)) {
|
|
|
|
|
return {
|
|
|
|
|
type: optionalContentType,
|
|
|
|
|
id: optionalContentGroups.toString(),
|
|
|
|
|
};
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
return null;
|
|
|
|
|
}
|
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
getOperatorList({
|
|
|
|
|
stream,
|
|
|
|
|
task,
|
|
|
|
|
resources,
|
|
|
|
|
operatorList,
|
|
|
|
|
initialState = null,
|
|
|
|
|
}) {
|
|
|
|
|
// Ensure that `resources`/`initialState` is correctly initialized,
|
|
|
|
|
// even if the provided parameter is e.g. `null`.
|
|
|
|
|
resources = resources || Dict.empty;
|
|
|
|
|
initialState = initialState || new EvalState();
|
2014-03-23 03:15:51 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (!operatorList) {
|
|
|
|
|
throw new Error('getOperatorList: missing "operatorList" parameter');
|
|
|
|
|
}
|
2017-09-17 20:35:18 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var self = this;
|
|
|
|
|
var xref = this.xref;
|
|
|
|
|
let parsingText = false;
|
|
|
|
|
const localImageCache = new LocalImageCache();
|
|
|
|
|
const localColorSpaceCache = new LocalColorSpaceCache();
|
2020-07-11 20:52:11 +09:00
|
|
|
|
const localGStateCache = new LocalGStateCache();
|
2020-10-09 00:33:23 +09:00
|
|
|
|
const localTilingPatternCache = new LocalTilingPatternCache();
|
Improve the *local* image caching in `PartialEvaluator.getOperatorList`
Currently the local `imageCache`, as used in `PartialEvaluator.getOperatorList`, will miss certain cases of repeated images because the caching is *only* done by name (usually using a format such as e.g. "Im0", "Im1", ...).
However, in some PDF documents the `/XObject` dictionaries many contain hundreds (or even thousands) of distinctly named images, despite them referring to only a handful of actual image objects (via the XRef table).
With these changes we'll now cache *local* images using both name and (where applicable) reference, thus improving re-usage of images resources even further.
This patch was tested using the PDF file from [bug 857031](https://bugzilla.mozilla.org/show_bug.cgi?id=857031), i.e. https://bug857031.bmoattachments.org/attachment.cgi?id=732270, with the following manifest file:
```
[
{ "id": "bug857031",
"file": "../web/pdfs/bug857031.pdf",
"md5": "",
"rounds": 250,
"lastPage": 1,
"type": "eq"
}
]
```
which gave the following results when comparing this patch against the `master` branch:
```
-- Grouped By browser, page, stat --
browser | page | stat | Count | Baseline(ms) | Current(ms) | +/- | % | Result(P<.05)
------- | ---- | ------------ | ----- | ------------ | ----------- | --- | ----- | -------------
firefox | 0 | Overall | 250 | 2749 | 2656 | -93 | -3.38 | faster
firefox | 0 | Page Request | 250 | 3 | 4 | 1 | 50.14 | slower
firefox | 0 | Rendering | 250 | 2746 | 2652 | -94 | -3.44 | faster
```
While this is certainly an improvement, since we now avoid re-parsing ~1000 images on the first page, all of the image resources are small enough that the total rendering time doesn't improve that much in this particular case.
In pathological cases, such as e.g. the PDF document in issue 4958, the improvements with this patch can be very significant. Looking for example at page 2, from issue 4958, the rendering time drops from ~60 seconds with `master` to ~30 seconds with this patch (obviously still slow, but it really showcases the potential of this patch nicely).
Finally, note that there's also potential for additional improvements by re-using `LocalImageCache` instances for e.g. /XObject data of the `Form`-type. However, given that recent changes in this area I purposely didn't want to complicate *this* patch more than necessary.
2020-05-23 20:55:31 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var xobjs = resources.get("XObject") || Dict.empty;
|
|
|
|
|
var patterns = resources.get("Pattern") || Dict.empty;
|
|
|
|
|
var stateManager = new StateManager(initialState);
|
|
|
|
|
var preprocessor = new EvaluatorPreprocessor(stream, xref, stateManager);
|
|
|
|
|
var timeSlotManager = new TimeSlotManager();
|
Improve the *local* image caching in `PartialEvaluator.getOperatorList`
Currently the local `imageCache`, as used in `PartialEvaluator.getOperatorList`, will miss certain cases of repeated images because the caching is *only* done by name (usually using a format such as e.g. "Im0", "Im1", ...).
However, in some PDF documents the `/XObject` dictionaries many contain hundreds (or even thousands) of distinctly named images, despite them referring to only a handful of actual image objects (via the XRef table).
With these changes we'll now cache *local* images using both name and (where applicable) reference, thus improving re-usage of images resources even further.
This patch was tested using the PDF file from [bug 857031](https://bugzilla.mozilla.org/show_bug.cgi?id=857031), i.e. https://bug857031.bmoattachments.org/attachment.cgi?id=732270, with the following manifest file:
```
[
{ "id": "bug857031",
"file": "../web/pdfs/bug857031.pdf",
"md5": "",
"rounds": 250,
"lastPage": 1,
"type": "eq"
}
]
```
which gave the following results when comparing this patch against the `master` branch:
```
-- Grouped By browser, page, stat --
browser | page | stat | Count | Baseline(ms) | Current(ms) | +/- | % | Result(P<.05)
------- | ---- | ------------ | ----- | ------------ | ----------- | --- | ----- | -------------
firefox | 0 | Overall | 250 | 2749 | 2656 | -93 | -3.38 | faster
firefox | 0 | Page Request | 250 | 3 | 4 | 1 | 50.14 | slower
firefox | 0 | Rendering | 250 | 2746 | 2652 | -94 | -3.44 | faster
```
While this is certainly an improvement, since we now avoid re-parsing ~1000 images on the first page, all of the image resources are small enough that the total rendering time doesn't improve that much in this particular case.
In pathological cases, such as e.g. the PDF document in issue 4958, the improvements with this patch can be very significant. Looking for example at page 2, from issue 4958, the rendering time drops from ~60 seconds with `master` to ~30 seconds with this patch (obviously still slow, but it really showcases the potential of this patch nicely).
Finally, note that there's also potential for additional improvements by re-using `LocalImageCache` instances for e.g. /XObject data of the `Form`-type. However, given that recent changes in this area I purposely didn't want to complicate *this* patch more than necessary.
2020-05-23 20:55:31 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
function closePendingRestoreOPS(argument) {
|
|
|
|
|
for (var i = 0, ii = preprocessor.savedStatesDepth; i < ii; i++) {
|
|
|
|
|
operatorList.addOp(OPS.restore, []);
|
|
|
|
|
}
|
|
|
|
|
}
|
Attempt to cache repeated images at the document, rather than the page, level (issue 11878)
Currently image resources, as opposed to e.g. font resources, are handled exclusively on a page-specific basis. Generally speaking this makes sense, since pages are separate from each other, however there's PDF documents where many (or even all) pages actually references exactly the same image resources (through the XRef table). Hence, in some cases, we're decoding the *same* images over and over for every page which is obviously slow and wasting both CPU and memory resources better used elsewhere.[1]
Obviously we cannot simply treat all image resources as-if they're used throughout the entire PDF document, since that would end up increasing memory usage too much.[2]
However, by introducing a `GlobalImageCache` in the worker we can track image resources that appear on more than one page. Hence we can switch image resources from being page-specific to being document-specific, once the image resource has been seen on more than a certain number of pages.
In many cases, such as e.g. the referenced issue, this patch will thus lead to reduced memory usage for image resources. Scrolling through all pages of the document, there's now only a few main-thread copies of the same image data, as opposed to one for each rendered page (i.e. there could theoretically be *twenty* copies of the image data).
While this obviously benefit both CPU and memory usage in this case, for *very* large image data this patch *may* possibly increase persistent main-thread memory usage a tiny bit. Thus to avoid negatively affecting memory usage too much in general, particularly on the main-thread, the `GlobalImageCache` will *only* cache a certain number of image resources at the document level and simply fallback to the default behaviour.
Unfortunately the asynchronous nature of the code, with ranged/streamed loading of data, actually makes all of this much more complicated than if all data could be assumed to be immediately available.[3]
*Please note:* The patch will lead to *small* movement in some existing test-cases, since we're now using the built-in PDF.js JPEG decoder more. This was done in order to simplify the overall implementation, especially on the main-thread, by limiting it to only the `OPS.paintImageXObject` operator.
---
[1] There's e.g. PDF documents that use the same image as background on all pages.
[2] Given that data stored in the `commonObjs`, on the main-thread, are only cleared manually through `PDFDocumentProxy.cleanup`. This as opposed to data stored in the `objs` of each page, which is automatically removed when the page is cleaned-up e.g. by being evicted from the cache in the default viewer.
[3] If the latter case were true, we could simply check for repeat images *before* parsing started and thus avoid handling *any* duplicate image resources.
2020-05-18 21:17:56 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return new Promise(function promiseBody(resolve, reject) {
|
|
|
|
|
const next = function (promise) {
|
|
|
|
|
Promise.all([promise, operatorList.ready]).then(function () {
|
|
|
|
|
try {
|
|
|
|
|
promiseBody(resolve, reject);
|
|
|
|
|
} catch (ex) {
|
|
|
|
|
reject(ex);
|
|
|
|
|
}
|
|
|
|
|
}, reject);
|
|
|
|
|
};
|
|
|
|
|
task.ensureNotTerminated();
|
|
|
|
|
timeSlotManager.reset();
|
|
|
|
|
var stop,
|
|
|
|
|
operation = {},
|
|
|
|
|
i,
|
|
|
|
|
ii,
|
2020-07-11 20:52:11 +09:00
|
|
|
|
cs,
|
|
|
|
|
name;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
while (!(stop = timeSlotManager.check())) {
|
|
|
|
|
// The arguments parsed by read() are used beyond this loop, so we
|
|
|
|
|
// cannot reuse the same array on each iteration. Therefore we pass
|
|
|
|
|
// in |null| as the initial value (see the comment on
|
|
|
|
|
// EvaluatorPreprocessor_read() for why).
|
|
|
|
|
operation.args = null;
|
|
|
|
|
if (!preprocessor.read(operation)) {
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
var args = operation.args;
|
|
|
|
|
var fn = operation.fn;
|
|
|
|
|
|
|
|
|
|
switch (fn | 0) {
|
|
|
|
|
case OPS.paintXObject:
|
|
|
|
|
// eagerly compile XForm objects
|
2020-07-11 20:52:11 +09:00
|
|
|
|
name = args[0].name;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (name) {
|
|
|
|
|
const localImage = localImageCache.getByName(name);
|
|
|
|
|
if (localImage) {
|
|
|
|
|
operatorList.addOp(localImage.fn, localImage.args);
|
|
|
|
|
args = null;
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
}
|
Improve the *local* image caching in `PartialEvaluator.getOperatorList`
Currently the local `imageCache`, as used in `PartialEvaluator.getOperatorList`, will miss certain cases of repeated images because the caching is *only* done by name (usually using a format such as e.g. "Im0", "Im1", ...).
However, in some PDF documents the `/XObject` dictionaries many contain hundreds (or even thousands) of distinctly named images, despite them referring to only a handful of actual image objects (via the XRef table).
With these changes we'll now cache *local* images using both name and (where applicable) reference, thus improving re-usage of images resources even further.
This patch was tested using the PDF file from [bug 857031](https://bugzilla.mozilla.org/show_bug.cgi?id=857031), i.e. https://bug857031.bmoattachments.org/attachment.cgi?id=732270, with the following manifest file:
```
[
{ "id": "bug857031",
"file": "../web/pdfs/bug857031.pdf",
"md5": "",
"rounds": 250,
"lastPage": 1,
"type": "eq"
}
]
```
which gave the following results when comparing this patch against the `master` branch:
```
-- Grouped By browser, page, stat --
browser | page | stat | Count | Baseline(ms) | Current(ms) | +/- | % | Result(P<.05)
------- | ---- | ------------ | ----- | ------------ | ----------- | --- | ----- | -------------
firefox | 0 | Overall | 250 | 2749 | 2656 | -93 | -3.38 | faster
firefox | 0 | Page Request | 250 | 3 | 4 | 1 | 50.14 | slower
firefox | 0 | Rendering | 250 | 2746 | 2652 | -94 | -3.44 | faster
```
While this is certainly an improvement, since we now avoid re-parsing ~1000 images on the first page, all of the image resources are small enough that the total rendering time doesn't improve that much in this particular case.
In pathological cases, such as e.g. the PDF document in issue 4958, the improvements with this patch can be very significant. Looking for example at page 2, from issue 4958, the rendering time drops from ~60 seconds with `master` to ~30 seconds with this patch (obviously still slow, but it really showcases the potential of this patch nicely).
Finally, note that there's also potential for additional improvements by re-using `LocalImageCache` instances for e.g. /XObject data of the `Form`-type. However, given that recent changes in this area I purposely didn't want to complicate *this* patch more than necessary.
2020-05-23 20:55:31 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
next(
|
|
|
|
|
new Promise(function (resolveXObject, rejectXObject) {
|
|
|
|
|
if (!name) {
|
|
|
|
|
throw new FormatError("XObject must be referred to by name.");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
let xobj = xobjs.getRaw(name);
|
|
|
|
|
if (xobj instanceof Ref) {
|
|
|
|
|
const localImage = localImageCache.getByRef(xobj);
|
|
|
|
|
if (localImage) {
|
|
|
|
|
operatorList.addOp(localImage.fn, localImage.args);
|
Attempt to cache repeated images at the document, rather than the page, level (issue 11878)
Currently image resources, as opposed to e.g. font resources, are handled exclusively on a page-specific basis. Generally speaking this makes sense, since pages are separate from each other, however there's PDF documents where many (or even all) pages actually references exactly the same image resources (through the XRef table). Hence, in some cases, we're decoding the *same* images over and over for every page which is obviously slow and wasting both CPU and memory resources better used elsewhere.[1]
Obviously we cannot simply treat all image resources as-if they're used throughout the entire PDF document, since that would end up increasing memory usage too much.[2]
However, by introducing a `GlobalImageCache` in the worker we can track image resources that appear on more than one page. Hence we can switch image resources from being page-specific to being document-specific, once the image resource has been seen on more than a certain number of pages.
In many cases, such as e.g. the referenced issue, this patch will thus lead to reduced memory usage for image resources. Scrolling through all pages of the document, there's now only a few main-thread copies of the same image data, as opposed to one for each rendered page (i.e. there could theoretically be *twenty* copies of the image data).
While this obviously benefit both CPU and memory usage in this case, for *very* large image data this patch *may* possibly increase persistent main-thread memory usage a tiny bit. Thus to avoid negatively affecting memory usage too much in general, particularly on the main-thread, the `GlobalImageCache` will *only* cache a certain number of image resources at the document level and simply fallback to the default behaviour.
Unfortunately the asynchronous nature of the code, with ranged/streamed loading of data, actually makes all of this much more complicated than if all data could be assumed to be immediately available.[3]
*Please note:* The patch will lead to *small* movement in some existing test-cases, since we're now using the built-in PDF.js JPEG decoder more. This was done in order to simplify the overall implementation, especially on the main-thread, by limiting it to only the `OPS.paintImageXObject` operator.
---
[1] There's e.g. PDF documents that use the same image as background on all pages.
[2] Given that data stored in the `commonObjs`, on the main-thread, are only cleared manually through `PDFDocumentProxy.cleanup`. This as opposed to data stored in the `objs` of each page, which is automatically removed when the page is cleaned-up e.g. by being evicted from the cache in the default viewer.
[3] If the latter case were true, we could simply check for repeat images *before* parsing started and thus avoid handling *any* duplicate image resources.
2020-05-18 21:17:56 +09:00
|
|
|
|
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
resolveXObject();
|
|
|
|
|
return;
|
|
|
|
|
}
|
2014-05-10 10:21:15 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
const globalImage = self.globalImageCache.getData(
|
|
|
|
|
xobj,
|
|
|
|
|
self.pageIndex
|
|
|
|
|
);
|
|
|
|
|
if (globalImage) {
|
|
|
|
|
operatorList.addDependency(globalImage.objId);
|
|
|
|
|
operatorList.addOp(globalImage.fn, globalImage.args);
|
2014-05-10 10:21:15 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
resolveXObject();
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
return;
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
|
|
|
|
|
xobj = xref.fetch(xobj);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (!isStream(xobj)) {
|
|
|
|
|
throw new FormatError("XObject should be a stream");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
const type = xobj.dict.get("Subtype");
|
|
|
|
|
if (!isName(type)) {
|
|
|
|
|
throw new FormatError("XObject should have a Name subtype");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (type.name === "Form") {
|
|
|
|
|
stateManager.save();
|
|
|
|
|
self
|
|
|
|
|
.buildFormXObject(
|
|
|
|
|
resources,
|
|
|
|
|
xobj,
|
|
|
|
|
null,
|
|
|
|
|
operatorList,
|
|
|
|
|
task,
|
|
|
|
|
stateManager.state.clone(),
|
|
|
|
|
localColorSpaceCache
|
|
|
|
|
)
|
|
|
|
|
.then(function () {
|
|
|
|
|
stateManager.restore();
|
|
|
|
|
resolveXObject();
|
|
|
|
|
}, rejectXObject);
|
|
|
|
|
return;
|
|
|
|
|
} else if (type.name === "Image") {
|
|
|
|
|
self
|
|
|
|
|
.buildPaintImageXObject({
|
|
|
|
|
resources,
|
|
|
|
|
image: xobj,
|
|
|
|
|
operatorList,
|
|
|
|
|
cacheKey: name,
|
|
|
|
|
localImageCache,
|
|
|
|
|
localColorSpaceCache,
|
|
|
|
|
})
|
|
|
|
|
.then(resolveXObject, rejectXObject);
|
|
|
|
|
return;
|
|
|
|
|
} else if (type.name === "PS") {
|
|
|
|
|
// PostScript XObjects are unused when viewing documents.
|
|
|
|
|
// See section 4.7.1 of Adobe's PDF reference.
|
|
|
|
|
info("Ignored XObject subtype PS");
|
|
|
|
|
} else {
|
|
|
|
|
throw new FormatError(
|
|
|
|
|
`Unhandled XObject subtype ${type.name}`
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
);
|
2014-05-10 10:21:15 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
resolveXObject();
|
|
|
|
|
}).catch(function (reason) {
|
|
|
|
|
if (reason instanceof AbortException) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
if (self.options.ignoreErrors) {
|
|
|
|
|
// Error(s) in the XObject -- sending unsupported feature
|
|
|
|
|
// notification and allow rendering to continue.
|
|
|
|
|
self.handler.send("UnsupportedFeature", {
|
|
|
|
|
featureId: UNSUPPORTED_FEATURES.errorXObject,
|
|
|
|
|
});
|
|
|
|
|
warn(`getOperatorList - ignoring XObject: "${reason}".`);
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
throw reason;
|
|
|
|
|
})
|
|
|
|
|
);
|
|
|
|
|
return;
|
|
|
|
|
case OPS.setFont:
|
|
|
|
|
var fontSize = args[1];
|
|
|
|
|
// eagerly collect all fonts
|
|
|
|
|
next(
|
|
|
|
|
self
|
|
|
|
|
.handleSetFont(
|
|
|
|
|
resources,
|
|
|
|
|
args,
|
|
|
|
|
null,
|
|
|
|
|
operatorList,
|
|
|
|
|
task,
|
|
|
|
|
stateManager.state
|
|
|
|
|
)
|
|
|
|
|
.then(function (loadedName) {
|
|
|
|
|
operatorList.addDependency(loadedName);
|
|
|
|
|
operatorList.addOp(OPS.setFont, [loadedName, fontSize]);
|
|
|
|
|
})
|
|
|
|
|
);
|
|
|
|
|
return;
|
|
|
|
|
case OPS.beginText:
|
|
|
|
|
parsingText = true;
|
|
|
|
|
break;
|
|
|
|
|
case OPS.endText:
|
|
|
|
|
parsingText = false;
|
|
|
|
|
break;
|
|
|
|
|
case OPS.endInlineImage:
|
|
|
|
|
var cacheKey = args[0].cacheKey;
|
|
|
|
|
if (cacheKey) {
|
|
|
|
|
const localImage = localImageCache.getByName(cacheKey);
|
|
|
|
|
if (localImage) {
|
|
|
|
|
operatorList.addOp(localImage.fn, localImage.args);
|
|
|
|
|
args = null;
|
2020-03-02 23:34:00 +09:00
|
|
|
|
continue;
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
next(
|
|
|
|
|
self.buildPaintImageXObject({
|
|
|
|
|
resources,
|
|
|
|
|
image: args[0],
|
|
|
|
|
isInline: true,
|
|
|
|
|
operatorList,
|
|
|
|
|
cacheKey,
|
|
|
|
|
localImageCache,
|
|
|
|
|
localColorSpaceCache,
|
|
|
|
|
})
|
|
|
|
|
);
|
|
|
|
|
return;
|
|
|
|
|
case OPS.showText:
|
|
|
|
|
if (!stateManager.state.font) {
|
|
|
|
|
self.ensureStateFont(stateManager.state);
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
args[0] = self.handleText(args[0], stateManager.state);
|
|
|
|
|
break;
|
|
|
|
|
case OPS.showSpacedText:
|
|
|
|
|
if (!stateManager.state.font) {
|
|
|
|
|
self.ensureStateFont(stateManager.state);
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
var arr = args[0];
|
|
|
|
|
var combinedGlyphs = [];
|
|
|
|
|
var arrLength = arr.length;
|
|
|
|
|
var state = stateManager.state;
|
|
|
|
|
for (i = 0; i < arrLength; ++i) {
|
|
|
|
|
var arrItem = arr[i];
|
|
|
|
|
if (isString(arrItem)) {
|
|
|
|
|
Array.prototype.push.apply(
|
|
|
|
|
combinedGlyphs,
|
|
|
|
|
self.handleText(arrItem, state)
|
|
|
|
|
);
|
|
|
|
|
} else if (isNum(arrItem)) {
|
|
|
|
|
combinedGlyphs.push(arrItem);
|
2020-03-02 23:34:00 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
args[0] = combinedGlyphs;
|
|
|
|
|
fn = OPS.showText;
|
|
|
|
|
break;
|
|
|
|
|
case OPS.nextLineShowText:
|
|
|
|
|
if (!stateManager.state.font) {
|
|
|
|
|
self.ensureStateFont(stateManager.state);
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
operatorList.addOp(OPS.nextLine);
|
|
|
|
|
args[0] = self.handleText(args[0], stateManager.state);
|
|
|
|
|
fn = OPS.showText;
|
|
|
|
|
break;
|
|
|
|
|
case OPS.nextLineSetSpacingShowText:
|
|
|
|
|
if (!stateManager.state.font) {
|
|
|
|
|
self.ensureStateFont(stateManager.state);
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
operatorList.addOp(OPS.nextLine);
|
|
|
|
|
operatorList.addOp(OPS.setWordSpacing, [args.shift()]);
|
|
|
|
|
operatorList.addOp(OPS.setCharSpacing, [args.shift()]);
|
|
|
|
|
args[0] = self.handleText(args[0], stateManager.state);
|
|
|
|
|
fn = OPS.showText;
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setTextRenderingMode:
|
|
|
|
|
stateManager.state.textRenderingMode = args[0];
|
|
|
|
|
break;
|
2014-05-22 02:47:42 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
case OPS.setFillColorSpace: {
|
|
|
|
|
const cachedColorSpace = ColorSpace.getCached(
|
|
|
|
|
args[0],
|
|
|
|
|
xref,
|
|
|
|
|
localColorSpaceCache
|
|
|
|
|
);
|
|
|
|
|
if (cachedColorSpace) {
|
|
|
|
|
stateManager.state.fillColorSpace = cachedColorSpace;
|
|
|
|
|
continue;
|
|
|
|
|
}
|
Add local caching of `ColorSpace`s, by name, in `PartialEvaluator.getOperatorList` (issue 2504)
By caching parsed `ColorSpace`s, we thus don't need to re-parse the same data over and over which saves CPU cycles *and* reduces peak memory usage. (Obviously persistent memory usage *may* increase a tiny bit, but since the caching is done per `PartialEvaluator.getOperatorList` invocation and given that `ColorSpace` instances generally hold very little data this shouldn't be much of an issue.)
Furthermore, by caching `ColorSpace`s we can also lookup the already parsed ones *synchronously* during the `OperatorList` building, instead of having to defer to the event loop/microtask queue since the parsing is done asynchronously (such that error handling is easier).
Possible future improvements:
- Cache/lookup parsed `ColorSpaces` used in `Pattern`s and `Image`s.
- Attempt to cache *local* `ColorSpace`s by reference as well, in addition to only by name, assuming that there's documents where that would be beneficial and that it's not too difficult to implement.
- Assuming there's documents that would benefit from it, also cache repeated `ColorSpace`s *globally* as well.
Given that we've never, until now, been doing *any* caching of parsed `ColorSpace`s and that even using a simple name-only *local* cache helps tremendously in pathological cases, I purposely decided against complicating the implementation too much initially.
Also, compared to parsing of `Image`s, simply creating a `ColorSpace` instance isn't that expensive (hence I'd be somewhat surprised if adding a *global* cache would help much).
---
This patch was tested using:
- The default `tracemonkey` PDF file, which was included mostly to show that "normal" documents aren't negatively affected by these changes.
- The PDF file from issue 2504, i.e. https://dl-ctlg.panasonic.com/jp/manual/sd/sd_rbm1000_0.pdf, where most pages will switch *thousands* of times between a handful of `ColorSpace`s.
with the following manifest file:
```
[
{ "id": "tracemonkey",
"file": "pdfs/tracemonkey.pdf",
"md5": "9a192d8b1a7dc652a19835f6f08098bd",
"rounds": 100,
"type": "eq"
},
{ "id": "issue2504",
"file": "../web/pdfs/issue2504.pdf",
"md5": "",
"rounds": 20,
"type": "eq"
}
]
```
which gave the following results when comparing this patch against the `master` branch:
- Overall
```
-- Grouped By browser, pdf, stat --
browser | pdf | stat | Count | Baseline(ms) | Current(ms) | +/- | % | Result(P<.05)
------- | ----------- | ------------ | ----- | ------------ | ----------- | ---- | ------ | -------------
firefox | issue2504 | Overall | 640 | 977 | 497 | -479 | -49.08 | faster
firefox | issue2504 | Page Request | 640 | 3 | 4 | 1 | 59.18 |
firefox | issue2504 | Rendering | 640 | 974 | 493 | -481 | -49.37 | faster
firefox | tracemonkey | Overall | 1400 | 116 | 111 | -5 | -4.43 |
firefox | tracemonkey | Page Request | 1400 | 2 | 2 | 0 | -2.86 |
firefox | tracemonkey | Rendering | 1400 | 114 | 109 | -5 | -4.47 |
```
- Page-specific
```
-- Grouped By browser, pdf, page, stat --
browser | pdf | page | stat | Count | Baseline(ms) | Current(ms) | +/- | % | Result(P<.05)
------- | ----------- | ---- | ------------ | ----- | ------------ | ----------- | ----- | ------- | -------------
firefox | issue2504 | 0 | Overall | 20 | 2295 | 1268 | -1027 | -44.76 | faster
firefox | issue2504 | 0 | Page Request | 20 | 6 | 7 | 1 | 15.32 |
firefox | issue2504 | 0 | Rendering | 20 | 2288 | 1260 | -1028 | -44.93 | faster
firefox | issue2504 | 1 | Overall | 20 | 3059 | 2806 | -252 | -8.25 | faster
firefox | issue2504 | 1 | Page Request | 20 | 11 | 14 | 3 | 23.25 | slower
firefox | issue2504 | 1 | Rendering | 20 | 3047 | 2792 | -255 | -8.37 | faster
firefox | issue2504 | 2 | Overall | 20 | 411 | 295 | -116 | -28.20 | faster
firefox | issue2504 | 2 | Page Request | 20 | 2 | 42 | 40 | 1897.62 |
firefox | issue2504 | 2 | Rendering | 20 | 409 | 253 | -156 | -38.09 | faster
firefox | issue2504 | 3 | Overall | 20 | 736 | 299 | -437 | -59.34 | faster
firefox | issue2504 | 3 | Page Request | 20 | 2 | 2 | 0 | 0.00 |
firefox | issue2504 | 3 | Rendering | 20 | 734 | 297 | -437 | -59.49 | faster
firefox | issue2504 | 4 | Overall | 20 | 356 | 458 | 102 | 28.63 |
firefox | issue2504 | 4 | Page Request | 20 | 1 | 2 | 1 | 57.14 | slower
firefox | issue2504 | 4 | Rendering | 20 | 354 | 455 | 101 | 28.53 |
firefox | issue2504 | 5 | Overall | 20 | 1381 | 765 | -616 | -44.59 | faster
firefox | issue2504 | 5 | Page Request | 20 | 3 | 5 | 2 | 50.00 | slower
firefox | issue2504 | 5 | Rendering | 20 | 1378 | 760 | -617 | -44.81 | faster
firefox | issue2504 | 6 | Overall | 20 | 757 | 299 | -459 | -60.57 | faster
firefox | issue2504 | 6 | Page Request | 20 | 2 | 5 | 3 | 150.00 | slower
firefox | issue2504 | 6 | Rendering | 20 | 755 | 294 | -462 | -61.11 | faster
firefox | issue2504 | 7 | Overall | 20 | 394 | 302 | -92 | -23.39 | faster
firefox | issue2504 | 7 | Page Request | 20 | 2 | 1 | -1 | -34.88 | faster
firefox | issue2504 | 7 | Rendering | 20 | 392 | 301 | -91 | -23.32 | faster
firefox | issue2504 | 8 | Overall | 20 | 2875 | 979 | -1896 | -65.95 | faster
firefox | issue2504 | 8 | Page Request | 20 | 1 | 2 | 0 | 11.11 |
firefox | issue2504 | 8 | Rendering | 20 | 2874 | 978 | -1896 | -65.99 | faster
firefox | issue2504 | 9 | Overall | 20 | 700 | 332 | -368 | -52.60 | faster
firefox | issue2504 | 9 | Page Request | 20 | 3 | 2 | 0 | -4.00 |
firefox | issue2504 | 9 | Rendering | 20 | 698 | 329 | -368 | -52.78 | faster
firefox | issue2504 | 10 | Overall | 20 | 3296 | 926 | -2370 | -71.91 | faster
firefox | issue2504 | 10 | Page Request | 20 | 2 | 2 | 0 | -18.75 |
firefox | issue2504 | 10 | Rendering | 20 | 3293 | 924 | -2370 | -71.96 | faster
firefox | issue2504 | 11 | Overall | 20 | 524 | 197 | -327 | -62.34 | faster
firefox | issue2504 | 11 | Page Request | 20 | 2 | 3 | 1 | 58.54 |
firefox | issue2504 | 11 | Rendering | 20 | 522 | 194 | -328 | -62.81 | faster
firefox | issue2504 | 12 | Overall | 20 | 752 | 369 | -384 | -50.98 | faster
firefox | issue2504 | 12 | Page Request | 20 | 3 | 2 | -1 | -36.51 | faster
firefox | issue2504 | 12 | Rendering | 20 | 749 | 367 | -382 | -51.05 | faster
firefox | issue2504 | 13 | Overall | 20 | 679 | 487 | -193 | -28.38 | faster
firefox | issue2504 | 13 | Page Request | 20 | 4 | 2 | -2 | -48.68 | faster
firefox | issue2504 | 13 | Rendering | 20 | 676 | 485 | -191 | -28.28 | faster
firefox | issue2504 | 14 | Overall | 20 | 474 | 283 | -191 | -40.26 | faster
firefox | issue2504 | 14 | Page Request | 20 | 2 | 4 | 2 | 78.57 |
firefox | issue2504 | 14 | Rendering | 20 | 471 | 279 | -192 | -40.79 | faster
firefox | issue2504 | 15 | Overall | 20 | 860 | 618 | -241 | -28.05 | faster
firefox | issue2504 | 15 | Page Request | 20 | 2 | 3 | 0 | 10.87 |
firefox | issue2504 | 15 | Rendering | 20 | 857 | 616 | -241 | -28.15 | faster
firefox | issue2504 | 16 | Overall | 20 | 389 | 243 | -147 | -37.71 | faster
firefox | issue2504 | 16 | Page Request | 20 | 2 | 2 | 0 | 2.33 |
firefox | issue2504 | 16 | Rendering | 20 | 387 | 240 | -147 | -37.94 | faster
firefox | issue2504 | 17 | Overall | 20 | 1484 | 672 | -812 | -54.70 | faster
firefox | issue2504 | 17 | Page Request | 20 | 2 | 3 | 1 | 37.21 |
firefox | issue2504 | 17 | Rendering | 20 | 1482 | 669 | -812 | -54.84 | faster
firefox | issue2504 | 18 | Overall | 20 | 575 | 252 | -323 | -56.12 | faster
firefox | issue2504 | 18 | Page Request | 20 | 2 | 2 | 0 | -16.22 |
firefox | issue2504 | 18 | Rendering | 20 | 573 | 251 | -322 | -56.24 | faster
firefox | issue2504 | 19 | Overall | 20 | 517 | 227 | -290 | -56.08 | faster
firefox | issue2504 | 19 | Page Request | 20 | 2 | 2 | 0 | 21.62 |
firefox | issue2504 | 19 | Rendering | 20 | 515 | 225 | -290 | -56.37 | faster
firefox | issue2504 | 20 | Overall | 20 | 668 | 670 | 2 | 0.31 |
firefox | issue2504 | 20 | Page Request | 20 | 4 | 2 | -1 | -34.29 |
firefox | issue2504 | 20 | Rendering | 20 | 664 | 667 | 3 | 0.49 |
firefox | issue2504 | 21 | Overall | 20 | 486 | 309 | -177 | -36.44 | faster
firefox | issue2504 | 21 | Page Request | 20 | 2 | 2 | 0 | 16.13 |
firefox | issue2504 | 21 | Rendering | 20 | 484 | 307 | -177 | -36.60 | faster
firefox | issue2504 | 22 | Overall | 20 | 543 | 267 | -276 | -50.85 | faster
firefox | issue2504 | 22 | Page Request | 20 | 2 | 2 | 0 | 10.26 |
firefox | issue2504 | 22 | Rendering | 20 | 541 | 265 | -276 | -51.07 | faster
firefox | issue2504 | 23 | Overall | 20 | 3246 | 871 | -2375 | -73.17 | faster
firefox | issue2504 | 23 | Page Request | 20 | 2 | 3 | 1 | 37.21 |
firefox | issue2504 | 23 | Rendering | 20 | 3243 | 868 | -2376 | -73.25 | faster
firefox | issue2504 | 24 | Overall | 20 | 379 | 156 | -223 | -58.83 | faster
firefox | issue2504 | 24 | Page Request | 20 | 2 | 2 | 0 | -2.86 |
firefox | issue2504 | 24 | Rendering | 20 | 378 | 154 | -223 | -59.10 | faster
firefox | issue2504 | 25 | Overall | 20 | 176 | 127 | -50 | -28.19 | faster
firefox | issue2504 | 25 | Page Request | 20 | 2 | 1 | 0 | -15.63 |
firefox | issue2504 | 25 | Rendering | 20 | 175 | 125 | -49 | -28.31 | faster
firefox | issue2504 | 26 | Overall | 20 | 181 | 108 | -74 | -40.67 | faster
firefox | issue2504 | 26 | Page Request | 20 | 3 | 2 | -1 | -39.13 | faster
firefox | issue2504 | 26 | Rendering | 20 | 178 | 105 | -72 | -40.69 | faster
firefox | issue2504 | 27 | Overall | 20 | 208 | 104 | -104 | -49.92 | faster
firefox | issue2504 | 27 | Page Request | 20 | 2 | 2 | 1 | 48.39 |
firefox | issue2504 | 27 | Rendering | 20 | 206 | 102 | -104 | -50.64 | faster
firefox | issue2504 | 28 | Overall | 20 | 241 | 111 | -131 | -54.16 | faster
firefox | issue2504 | 28 | Page Request | 20 | 2 | 2 | -1 | -33.33 |
firefox | issue2504 | 28 | Rendering | 20 | 239 | 109 | -130 | -54.39 | faster
firefox | issue2504 | 29 | Overall | 20 | 321 | 196 | -125 | -39.05 | faster
firefox | issue2504 | 29 | Page Request | 20 | 1 | 2 | 0 | 17.86 |
firefox | issue2504 | 29 | Rendering | 20 | 319 | 194 | -126 | -39.35 | faster
firefox | issue2504 | 30 | Overall | 20 | 651 | 271 | -380 | -58.41 | faster
firefox | issue2504 | 30 | Page Request | 20 | 1 | 2 | 1 | 50.00 |
firefox | issue2504 | 30 | Rendering | 20 | 649 | 269 | -381 | -58.60 | faster
firefox | issue2504 | 31 | Overall | 20 | 1635 | 647 | -988 | -60.42 | faster
firefox | issue2504 | 31 | Page Request | 20 | 1 | 2 | 0 | 30.43 |
firefox | issue2504 | 31 | Rendering | 20 | 1634 | 645 | -988 | -60.49 | faster
firefox | tracemonkey | 0 | Overall | 100 | 51 | 51 | 0 | 0.02 |
firefox | tracemonkey | 0 | Page Request | 100 | 1 | 1 | 0 | -4.76 |
firefox | tracemonkey | 0 | Rendering | 100 | 50 | 50 | 0 | 0.12 |
firefox | tracemonkey | 1 | Overall | 100 | 97 | 91 | -5 | -5.52 | faster
firefox | tracemonkey | 1 | Page Request | 100 | 3 | 3 | 0 | -1.32 |
firefox | tracemonkey | 1 | Rendering | 100 | 94 | 88 | -5 | -5.73 | faster
firefox | tracemonkey | 2 | Overall | 100 | 40 | 40 | 0 | 0.50 |
firefox | tracemonkey | 2 | Page Request | 100 | 1 | 1 | 0 | 3.16 |
firefox | tracemonkey | 2 | Rendering | 100 | 39 | 39 | 0 | 0.54 |
firefox | tracemonkey | 3 | Overall | 100 | 62 | 62 | -1 | -0.94 |
firefox | tracemonkey | 3 | Page Request | 100 | 1 | 1 | 0 | 17.05 |
firefox | tracemonkey | 3 | Rendering | 100 | 61 | 61 | -1 | -1.11 |
firefox | tracemonkey | 4 | Overall | 100 | 56 | 58 | 2 | 3.41 |
firefox | tracemonkey | 4 | Page Request | 100 | 1 | 1 | 0 | 15.31 |
firefox | tracemonkey | 4 | Rendering | 100 | 55 | 57 | 2 | 3.23 |
firefox | tracemonkey | 5 | Overall | 100 | 73 | 71 | -2 | -2.28 |
firefox | tracemonkey | 5 | Page Request | 100 | 2 | 2 | 0 | 12.20 |
firefox | tracemonkey | 5 | Rendering | 100 | 71 | 69 | -2 | -2.69 |
firefox | tracemonkey | 6 | Overall | 100 | 85 | 69 | -16 | -18.73 | faster
firefox | tracemonkey | 6 | Page Request | 100 | 2 | 2 | 0 | -9.90 |
firefox | tracemonkey | 6 | Rendering | 100 | 83 | 67 | -16 | -18.97 | faster
firefox | tracemonkey | 7 | Overall | 100 | 65 | 64 | 0 | -0.37 |
firefox | tracemonkey | 7 | Page Request | 100 | 1 | 1 | 0 | -11.94 |
firefox | tracemonkey | 7 | Rendering | 100 | 63 | 63 | 0 | -0.05 |
firefox | tracemonkey | 8 | Overall | 100 | 53 | 54 | 1 | 2.04 |
firefox | tracemonkey | 8 | Page Request | 100 | 1 | 1 | 0 | 17.02 |
firefox | tracemonkey | 8 | Rendering | 100 | 52 | 53 | 1 | 1.82 |
firefox | tracemonkey | 9 | Overall | 100 | 79 | 73 | -6 | -7.86 | faster
firefox | tracemonkey | 9 | Page Request | 100 | 2 | 2 | 0 | -15.14 |
firefox | tracemonkey | 9 | Rendering | 100 | 77 | 71 | -6 | -7.86 | faster
firefox | tracemonkey | 10 | Overall | 100 | 545 | 519 | -27 | -4.86 | faster
firefox | tracemonkey | 10 | Page Request | 100 | 14 | 13 | 0 | -3.56 |
firefox | tracemonkey | 10 | Rendering | 100 | 532 | 506 | -26 | -4.90 | faster
firefox | tracemonkey | 11 | Overall | 100 | 42 | 41 | -1 | -2.50 |
firefox | tracemonkey | 11 | Page Request | 100 | 1 | 1 | 0 | -27.42 | faster
firefox | tracemonkey | 11 | Rendering | 100 | 41 | 40 | -1 | -1.75 |
firefox | tracemonkey | 12 | Overall | 100 | 350 | 332 | -18 | -5.16 | faster
firefox | tracemonkey | 12 | Page Request | 100 | 3 | 3 | 0 | -5.17 |
firefox | tracemonkey | 12 | Rendering | 100 | 347 | 329 | -18 | -5.15 | faster
firefox | tracemonkey | 13 | Overall | 100 | 31 | 31 | 0 | 0.52 |
firefox | tracemonkey | 13 | Page Request | 100 | 1 | 1 | 0 | 4.95 |
firefox | tracemonkey | 13 | Rendering | 100 | 30 | 30 | 0 | 0.20 |
```
2020-06-13 21:12:40 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
next(
|
|
|
|
|
self
|
|
|
|
|
.parseColorSpace({
|
|
|
|
|
cs: args[0],
|
|
|
|
|
resources,
|
|
|
|
|
localColorSpaceCache,
|
|
|
|
|
})
|
|
|
|
|
.then(function (colorSpace) {
|
|
|
|
|
if (colorSpace) {
|
|
|
|
|
stateManager.state.fillColorSpace = colorSpace;
|
|
|
|
|
}
|
|
|
|
|
})
|
|
|
|
|
);
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
case OPS.setStrokeColorSpace: {
|
|
|
|
|
const cachedColorSpace = ColorSpace.getCached(
|
|
|
|
|
args[0],
|
|
|
|
|
xref,
|
|
|
|
|
localColorSpaceCache
|
|
|
|
|
);
|
|
|
|
|
if (cachedColorSpace) {
|
|
|
|
|
stateManager.state.strokeColorSpace = cachedColorSpace;
|
|
|
|
|
continue;
|
Add local caching of `ColorSpace`s, by name, in `PartialEvaluator.getOperatorList` (issue 2504)
By caching parsed `ColorSpace`s, we thus don't need to re-parse the same data over and over which saves CPU cycles *and* reduces peak memory usage. (Obviously persistent memory usage *may* increase a tiny bit, but since the caching is done per `PartialEvaluator.getOperatorList` invocation and given that `ColorSpace` instances generally hold very little data this shouldn't be much of an issue.)
Furthermore, by caching `ColorSpace`s we can also lookup the already parsed ones *synchronously* during the `OperatorList` building, instead of having to defer to the event loop/microtask queue since the parsing is done asynchronously (such that error handling is easier).
Possible future improvements:
- Cache/lookup parsed `ColorSpaces` used in `Pattern`s and `Image`s.
- Attempt to cache *local* `ColorSpace`s by reference as well, in addition to only by name, assuming that there's documents where that would be beneficial and that it's not too difficult to implement.
- Assuming there's documents that would benefit from it, also cache repeated `ColorSpace`s *globally* as well.
Given that we've never, until now, been doing *any* caching of parsed `ColorSpace`s and that even using a simple name-only *local* cache helps tremendously in pathological cases, I purposely decided against complicating the implementation too much initially.
Also, compared to parsing of `Image`s, simply creating a `ColorSpace` instance isn't that expensive (hence I'd be somewhat surprised if adding a *global* cache would help much).
---
This patch was tested using:
- The default `tracemonkey` PDF file, which was included mostly to show that "normal" documents aren't negatively affected by these changes.
- The PDF file from issue 2504, i.e. https://dl-ctlg.panasonic.com/jp/manual/sd/sd_rbm1000_0.pdf, where most pages will switch *thousands* of times between a handful of `ColorSpace`s.
with the following manifest file:
```
[
{ "id": "tracemonkey",
"file": "pdfs/tracemonkey.pdf",
"md5": "9a192d8b1a7dc652a19835f6f08098bd",
"rounds": 100,
"type": "eq"
},
{ "id": "issue2504",
"file": "../web/pdfs/issue2504.pdf",
"md5": "",
"rounds": 20,
"type": "eq"
}
]
```
which gave the following results when comparing this patch against the `master` branch:
- Overall
```
-- Grouped By browser, pdf, stat --
browser | pdf | stat | Count | Baseline(ms) | Current(ms) | +/- | % | Result(P<.05)
------- | ----------- | ------------ | ----- | ------------ | ----------- | ---- | ------ | -------------
firefox | issue2504 | Overall | 640 | 977 | 497 | -479 | -49.08 | faster
firefox | issue2504 | Page Request | 640 | 3 | 4 | 1 | 59.18 |
firefox | issue2504 | Rendering | 640 | 974 | 493 | -481 | -49.37 | faster
firefox | tracemonkey | Overall | 1400 | 116 | 111 | -5 | -4.43 |
firefox | tracemonkey | Page Request | 1400 | 2 | 2 | 0 | -2.86 |
firefox | tracemonkey | Rendering | 1400 | 114 | 109 | -5 | -4.47 |
```
- Page-specific
```
-- Grouped By browser, pdf, page, stat --
browser | pdf | page | stat | Count | Baseline(ms) | Current(ms) | +/- | % | Result(P<.05)
------- | ----------- | ---- | ------------ | ----- | ------------ | ----------- | ----- | ------- | -------------
firefox | issue2504 | 0 | Overall | 20 | 2295 | 1268 | -1027 | -44.76 | faster
firefox | issue2504 | 0 | Page Request | 20 | 6 | 7 | 1 | 15.32 |
firefox | issue2504 | 0 | Rendering | 20 | 2288 | 1260 | -1028 | -44.93 | faster
firefox | issue2504 | 1 | Overall | 20 | 3059 | 2806 | -252 | -8.25 | faster
firefox | issue2504 | 1 | Page Request | 20 | 11 | 14 | 3 | 23.25 | slower
firefox | issue2504 | 1 | Rendering | 20 | 3047 | 2792 | -255 | -8.37 | faster
firefox | issue2504 | 2 | Overall | 20 | 411 | 295 | -116 | -28.20 | faster
firefox | issue2504 | 2 | Page Request | 20 | 2 | 42 | 40 | 1897.62 |
firefox | issue2504 | 2 | Rendering | 20 | 409 | 253 | -156 | -38.09 | faster
firefox | issue2504 | 3 | Overall | 20 | 736 | 299 | -437 | -59.34 | faster
firefox | issue2504 | 3 | Page Request | 20 | 2 | 2 | 0 | 0.00 |
firefox | issue2504 | 3 | Rendering | 20 | 734 | 297 | -437 | -59.49 | faster
firefox | issue2504 | 4 | Overall | 20 | 356 | 458 | 102 | 28.63 |
firefox | issue2504 | 4 | Page Request | 20 | 1 | 2 | 1 | 57.14 | slower
firefox | issue2504 | 4 | Rendering | 20 | 354 | 455 | 101 | 28.53 |
firefox | issue2504 | 5 | Overall | 20 | 1381 | 765 | -616 | -44.59 | faster
firefox | issue2504 | 5 | Page Request | 20 | 3 | 5 | 2 | 50.00 | slower
firefox | issue2504 | 5 | Rendering | 20 | 1378 | 760 | -617 | -44.81 | faster
firefox | issue2504 | 6 | Overall | 20 | 757 | 299 | -459 | -60.57 | faster
firefox | issue2504 | 6 | Page Request | 20 | 2 | 5 | 3 | 150.00 | slower
firefox | issue2504 | 6 | Rendering | 20 | 755 | 294 | -462 | -61.11 | faster
firefox | issue2504 | 7 | Overall | 20 | 394 | 302 | -92 | -23.39 | faster
firefox | issue2504 | 7 | Page Request | 20 | 2 | 1 | -1 | -34.88 | faster
firefox | issue2504 | 7 | Rendering | 20 | 392 | 301 | -91 | -23.32 | faster
firefox | issue2504 | 8 | Overall | 20 | 2875 | 979 | -1896 | -65.95 | faster
firefox | issue2504 | 8 | Page Request | 20 | 1 | 2 | 0 | 11.11 |
firefox | issue2504 | 8 | Rendering | 20 | 2874 | 978 | -1896 | -65.99 | faster
firefox | issue2504 | 9 | Overall | 20 | 700 | 332 | -368 | -52.60 | faster
firefox | issue2504 | 9 | Page Request | 20 | 3 | 2 | 0 | -4.00 |
firefox | issue2504 | 9 | Rendering | 20 | 698 | 329 | -368 | -52.78 | faster
firefox | issue2504 | 10 | Overall | 20 | 3296 | 926 | -2370 | -71.91 | faster
firefox | issue2504 | 10 | Page Request | 20 | 2 | 2 | 0 | -18.75 |
firefox | issue2504 | 10 | Rendering | 20 | 3293 | 924 | -2370 | -71.96 | faster
firefox | issue2504 | 11 | Overall | 20 | 524 | 197 | -327 | -62.34 | faster
firefox | issue2504 | 11 | Page Request | 20 | 2 | 3 | 1 | 58.54 |
firefox | issue2504 | 11 | Rendering | 20 | 522 | 194 | -328 | -62.81 | faster
firefox | issue2504 | 12 | Overall | 20 | 752 | 369 | -384 | -50.98 | faster
firefox | issue2504 | 12 | Page Request | 20 | 3 | 2 | -1 | -36.51 | faster
firefox | issue2504 | 12 | Rendering | 20 | 749 | 367 | -382 | -51.05 | faster
firefox | issue2504 | 13 | Overall | 20 | 679 | 487 | -193 | -28.38 | faster
firefox | issue2504 | 13 | Page Request | 20 | 4 | 2 | -2 | -48.68 | faster
firefox | issue2504 | 13 | Rendering | 20 | 676 | 485 | -191 | -28.28 | faster
firefox | issue2504 | 14 | Overall | 20 | 474 | 283 | -191 | -40.26 | faster
firefox | issue2504 | 14 | Page Request | 20 | 2 | 4 | 2 | 78.57 |
firefox | issue2504 | 14 | Rendering | 20 | 471 | 279 | -192 | -40.79 | faster
firefox | issue2504 | 15 | Overall | 20 | 860 | 618 | -241 | -28.05 | faster
firefox | issue2504 | 15 | Page Request | 20 | 2 | 3 | 0 | 10.87 |
firefox | issue2504 | 15 | Rendering | 20 | 857 | 616 | -241 | -28.15 | faster
firefox | issue2504 | 16 | Overall | 20 | 389 | 243 | -147 | -37.71 | faster
firefox | issue2504 | 16 | Page Request | 20 | 2 | 2 | 0 | 2.33 |
firefox | issue2504 | 16 | Rendering | 20 | 387 | 240 | -147 | -37.94 | faster
firefox | issue2504 | 17 | Overall | 20 | 1484 | 672 | -812 | -54.70 | faster
firefox | issue2504 | 17 | Page Request | 20 | 2 | 3 | 1 | 37.21 |
firefox | issue2504 | 17 | Rendering | 20 | 1482 | 669 | -812 | -54.84 | faster
firefox | issue2504 | 18 | Overall | 20 | 575 | 252 | -323 | -56.12 | faster
firefox | issue2504 | 18 | Page Request | 20 | 2 | 2 | 0 | -16.22 |
firefox | issue2504 | 18 | Rendering | 20 | 573 | 251 | -322 | -56.24 | faster
firefox | issue2504 | 19 | Overall | 20 | 517 | 227 | -290 | -56.08 | faster
firefox | issue2504 | 19 | Page Request | 20 | 2 | 2 | 0 | 21.62 |
firefox | issue2504 | 19 | Rendering | 20 | 515 | 225 | -290 | -56.37 | faster
firefox | issue2504 | 20 | Overall | 20 | 668 | 670 | 2 | 0.31 |
firefox | issue2504 | 20 | Page Request | 20 | 4 | 2 | -1 | -34.29 |
firefox | issue2504 | 20 | Rendering | 20 | 664 | 667 | 3 | 0.49 |
firefox | issue2504 | 21 | Overall | 20 | 486 | 309 | -177 | -36.44 | faster
firefox | issue2504 | 21 | Page Request | 20 | 2 | 2 | 0 | 16.13 |
firefox | issue2504 | 21 | Rendering | 20 | 484 | 307 | -177 | -36.60 | faster
firefox | issue2504 | 22 | Overall | 20 | 543 | 267 | -276 | -50.85 | faster
firefox | issue2504 | 22 | Page Request | 20 | 2 | 2 | 0 | 10.26 |
firefox | issue2504 | 22 | Rendering | 20 | 541 | 265 | -276 | -51.07 | faster
firefox | issue2504 | 23 | Overall | 20 | 3246 | 871 | -2375 | -73.17 | faster
firefox | issue2504 | 23 | Page Request | 20 | 2 | 3 | 1 | 37.21 |
firefox | issue2504 | 23 | Rendering | 20 | 3243 | 868 | -2376 | -73.25 | faster
firefox | issue2504 | 24 | Overall | 20 | 379 | 156 | -223 | -58.83 | faster
firefox | issue2504 | 24 | Page Request | 20 | 2 | 2 | 0 | -2.86 |
firefox | issue2504 | 24 | Rendering | 20 | 378 | 154 | -223 | -59.10 | faster
firefox | issue2504 | 25 | Overall | 20 | 176 | 127 | -50 | -28.19 | faster
firefox | issue2504 | 25 | Page Request | 20 | 2 | 1 | 0 | -15.63 |
firefox | issue2504 | 25 | Rendering | 20 | 175 | 125 | -49 | -28.31 | faster
firefox | issue2504 | 26 | Overall | 20 | 181 | 108 | -74 | -40.67 | faster
firefox | issue2504 | 26 | Page Request | 20 | 3 | 2 | -1 | -39.13 | faster
firefox | issue2504 | 26 | Rendering | 20 | 178 | 105 | -72 | -40.69 | faster
firefox | issue2504 | 27 | Overall | 20 | 208 | 104 | -104 | -49.92 | faster
firefox | issue2504 | 27 | Page Request | 20 | 2 | 2 | 1 | 48.39 |
firefox | issue2504 | 27 | Rendering | 20 | 206 | 102 | -104 | -50.64 | faster
firefox | issue2504 | 28 | Overall | 20 | 241 | 111 | -131 | -54.16 | faster
firefox | issue2504 | 28 | Page Request | 20 | 2 | 2 | -1 | -33.33 |
firefox | issue2504 | 28 | Rendering | 20 | 239 | 109 | -130 | -54.39 | faster
firefox | issue2504 | 29 | Overall | 20 | 321 | 196 | -125 | -39.05 | faster
firefox | issue2504 | 29 | Page Request | 20 | 1 | 2 | 0 | 17.86 |
firefox | issue2504 | 29 | Rendering | 20 | 319 | 194 | -126 | -39.35 | faster
firefox | issue2504 | 30 | Overall | 20 | 651 | 271 | -380 | -58.41 | faster
firefox | issue2504 | 30 | Page Request | 20 | 1 | 2 | 1 | 50.00 |
firefox | issue2504 | 30 | Rendering | 20 | 649 | 269 | -381 | -58.60 | faster
firefox | issue2504 | 31 | Overall | 20 | 1635 | 647 | -988 | -60.42 | faster
firefox | issue2504 | 31 | Page Request | 20 | 1 | 2 | 0 | 30.43 |
firefox | issue2504 | 31 | Rendering | 20 | 1634 | 645 | -988 | -60.49 | faster
firefox | tracemonkey | 0 | Overall | 100 | 51 | 51 | 0 | 0.02 |
firefox | tracemonkey | 0 | Page Request | 100 | 1 | 1 | 0 | -4.76 |
firefox | tracemonkey | 0 | Rendering | 100 | 50 | 50 | 0 | 0.12 |
firefox | tracemonkey | 1 | Overall | 100 | 97 | 91 | -5 | -5.52 | faster
firefox | tracemonkey | 1 | Page Request | 100 | 3 | 3 | 0 | -1.32 |
firefox | tracemonkey | 1 | Rendering | 100 | 94 | 88 | -5 | -5.73 | faster
firefox | tracemonkey | 2 | Overall | 100 | 40 | 40 | 0 | 0.50 |
firefox | tracemonkey | 2 | Page Request | 100 | 1 | 1 | 0 | 3.16 |
firefox | tracemonkey | 2 | Rendering | 100 | 39 | 39 | 0 | 0.54 |
firefox | tracemonkey | 3 | Overall | 100 | 62 | 62 | -1 | -0.94 |
firefox | tracemonkey | 3 | Page Request | 100 | 1 | 1 | 0 | 17.05 |
firefox | tracemonkey | 3 | Rendering | 100 | 61 | 61 | -1 | -1.11 |
firefox | tracemonkey | 4 | Overall | 100 | 56 | 58 | 2 | 3.41 |
firefox | tracemonkey | 4 | Page Request | 100 | 1 | 1 | 0 | 15.31 |
firefox | tracemonkey | 4 | Rendering | 100 | 55 | 57 | 2 | 3.23 |
firefox | tracemonkey | 5 | Overall | 100 | 73 | 71 | -2 | -2.28 |
firefox | tracemonkey | 5 | Page Request | 100 | 2 | 2 | 0 | 12.20 |
firefox | tracemonkey | 5 | Rendering | 100 | 71 | 69 | -2 | -2.69 |
firefox | tracemonkey | 6 | Overall | 100 | 85 | 69 | -16 | -18.73 | faster
firefox | tracemonkey | 6 | Page Request | 100 | 2 | 2 | 0 | -9.90 |
firefox | tracemonkey | 6 | Rendering | 100 | 83 | 67 | -16 | -18.97 | faster
firefox | tracemonkey | 7 | Overall | 100 | 65 | 64 | 0 | -0.37 |
firefox | tracemonkey | 7 | Page Request | 100 | 1 | 1 | 0 | -11.94 |
firefox | tracemonkey | 7 | Rendering | 100 | 63 | 63 | 0 | -0.05 |
firefox | tracemonkey | 8 | Overall | 100 | 53 | 54 | 1 | 2.04 |
firefox | tracemonkey | 8 | Page Request | 100 | 1 | 1 | 0 | 17.02 |
firefox | tracemonkey | 8 | Rendering | 100 | 52 | 53 | 1 | 1.82 |
firefox | tracemonkey | 9 | Overall | 100 | 79 | 73 | -6 | -7.86 | faster
firefox | tracemonkey | 9 | Page Request | 100 | 2 | 2 | 0 | -15.14 |
firefox | tracemonkey | 9 | Rendering | 100 | 77 | 71 | -6 | -7.86 | faster
firefox | tracemonkey | 10 | Overall | 100 | 545 | 519 | -27 | -4.86 | faster
firefox | tracemonkey | 10 | Page Request | 100 | 14 | 13 | 0 | -3.56 |
firefox | tracemonkey | 10 | Rendering | 100 | 532 | 506 | -26 | -4.90 | faster
firefox | tracemonkey | 11 | Overall | 100 | 42 | 41 | -1 | -2.50 |
firefox | tracemonkey | 11 | Page Request | 100 | 1 | 1 | 0 | -27.42 | faster
firefox | tracemonkey | 11 | Rendering | 100 | 41 | 40 | -1 | -1.75 |
firefox | tracemonkey | 12 | Overall | 100 | 350 | 332 | -18 | -5.16 | faster
firefox | tracemonkey | 12 | Page Request | 100 | 3 | 3 | 0 | -5.17 |
firefox | tracemonkey | 12 | Rendering | 100 | 347 | 329 | -18 | -5.15 | faster
firefox | tracemonkey | 13 | Overall | 100 | 31 | 31 | 0 | 0.52 |
firefox | tracemonkey | 13 | Page Request | 100 | 1 | 1 | 0 | 4.95 |
firefox | tracemonkey | 13 | Rendering | 100 | 30 | 30 | 0 | 0.20 |
```
2020-06-13 21:12:40 +09:00
|
|
|
|
}
|
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
next(
|
|
|
|
|
self
|
|
|
|
|
.parseColorSpace({
|
|
|
|
|
cs: args[0],
|
|
|
|
|
resources,
|
|
|
|
|
localColorSpaceCache,
|
|
|
|
|
})
|
|
|
|
|
.then(function (colorSpace) {
|
|
|
|
|
if (colorSpace) {
|
|
|
|
|
stateManager.state.strokeColorSpace = colorSpace;
|
|
|
|
|
}
|
|
|
|
|
})
|
|
|
|
|
);
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
case OPS.setFillColor:
|
|
|
|
|
cs = stateManager.state.fillColorSpace;
|
|
|
|
|
args = cs.getRgb(args, 0);
|
|
|
|
|
fn = OPS.setFillRGBColor;
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setStrokeColor:
|
|
|
|
|
cs = stateManager.state.strokeColorSpace;
|
|
|
|
|
args = cs.getRgb(args, 0);
|
|
|
|
|
fn = OPS.setStrokeRGBColor;
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setFillGray:
|
|
|
|
|
stateManager.state.fillColorSpace = ColorSpace.singletons.gray;
|
|
|
|
|
args = ColorSpace.singletons.gray.getRgb(args, 0);
|
|
|
|
|
fn = OPS.setFillRGBColor;
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setStrokeGray:
|
|
|
|
|
stateManager.state.strokeColorSpace = ColorSpace.singletons.gray;
|
|
|
|
|
args = ColorSpace.singletons.gray.getRgb(args, 0);
|
|
|
|
|
fn = OPS.setStrokeRGBColor;
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setFillCMYKColor:
|
|
|
|
|
stateManager.state.fillColorSpace = ColorSpace.singletons.cmyk;
|
|
|
|
|
args = ColorSpace.singletons.cmyk.getRgb(args, 0);
|
|
|
|
|
fn = OPS.setFillRGBColor;
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setStrokeCMYKColor:
|
|
|
|
|
stateManager.state.strokeColorSpace = ColorSpace.singletons.cmyk;
|
|
|
|
|
args = ColorSpace.singletons.cmyk.getRgb(args, 0);
|
|
|
|
|
fn = OPS.setStrokeRGBColor;
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setFillRGBColor:
|
|
|
|
|
stateManager.state.fillColorSpace = ColorSpace.singletons.rgb;
|
|
|
|
|
args = ColorSpace.singletons.rgb.getRgb(args, 0);
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setStrokeRGBColor:
|
|
|
|
|
stateManager.state.strokeColorSpace = ColorSpace.singletons.rgb;
|
|
|
|
|
args = ColorSpace.singletons.rgb.getRgb(args, 0);
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setFillColorN:
|
|
|
|
|
cs = stateManager.state.fillColorSpace;
|
|
|
|
|
if (cs.name === "Pattern") {
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
next(
|
2020-07-05 19:20:10 +09:00
|
|
|
|
self.handleColorN(
|
|
|
|
|
operatorList,
|
|
|
|
|
OPS.setFillColorN,
|
|
|
|
|
args,
|
|
|
|
|
cs,
|
|
|
|
|
patterns,
|
|
|
|
|
resources,
|
|
|
|
|
task,
|
2020-10-09 00:33:23 +09:00
|
|
|
|
localColorSpaceCache,
|
|
|
|
|
localTilingPatternCache
|
2020-07-05 19:20:10 +09:00
|
|
|
|
)
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
);
|
2019-10-31 23:53:51 +09:00
|
|
|
|
return;
|
Add local caching of `ColorSpace`s, by name, in `PartialEvaluator.getOperatorList` (issue 2504)
By caching parsed `ColorSpace`s, we thus don't need to re-parse the same data over and over which saves CPU cycles *and* reduces peak memory usage. (Obviously persistent memory usage *may* increase a tiny bit, but since the caching is done per `PartialEvaluator.getOperatorList` invocation and given that `ColorSpace` instances generally hold very little data this shouldn't be much of an issue.)
Furthermore, by caching `ColorSpace`s we can also lookup the already parsed ones *synchronously* during the `OperatorList` building, instead of having to defer to the event loop/microtask queue since the parsing is done asynchronously (such that error handling is easier).
Possible future improvements:
- Cache/lookup parsed `ColorSpaces` used in `Pattern`s and `Image`s.
- Attempt to cache *local* `ColorSpace`s by reference as well, in addition to only by name, assuming that there's documents where that would be beneficial and that it's not too difficult to implement.
- Assuming there's documents that would benefit from it, also cache repeated `ColorSpace`s *globally* as well.
Given that we've never, until now, been doing *any* caching of parsed `ColorSpace`s and that even using a simple name-only *local* cache helps tremendously in pathological cases, I purposely decided against complicating the implementation too much initially.
Also, compared to parsing of `Image`s, simply creating a `ColorSpace` instance isn't that expensive (hence I'd be somewhat surprised if adding a *global* cache would help much).
---
This patch was tested using:
- The default `tracemonkey` PDF file, which was included mostly to show that "normal" documents aren't negatively affected by these changes.
- The PDF file from issue 2504, i.e. https://dl-ctlg.panasonic.com/jp/manual/sd/sd_rbm1000_0.pdf, where most pages will switch *thousands* of times between a handful of `ColorSpace`s.
with the following manifest file:
```
[
{ "id": "tracemonkey",
"file": "pdfs/tracemonkey.pdf",
"md5": "9a192d8b1a7dc652a19835f6f08098bd",
"rounds": 100,
"type": "eq"
},
{ "id": "issue2504",
"file": "../web/pdfs/issue2504.pdf",
"md5": "",
"rounds": 20,
"type": "eq"
}
]
```
which gave the following results when comparing this patch against the `master` branch:
- Overall
```
-- Grouped By browser, pdf, stat --
browser | pdf | stat | Count | Baseline(ms) | Current(ms) | +/- | % | Result(P<.05)
------- | ----------- | ------------ | ----- | ------------ | ----------- | ---- | ------ | -------------
firefox | issue2504 | Overall | 640 | 977 | 497 | -479 | -49.08 | faster
firefox | issue2504 | Page Request | 640 | 3 | 4 | 1 | 59.18 |
firefox | issue2504 | Rendering | 640 | 974 | 493 | -481 | -49.37 | faster
firefox | tracemonkey | Overall | 1400 | 116 | 111 | -5 | -4.43 |
firefox | tracemonkey | Page Request | 1400 | 2 | 2 | 0 | -2.86 |
firefox | tracemonkey | Rendering | 1400 | 114 | 109 | -5 | -4.47 |
```
- Page-specific
```
-- Grouped By browser, pdf, page, stat --
browser | pdf | page | stat | Count | Baseline(ms) | Current(ms) | +/- | % | Result(P<.05)
------- | ----------- | ---- | ------------ | ----- | ------------ | ----------- | ----- | ------- | -------------
firefox | issue2504 | 0 | Overall | 20 | 2295 | 1268 | -1027 | -44.76 | faster
firefox | issue2504 | 0 | Page Request | 20 | 6 | 7 | 1 | 15.32 |
firefox | issue2504 | 0 | Rendering | 20 | 2288 | 1260 | -1028 | -44.93 | faster
firefox | issue2504 | 1 | Overall | 20 | 3059 | 2806 | -252 | -8.25 | faster
firefox | issue2504 | 1 | Page Request | 20 | 11 | 14 | 3 | 23.25 | slower
firefox | issue2504 | 1 | Rendering | 20 | 3047 | 2792 | -255 | -8.37 | faster
firefox | issue2504 | 2 | Overall | 20 | 411 | 295 | -116 | -28.20 | faster
firefox | issue2504 | 2 | Page Request | 20 | 2 | 42 | 40 | 1897.62 |
firefox | issue2504 | 2 | Rendering | 20 | 409 | 253 | -156 | -38.09 | faster
firefox | issue2504 | 3 | Overall | 20 | 736 | 299 | -437 | -59.34 | faster
firefox | issue2504 | 3 | Page Request | 20 | 2 | 2 | 0 | 0.00 |
firefox | issue2504 | 3 | Rendering | 20 | 734 | 297 | -437 | -59.49 | faster
firefox | issue2504 | 4 | Overall | 20 | 356 | 458 | 102 | 28.63 |
firefox | issue2504 | 4 | Page Request | 20 | 1 | 2 | 1 | 57.14 | slower
firefox | issue2504 | 4 | Rendering | 20 | 354 | 455 | 101 | 28.53 |
firefox | issue2504 | 5 | Overall | 20 | 1381 | 765 | -616 | -44.59 | faster
firefox | issue2504 | 5 | Page Request | 20 | 3 | 5 | 2 | 50.00 | slower
firefox | issue2504 | 5 | Rendering | 20 | 1378 | 760 | -617 | -44.81 | faster
firefox | issue2504 | 6 | Overall | 20 | 757 | 299 | -459 | -60.57 | faster
firefox | issue2504 | 6 | Page Request | 20 | 2 | 5 | 3 | 150.00 | slower
firefox | issue2504 | 6 | Rendering | 20 | 755 | 294 | -462 | -61.11 | faster
firefox | issue2504 | 7 | Overall | 20 | 394 | 302 | -92 | -23.39 | faster
firefox | issue2504 | 7 | Page Request | 20 | 2 | 1 | -1 | -34.88 | faster
firefox | issue2504 | 7 | Rendering | 20 | 392 | 301 | -91 | -23.32 | faster
firefox | issue2504 | 8 | Overall | 20 | 2875 | 979 | -1896 | -65.95 | faster
firefox | issue2504 | 8 | Page Request | 20 | 1 | 2 | 0 | 11.11 |
firefox | issue2504 | 8 | Rendering | 20 | 2874 | 978 | -1896 | -65.99 | faster
firefox | issue2504 | 9 | Overall | 20 | 700 | 332 | -368 | -52.60 | faster
firefox | issue2504 | 9 | Page Request | 20 | 3 | 2 | 0 | -4.00 |
firefox | issue2504 | 9 | Rendering | 20 | 698 | 329 | -368 | -52.78 | faster
firefox | issue2504 | 10 | Overall | 20 | 3296 | 926 | -2370 | -71.91 | faster
firefox | issue2504 | 10 | Page Request | 20 | 2 | 2 | 0 | -18.75 |
firefox | issue2504 | 10 | Rendering | 20 | 3293 | 924 | -2370 | -71.96 | faster
firefox | issue2504 | 11 | Overall | 20 | 524 | 197 | -327 | -62.34 | faster
firefox | issue2504 | 11 | Page Request | 20 | 2 | 3 | 1 | 58.54 |
firefox | issue2504 | 11 | Rendering | 20 | 522 | 194 | -328 | -62.81 | faster
firefox | issue2504 | 12 | Overall | 20 | 752 | 369 | -384 | -50.98 | faster
firefox | issue2504 | 12 | Page Request | 20 | 3 | 2 | -1 | -36.51 | faster
firefox | issue2504 | 12 | Rendering | 20 | 749 | 367 | -382 | -51.05 | faster
firefox | issue2504 | 13 | Overall | 20 | 679 | 487 | -193 | -28.38 | faster
firefox | issue2504 | 13 | Page Request | 20 | 4 | 2 | -2 | -48.68 | faster
firefox | issue2504 | 13 | Rendering | 20 | 676 | 485 | -191 | -28.28 | faster
firefox | issue2504 | 14 | Overall | 20 | 474 | 283 | -191 | -40.26 | faster
firefox | issue2504 | 14 | Page Request | 20 | 2 | 4 | 2 | 78.57 |
firefox | issue2504 | 14 | Rendering | 20 | 471 | 279 | -192 | -40.79 | faster
firefox | issue2504 | 15 | Overall | 20 | 860 | 618 | -241 | -28.05 | faster
firefox | issue2504 | 15 | Page Request | 20 | 2 | 3 | 0 | 10.87 |
firefox | issue2504 | 15 | Rendering | 20 | 857 | 616 | -241 | -28.15 | faster
firefox | issue2504 | 16 | Overall | 20 | 389 | 243 | -147 | -37.71 | faster
firefox | issue2504 | 16 | Page Request | 20 | 2 | 2 | 0 | 2.33 |
firefox | issue2504 | 16 | Rendering | 20 | 387 | 240 | -147 | -37.94 | faster
firefox | issue2504 | 17 | Overall | 20 | 1484 | 672 | -812 | -54.70 | faster
firefox | issue2504 | 17 | Page Request | 20 | 2 | 3 | 1 | 37.21 |
firefox | issue2504 | 17 | Rendering | 20 | 1482 | 669 | -812 | -54.84 | faster
firefox | issue2504 | 18 | Overall | 20 | 575 | 252 | -323 | -56.12 | faster
firefox | issue2504 | 18 | Page Request | 20 | 2 | 2 | 0 | -16.22 |
firefox | issue2504 | 18 | Rendering | 20 | 573 | 251 | -322 | -56.24 | faster
firefox | issue2504 | 19 | Overall | 20 | 517 | 227 | -290 | -56.08 | faster
firefox | issue2504 | 19 | Page Request | 20 | 2 | 2 | 0 | 21.62 |
firefox | issue2504 | 19 | Rendering | 20 | 515 | 225 | -290 | -56.37 | faster
firefox | issue2504 | 20 | Overall | 20 | 668 | 670 | 2 | 0.31 |
firefox | issue2504 | 20 | Page Request | 20 | 4 | 2 | -1 | -34.29 |
firefox | issue2504 | 20 | Rendering | 20 | 664 | 667 | 3 | 0.49 |
firefox | issue2504 | 21 | Overall | 20 | 486 | 309 | -177 | -36.44 | faster
firefox | issue2504 | 21 | Page Request | 20 | 2 | 2 | 0 | 16.13 |
firefox | issue2504 | 21 | Rendering | 20 | 484 | 307 | -177 | -36.60 | faster
firefox | issue2504 | 22 | Overall | 20 | 543 | 267 | -276 | -50.85 | faster
firefox | issue2504 | 22 | Page Request | 20 | 2 | 2 | 0 | 10.26 |
firefox | issue2504 | 22 | Rendering | 20 | 541 | 265 | -276 | -51.07 | faster
firefox | issue2504 | 23 | Overall | 20 | 3246 | 871 | -2375 | -73.17 | faster
firefox | issue2504 | 23 | Page Request | 20 | 2 | 3 | 1 | 37.21 |
firefox | issue2504 | 23 | Rendering | 20 | 3243 | 868 | -2376 | -73.25 | faster
firefox | issue2504 | 24 | Overall | 20 | 379 | 156 | -223 | -58.83 | faster
firefox | issue2504 | 24 | Page Request | 20 | 2 | 2 | 0 | -2.86 |
firefox | issue2504 | 24 | Rendering | 20 | 378 | 154 | -223 | -59.10 | faster
firefox | issue2504 | 25 | Overall | 20 | 176 | 127 | -50 | -28.19 | faster
firefox | issue2504 | 25 | Page Request | 20 | 2 | 1 | 0 | -15.63 |
firefox | issue2504 | 25 | Rendering | 20 | 175 | 125 | -49 | -28.31 | faster
firefox | issue2504 | 26 | Overall | 20 | 181 | 108 | -74 | -40.67 | faster
firefox | issue2504 | 26 | Page Request | 20 | 3 | 2 | -1 | -39.13 | faster
firefox | issue2504 | 26 | Rendering | 20 | 178 | 105 | -72 | -40.69 | faster
firefox | issue2504 | 27 | Overall | 20 | 208 | 104 | -104 | -49.92 | faster
firefox | issue2504 | 27 | Page Request | 20 | 2 | 2 | 1 | 48.39 |
firefox | issue2504 | 27 | Rendering | 20 | 206 | 102 | -104 | -50.64 | faster
firefox | issue2504 | 28 | Overall | 20 | 241 | 111 | -131 | -54.16 | faster
firefox | issue2504 | 28 | Page Request | 20 | 2 | 2 | -1 | -33.33 |
firefox | issue2504 | 28 | Rendering | 20 | 239 | 109 | -130 | -54.39 | faster
firefox | issue2504 | 29 | Overall | 20 | 321 | 196 | -125 | -39.05 | faster
firefox | issue2504 | 29 | Page Request | 20 | 1 | 2 | 0 | 17.86 |
firefox | issue2504 | 29 | Rendering | 20 | 319 | 194 | -126 | -39.35 | faster
firefox | issue2504 | 30 | Overall | 20 | 651 | 271 | -380 | -58.41 | faster
firefox | issue2504 | 30 | Page Request | 20 | 1 | 2 | 1 | 50.00 |
firefox | issue2504 | 30 | Rendering | 20 | 649 | 269 | -381 | -58.60 | faster
firefox | issue2504 | 31 | Overall | 20 | 1635 | 647 | -988 | -60.42 | faster
firefox | issue2504 | 31 | Page Request | 20 | 1 | 2 | 0 | 30.43 |
firefox | issue2504 | 31 | Rendering | 20 | 1634 | 645 | -988 | -60.49 | faster
firefox | tracemonkey | 0 | Overall | 100 | 51 | 51 | 0 | 0.02 |
firefox | tracemonkey | 0 | Page Request | 100 | 1 | 1 | 0 | -4.76 |
firefox | tracemonkey | 0 | Rendering | 100 | 50 | 50 | 0 | 0.12 |
firefox | tracemonkey | 1 | Overall | 100 | 97 | 91 | -5 | -5.52 | faster
firefox | tracemonkey | 1 | Page Request | 100 | 3 | 3 | 0 | -1.32 |
firefox | tracemonkey | 1 | Rendering | 100 | 94 | 88 | -5 | -5.73 | faster
firefox | tracemonkey | 2 | Overall | 100 | 40 | 40 | 0 | 0.50 |
firefox | tracemonkey | 2 | Page Request | 100 | 1 | 1 | 0 | 3.16 |
firefox | tracemonkey | 2 | Rendering | 100 | 39 | 39 | 0 | 0.54 |
firefox | tracemonkey | 3 | Overall | 100 | 62 | 62 | -1 | -0.94 |
firefox | tracemonkey | 3 | Page Request | 100 | 1 | 1 | 0 | 17.05 |
firefox | tracemonkey | 3 | Rendering | 100 | 61 | 61 | -1 | -1.11 |
firefox | tracemonkey | 4 | Overall | 100 | 56 | 58 | 2 | 3.41 |
firefox | tracemonkey | 4 | Page Request | 100 | 1 | 1 | 0 | 15.31 |
firefox | tracemonkey | 4 | Rendering | 100 | 55 | 57 | 2 | 3.23 |
firefox | tracemonkey | 5 | Overall | 100 | 73 | 71 | -2 | -2.28 |
firefox | tracemonkey | 5 | Page Request | 100 | 2 | 2 | 0 | 12.20 |
firefox | tracemonkey | 5 | Rendering | 100 | 71 | 69 | -2 | -2.69 |
firefox | tracemonkey | 6 | Overall | 100 | 85 | 69 | -16 | -18.73 | faster
firefox | tracemonkey | 6 | Page Request | 100 | 2 | 2 | 0 | -9.90 |
firefox | tracemonkey | 6 | Rendering | 100 | 83 | 67 | -16 | -18.97 | faster
firefox | tracemonkey | 7 | Overall | 100 | 65 | 64 | 0 | -0.37 |
firefox | tracemonkey | 7 | Page Request | 100 | 1 | 1 | 0 | -11.94 |
firefox | tracemonkey | 7 | Rendering | 100 | 63 | 63 | 0 | -0.05 |
firefox | tracemonkey | 8 | Overall | 100 | 53 | 54 | 1 | 2.04 |
firefox | tracemonkey | 8 | Page Request | 100 | 1 | 1 | 0 | 17.02 |
firefox | tracemonkey | 8 | Rendering | 100 | 52 | 53 | 1 | 1.82 |
firefox | tracemonkey | 9 | Overall | 100 | 79 | 73 | -6 | -7.86 | faster
firefox | tracemonkey | 9 | Page Request | 100 | 2 | 2 | 0 | -15.14 |
firefox | tracemonkey | 9 | Rendering | 100 | 77 | 71 | -6 | -7.86 | faster
firefox | tracemonkey | 10 | Overall | 100 | 545 | 519 | -27 | -4.86 | faster
firefox | tracemonkey | 10 | Page Request | 100 | 14 | 13 | 0 | -3.56 |
firefox | tracemonkey | 10 | Rendering | 100 | 532 | 506 | -26 | -4.90 | faster
firefox | tracemonkey | 11 | Overall | 100 | 42 | 41 | -1 | -2.50 |
firefox | tracemonkey | 11 | Page Request | 100 | 1 | 1 | 0 | -27.42 | faster
firefox | tracemonkey | 11 | Rendering | 100 | 41 | 40 | -1 | -1.75 |
firefox | tracemonkey | 12 | Overall | 100 | 350 | 332 | -18 | -5.16 | faster
firefox | tracemonkey | 12 | Page Request | 100 | 3 | 3 | 0 | -5.17 |
firefox | tracemonkey | 12 | Rendering | 100 | 347 | 329 | -18 | -5.15 | faster
firefox | tracemonkey | 13 | Overall | 100 | 31 | 31 | 0 | 0.52 |
firefox | tracemonkey | 13 | Page Request | 100 | 1 | 1 | 0 | 4.95 |
firefox | tracemonkey | 13 | Rendering | 100 | 30 | 30 | 0 | 0.20 |
```
2020-06-13 21:12:40 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
args = cs.getRgb(args, 0);
|
|
|
|
|
fn = OPS.setFillRGBColor;
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setStrokeColorN:
|
|
|
|
|
cs = stateManager.state.strokeColorSpace;
|
|
|
|
|
if (cs.name === "Pattern") {
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
next(
|
2020-07-05 19:20:10 +09:00
|
|
|
|
self.handleColorN(
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
operatorList,
|
2020-07-05 19:20:10 +09:00
|
|
|
|
OPS.setStrokeColorN,
|
|
|
|
|
args,
|
|
|
|
|
cs,
|
|
|
|
|
patterns,
|
|
|
|
|
resources,
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
task,
|
2020-10-09 00:33:23 +09:00
|
|
|
|
localColorSpaceCache,
|
|
|
|
|
localTilingPatternCache
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
)
|
|
|
|
|
);
|
2016-03-11 22:59:09 +09:00
|
|
|
|
return;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
args = cs.getRgb(args, 0);
|
|
|
|
|
fn = OPS.setStrokeRGBColor;
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
case OPS.shadingFill:
|
|
|
|
|
var shadingRes = resources.get("Shading");
|
|
|
|
|
if (!shadingRes) {
|
|
|
|
|
throw new FormatError("No shading resource found");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
var shading = shadingRes.get(args[0].name);
|
|
|
|
|
if (!shading) {
|
|
|
|
|
throw new FormatError("No shading object found");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
var shadingFill = Pattern.parseShading(
|
|
|
|
|
shading,
|
|
|
|
|
null,
|
|
|
|
|
xref,
|
|
|
|
|
resources,
|
|
|
|
|
self.handler,
|
|
|
|
|
self._pdfFunctionFactory,
|
|
|
|
|
localColorSpaceCache
|
|
|
|
|
);
|
|
|
|
|
var patternIR = shadingFill.getIR();
|
|
|
|
|
args = [patternIR];
|
|
|
|
|
fn = OPS.shadingFill;
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setGState:
|
2020-07-11 20:52:11 +09:00
|
|
|
|
name = args[0].name;
|
|
|
|
|
if (name) {
|
|
|
|
|
const localGStateObj = localGStateCache.getByName(name);
|
|
|
|
|
if (localGStateObj) {
|
|
|
|
|
if (localGStateObj.length > 0) {
|
|
|
|
|
operatorList.addOp(OPS.setGState, [localGStateObj]);
|
|
|
|
|
}
|
|
|
|
|
args = null;
|
|
|
|
|
continue;
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
next(
|
2020-07-11 20:52:11 +09:00
|
|
|
|
new Promise(function (resolveGState, rejectGState) {
|
|
|
|
|
if (!name) {
|
|
|
|
|
throw new FormatError("GState must be referred to by name.");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
const extGState = resources.get("ExtGState");
|
|
|
|
|
if (!(extGState instanceof Dict)) {
|
|
|
|
|
throw new FormatError("ExtGState should be a dictionary.");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
const gState = extGState.get(name);
|
|
|
|
|
// TODO: Attempt to lookup cached GStates by reference as well,
|
|
|
|
|
// if and only if there are PDF documents where doing so
|
|
|
|
|
// would significantly improve performance.
|
|
|
|
|
if (!(gState instanceof Dict)) {
|
|
|
|
|
throw new FormatError("GState should be a dictionary.");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
self
|
|
|
|
|
.setGState({
|
|
|
|
|
resources,
|
|
|
|
|
gState,
|
|
|
|
|
operatorList,
|
|
|
|
|
cacheKey: name,
|
|
|
|
|
task,
|
|
|
|
|
stateManager,
|
|
|
|
|
localGStateCache,
|
|
|
|
|
localColorSpaceCache,
|
|
|
|
|
})
|
|
|
|
|
.then(resolveGState, rejectGState);
|
|
|
|
|
}).catch(function (reason) {
|
|
|
|
|
if (reason instanceof AbortException) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
if (self.options.ignoreErrors) {
|
|
|
|
|
// Error(s) in the ExtGState -- sending unsupported feature
|
|
|
|
|
// notification and allow parsing/rendering to continue.
|
|
|
|
|
self.handler.send("UnsupportedFeature", {
|
|
|
|
|
featureId: UNSUPPORTED_FEATURES.errorExtGState,
|
|
|
|
|
});
|
|
|
|
|
warn(`getOperatorList - ignoring ExtGState: "${reason}".`);
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
throw reason;
|
|
|
|
|
})
|
2020-07-05 19:20:10 +09:00
|
|
|
|
);
|
|
|
|
|
return;
|
|
|
|
|
case OPS.moveTo:
|
|
|
|
|
case OPS.lineTo:
|
|
|
|
|
case OPS.curveTo:
|
|
|
|
|
case OPS.curveTo2:
|
|
|
|
|
case OPS.curveTo3:
|
|
|
|
|
case OPS.closePath:
|
|
|
|
|
case OPS.rectangle:
|
|
|
|
|
self.buildPath(operatorList, fn, args, parsingText);
|
|
|
|
|
continue;
|
|
|
|
|
case OPS.markPoint:
|
|
|
|
|
case OPS.markPointProps:
|
|
|
|
|
case OPS.beginCompat:
|
|
|
|
|
case OPS.endCompat:
|
|
|
|
|
// Ignore operators where the corresponding handlers are known to
|
|
|
|
|
// be no-op in CanvasGraphics (display/canvas.js). This prevents
|
|
|
|
|
// serialization errors and is also a bit more efficient.
|
|
|
|
|
// We could also try to serialize all objects in a general way,
|
|
|
|
|
// e.g. as done in https://github.com/mozilla/pdf.js/pull/6266,
|
|
|
|
|
// but doing so is meaningless without knowing the semantics.
|
|
|
|
|
continue;
|
2020-07-15 07:17:27 +09:00
|
|
|
|
case OPS.beginMarkedContentProps:
|
|
|
|
|
if (!isName(args[0])) {
|
|
|
|
|
warn(`Expected name for beginMarkedContentProps arg0=${args[0]}`);
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
if (args[0].name === "OC") {
|
|
|
|
|
next(
|
|
|
|
|
self
|
|
|
|
|
.parseMarkedContentProps(args[1], resources)
|
|
|
|
|
.then(data => {
|
|
|
|
|
operatorList.addOp(OPS.beginMarkedContentProps, [
|
|
|
|
|
"OC",
|
|
|
|
|
data,
|
|
|
|
|
]);
|
|
|
|
|
})
|
|
|
|
|
.catch(reason => {
|
|
|
|
|
if (reason instanceof AbortException) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
if (self.options.ignoreErrors) {
|
|
|
|
|
self.handler.send("UnsupportedFeature", {
|
|
|
|
|
featureId: UNSUPPORTED_FEATURES.errorMarkedContent,
|
|
|
|
|
});
|
|
|
|
|
warn(
|
|
|
|
|
`getOperatorList - ignoring beginMarkedContentProps: "${reason}".`
|
|
|
|
|
);
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
throw reason;
|
|
|
|
|
})
|
|
|
|
|
);
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
// Other marked content types aren't supported yet.
|
|
|
|
|
args = [args[0].name];
|
|
|
|
|
|
|
|
|
|
break;
|
|
|
|
|
case OPS.beginMarkedContent:
|
|
|
|
|
case OPS.endMarkedContent:
|
2020-07-05 19:20:10 +09:00
|
|
|
|
default:
|
|
|
|
|
// Note: Ignore the operator if it has `Dict` arguments, since
|
|
|
|
|
// those are non-serializable, otherwise postMessage will throw
|
|
|
|
|
// "An object could not be cloned.".
|
|
|
|
|
if (args !== null) {
|
|
|
|
|
for (i = 0, ii = args.length; i < ii; i++) {
|
|
|
|
|
if (args[i] instanceof Dict) {
|
|
|
|
|
break;
|
2016-02-13 02:15:49 +09:00
|
|
|
|
}
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (i < ii) {
|
|
|
|
|
warn("getOperatorList - ignoring operator: " + fn);
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
}
|
[api-minor] Always allow e.g. rendering to continue even if there are errors, and add a `stopAtErrors` parameter to `getDocument` to opt-out of this behaviour (issue 6342, issue 3795, bug 1130815)
Other PDF readers, e.g. Adobe Reader and PDFium (in Chrome), will attempt to render as much of a page as possible even if there are errors present.
Currently we just bail as soon the first error is hit, which means that we'll usually not render anything in these cases and just display a blank page instead.
NOTE: This patch changes the default behaviour of the PDF.js API to always attempt to recover as much data as possible, even when encountering errors during e.g. `getOperatorList`/`getTextContent`, which thus improve our handling of corrupt PDF files and allow the default viewer to handle errors slightly more gracefully.
In the event that an API consumer wishes to use the old behaviour, where we stop parsing as soon as an error is encountered, the `stopAtErrors` parameter can be set at `getDocument`.
Fixes, inasmuch it's possible since the PDF files are corrupt, e.g. issue 6342, issue 3795, and [bug 1130815](https://bugzilla.mozilla.org/show_bug.cgi?id=1130815) (and probably others too).
2017-02-19 22:03:08 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
operatorList.addOp(fn, args);
|
|
|
|
|
}
|
|
|
|
|
if (stop) {
|
|
|
|
|
next(deferred);
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
// Some PDFs don't close all restores inside object/form.
|
|
|
|
|
// Closing those for them.
|
|
|
|
|
closePendingRestoreOPS();
|
|
|
|
|
resolve();
|
|
|
|
|
}).catch(reason => {
|
|
|
|
|
if (reason instanceof AbortException) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
if (this.options.ignoreErrors) {
|
|
|
|
|
// Error(s) in the OperatorList -- sending unsupported feature
|
|
|
|
|
// notification and allow rendering to continue.
|
|
|
|
|
this.handler.send("UnsupportedFeature", {
|
|
|
|
|
featureId: UNSUPPORTED_FEATURES.errorOperatorList,
|
|
|
|
|
});
|
|
|
|
|
warn(
|
|
|
|
|
`getOperatorList - ignoring errors during "${task.name}" ` +
|
|
|
|
|
`task: "${reason}".`
|
|
|
|
|
);
|
2011-12-11 08:24:54 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
closePendingRestoreOPS();
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
throw reason;
|
|
|
|
|
});
|
|
|
|
|
}
|
2013-08-01 03:17:36 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
getTextContent({
|
|
|
|
|
stream,
|
|
|
|
|
task,
|
|
|
|
|
resources,
|
|
|
|
|
stateManager = null,
|
|
|
|
|
normalizeWhitespace = false,
|
|
|
|
|
combineTextItems = false,
|
|
|
|
|
sink,
|
|
|
|
|
seenStyles = Object.create(null),
|
|
|
|
|
}) {
|
|
|
|
|
// Ensure that `resources`/`stateManager` is correctly initialized,
|
|
|
|
|
// even if the provided parameter is e.g. `null`.
|
|
|
|
|
resources = resources || Dict.empty;
|
|
|
|
|
stateManager = stateManager || new StateManager(new TextState());
|
2011-12-11 08:24:54 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var WhitespaceRegexp = /\s/g;
|
2013-04-09 07:14:56 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var textContent = {
|
|
|
|
|
items: [],
|
|
|
|
|
styles: Object.create(null),
|
|
|
|
|
};
|
|
|
|
|
var textContentItem = {
|
|
|
|
|
initialized: false,
|
|
|
|
|
str: [],
|
|
|
|
|
width: 0,
|
|
|
|
|
height: 0,
|
|
|
|
|
vertical: false,
|
|
|
|
|
lastAdvanceWidth: 0,
|
|
|
|
|
lastAdvanceHeight: 0,
|
|
|
|
|
textAdvanceScale: 0,
|
|
|
|
|
spaceWidth: 0,
|
|
|
|
|
fakeSpaceMin: Infinity,
|
|
|
|
|
fakeMultiSpaceMin: Infinity,
|
|
|
|
|
fakeMultiSpaceMax: -0,
|
|
|
|
|
textRunBreakAllowed: false,
|
|
|
|
|
transform: null,
|
|
|
|
|
fontName: null,
|
|
|
|
|
};
|
|
|
|
|
var SPACE_FACTOR = 0.3;
|
|
|
|
|
var MULTI_SPACE_FACTOR = 1.5;
|
|
|
|
|
var MULTI_SPACE_FACTOR_MAX = 4;
|
2014-04-10 08:44:07 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var self = this;
|
|
|
|
|
var xref = this.xref;
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// The xobj is parsed iff it's needed, e.g. if there is a `DO` cmd.
|
|
|
|
|
var xobjs = null;
|
|
|
|
|
const emptyXObjectCache = new LocalImageCache();
|
2020-07-11 21:05:53 +09:00
|
|
|
|
const emptyGStateCache = new LocalGStateCache();
|
2015-11-04 01:12:41 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var preprocessor = new EvaluatorPreprocessor(stream, xref, stateManager);
|
2015-11-04 01:12:41 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var textState;
|
2015-11-06 23:40:44 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
function ensureTextContentItem() {
|
|
|
|
|
if (textContentItem.initialized) {
|
2015-11-04 01:12:41 +09:00
|
|
|
|
return textContentItem;
|
2014-04-10 08:44:07 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var font = textState.font;
|
|
|
|
|
if (!(font.loadedName in seenStyles)) {
|
|
|
|
|
seenStyles[font.loadedName] = true;
|
|
|
|
|
textContent.styles[font.loadedName] = {
|
|
|
|
|
fontFamily: font.fallbackName,
|
|
|
|
|
ascent: font.ascent,
|
|
|
|
|
descent: font.descent,
|
|
|
|
|
vertical: font.vertical,
|
|
|
|
|
};
|
|
|
|
|
}
|
|
|
|
|
textContentItem.fontName = font.loadedName;
|
|
|
|
|
|
|
|
|
|
// 9.4.4 Text Space Details
|
|
|
|
|
var tsm = [
|
|
|
|
|
textState.fontSize * textState.textHScale,
|
|
|
|
|
0,
|
|
|
|
|
0,
|
|
|
|
|
textState.fontSize,
|
|
|
|
|
0,
|
|
|
|
|
textState.textRise,
|
|
|
|
|
];
|
2014-04-10 08:44:07 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (
|
|
|
|
|
font.isType3Font &&
|
|
|
|
|
textState.fontSize <= 1 &&
|
|
|
|
|
!isArrayEqual(textState.fontMatrix, FONT_IDENTITY_MATRIX)
|
|
|
|
|
) {
|
|
|
|
|
const glyphHeight = font.bbox[3] - font.bbox[1];
|
|
|
|
|
if (glyphHeight > 0) {
|
|
|
|
|
tsm[3] *= glyphHeight * textState.fontMatrix[3];
|
2015-11-24 00:57:43 +09:00
|
|
|
|
}
|
2014-04-10 08:44:07 +09:00
|
|
|
|
}
|
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var trm = Util.transform(
|
|
|
|
|
textState.ctm,
|
|
|
|
|
Util.transform(textState.textMatrix, tsm)
|
|
|
|
|
);
|
|
|
|
|
textContentItem.transform = trm;
|
|
|
|
|
if (!font.vertical) {
|
|
|
|
|
textContentItem.width = 0;
|
|
|
|
|
textContentItem.height = Math.sqrt(trm[2] * trm[2] + trm[3] * trm[3]);
|
|
|
|
|
textContentItem.vertical = false;
|
|
|
|
|
} else {
|
|
|
|
|
textContentItem.width = Math.sqrt(trm[0] * trm[0] + trm[1] * trm[1]);
|
|
|
|
|
textContentItem.height = 0;
|
|
|
|
|
textContentItem.vertical = true;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
var a = textState.textLineMatrix[0];
|
|
|
|
|
var b = textState.textLineMatrix[1];
|
|
|
|
|
var scaleLineX = Math.sqrt(a * a + b * b);
|
|
|
|
|
a = textState.ctm[0];
|
|
|
|
|
b = textState.ctm[1];
|
|
|
|
|
var scaleCtmX = Math.sqrt(a * a + b * b);
|
|
|
|
|
textContentItem.textAdvanceScale = scaleCtmX * scaleLineX;
|
|
|
|
|
textContentItem.lastAdvanceWidth = 0;
|
|
|
|
|
textContentItem.lastAdvanceHeight = 0;
|
|
|
|
|
|
|
|
|
|
var spaceWidth = (font.spaceWidth / 1000) * textState.fontSize;
|
|
|
|
|
if (spaceWidth) {
|
|
|
|
|
textContentItem.spaceWidth = spaceWidth;
|
|
|
|
|
textContentItem.fakeSpaceMin = spaceWidth * SPACE_FACTOR;
|
|
|
|
|
textContentItem.fakeMultiSpaceMin = spaceWidth * MULTI_SPACE_FACTOR;
|
|
|
|
|
textContentItem.fakeMultiSpaceMax = spaceWidth * MULTI_SPACE_FACTOR_MAX;
|
|
|
|
|
// It's okay for monospace fonts to fake as much space as needed.
|
|
|
|
|
textContentItem.textRunBreakAllowed = !font.isMonospace;
|
|
|
|
|
} else {
|
|
|
|
|
textContentItem.spaceWidth = 0;
|
|
|
|
|
textContentItem.fakeSpaceMin = Infinity;
|
|
|
|
|
textContentItem.fakeMultiSpaceMin = Infinity;
|
|
|
|
|
textContentItem.fakeMultiSpaceMax = 0;
|
|
|
|
|
textContentItem.textRunBreakAllowed = false;
|
2014-04-10 08:44:07 +09:00
|
|
|
|
}
|
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
textContentItem.initialized = true;
|
|
|
|
|
return textContentItem;
|
|
|
|
|
}
|
2014-04-10 08:44:07 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
function replaceWhitespace(str) {
|
|
|
|
|
// Replaces all whitespaces with standard spaces (0x20), to avoid
|
|
|
|
|
// alignment issues between the textLayer and the canvas if the text
|
|
|
|
|
// contains e.g. tabs (fixes issue6612.pdf).
|
|
|
|
|
var i = 0,
|
|
|
|
|
ii = str.length,
|
|
|
|
|
code;
|
|
|
|
|
while (i < ii && (code = str.charCodeAt(i)) >= 0x20 && code <= 0x7f) {
|
|
|
|
|
i++;
|
|
|
|
|
}
|
|
|
|
|
return i < ii ? str.replace(WhitespaceRegexp, " ") : str;
|
|
|
|
|
}
|
2015-05-10 18:28:15 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
function runBidiTransform(textChunk) {
|
|
|
|
|
var str = textChunk.str.join("");
|
|
|
|
|
var bidiResult = bidi(str, -1, textChunk.vertical);
|
|
|
|
|
return {
|
|
|
|
|
str: normalizeWhitespace
|
|
|
|
|
? replaceWhitespace(bidiResult.str)
|
|
|
|
|
: bidiResult.str,
|
|
|
|
|
dir: bidiResult.dir,
|
|
|
|
|
width: textChunk.width,
|
|
|
|
|
height: textChunk.height,
|
|
|
|
|
transform: textChunk.transform,
|
|
|
|
|
fontName: textChunk.fontName,
|
|
|
|
|
};
|
|
|
|
|
}
|
2014-04-10 08:44:07 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
function handleSetFont(fontName, fontRef) {
|
|
|
|
|
return self
|
|
|
|
|
.loadFont(fontName, fontRef, resources)
|
|
|
|
|
.then(function (translated) {
|
|
|
|
|
textState.font = translated.font;
|
|
|
|
|
textState.fontMatrix =
|
|
|
|
|
translated.font.fontMatrix || FONT_IDENTITY_MATRIX;
|
|
|
|
|
});
|
|
|
|
|
}
|
2014-04-10 08:44:07 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
function buildTextContentItem(chars) {
|
|
|
|
|
var font = textState.font;
|
|
|
|
|
var textChunk = ensureTextContentItem();
|
|
|
|
|
var width = 0;
|
|
|
|
|
var height = 0;
|
|
|
|
|
var glyphs = font.charsToGlyphs(chars);
|
|
|
|
|
for (var i = 0; i < glyphs.length; i++) {
|
|
|
|
|
var glyph = glyphs[i];
|
|
|
|
|
var glyphWidth = null;
|
|
|
|
|
if (font.vertical && glyph.vmetric) {
|
|
|
|
|
glyphWidth = glyph.vmetric[0];
|
2014-04-10 08:44:07 +09:00
|
|
|
|
} else {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
glyphWidth = glyph.width;
|
2014-04-10 08:44:07 +09:00
|
|
|
|
}
|
2015-11-06 23:40:44 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var glyphUnicode = glyph.unicode;
|
|
|
|
|
var NormalizedUnicodes = getNormalizedUnicodes();
|
|
|
|
|
if (NormalizedUnicodes[glyphUnicode] !== undefined) {
|
|
|
|
|
glyphUnicode = NormalizedUnicodes[glyphUnicode];
|
2015-11-02 23:54:15 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
glyphUnicode = reverseIfRtl(glyphUnicode);
|
2015-11-02 23:54:15 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var charSpacing = textState.charSpacing;
|
|
|
|
|
if (glyph.isSpace) {
|
|
|
|
|
var wordSpacing = textState.wordSpacing;
|
|
|
|
|
charSpacing += wordSpacing;
|
|
|
|
|
if (wordSpacing > 0) {
|
|
|
|
|
addFakeSpaces(wordSpacing, textChunk.str);
|
|
|
|
|
}
|
2015-11-04 01:12:41 +09:00
|
|
|
|
}
|
2016-12-07 07:07:16 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var tx = 0;
|
|
|
|
|
var ty = 0;
|
|
|
|
|
if (!font.vertical) {
|
|
|
|
|
var w0 = glyphWidth * textState.fontMatrix[0];
|
|
|
|
|
tx = (w0 * textState.fontSize + charSpacing) * textState.textHScale;
|
|
|
|
|
width += tx;
|
2019-01-29 22:24:48 +09:00
|
|
|
|
} else {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var w1 = glyphWidth * textState.fontMatrix[0];
|
|
|
|
|
ty = w1 * textState.fontSize + charSpacing;
|
|
|
|
|
height += ty;
|
2019-01-29 22:24:48 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
textState.translateTextMatrix(tx, ty);
|
2015-11-04 01:12:41 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
textChunk.str.push(glyphUnicode);
|
2015-11-04 01:12:41 +09:00
|
|
|
|
}
|
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (!font.vertical) {
|
|
|
|
|
textChunk.lastAdvanceWidth = width;
|
|
|
|
|
textChunk.width += width;
|
|
|
|
|
} else {
|
|
|
|
|
textChunk.lastAdvanceHeight = height;
|
|
|
|
|
textChunk.height += Math.abs(height);
|
2017-04-17 21:46:53 +09:00
|
|
|
|
}
|
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return textChunk;
|
|
|
|
|
}
|
2014-05-10 10:41:03 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
function addFakeSpaces(width, strBuf) {
|
|
|
|
|
if (width < textContentItem.fakeSpaceMin) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
if (width < textContentItem.fakeMultiSpaceMin) {
|
|
|
|
|
strBuf.push(" ");
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
var fakeSpaces = Math.round(width / textContentItem.spaceWidth);
|
|
|
|
|
while (fakeSpaces-- > 0) {
|
|
|
|
|
strBuf.push(" ");
|
|
|
|
|
}
|
|
|
|
|
}
|
2016-06-01 06:01:35 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
function flushTextContentItem() {
|
|
|
|
|
if (!textContentItem.initialized) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Do final text scaling.
|
|
|
|
|
if (!textContentItem.vertical) {
|
|
|
|
|
textContentItem.width *= textContentItem.textAdvanceScale;
|
|
|
|
|
} else {
|
|
|
|
|
textContentItem.height *= textContentItem.textAdvanceScale;
|
|
|
|
|
}
|
|
|
|
|
textContent.items.push(runBidiTransform(textContentItem));
|
|
|
|
|
|
|
|
|
|
textContentItem.initialized = false;
|
|
|
|
|
textContentItem.str.length = 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
function enqueueChunk() {
|
|
|
|
|
const length = textContent.items.length;
|
|
|
|
|
if (length > 0) {
|
|
|
|
|
sink.enqueue(textContent, length);
|
|
|
|
|
textContent.items = [];
|
|
|
|
|
textContent.styles = Object.create(null);
|
|
|
|
|
}
|
|
|
|
|
}
|
2015-11-04 01:12:41 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var timeSlotManager = new TimeSlotManager();
|
|
|
|
|
|
|
|
|
|
return new Promise(function promiseBody(resolve, reject) {
|
|
|
|
|
const next = function (promise) {
|
|
|
|
|
enqueueChunk();
|
|
|
|
|
Promise.all([promise, sink.ready]).then(function () {
|
|
|
|
|
try {
|
|
|
|
|
promiseBody(resolve, reject);
|
|
|
|
|
} catch (ex) {
|
|
|
|
|
reject(ex);
|
|
|
|
|
}
|
|
|
|
|
}, reject);
|
|
|
|
|
};
|
|
|
|
|
task.ensureNotTerminated();
|
|
|
|
|
timeSlotManager.reset();
|
|
|
|
|
var stop,
|
|
|
|
|
operation = {},
|
|
|
|
|
args = [];
|
|
|
|
|
while (!(stop = timeSlotManager.check())) {
|
|
|
|
|
// The arguments parsed by read() are not used beyond this loop, so
|
|
|
|
|
// we can reuse the same array on every iteration, thus avoiding
|
|
|
|
|
// unnecessary allocations.
|
|
|
|
|
args.length = 0;
|
|
|
|
|
operation.args = args;
|
|
|
|
|
if (!preprocessor.read(operation)) {
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
textState = stateManager.state;
|
|
|
|
|
var fn = operation.fn;
|
|
|
|
|
args = operation.args;
|
|
|
|
|
var advance, diff;
|
|
|
|
|
|
|
|
|
|
switch (fn | 0) {
|
|
|
|
|
case OPS.setFont:
|
|
|
|
|
// Optimization to ignore multiple identical Tf commands.
|
|
|
|
|
var fontNameArg = args[0].name,
|
|
|
|
|
fontSizeArg = args[1];
|
|
|
|
|
if (
|
|
|
|
|
textState.font &&
|
|
|
|
|
fontNameArg === textState.fontName &&
|
|
|
|
|
fontSizeArg === textState.fontSize
|
|
|
|
|
) {
|
2014-05-10 10:21:15 +09:00
|
|
|
|
break;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
flushTextContentItem();
|
|
|
|
|
textState.fontName = fontNameArg;
|
|
|
|
|
textState.fontSize = fontSizeArg;
|
|
|
|
|
next(handleSetFont(fontNameArg, null));
|
|
|
|
|
return;
|
|
|
|
|
case OPS.setTextRise:
|
|
|
|
|
flushTextContentItem();
|
|
|
|
|
textState.textRise = args[0];
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setHScale:
|
|
|
|
|
flushTextContentItem();
|
|
|
|
|
textState.textHScale = args[0] / 100;
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setLeading:
|
|
|
|
|
flushTextContentItem();
|
|
|
|
|
textState.leading = args[0];
|
|
|
|
|
break;
|
|
|
|
|
case OPS.moveText:
|
|
|
|
|
// Optimization to treat same line movement as advance
|
|
|
|
|
var isSameTextLine = !textState.font
|
|
|
|
|
? false
|
|
|
|
|
: (textState.font.vertical ? args[0] : args[1]) === 0;
|
|
|
|
|
advance = args[0] - args[1];
|
|
|
|
|
if (
|
|
|
|
|
combineTextItems &&
|
|
|
|
|
isSameTextLine &&
|
|
|
|
|
textContentItem.initialized &&
|
|
|
|
|
advance > 0 &&
|
|
|
|
|
advance <= textContentItem.fakeMultiSpaceMax
|
|
|
|
|
) {
|
2014-05-10 10:21:15 +09:00
|
|
|
|
textState.translateTextLineMatrix(args[0], args[1]);
|
2020-07-05 19:20:10 +09:00
|
|
|
|
textContentItem.width +=
|
|
|
|
|
args[0] - textContentItem.lastAdvanceWidth;
|
|
|
|
|
textContentItem.height +=
|
|
|
|
|
args[1] - textContentItem.lastAdvanceHeight;
|
|
|
|
|
diff =
|
|
|
|
|
args[0] -
|
|
|
|
|
textContentItem.lastAdvanceWidth -
|
|
|
|
|
(args[1] - textContentItem.lastAdvanceHeight);
|
|
|
|
|
addFakeSpaces(diff, textContentItem.str);
|
2014-05-10 10:21:15 +09:00
|
|
|
|
break;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
2016-05-15 05:13:12 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
flushTextContentItem();
|
|
|
|
|
textState.translateTextLineMatrix(args[0], args[1]);
|
|
|
|
|
textState.textMatrix = textState.textLineMatrix.slice();
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setLeadingMoveText:
|
|
|
|
|
flushTextContentItem();
|
|
|
|
|
textState.leading = -args[1];
|
|
|
|
|
textState.translateTextLineMatrix(args[0], args[1]);
|
|
|
|
|
textState.textMatrix = textState.textLineMatrix.slice();
|
|
|
|
|
break;
|
|
|
|
|
case OPS.nextLine:
|
|
|
|
|
flushTextContentItem();
|
|
|
|
|
textState.carriageReturn();
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setTextMatrix:
|
|
|
|
|
// Optimization to treat same line movement as advance.
|
|
|
|
|
advance = textState.calcTextLineMatrixAdvance(
|
|
|
|
|
args[0],
|
|
|
|
|
args[1],
|
|
|
|
|
args[2],
|
|
|
|
|
args[3],
|
|
|
|
|
args[4],
|
|
|
|
|
args[5]
|
|
|
|
|
);
|
|
|
|
|
if (
|
|
|
|
|
combineTextItems &&
|
|
|
|
|
advance !== null &&
|
|
|
|
|
textContentItem.initialized &&
|
|
|
|
|
advance.value > 0 &&
|
|
|
|
|
advance.value <= textContentItem.fakeMultiSpaceMax
|
|
|
|
|
) {
|
|
|
|
|
textState.translateTextLineMatrix(advance.width, advance.height);
|
|
|
|
|
textContentItem.width +=
|
|
|
|
|
advance.width - textContentItem.lastAdvanceWidth;
|
|
|
|
|
textContentItem.height +=
|
|
|
|
|
advance.height - textContentItem.lastAdvanceHeight;
|
|
|
|
|
diff =
|
|
|
|
|
advance.width -
|
|
|
|
|
textContentItem.lastAdvanceWidth -
|
|
|
|
|
(advance.height - textContentItem.lastAdvanceHeight);
|
|
|
|
|
addFakeSpaces(diff, textContentItem.str);
|
2014-05-10 10:21:15 +09:00
|
|
|
|
break;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
flushTextContentItem();
|
|
|
|
|
textState.setTextMatrix(
|
|
|
|
|
args[0],
|
|
|
|
|
args[1],
|
|
|
|
|
args[2],
|
|
|
|
|
args[3],
|
|
|
|
|
args[4],
|
|
|
|
|
args[5]
|
|
|
|
|
);
|
|
|
|
|
textState.setTextLineMatrix(
|
|
|
|
|
args[0],
|
|
|
|
|
args[1],
|
|
|
|
|
args[2],
|
|
|
|
|
args[3],
|
|
|
|
|
args[4],
|
|
|
|
|
args[5]
|
|
|
|
|
);
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setCharSpacing:
|
|
|
|
|
textState.charSpacing = args[0];
|
|
|
|
|
break;
|
|
|
|
|
case OPS.setWordSpacing:
|
|
|
|
|
textState.wordSpacing = args[0];
|
|
|
|
|
break;
|
|
|
|
|
case OPS.beginText:
|
|
|
|
|
flushTextContentItem();
|
|
|
|
|
textState.textMatrix = IDENTITY_MATRIX.slice();
|
|
|
|
|
textState.textLineMatrix = IDENTITY_MATRIX.slice();
|
|
|
|
|
break;
|
|
|
|
|
case OPS.showSpacedText:
|
|
|
|
|
if (!stateManager.state.font) {
|
|
|
|
|
self.ensureStateFont(stateManager.state);
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
var items = args[0];
|
|
|
|
|
var offset;
|
|
|
|
|
for (var j = 0, jj = items.length; j < jj; j++) {
|
|
|
|
|
if (typeof items[j] === "string") {
|
|
|
|
|
buildTextContentItem(items[j]);
|
|
|
|
|
} else if (isNum(items[j])) {
|
|
|
|
|
ensureTextContentItem();
|
|
|
|
|
|
|
|
|
|
// PDF Specification 5.3.2 states:
|
|
|
|
|
// The number is expressed in thousandths of a unit of text
|
|
|
|
|
// space.
|
|
|
|
|
// This amount is subtracted from the current horizontal or
|
|
|
|
|
// vertical coordinate, depending on the writing mode.
|
|
|
|
|
// In the default coordinate system, a positive adjustment
|
|
|
|
|
// has the effect of moving the next glyph painted either to
|
|
|
|
|
// the left or down by the given amount.
|
|
|
|
|
advance = (items[j] * textState.fontSize) / 1000;
|
|
|
|
|
var breakTextRun = false;
|
|
|
|
|
if (textState.font.vertical) {
|
|
|
|
|
offset = advance;
|
|
|
|
|
textState.translateTextMatrix(0, offset);
|
|
|
|
|
breakTextRun =
|
|
|
|
|
textContentItem.textRunBreakAllowed &&
|
|
|
|
|
advance > textContentItem.fakeMultiSpaceMax;
|
|
|
|
|
if (!breakTextRun) {
|
|
|
|
|
// Value needs to be added to height to paint down.
|
|
|
|
|
textContentItem.height += offset;
|
2014-05-10 10:21:15 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
} else {
|
|
|
|
|
advance = -advance;
|
|
|
|
|
offset = advance * textState.textHScale;
|
|
|
|
|
textState.translateTextMatrix(offset, 0);
|
|
|
|
|
breakTextRun =
|
|
|
|
|
textContentItem.textRunBreakAllowed &&
|
|
|
|
|
advance > textContentItem.fakeMultiSpaceMax;
|
|
|
|
|
if (!breakTextRun) {
|
|
|
|
|
// Value needs to be subtracted from width to paint left.
|
|
|
|
|
textContentItem.width += offset;
|
2012-09-15 02:53:06 +09:00
|
|
|
|
}
|
2013-06-05 09:57:52 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (breakTextRun) {
|
|
|
|
|
flushTextContentItem();
|
|
|
|
|
} else if (advance > 0) {
|
|
|
|
|
addFakeSpaces(advance, textContentItem.str);
|
|
|
|
|
}
|
2013-08-01 03:17:36 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
break;
|
|
|
|
|
case OPS.showText:
|
|
|
|
|
if (!stateManager.state.font) {
|
|
|
|
|
self.ensureStateFont(stateManager.state);
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
buildTextContentItem(args[0]);
|
|
|
|
|
break;
|
|
|
|
|
case OPS.nextLineShowText:
|
|
|
|
|
if (!stateManager.state.font) {
|
|
|
|
|
self.ensureStateFont(stateManager.state);
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
flushTextContentItem();
|
|
|
|
|
textState.carriageReturn();
|
|
|
|
|
buildTextContentItem(args[0]);
|
|
|
|
|
break;
|
|
|
|
|
case OPS.nextLineSetSpacingShowText:
|
|
|
|
|
if (!stateManager.state.font) {
|
|
|
|
|
self.ensureStateFont(stateManager.state);
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
flushTextContentItem();
|
|
|
|
|
textState.wordSpacing = args[0];
|
|
|
|
|
textState.charSpacing = args[1];
|
|
|
|
|
textState.carriageReturn();
|
|
|
|
|
buildTextContentItem(args[2]);
|
|
|
|
|
break;
|
|
|
|
|
case OPS.paintXObject:
|
|
|
|
|
flushTextContentItem();
|
|
|
|
|
if (!xobjs) {
|
|
|
|
|
xobjs = resources.get("XObject") || Dict.empty;
|
|
|
|
|
}
|
[api-minor] Always allow e.g. rendering to continue even if there are errors, and add a `stopAtErrors` parameter to `getDocument` to opt-out of this behaviour (issue 6342, issue 3795, bug 1130815)
Other PDF readers, e.g. Adobe Reader and PDFium (in Chrome), will attempt to render as much of a page as possible even if there are errors present.
Currently we just bail as soon the first error is hit, which means that we'll usually not render anything in these cases and just display a blank page instead.
NOTE: This patch changes the default behaviour of the PDF.js API to always attempt to recover as much data as possible, even when encountering errors during e.g. `getOperatorList`/`getTextContent`, which thus improve our handling of corrupt PDF files and allow the default viewer to handle errors slightly more gracefully.
In the event that an API consumer wishes to use the old behaviour, where we stop parsing as soon as an error is encountered, the `stopAtErrors` parameter can be set at `getDocument`.
Fixes, inasmuch it's possible since the PDF files are corrupt, e.g. issue 6342, issue 3795, and [bug 1130815](https://bugzilla.mozilla.org/show_bug.cgi?id=1130815) (and probably others too).
2017-02-19 22:03:08 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var name = args[0].name;
|
|
|
|
|
if (name && emptyXObjectCache.getByName(name)) {
|
|
|
|
|
break;
|
|
|
|
|
}
|
2020-05-26 16:47:59 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
next(
|
|
|
|
|
new Promise(function (resolveXObject, rejectXObject) {
|
|
|
|
|
if (!name) {
|
|
|
|
|
throw new FormatError("XObject must be referred to by name.");
|
|
|
|
|
}
|
2020-05-26 16:47:59 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
let xobj = xobjs.getRaw(name);
|
|
|
|
|
if (xobj instanceof Ref) {
|
|
|
|
|
if (emptyXObjectCache.getByRef(xobj)) {
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
resolveXObject();
|
|
|
|
|
return;
|
|
|
|
|
}
|
2014-04-10 08:44:07 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
xobj = xref.fetch(xobj);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (!isStream(xobj)) {
|
|
|
|
|
throw new FormatError("XObject should be a stream");
|
|
|
|
|
}
|
2017-09-17 20:35:18 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
const type = xobj.dict.get("Subtype");
|
|
|
|
|
if (!isName(type)) {
|
|
|
|
|
throw new FormatError("XObject should have a Name subtype");
|
|
|
|
|
}
|
2020-05-26 16:47:59 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (type.name !== "Form") {
|
|
|
|
|
emptyXObjectCache.set(name, xobj.dict.objId, true);
|
2017-09-17 20:35:18 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
resolveXObject();
|
|
|
|
|
return;
|
|
|
|
|
}
|
2017-09-17 20:35:18 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Use a new `StateManager` to prevent incorrect positioning
|
|
|
|
|
// of textItems *after* the Form XObject, since errors in the
|
|
|
|
|
// data can otherwise prevent `restore` operators from
|
|
|
|
|
// executing.
|
|
|
|
|
// NOTE: Only an issue when `options.ignoreErrors === true`.
|
|
|
|
|
const currentState = stateManager.state.clone();
|
|
|
|
|
const xObjStateManager = new StateManager(currentState);
|
|
|
|
|
|
|
|
|
|
const matrix = xobj.dict.getArray("Matrix");
|
|
|
|
|
if (Array.isArray(matrix) && matrix.length === 6) {
|
|
|
|
|
xObjStateManager.transform(matrix);
|
|
|
|
|
}
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Enqueue the `textContent` chunk before parsing the /Form
|
|
|
|
|
// XObject.
|
|
|
|
|
enqueueChunk();
|
|
|
|
|
const sinkWrapper = {
|
|
|
|
|
enqueueInvoked: false,
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
enqueue(chunk, size) {
|
|
|
|
|
this.enqueueInvoked = true;
|
|
|
|
|
sink.enqueue(chunk, size);
|
|
|
|
|
},
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
get desiredSize() {
|
|
|
|
|
return sink.desiredSize;
|
|
|
|
|
},
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
get ready() {
|
|
|
|
|
return sink.ready;
|
|
|
|
|
},
|
|
|
|
|
};
|
2014-03-23 03:15:51 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
self
|
|
|
|
|
.getTextContent({
|
|
|
|
|
stream: xobj,
|
|
|
|
|
task,
|
|
|
|
|
resources: xobj.dict.get("Resources") || resources,
|
|
|
|
|
stateManager: xObjStateManager,
|
|
|
|
|
normalizeWhitespace,
|
|
|
|
|
combineTextItems,
|
|
|
|
|
sink: sinkWrapper,
|
|
|
|
|
seenStyles,
|
|
|
|
|
})
|
|
|
|
|
.then(function () {
|
|
|
|
|
if (!sinkWrapper.enqueueInvoked) {
|
|
|
|
|
emptyXObjectCache.set(name, xobj.dict.objId, true);
|
|
|
|
|
}
|
|
|
|
|
resolveXObject();
|
|
|
|
|
}, rejectXObject);
|
|
|
|
|
}).catch(function (reason) {
|
|
|
|
|
if (reason instanceof AbortException) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
if (self.options.ignoreErrors) {
|
|
|
|
|
// Error(s) in the XObject -- allow text-extraction to
|
|
|
|
|
// continue.
|
|
|
|
|
warn(`getTextContent - ignoring XObject: "${reason}".`);
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
throw reason;
|
|
|
|
|
})
|
|
|
|
|
);
|
|
|
|
|
return;
|
|
|
|
|
case OPS.setGState:
|
2020-07-11 21:05:53 +09:00
|
|
|
|
name = args[0].name;
|
|
|
|
|
if (name && emptyGStateCache.getByName(name)) {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
break;
|
|
|
|
|
}
|
2020-07-11 21:05:53 +09:00
|
|
|
|
|
|
|
|
|
next(
|
|
|
|
|
new Promise(function (resolveGState, rejectGState) {
|
|
|
|
|
if (!name) {
|
|
|
|
|
throw new FormatError("GState must be referred to by name.");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
const extGState = resources.get("ExtGState");
|
|
|
|
|
if (!(extGState instanceof Dict)) {
|
|
|
|
|
throw new FormatError("ExtGState should be a dictionary.");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
const gState = extGState.get(name);
|
|
|
|
|
// TODO: Attempt to lookup cached GStates by reference as well,
|
|
|
|
|
// if and only if there are PDF documents where doing so
|
|
|
|
|
// would significantly improve performance.
|
|
|
|
|
if (!(gState instanceof Dict)) {
|
|
|
|
|
throw new FormatError("GState should be a dictionary.");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
const gStateFont = gState.get("Font");
|
|
|
|
|
if (!gStateFont) {
|
|
|
|
|
emptyGStateCache.set(name, gState.objId, true);
|
|
|
|
|
|
|
|
|
|
resolveGState();
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
flushTextContentItem();
|
|
|
|
|
|
|
|
|
|
textState.fontName = null;
|
|
|
|
|
textState.fontSize = gStateFont[1];
|
|
|
|
|
handleSetFont(null, gStateFont[0]).then(
|
|
|
|
|
resolveGState,
|
|
|
|
|
rejectGState
|
|
|
|
|
);
|
|
|
|
|
}).catch(function (reason) {
|
|
|
|
|
if (reason instanceof AbortException) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
if (self.options.ignoreErrors) {
|
|
|
|
|
// Error(s) in the ExtGState -- allow text-extraction to
|
|
|
|
|
// continue.
|
|
|
|
|
warn(`getTextContent - ignoring ExtGState: "${reason}".`);
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
throw reason;
|
|
|
|
|
})
|
|
|
|
|
);
|
|
|
|
|
return;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
} // switch
|
|
|
|
|
if (textContent.items.length >= sink.desiredSize) {
|
|
|
|
|
// Wait for ready, if we reach highWaterMark.
|
|
|
|
|
stop = true;
|
|
|
|
|
break;
|
2014-05-10 10:41:03 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
} // while
|
|
|
|
|
if (stop) {
|
|
|
|
|
next(deferred);
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
flushTextContentItem();
|
|
|
|
|
enqueueChunk();
|
|
|
|
|
resolve();
|
|
|
|
|
}).catch(reason => {
|
|
|
|
|
if (reason instanceof AbortException) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
if (this.options.ignoreErrors) {
|
|
|
|
|
// Error(s) in the TextContent -- allow text-extraction to continue.
|
|
|
|
|
warn(
|
|
|
|
|
`getTextContent - ignoring errors during "${task.name}" ` +
|
|
|
|
|
`task: "${reason}".`
|
|
|
|
|
);
|
|
|
|
|
|
2015-11-04 01:12:41 +09:00
|
|
|
|
flushTextContentItem();
|
2017-04-17 21:46:53 +09:00
|
|
|
|
enqueueChunk();
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
throw reason;
|
|
|
|
|
});
|
|
|
|
|
}
|
2011-12-11 08:24:54 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
extractDataStructures(dict, baseDict, properties) {
|
|
|
|
|
const xref = this.xref;
|
|
|
|
|
let cidToGidBytes;
|
|
|
|
|
// 9.10.2
|
|
|
|
|
var toUnicode = dict.get("ToUnicode") || baseDict.get("ToUnicode");
|
|
|
|
|
var toUnicodePromise = toUnicode
|
|
|
|
|
? this.readToUnicode(toUnicode)
|
|
|
|
|
: Promise.resolve(undefined);
|
|
|
|
|
|
|
|
|
|
if (properties.composite) {
|
|
|
|
|
// CIDSystemInfo helps to match CID to glyphs
|
|
|
|
|
var cidSystemInfo = dict.get("CIDSystemInfo");
|
|
|
|
|
if (isDict(cidSystemInfo)) {
|
|
|
|
|
properties.cidSystemInfo = {
|
|
|
|
|
registry: stringToPDFString(cidSystemInfo.get("Registry")),
|
|
|
|
|
ordering: stringToPDFString(cidSystemInfo.get("Ordering")),
|
|
|
|
|
supplement: cidSystemInfo.get("Supplement"),
|
|
|
|
|
};
|
|
|
|
|
}
|
2011-10-25 08:55:23 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var cidToGidMap = dict.get("CIDToGIDMap");
|
|
|
|
|
if (isStream(cidToGidMap)) {
|
|
|
|
|
cidToGidBytes = cidToGidMap.getBytes();
|
2011-10-25 08:55:23 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
2011-10-25 08:55:23 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Based on 9.6.6 of the spec the encoding can come from multiple places
|
|
|
|
|
// and depends on the font type. The base encoding and differences are
|
|
|
|
|
// read here, but the encoding that is actually used is chosen during
|
|
|
|
|
// glyph mapping in the font.
|
|
|
|
|
// TODO: Loading the built in encoding in the font would allow the
|
|
|
|
|
// differences to be merged in here not require us to hold on to it.
|
|
|
|
|
var differences = [];
|
|
|
|
|
var baseEncodingName = null;
|
|
|
|
|
var encoding;
|
|
|
|
|
if (dict.has("Encoding")) {
|
|
|
|
|
encoding = dict.get("Encoding");
|
|
|
|
|
if (isDict(encoding)) {
|
|
|
|
|
baseEncodingName = encoding.get("BaseEncoding");
|
|
|
|
|
baseEncodingName = isName(baseEncodingName)
|
|
|
|
|
? baseEncodingName.name
|
|
|
|
|
: null;
|
|
|
|
|
// Load the differences between the base and original
|
|
|
|
|
if (encoding.has("Differences")) {
|
|
|
|
|
var diffEncoding = encoding.get("Differences");
|
|
|
|
|
var index = 0;
|
|
|
|
|
for (var j = 0, jj = diffEncoding.length; j < jj; j++) {
|
|
|
|
|
var data = xref.fetchIfRef(diffEncoding[j]);
|
|
|
|
|
if (isNum(data)) {
|
|
|
|
|
index = data;
|
|
|
|
|
} else if (isName(data)) {
|
|
|
|
|
differences[index++] = data.name;
|
|
|
|
|
} else {
|
|
|
|
|
throw new FormatError(
|
|
|
|
|
`Invalid entry in 'Differences' array: ${data}`
|
|
|
|
|
);
|
2011-10-25 08:55:23 +09:00
|
|
|
|
}
|
|
|
|
|
}
|
2014-02-12 03:27:09 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
} else if (isName(encoding)) {
|
|
|
|
|
baseEncodingName = encoding.name;
|
|
|
|
|
} else {
|
|
|
|
|
throw new FormatError("Encoding is not a Name nor a Dict");
|
|
|
|
|
}
|
|
|
|
|
// According to table 114 if the encoding is a named encoding it must be
|
|
|
|
|
// one of these predefined encodings.
|
|
|
|
|
if (
|
|
|
|
|
baseEncodingName !== "MacRomanEncoding" &&
|
|
|
|
|
baseEncodingName !== "MacExpertEncoding" &&
|
|
|
|
|
baseEncodingName !== "WinAnsiEncoding"
|
|
|
|
|
) {
|
|
|
|
|
baseEncodingName = null;
|
2014-02-12 03:27:09 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
2014-02-12 03:27:09 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (baseEncodingName) {
|
|
|
|
|
properties.defaultEncoding = getEncoding(baseEncodingName).slice();
|
|
|
|
|
} else {
|
|
|
|
|
var isSymbolicFont = !!(properties.flags & FontFlags.Symbolic);
|
|
|
|
|
var isNonsymbolicFont = !!(properties.flags & FontFlags.Nonsymbolic);
|
|
|
|
|
// According to "Table 114" in section "9.6.6.1 General" (under
|
|
|
|
|
// "9.6.6 Character Encoding") of the PDF specification, a Nonsymbolic
|
|
|
|
|
// font should use the `StandardEncoding` if no encoding is specified.
|
|
|
|
|
encoding = StandardEncoding;
|
|
|
|
|
if (properties.type === "TrueType" && !isNonsymbolicFont) {
|
|
|
|
|
encoding = WinAnsiEncoding;
|
|
|
|
|
}
|
|
|
|
|
// The Symbolic attribute can be misused for regular fonts
|
|
|
|
|
// Heuristic: we have to check if the font is a standard one also
|
|
|
|
|
if (isSymbolicFont) {
|
|
|
|
|
encoding = MacRomanEncoding;
|
|
|
|
|
if (!properties.file) {
|
|
|
|
|
if (/Symbol/i.test(properties.name)) {
|
|
|
|
|
encoding = SymbolSetEncoding;
|
|
|
|
|
} else if (/Dingbats|Wingdings/i.test(properties.name)) {
|
|
|
|
|
encoding = ZapfDingbatsEncoding;
|
2014-09-01 10:22:24 +09:00
|
|
|
|
}
|
2014-02-12 03:27:09 +09:00
|
|
|
|
}
|
2011-10-25 08:55:23 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
properties.defaultEncoding = encoding;
|
|
|
|
|
}
|
2011-11-25 00:38:09 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
properties.differences = differences;
|
|
|
|
|
properties.baseEncodingName = baseEncodingName;
|
|
|
|
|
properties.hasEncoding = !!baseEncodingName || differences.length > 0;
|
|
|
|
|
properties.dict = dict;
|
|
|
|
|
return toUnicodePromise
|
|
|
|
|
.then(readToUnicode => {
|
|
|
|
|
properties.toUnicode = readToUnicode;
|
|
|
|
|
return this.buildToUnicode(properties);
|
|
|
|
|
})
|
|
|
|
|
.then(builtToUnicode => {
|
|
|
|
|
properties.toUnicode = builtToUnicode;
|
|
|
|
|
if (cidToGidBytes) {
|
|
|
|
|
properties.cidToGidMap = this.readCidToGidMap(
|
|
|
|
|
cidToGidBytes,
|
|
|
|
|
builtToUnicode
|
|
|
|
|
);
|
2017-11-26 20:53:06 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return properties;
|
|
|
|
|
});
|
|
|
|
|
}
|
2019-09-30 06:50:58 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
/**
|
|
|
|
|
* @returns {ToUnicodeMap}
|
|
|
|
|
* @private
|
|
|
|
|
*/
|
|
|
|
|
_buildSimpleFontToUnicode(properties, forceGlyphs = false) {
|
|
|
|
|
assert(!properties.composite, "Must be a simple font.");
|
|
|
|
|
|
|
|
|
|
const toUnicode = [];
|
|
|
|
|
const encoding = properties.defaultEncoding.slice();
|
|
|
|
|
const baseEncodingName = properties.baseEncodingName;
|
|
|
|
|
// Merge in the differences array.
|
|
|
|
|
const differences = properties.differences;
|
|
|
|
|
for (const charcode in differences) {
|
|
|
|
|
const glyphName = differences[charcode];
|
|
|
|
|
if (glyphName === ".notdef") {
|
|
|
|
|
// Skip .notdef to prevent rendering errors, e.g. boxes appearing
|
|
|
|
|
// where there should be spaces (fixes issue5256.pdf).
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
encoding[charcode] = glyphName;
|
|
|
|
|
}
|
|
|
|
|
const glyphsUnicodeMap = getGlyphsUnicode();
|
|
|
|
|
for (const charcode in encoding) {
|
|
|
|
|
// a) Map the character code to a character name.
|
|
|
|
|
let glyphName = encoding[charcode];
|
|
|
|
|
// b) Look up the character name in the Adobe Glyph List (see the
|
|
|
|
|
// Bibliography) to obtain the corresponding Unicode value.
|
|
|
|
|
if (glyphName === "") {
|
|
|
|
|
continue;
|
|
|
|
|
} else if (glyphsUnicodeMap[glyphName] === undefined) {
|
|
|
|
|
// (undocumented) c) Few heuristics to recognize unknown glyphs
|
|
|
|
|
// NOTE: Adobe Reader does not do this step, but OSX Preview does
|
|
|
|
|
let code = 0;
|
|
|
|
|
switch (glyphName[0]) {
|
|
|
|
|
case "G": // Gxx glyph
|
|
|
|
|
if (glyphName.length === 3) {
|
|
|
|
|
code = parseInt(glyphName.substring(1), 16);
|
|
|
|
|
}
|
|
|
|
|
break;
|
|
|
|
|
case "g": // g00xx glyph
|
|
|
|
|
if (glyphName.length === 5) {
|
|
|
|
|
code = parseInt(glyphName.substring(1), 16);
|
|
|
|
|
}
|
|
|
|
|
break;
|
|
|
|
|
case "C": // Cdd{d} glyph
|
|
|
|
|
case "c": // cdd{d} glyph
|
|
|
|
|
if (glyphName.length >= 3 && glyphName.length <= 4) {
|
|
|
|
|
const codeStr = glyphName.substring(1);
|
|
|
|
|
|
|
|
|
|
if (forceGlyphs) {
|
|
|
|
|
code = parseInt(codeStr, 16);
|
|
|
|
|
break;
|
2017-11-26 20:53:06 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Normally the Cdd{d}/cdd{d} glyphName format will contain
|
|
|
|
|
// regular, i.e. base 10, charCodes (see issue4550.pdf)...
|
|
|
|
|
code = +codeStr;
|
|
|
|
|
|
|
|
|
|
// ... however some PDF generators violate that assumption by
|
|
|
|
|
// containing glyph, i.e. base 16, codes instead.
|
|
|
|
|
// In that case we need to re-parse the *entire* encoding to
|
|
|
|
|
// prevent broken text-selection (fixes issue9655_reduced.pdf).
|
|
|
|
|
if (
|
|
|
|
|
Number.isNaN(code) &&
|
|
|
|
|
Number.isInteger(parseInt(codeStr, 16))
|
|
|
|
|
) {
|
|
|
|
|
return this._buildSimpleFontToUnicode(
|
|
|
|
|
properties,
|
|
|
|
|
/* forceGlyphs */ true
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
);
|
2017-11-26 20:53:06 +09:00
|
|
|
|
}
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
break;
|
|
|
|
|
default:
|
|
|
|
|
// 'uniXXXX'/'uXXXX{XX}' glyphs
|
|
|
|
|
const unicode = getUnicodeForGlyph(glyphName, glyphsUnicodeMap);
|
|
|
|
|
if (unicode !== -1) {
|
|
|
|
|
code = unicode;
|
|
|
|
|
}
|
2017-11-26 20:53:06 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (code > 0 && code <= 0x10ffff && Number.isInteger(code)) {
|
|
|
|
|
// If `baseEncodingName` is one the predefined encodings, and `code`
|
|
|
|
|
// equals `charcode`, using the glyph defined in the baseEncoding
|
|
|
|
|
// seems to yield a better `toUnicode` mapping (fixes issue 5070).
|
|
|
|
|
if (baseEncodingName && code === +charcode) {
|
|
|
|
|
const baseEncoding = getEncoding(baseEncodingName);
|
|
|
|
|
if (baseEncoding && (glyphName = baseEncoding[charcode])) {
|
|
|
|
|
toUnicode[charcode] = String.fromCharCode(
|
|
|
|
|
glyphsUnicodeMap[glyphName]
|
|
|
|
|
);
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
toUnicode[charcode] = String.fromCodePoint(code);
|
Build a fallback `ToUnicode` map for simple fonts (issue 8229)
In some fonts, the included `ToUnicode` data is incomplete causing text-selection to not work properly. For simple fonts that contain encoding data, we can manually build a `ToUnicode` map to attempt to improve things.
Please note that since we're currently using the `ToUnicode` data during glyph mapping, in an attempt to avoid rendering regressions, I purposely didn't want to amend to original `ToUnicode` data for this text-selection edge-case.
Instead, I opted for the current solution, which will (hopefully) give slightly better text-extraction results in PDF file with incomplete `ToUnicode` data.
According to the PDF specification, see [section 9.10.2](http://www.adobe.com/content/dam/acom/en/devnet/acrobat/pdfs/PDF32000_2008.pdf#G8.1873172):
> A conforming reader can use these methods, in the priority given, to map a character code to a Unicode value.
> ...
Reading that paragraph literally, it doesn't seem too unreasonable to use *different* methods for different charcodes.
Fixes 8229.
2017-11-26 21:29:43 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
continue;
|
2016-02-29 01:20:29 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
toUnicode[charcode] = String.fromCharCode(glyphsUnicodeMap[glyphName]);
|
|
|
|
|
}
|
|
|
|
|
return new ToUnicodeMap(toUnicode);
|
|
|
|
|
}
|
2017-11-26 20:53:06 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
/**
|
|
|
|
|
* Builds a char code to unicode map based on section 9.10 of the spec.
|
|
|
|
|
* @param {Object} properties Font properties object.
|
|
|
|
|
* @returns {Promise} A Promise that is resolved with a
|
|
|
|
|
* {ToUnicodeMap|IdentityToUnicodeMap} object.
|
|
|
|
|
*/
|
|
|
|
|
buildToUnicode(properties) {
|
|
|
|
|
properties.hasIncludedToUnicodeMap =
|
|
|
|
|
!!properties.toUnicode && properties.toUnicode.length > 0;
|
|
|
|
|
|
|
|
|
|
// Section 9.10.2 Mapping Character Codes to Unicode Values
|
|
|
|
|
if (properties.hasIncludedToUnicodeMap) {
|
|
|
|
|
// Some fonts contain incomplete ToUnicode data, causing issues with
|
|
|
|
|
// text-extraction. For simple fonts, containing encoding information,
|
|
|
|
|
// use a fallback ToUnicode map to improve this (fixes issue8229.pdf).
|
|
|
|
|
if (!properties.composite && properties.hasEncoding) {
|
|
|
|
|
properties.fallbackToUnicode = this._buildSimpleFontToUnicode(
|
|
|
|
|
properties
|
|
|
|
|
);
|
2016-02-29 01:20:29 +09:00
|
|
|
|
}
|
2017-11-26 20:53:06 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return Promise.resolve(properties.toUnicode);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// According to the spec if the font is a simple font we should only map
|
|
|
|
|
// to unicode if the base encoding is MacRoman, MacExpert, or WinAnsi or
|
|
|
|
|
// the differences array only contains adobe standard or symbol set names,
|
|
|
|
|
// in pratice it seems better to always try to create a toUnicode map
|
|
|
|
|
// based of the default encoding.
|
|
|
|
|
if (!properties.composite /* is simple font */) {
|
|
|
|
|
return Promise.resolve(this._buildSimpleFontToUnicode(properties));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// If the font is a composite font that uses one of the predefined CMaps
|
|
|
|
|
// listed in Table 118 (except Identity–H and Identity–V) or whose
|
|
|
|
|
// descendant CIDFont uses the Adobe-GB1, Adobe-CNS1, Adobe-Japan1, or
|
|
|
|
|
// Adobe-Korea1 character collection:
|
|
|
|
|
if (
|
|
|
|
|
properties.composite &&
|
|
|
|
|
((properties.cMap.builtInCMap &&
|
|
|
|
|
!(properties.cMap instanceof IdentityCMap)) ||
|
|
|
|
|
(properties.cidSystemInfo.registry === "Adobe" &&
|
|
|
|
|
(properties.cidSystemInfo.ordering === "GB1" ||
|
|
|
|
|
properties.cidSystemInfo.ordering === "CNS1" ||
|
|
|
|
|
properties.cidSystemInfo.ordering === "Japan1" ||
|
|
|
|
|
properties.cidSystemInfo.ordering === "Korea1")))
|
|
|
|
|
) {
|
|
|
|
|
// Then:
|
|
|
|
|
// a) Map the character code to a character identifier (CID) according
|
|
|
|
|
// to the font’s CMap.
|
|
|
|
|
// b) Obtain the registry and ordering of the character collection used
|
|
|
|
|
// by the font’s CMap (for example, Adobe and Japan1) from its
|
|
|
|
|
// CIDSystemInfo dictionary.
|
|
|
|
|
const registry = properties.cidSystemInfo.registry;
|
|
|
|
|
const ordering = properties.cidSystemInfo.ordering;
|
|
|
|
|
// c) Construct a second CMap name by concatenating the registry and
|
|
|
|
|
// ordering obtained in step (b) in the format registry–ordering–UCS2
|
|
|
|
|
// (for example, Adobe–Japan1–UCS2).
|
|
|
|
|
const ucs2CMapName = Name.get(registry + "-" + ordering + "-UCS2");
|
|
|
|
|
// d) Obtain the CMap with the name constructed in step (c) (available
|
|
|
|
|
// from the ASN Web site; see the Bibliography).
|
|
|
|
|
return CMapFactory.create({
|
|
|
|
|
encoding: ucs2CMapName,
|
|
|
|
|
fetchBuiltInCMap: this._fetchBuiltInCMapBound,
|
|
|
|
|
useCMap: null,
|
|
|
|
|
}).then(function (ucs2CMap) {
|
|
|
|
|
const cMap = properties.cMap;
|
|
|
|
|
const toUnicode = [];
|
|
|
|
|
cMap.forEach(function (charcode, cid) {
|
|
|
|
|
if (cid > 0xffff) {
|
|
|
|
|
throw new FormatError("Max size of CID is 65,535");
|
|
|
|
|
}
|
|
|
|
|
// e) Map the CID obtained in step (a) according to the CMap
|
|
|
|
|
// obtained in step (d), producing a Unicode value.
|
|
|
|
|
const ucs2 = ucs2CMap.lookup(cid);
|
|
|
|
|
if (ucs2) {
|
|
|
|
|
toUnicode[charcode] = String.fromCharCode(
|
|
|
|
|
(ucs2.charCodeAt(0) << 8) + ucs2.charCodeAt(1)
|
|
|
|
|
);
|
|
|
|
|
}
|
2016-02-29 01:20:29 +09:00
|
|
|
|
});
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return new ToUnicodeMap(toUnicode);
|
|
|
|
|
});
|
|
|
|
|
}
|
2016-02-29 01:20:29 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// The viewer's choice, just use an identity map.
|
|
|
|
|
return Promise.resolve(
|
|
|
|
|
new IdentityToUnicodeMap(properties.firstChar, properties.lastChar)
|
|
|
|
|
);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
readToUnicode(toUnicode) {
|
|
|
|
|
var cmapObj = toUnicode;
|
|
|
|
|
if (isName(cmapObj)) {
|
|
|
|
|
return CMapFactory.create({
|
|
|
|
|
encoding: cmapObj,
|
|
|
|
|
fetchBuiltInCMap: this._fetchBuiltInCMapBound,
|
|
|
|
|
useCMap: null,
|
|
|
|
|
}).then(function (cmap) {
|
|
|
|
|
if (cmap instanceof IdentityCMap) {
|
|
|
|
|
return new IdentityToUnicodeMap(0, 0xffff);
|
|
|
|
|
}
|
|
|
|
|
return new ToUnicodeMap(cmap.getMap());
|
|
|
|
|
});
|
|
|
|
|
} else if (isStream(cmapObj)) {
|
|
|
|
|
return CMapFactory.create({
|
|
|
|
|
encoding: cmapObj,
|
|
|
|
|
fetchBuiltInCMap: this._fetchBuiltInCMapBound,
|
|
|
|
|
useCMap: null,
|
|
|
|
|
}).then(
|
|
|
|
|
function (cmap) {
|
2016-02-29 01:20:29 +09:00
|
|
|
|
if (cmap instanceof IdentityCMap) {
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
return new IdentityToUnicodeMap(0, 0xffff);
|
2016-02-29 01:20:29 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var map = new Array(cmap.length);
|
|
|
|
|
// Convert UTF-16BE
|
|
|
|
|
// NOTE: cmap can be a sparse array, so use forEach instead of
|
|
|
|
|
// `for(;;)` to iterate over all keys.
|
|
|
|
|
cmap.forEach(function (charCode, token) {
|
|
|
|
|
var str = [];
|
|
|
|
|
for (var k = 0; k < token.length; k += 2) {
|
|
|
|
|
var w1 = (token.charCodeAt(k) << 8) | token.charCodeAt(k + 1);
|
|
|
|
|
if ((w1 & 0xf800) !== 0xd800) {
|
|
|
|
|
// w1 < 0xD800 || w1 > 0xDFFF
|
|
|
|
|
str.push(w1);
|
|
|
|
|
continue;
|
2016-02-29 01:20:29 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
k += 2;
|
|
|
|
|
var w2 = (token.charCodeAt(k) << 8) | token.charCodeAt(k + 1);
|
|
|
|
|
str.push(((w1 & 0x3ff) << 10) + (w2 & 0x3ff) + 0x10000);
|
2020-01-30 21:13:51 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
map[charCode] = String.fromCodePoint.apply(String, str);
|
|
|
|
|
});
|
|
|
|
|
return new ToUnicodeMap(map);
|
|
|
|
|
},
|
|
|
|
|
reason => {
|
|
|
|
|
if (reason instanceof AbortException) {
|
|
|
|
|
return null;
|
|
|
|
|
}
|
|
|
|
|
if (this.options.ignoreErrors) {
|
|
|
|
|
// Error in the ToUnicode data -- sending unsupported feature
|
|
|
|
|
// notification and allow font parsing to continue.
|
|
|
|
|
this.handler.send("UnsupportedFeature", {
|
|
|
|
|
featureId: UNSUPPORTED_FEATURES.errorFontToUnicode,
|
|
|
|
|
});
|
|
|
|
|
warn(`readToUnicode - ignoring ToUnicode data: "${reason}".`);
|
|
|
|
|
return null;
|
2020-01-30 21:13:51 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
throw reason;
|
|
|
|
|
}
|
|
|
|
|
);
|
|
|
|
|
}
|
|
|
|
|
return Promise.resolve(null);
|
|
|
|
|
}
|
2014-03-23 03:15:51 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
readCidToGidMap(glyphsData, toUnicode) {
|
|
|
|
|
// Extract the encoding from the CIDToGIDMap
|
2011-10-29 10:38:31 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Set encoding 0 to later verify the font has an encoding
|
|
|
|
|
var result = [];
|
|
|
|
|
for (var j = 0, jj = glyphsData.length; j < jj; j++) {
|
|
|
|
|
var glyphID = (glyphsData[j++] << 8) | glyphsData[j];
|
|
|
|
|
const code = j >> 1;
|
|
|
|
|
if (glyphID === 0 && !toUnicode.has(code)) {
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
result[code] = glyphID;
|
|
|
|
|
}
|
|
|
|
|
return result;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
extractWidths(dict, descriptor, properties) {
|
|
|
|
|
var xref = this.xref;
|
|
|
|
|
var glyphsWidths = [];
|
|
|
|
|
var defaultWidth = 0;
|
|
|
|
|
var glyphsVMetrics = [];
|
|
|
|
|
var defaultVMetrics;
|
|
|
|
|
var i, ii, j, jj, start, code, widths;
|
|
|
|
|
if (properties.composite) {
|
|
|
|
|
defaultWidth = dict.has("DW") ? dict.get("DW") : 1000;
|
|
|
|
|
|
|
|
|
|
widths = dict.get("W");
|
|
|
|
|
if (widths) {
|
|
|
|
|
for (i = 0, ii = widths.length; i < ii; i++) {
|
|
|
|
|
start = xref.fetchIfRef(widths[i++]);
|
|
|
|
|
code = xref.fetchIfRef(widths[i]);
|
|
|
|
|
if (Array.isArray(code)) {
|
|
|
|
|
for (j = 0, jj = code.length; j < jj; j++) {
|
|
|
|
|
glyphsWidths[start++] = xref.fetchIfRef(code[j]);
|
|
|
|
|
}
|
|
|
|
|
} else {
|
|
|
|
|
var width = xref.fetchIfRef(widths[++i]);
|
|
|
|
|
for (j = start; j <= code; j++) {
|
|
|
|
|
glyphsWidths[j] = width;
|
|
|
|
|
}
|
|
|
|
|
}
|
2014-03-23 03:15:51 +09:00
|
|
|
|
}
|
2011-10-25 08:55:23 +09:00
|
|
|
|
}
|
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (properties.vertical) {
|
|
|
|
|
var vmetrics = dict.getArray("DW2") || [880, -1000];
|
|
|
|
|
defaultVMetrics = [vmetrics[1], defaultWidth * 0.5, vmetrics[0]];
|
|
|
|
|
vmetrics = dict.get("W2");
|
|
|
|
|
if (vmetrics) {
|
|
|
|
|
for (i = 0, ii = vmetrics.length; i < ii; i++) {
|
|
|
|
|
start = xref.fetchIfRef(vmetrics[i++]);
|
|
|
|
|
code = xref.fetchIfRef(vmetrics[i]);
|
2017-09-02 03:27:13 +09:00
|
|
|
|
if (Array.isArray(code)) {
|
2014-04-08 06:42:54 +09:00
|
|
|
|
for (j = 0, jj = code.length; j < jj; j++) {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
glyphsVMetrics[start++] = [
|
|
|
|
|
xref.fetchIfRef(code[j++]),
|
|
|
|
|
xref.fetchIfRef(code[j++]),
|
|
|
|
|
xref.fetchIfRef(code[j]),
|
|
|
|
|
];
|
2014-03-23 03:15:51 +09:00
|
|
|
|
}
|
2013-01-30 07:19:08 +09:00
|
|
|
|
} else {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var vmetric = [
|
|
|
|
|
xref.fetchIfRef(vmetrics[++i]),
|
|
|
|
|
xref.fetchIfRef(vmetrics[++i]),
|
|
|
|
|
xref.fetchIfRef(vmetrics[++i]),
|
|
|
|
|
];
|
2014-04-08 06:42:54 +09:00
|
|
|
|
for (j = start; j <= code; j++) {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
glyphsVMetrics[j] = vmetric;
|
2014-03-23 03:15:51 +09:00
|
|
|
|
}
|
2011-10-29 10:38:31 +09:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
} else {
|
|
|
|
|
var firstChar = properties.firstChar;
|
|
|
|
|
widths = dict.get("Widths");
|
|
|
|
|
if (widths) {
|
|
|
|
|
j = firstChar;
|
|
|
|
|
for (i = 0, ii = widths.length; i < ii; i++) {
|
|
|
|
|
glyphsWidths[j++] = xref.fetchIfRef(widths[i]);
|
2013-02-08 21:29:22 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
defaultWidth = parseFloat(descriptor.get("MissingWidth")) || 0;
|
2011-10-29 10:38:31 +09:00
|
|
|
|
} else {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Trying get the BaseFont metrics (see comment above).
|
|
|
|
|
var baseFontName = dict.get("BaseFont");
|
|
|
|
|
if (isName(baseFontName)) {
|
|
|
|
|
var metrics = this.getBaseFontMetrics(baseFontName.name);
|
|
|
|
|
|
|
|
|
|
glyphsWidths = this.buildCharCodeToWidth(metrics.widths, properties);
|
|
|
|
|
defaultWidth = metrics.defaultWidth;
|
2011-10-29 10:38:31 +09:00
|
|
|
|
}
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
2011-10-29 10:38:31 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Heuristic: detection of monospace font by checking all non-zero widths
|
|
|
|
|
var isMonospace = true;
|
|
|
|
|
var firstWidth = defaultWidth;
|
|
|
|
|
for (var glyph in glyphsWidths) {
|
|
|
|
|
var glyphWidth = glyphsWidths[glyph];
|
|
|
|
|
if (!glyphWidth) {
|
|
|
|
|
continue;
|
2012-09-17 04:38:30 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (!firstWidth) {
|
|
|
|
|
firstWidth = glyphWidth;
|
|
|
|
|
continue;
|
2014-03-23 03:15:51 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (firstWidth !== glyphWidth) {
|
|
|
|
|
isMonospace = false;
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
if (isMonospace) {
|
|
|
|
|
properties.flags |= FontFlags.FixedPitch;
|
|
|
|
|
}
|
2012-09-17 04:38:30 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
properties.defaultWidth = defaultWidth;
|
|
|
|
|
properties.widths = glyphsWidths;
|
|
|
|
|
properties.defaultVMetrics = defaultVMetrics;
|
|
|
|
|
properties.vmetrics = glyphsVMetrics;
|
|
|
|
|
}
|
2011-10-29 10:38:31 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
isSerifFont(baseFontName) {
|
|
|
|
|
// Simulating descriptor flags attribute
|
|
|
|
|
var fontNameWoStyle = baseFontName.split("-")[0];
|
|
|
|
|
return (
|
|
|
|
|
fontNameWoStyle in getSerifFonts() ||
|
|
|
|
|
fontNameWoStyle.search(/serif/gi) !== -1
|
|
|
|
|
);
|
|
|
|
|
}
|
2013-01-12 04:04:56 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
getBaseFontMetrics(name) {
|
|
|
|
|
var defaultWidth = 0;
|
|
|
|
|
var widths = [];
|
|
|
|
|
var monospace = false;
|
|
|
|
|
var stdFontMap = getStdFontMap();
|
|
|
|
|
var lookupName = stdFontMap[name] || name;
|
|
|
|
|
var Metrics = getMetrics();
|
|
|
|
|
|
|
|
|
|
if (!(lookupName in Metrics)) {
|
|
|
|
|
// Use default fonts for looking up font metrics if the passed
|
|
|
|
|
// font is not a base font
|
|
|
|
|
if (this.isSerifFont(name)) {
|
|
|
|
|
lookupName = "Times-Roman";
|
2011-10-29 10:38:31 +09:00
|
|
|
|
} else {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
lookupName = "Helvetica";
|
2011-10-25 08:55:23 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
var glyphWidths = Metrics[lookupName];
|
2011-10-25 08:55:23 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (isNum(glyphWidths)) {
|
|
|
|
|
defaultWidth = glyphWidths;
|
|
|
|
|
monospace = true;
|
|
|
|
|
} else {
|
|
|
|
|
widths = glyphWidths(); // expand lazy widths array
|
|
|
|
|
}
|
2011-10-25 08:55:23 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return {
|
|
|
|
|
defaultWidth,
|
|
|
|
|
monospace,
|
|
|
|
|
widths,
|
|
|
|
|
};
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
buildCharCodeToWidth(widthsByGlyphName, properties) {
|
|
|
|
|
var widths = Object.create(null);
|
|
|
|
|
var differences = properties.differences;
|
|
|
|
|
var encoding = properties.defaultEncoding;
|
|
|
|
|
for (var charCode = 0; charCode < 256; charCode++) {
|
|
|
|
|
if (charCode in differences && widthsByGlyphName[differences[charCode]]) {
|
|
|
|
|
widths[charCode] = widthsByGlyphName[differences[charCode]];
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
if (charCode in encoding && widthsByGlyphName[encoding[charCode]]) {
|
|
|
|
|
widths[charCode] = widthsByGlyphName[encoding[charCode]];
|
|
|
|
|
continue;
|
2014-02-12 03:27:09 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
return widths;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
preEvaluateFont(dict) {
|
|
|
|
|
var baseDict = dict;
|
|
|
|
|
var type = dict.get("Subtype");
|
|
|
|
|
if (!isName(type)) {
|
|
|
|
|
throw new FormatError("invalid font Subtype");
|
|
|
|
|
}
|
2014-02-12 03:27:09 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var composite = false;
|
|
|
|
|
var uint8array;
|
|
|
|
|
if (type.name === "Type0") {
|
|
|
|
|
// If font is a composite
|
|
|
|
|
// - get the descendant font
|
|
|
|
|
// - set the type according to the descendant font
|
|
|
|
|
// - get the FontDescriptor from the descendant font
|
|
|
|
|
var df = dict.get("DescendantFonts");
|
|
|
|
|
if (!df) {
|
|
|
|
|
throw new FormatError("Descendant fonts are not specified");
|
|
|
|
|
}
|
|
|
|
|
dict = Array.isArray(df) ? this.xref.fetchIfRef(df[0]) : df;
|
|
|
|
|
|
|
|
|
|
type = dict.get("Subtype");
|
2017-07-20 21:04:54 +09:00
|
|
|
|
if (!isName(type)) {
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
throw new FormatError("invalid font Subtype");
|
2017-07-20 21:04:54 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
composite = true;
|
|
|
|
|
}
|
2011-10-25 08:55:23 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var descriptor = dict.get("FontDescriptor");
|
|
|
|
|
if (descriptor) {
|
|
|
|
|
var hash = new MurmurHash3_64();
|
|
|
|
|
var encoding = baseDict.getRaw("Encoding");
|
|
|
|
|
if (isName(encoding)) {
|
|
|
|
|
hash.update(encoding.name);
|
|
|
|
|
} else if (isRef(encoding)) {
|
|
|
|
|
hash.update(encoding.toString());
|
|
|
|
|
} else if (isDict(encoding)) {
|
2020-07-17 19:57:34 +09:00
|
|
|
|
for (const entry of encoding.getRawValues()) {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (isName(entry)) {
|
|
|
|
|
hash.update(entry.name);
|
|
|
|
|
} else if (isRef(entry)) {
|
|
|
|
|
hash.update(entry.toString());
|
|
|
|
|
} else if (Array.isArray(entry)) {
|
|
|
|
|
// 'Differences' array (fixes bug1157493.pdf).
|
|
|
|
|
var diffLength = entry.length,
|
|
|
|
|
diffBuf = new Array(diffLength);
|
|
|
|
|
|
|
|
|
|
for (var j = 0; j < diffLength; j++) {
|
|
|
|
|
var diffEntry = entry[j];
|
|
|
|
|
if (isName(diffEntry)) {
|
|
|
|
|
diffBuf[j] = diffEntry.name;
|
|
|
|
|
} else if (isNum(diffEntry) || isRef(diffEntry)) {
|
|
|
|
|
diffBuf[j] = diffEntry.toString();
|
2016-12-28 08:06:54 +09:00
|
|
|
|
}
|
2015-04-25 20:27:10 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
hash.update(diffBuf.join());
|
2015-04-25 20:27:10 +09:00
|
|
|
|
}
|
2014-03-04 02:44:45 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
2014-03-04 02:44:45 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
const firstChar = dict.get("FirstChar") || 0;
|
|
|
|
|
const lastChar = dict.get("LastChar") || (composite ? 0xffff : 0xff);
|
|
|
|
|
hash.update(`${firstChar}-${lastChar}`);
|
2014-03-04 02:44:45 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var toUnicode = dict.get("ToUnicode") || baseDict.get("ToUnicode");
|
|
|
|
|
if (isStream(toUnicode)) {
|
|
|
|
|
var stream = toUnicode.str || toUnicode;
|
|
|
|
|
uint8array = stream.buffer
|
|
|
|
|
? new Uint8Array(stream.buffer.buffer, 0, stream.bufferLength)
|
|
|
|
|
: new Uint8Array(
|
|
|
|
|
stream.bytes.buffer,
|
|
|
|
|
stream.start,
|
|
|
|
|
stream.end - stream.start
|
|
|
|
|
);
|
|
|
|
|
hash.update(uint8array);
|
|
|
|
|
} else if (isName(toUnicode)) {
|
|
|
|
|
hash.update(toUnicode.name);
|
2014-03-04 02:44:45 +09:00
|
|
|
|
}
|
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var widths = dict.get("Widths") || baseDict.get("Widths");
|
|
|
|
|
if (widths) {
|
|
|
|
|
uint8array = new Uint8Array(new Uint32Array(widths).buffer);
|
|
|
|
|
hash.update(uint8array);
|
|
|
|
|
}
|
|
|
|
|
}
|
2014-03-04 02:44:45 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return {
|
|
|
|
|
descriptor,
|
|
|
|
|
dict,
|
|
|
|
|
baseDict,
|
|
|
|
|
composite,
|
|
|
|
|
type: type.name,
|
|
|
|
|
hash: hash ? hash.hexdigest() : "",
|
|
|
|
|
};
|
|
|
|
|
}
|
2011-10-25 08:55:23 +09:00
|
|
|
|
|
2020-10-15 16:30:54 +09:00
|
|
|
|
async translateFont(preEvaluatedFont) {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var baseDict = preEvaluatedFont.baseDict;
|
|
|
|
|
var dict = preEvaluatedFont.dict;
|
|
|
|
|
var composite = preEvaluatedFont.composite;
|
|
|
|
|
var descriptor = preEvaluatedFont.descriptor;
|
|
|
|
|
var type = preEvaluatedFont.type;
|
|
|
|
|
var maxCharIndex = composite ? 0xffff : 0xff;
|
|
|
|
|
var properties;
|
|
|
|
|
const firstChar = dict.get("FirstChar") || 0;
|
|
|
|
|
const lastChar = dict.get("LastChar") || maxCharIndex;
|
|
|
|
|
|
|
|
|
|
if (!descriptor) {
|
|
|
|
|
if (type === "Type3") {
|
|
|
|
|
// FontDescriptor is only required for Type3 fonts when the document
|
|
|
|
|
// is a tagged pdf. Create a barbebones one to get by.
|
|
|
|
|
descriptor = new Dict(null);
|
|
|
|
|
descriptor.set("FontName", Name.get(type));
|
|
|
|
|
descriptor.set("FontBBox", dict.getArray("FontBBox") || [0, 0, 0, 0]);
|
|
|
|
|
} else {
|
|
|
|
|
// Before PDF 1.5 if the font was one of the base 14 fonts, having a
|
|
|
|
|
// FontDescriptor was not required.
|
|
|
|
|
// This case is here for compatibility.
|
|
|
|
|
var baseFontName = dict.get("BaseFont");
|
|
|
|
|
if (!isName(baseFontName)) {
|
|
|
|
|
throw new FormatError("Base font is not specified");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Using base font name as a font name.
|
|
|
|
|
baseFontName = baseFontName.name.replace(/[,_]/g, "-");
|
|
|
|
|
var metrics = this.getBaseFontMetrics(baseFontName);
|
|
|
|
|
|
|
|
|
|
// Simulating descriptor flags attribute
|
|
|
|
|
var fontNameWoStyle = baseFontName.split("-")[0];
|
|
|
|
|
var flags =
|
|
|
|
|
(this.isSerifFont(fontNameWoStyle) ? FontFlags.Serif : 0) |
|
|
|
|
|
(metrics.monospace ? FontFlags.FixedPitch : 0) |
|
|
|
|
|
(getSymbolsFonts()[fontNameWoStyle]
|
|
|
|
|
? FontFlags.Symbolic
|
|
|
|
|
: FontFlags.Nonsymbolic);
|
|
|
|
|
|
|
|
|
|
properties = {
|
|
|
|
|
type,
|
|
|
|
|
name: baseFontName,
|
|
|
|
|
widths: metrics.widths,
|
|
|
|
|
defaultWidth: metrics.defaultWidth,
|
|
|
|
|
flags,
|
|
|
|
|
firstChar,
|
|
|
|
|
lastChar,
|
|
|
|
|
};
|
|
|
|
|
const widths = dict.get("Widths");
|
|
|
|
|
return this.extractDataStructures(dict, dict, properties).then(
|
|
|
|
|
newProperties => {
|
|
|
|
|
if (widths) {
|
|
|
|
|
const glyphWidths = [];
|
|
|
|
|
let j = firstChar;
|
|
|
|
|
for (let i = 0, ii = widths.length; i < ii; i++) {
|
|
|
|
|
glyphWidths[j++] = this.xref.fetchIfRef(widths[i]);
|
2020-01-22 03:36:41 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
newProperties.widths = glyphWidths;
|
|
|
|
|
} else {
|
|
|
|
|
newProperties.widths = this.buildCharCodeToWidth(
|
|
|
|
|
metrics.widths,
|
|
|
|
|
newProperties
|
|
|
|
|
);
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return new Font(baseFontName, null, newProperties);
|
|
|
|
|
}
|
|
|
|
|
);
|
2011-10-25 08:55:23 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
2011-10-25 08:55:23 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// According to the spec if 'FontDescriptor' is declared, 'FirstChar',
|
|
|
|
|
// 'LastChar' and 'Widths' should exist too, but some PDF encoders seem
|
|
|
|
|
// to ignore this rule when a variant of a standard font is used.
|
|
|
|
|
// TODO Fill the width array depending on which of the base font this is
|
|
|
|
|
// a variant.
|
|
|
|
|
|
|
|
|
|
var fontName = descriptor.get("FontName");
|
|
|
|
|
var baseFont = dict.get("BaseFont");
|
|
|
|
|
// Some bad PDFs have a string as the font name.
|
|
|
|
|
if (isString(fontName)) {
|
|
|
|
|
fontName = Name.get(fontName);
|
|
|
|
|
}
|
|
|
|
|
if (isString(baseFont)) {
|
|
|
|
|
baseFont = Name.get(baseFont);
|
|
|
|
|
}
|
2013-01-12 10:10:09 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (type !== "Type3") {
|
|
|
|
|
var fontNameStr = fontName && fontName.name;
|
|
|
|
|
var baseFontStr = baseFont && baseFont.name;
|
|
|
|
|
if (fontNameStr !== baseFontStr) {
|
|
|
|
|
info(
|
|
|
|
|
`The FontDescriptor\'s FontName is "${fontNameStr}" but ` +
|
|
|
|
|
`should be the same as the Font\'s BaseFont "${baseFontStr}".`
|
|
|
|
|
);
|
|
|
|
|
// Workaround for cases where e.g. fontNameStr = 'Arial' and
|
|
|
|
|
// baseFontStr = 'Arial,Bold' (needed when no font file is embedded).
|
|
|
|
|
if (fontNameStr && baseFontStr && baseFontStr.startsWith(fontNameStr)) {
|
|
|
|
|
fontName = baseFont;
|
2013-03-03 21:30:08 +09:00
|
|
|
|
}
|
2013-01-12 10:10:09 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
fontName = fontName || baseFont;
|
2013-01-12 10:10:09 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (!isName(fontName)) {
|
|
|
|
|
throw new FormatError("invalid font name");
|
|
|
|
|
}
|
2011-10-25 08:55:23 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var fontFile = descriptor.get("FontFile", "FontFile2", "FontFile3");
|
|
|
|
|
if (fontFile) {
|
|
|
|
|
if (fontFile.dict) {
|
|
|
|
|
var subtype = fontFile.dict.get("Subtype");
|
|
|
|
|
if (subtype) {
|
|
|
|
|
subtype = subtype.name;
|
2011-10-25 08:55:23 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var length1 = fontFile.dict.get("Length1");
|
|
|
|
|
var length2 = fontFile.dict.get("Length2");
|
|
|
|
|
var length3 = fontFile.dict.get("Length3");
|
2011-10-25 08:55:23 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
2011-10-25 08:55:23 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
properties = {
|
|
|
|
|
type,
|
|
|
|
|
name: fontName.name,
|
|
|
|
|
subtype,
|
|
|
|
|
file: fontFile,
|
|
|
|
|
length1,
|
|
|
|
|
length2,
|
|
|
|
|
length3,
|
|
|
|
|
loadedName: baseDict.loadedName,
|
|
|
|
|
composite,
|
|
|
|
|
fixedPitch: false,
|
|
|
|
|
fontMatrix: dict.getArray("FontMatrix") || FONT_IDENTITY_MATRIX,
|
|
|
|
|
firstChar: firstChar || 0,
|
|
|
|
|
lastChar: lastChar || maxCharIndex,
|
|
|
|
|
bbox: descriptor.getArray("FontBBox"),
|
|
|
|
|
ascent: descriptor.get("Ascent"),
|
|
|
|
|
descent: descriptor.get("Descent"),
|
|
|
|
|
xHeight: descriptor.get("XHeight"),
|
|
|
|
|
capHeight: descriptor.get("CapHeight"),
|
|
|
|
|
flags: descriptor.get("Flags"),
|
|
|
|
|
italicAngle: descriptor.get("ItalicAngle"),
|
|
|
|
|
isType3Font: false,
|
|
|
|
|
};
|
2013-02-08 21:29:22 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (composite) {
|
2020-10-15 16:30:54 +09:00
|
|
|
|
const cidEncoding = baseDict.get("Encoding");
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (isName(cidEncoding)) {
|
|
|
|
|
properties.cidEncoding = cidEncoding.name;
|
|
|
|
|
}
|
2020-10-15 16:30:54 +09:00
|
|
|
|
const cMap = await CMapFactory.create({
|
2020-07-05 19:20:10 +09:00
|
|
|
|
encoding: cidEncoding,
|
|
|
|
|
fetchBuiltInCMap: this._fetchBuiltInCMapBound,
|
|
|
|
|
useCMap: null,
|
|
|
|
|
});
|
2020-10-15 16:30:54 +09:00
|
|
|
|
properties.cMap = cMap;
|
|
|
|
|
properties.vertical = properties.cMap.vertical;
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
2011-10-25 08:55:23 +09:00
|
|
|
|
|
2020-10-15 16:30:54 +09:00
|
|
|
|
return this.extractDataStructures(dict, baseDict, properties).then(
|
|
|
|
|
newProperties => {
|
2020-07-05 19:20:10 +09:00
|
|
|
|
this.extractWidths(dict, descriptor, newProperties);
|
2016-02-29 01:20:29 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (type === "Type3") {
|
|
|
|
|
newProperties.isType3Font = true;
|
|
|
|
|
}
|
|
|
|
|
return new Font(fontName.name, fontFile, newProperties);
|
2020-10-15 16:30:54 +09:00
|
|
|
|
}
|
|
|
|
|
);
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
2011-10-25 08:55:23 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
static buildFontPaths(font, glyphs, handler) {
|
Fallback to the built-in font renderer when font loading fails
After PR 9340 all glyphs are now re-mapped to a Private Use Area (PUA) which means that if a font fails to load, for whatever reason[1], all glyphs in the font will now render as Unicode glyph outlines.
This obviously doesn't look good, to say the least, and might be seen as a "regression" since previously many glyphs were left in their original positions which provided a slightly better fallback[2].
Hence this patch, which implements a *general* fallback to the PDF.js built-in font renderer for fonts that fail to load (i.e. are rejected by the sanitizer). One caveat here is that this only works for the Font Loading API, since it's easy to handle errors in that case[3].
The solution implemented in this patch does *not* in any way delay the loading of valid fonts, which was the problem with my previous attempt at a solution, and will only require a bit of extra work/waiting for those fonts that actually fail to load.
*Please note:* This patch doesn't fix any of the underlying PDF.js font conversion bugs that's responsible for creating corrupt font files, however it does *improve* rendering in a number of cases; refer to this possibly incomplete list:
[Bug 1524888](https://bugzilla.mozilla.org/show_bug.cgi?id=1524888)
Issue 10175
Issue 10232
---
[1] Usually because the PDF.js font conversion code wasn't able to parse the font file correctly.
[2] Glyphs fell back to some default font, which while not accurate was more useful than the current state.
[3] Furthermore I'm not sure how to implement this generally, assuming that's even possible, and don't really have time/interest to look into it either.
2019-02-11 08:47:56 +09:00
|
|
|
|
function buildPath(fontChar) {
|
|
|
|
|
if (font.renderer.hasBuiltPath(fontChar)) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
handler.send("commonobj", [
|
Fallback to the built-in font renderer when font loading fails
After PR 9340 all glyphs are now re-mapped to a Private Use Area (PUA) which means that if a font fails to load, for whatever reason[1], all glyphs in the font will now render as Unicode glyph outlines.
This obviously doesn't look good, to say the least, and might be seen as a "regression" since previously many glyphs were left in their original positions which provided a slightly better fallback[2].
Hence this patch, which implements a *general* fallback to the PDF.js built-in font renderer for fonts that fail to load (i.e. are rejected by the sanitizer). One caveat here is that this only works for the Font Loading API, since it's easy to handle errors in that case[3].
The solution implemented in this patch does *not* in any way delay the loading of valid fonts, which was the problem with my previous attempt at a solution, and will only require a bit of extra work/waiting for those fonts that actually fail to load.
*Please note:* This patch doesn't fix any of the underlying PDF.js font conversion bugs that's responsible for creating corrupt font files, however it does *improve* rendering in a number of cases; refer to this possibly incomplete list:
[Bug 1524888](https://bugzilla.mozilla.org/show_bug.cgi?id=1524888)
Issue 10175
Issue 10232
---
[1] Usually because the PDF.js font conversion code wasn't able to parse the font file correctly.
[2] Glyphs fell back to some default font, which while not accurate was more useful than the current state.
[3] Furthermore I'm not sure how to implement this generally, assuming that's even possible, and don't really have time/interest to look into it either.
2019-02-11 08:47:56 +09:00
|
|
|
|
`${font.loadedName}_path_${fontChar}`,
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
"FontPath",
|
Fallback to the built-in font renderer when font loading fails
After PR 9340 all glyphs are now re-mapped to a Private Use Area (PUA) which means that if a font fails to load, for whatever reason[1], all glyphs in the font will now render as Unicode glyph outlines.
This obviously doesn't look good, to say the least, and might be seen as a "regression" since previously many glyphs were left in their original positions which provided a slightly better fallback[2].
Hence this patch, which implements a *general* fallback to the PDF.js built-in font renderer for fonts that fail to load (i.e. are rejected by the sanitizer). One caveat here is that this only works for the Font Loading API, since it's easy to handle errors in that case[3].
The solution implemented in this patch does *not* in any way delay the loading of valid fonts, which was the problem with my previous attempt at a solution, and will only require a bit of extra work/waiting for those fonts that actually fail to load.
*Please note:* This patch doesn't fix any of the underlying PDF.js font conversion bugs that's responsible for creating corrupt font files, however it does *improve* rendering in a number of cases; refer to this possibly incomplete list:
[Bug 1524888](https://bugzilla.mozilla.org/show_bug.cgi?id=1524888)
Issue 10175
Issue 10232
---
[1] Usually because the PDF.js font conversion code wasn't able to parse the font file correctly.
[2] Glyphs fell back to some default font, which while not accurate was more useful than the current state.
[3] Furthermore I'm not sure how to implement this generally, assuming that's even possible, and don't really have time/interest to look into it either.
2019-02-11 08:47:56 +09:00
|
|
|
|
font.renderer.getPathJs(fontChar),
|
|
|
|
|
]);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
for (const glyph of glyphs) {
|
|
|
|
|
buildPath(glyph.fontChar);
|
|
|
|
|
|
|
|
|
|
// If the glyph has an accent we need to build a path for its
|
|
|
|
|
// fontChar too, otherwise CanvasGraphics_paintChar will fail.
|
|
|
|
|
const accent = glyph.accent;
|
|
|
|
|
if (accent && accent.fontChar) {
|
|
|
|
|
buildPath(accent.fontChar);
|
|
|
|
|
}
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
Fallback to the built-in font renderer when font loading fails
After PR 9340 all glyphs are now re-mapped to a Private Use Area (PUA) which means that if a font fails to load, for whatever reason[1], all glyphs in the font will now render as Unicode glyph outlines.
This obviously doesn't look good, to say the least, and might be seen as a "regression" since previously many glyphs were left in their original positions which provided a slightly better fallback[2].
Hence this patch, which implements a *general* fallback to the PDF.js built-in font renderer for fonts that fail to load (i.e. are rejected by the sanitizer). One caveat here is that this only works for the Font Loading API, since it's easy to handle errors in that case[3].
The solution implemented in this patch does *not* in any way delay the loading of valid fonts, which was the problem with my previous attempt at a solution, and will only require a bit of extra work/waiting for those fonts that actually fail to load.
*Please note:* This patch doesn't fix any of the underlying PDF.js font conversion bugs that's responsible for creating corrupt font files, however it does *improve* rendering in a number of cases; refer to this possibly incomplete list:
[Bug 1524888](https://bugzilla.mozilla.org/show_bug.cgi?id=1524888)
Issue 10175
Issue 10232
---
[1] Usually because the PDF.js font conversion code wasn't able to parse the font file correctly.
[2] Glyphs fell back to some default font, which while not accurate was more useful than the current state.
[3] Furthermore I'm not sure how to implement this generally, assuming that's even possible, and don't really have time/interest to look into it either.
2019-02-11 08:47:56 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
static get fallbackFontDict() {
|
2020-01-16 23:08:25 +09:00
|
|
|
|
const dict = new Dict();
|
|
|
|
|
dict.set("BaseFont", Name.get("PDFJS-FallbackFont"));
|
|
|
|
|
dict.set("Type", Name.get("FallbackType"));
|
|
|
|
|
dict.set("Subtype", Name.get("FallbackType"));
|
|
|
|
|
dict.set("Encoding", Name.get("WinAnsiEncoding"));
|
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
return shadow(this, "fallbackFontDict", dict);
|
|
|
|
|
}
|
|
|
|
|
}
|
2011-10-25 08:55:23 +09:00
|
|
|
|
|
2020-04-03 17:19:02 +09:00
|
|
|
|
class TranslatedFont {
|
[api-minor] Change `Font.exportData` to, by default, stop exporting properties which are completely unused on the main-thread and/or in the API (PR 11773 follow-up)
For years now, the `Font.exportData` method has (because of its previous implementation) been exporting many properties despite them being completely unused on the main-thread and/or in the API.
This is unfortunate, since among those properties there's a number of potentially very large data-structures, containing e.g. Arrays and Objects, which thus have to be first structured cloned and then stored on the main-thread.
With the changes in this patch, we'll thus by default save memory for *every* `Font` instance created (there can be a lot in longer documents). The memory savings obviously depends a lot on the actual font data, but some approximate figures are: For non-embedded fonts it can save a couple of kilobytes, for simple embedded fonts a handful of kilobytes, and for composite fonts the size of this auxiliary can even be larger than the actual font program itself.
All-in-all, there's no good reason to keep exporting these properties by default when they're unused. However, since we cannot be sure that every property is unused in custom implementations of the PDF.js library, this patch adds a new `getDocument` option (named `fontExtraProperties`) that still allows access to the following properties:
- "cMap": An internal data structure, only used with composite fonts and never really intended to be exposed on the main-thread and/or in the API.
Note also that the `CMap`/`IdentityCMap` classes are a lot more complex than simple Objects, but only their "internal" properties survive the structured cloning used to send data to the main-thread. Given that CMaps can often be *very* large, not exporting them can also save a fair bit of memory.
- "defaultEncoding": An internal property used with simple fonts, and used when building the glyph mapping on the worker-thread. Considering how complex that topic is, and given that not all font types are handled identically, exposing this on the main-thread and/or in the API most likely isn't useful.
- "differences": An internal property used with simple fonts, and used when building the glyph mapping on the worker-thread. Considering how complex that topic is, and given that not all font types are handled identically, exposing this on the main-thread and/or in the API most likely isn't useful.
- "isSymbolicFont": An internal property, used during font parsing and building of the glyph mapping on the worker-thread.
- "seacMap": An internal map, only potentially used with *some* Type1/CFF fonts and never intended to be exposed in the API. The existing `Font.{charToGlyph, charToGlyphs}` functionality already takes this data into account when handling text.
- "toFontChar": The glyph map, necessary for mapping characters to glyphs in the font, which is built upon the various encoding information contained in the font dictionary and/or font program. This is not directly used on the main-thread and/or in the API.
- "toUnicode": The unicode map, necessary for text-extraction to work correctly, which is built upon the ToUnicode/CMap information contained in the font dictionary, but not directly used on the main-thread and/or in the API.
- "vmetrics": An array of width data used with fonts which are composite *and* vertical, but not directly used on the main-thread and/or in the API.
- "widths": An array of width data used with most fonts, but not directly used on the main-thread and/or in the API.
2020-04-03 18:51:46 +09:00
|
|
|
|
constructor({ loadedName, font, dict, extraProperties = false }) {
|
2014-05-20 06:27:54 +09:00
|
|
|
|
this.loadedName = loadedName;
|
|
|
|
|
this.font = font;
|
|
|
|
|
this.dict = dict;
|
[api-minor] Change `Font.exportData` to, by default, stop exporting properties which are completely unused on the main-thread and/or in the API (PR 11773 follow-up)
For years now, the `Font.exportData` method has (because of its previous implementation) been exporting many properties despite them being completely unused on the main-thread and/or in the API.
This is unfortunate, since among those properties there's a number of potentially very large data-structures, containing e.g. Arrays and Objects, which thus have to be first structured cloned and then stored on the main-thread.
With the changes in this patch, we'll thus by default save memory for *every* `Font` instance created (there can be a lot in longer documents). The memory savings obviously depends a lot on the actual font data, but some approximate figures are: For non-embedded fonts it can save a couple of kilobytes, for simple embedded fonts a handful of kilobytes, and for composite fonts the size of this auxiliary can even be larger than the actual font program itself.
All-in-all, there's no good reason to keep exporting these properties by default when they're unused. However, since we cannot be sure that every property is unused in custom implementations of the PDF.js library, this patch adds a new `getDocument` option (named `fontExtraProperties`) that still allows access to the following properties:
- "cMap": An internal data structure, only used with composite fonts and never really intended to be exposed on the main-thread and/or in the API.
Note also that the `CMap`/`IdentityCMap` classes are a lot more complex than simple Objects, but only their "internal" properties survive the structured cloning used to send data to the main-thread. Given that CMaps can often be *very* large, not exporting them can also save a fair bit of memory.
- "defaultEncoding": An internal property used with simple fonts, and used when building the glyph mapping on the worker-thread. Considering how complex that topic is, and given that not all font types are handled identically, exposing this on the main-thread and/or in the API most likely isn't useful.
- "differences": An internal property used with simple fonts, and used when building the glyph mapping on the worker-thread. Considering how complex that topic is, and given that not all font types are handled identically, exposing this on the main-thread and/or in the API most likely isn't useful.
- "isSymbolicFont": An internal property, used during font parsing and building of the glyph mapping on the worker-thread.
- "seacMap": An internal map, only potentially used with *some* Type1/CFF fonts and never intended to be exposed in the API. The existing `Font.{charToGlyph, charToGlyphs}` functionality already takes this data into account when handling text.
- "toFontChar": The glyph map, necessary for mapping characters to glyphs in the font, which is built upon the various encoding information contained in the font dictionary and/or font program. This is not directly used on the main-thread and/or in the API.
- "toUnicode": The unicode map, necessary for text-extraction to work correctly, which is built upon the ToUnicode/CMap information contained in the font dictionary, but not directly used on the main-thread and/or in the API.
- "vmetrics": An array of width data used with fonts which are composite *and* vertical, but not directly used on the main-thread and/or in the API.
- "widths": An array of width data used with most fonts, but not directly used on the main-thread and/or in the API.
2020-04-03 18:51:46 +09:00
|
|
|
|
this._extraProperties = extraProperties;
|
2014-05-20 06:27:54 +09:00
|
|
|
|
this.type3Loaded = null;
|
Improve how Type3-fonts with dependencies are handled
While the `CharProcs` streams of Type3-fonts *usually* don't rely on dependencies, such as e.g. images, it does happen in some cases.
Currently any dependencies are simply appended to the parent operatorList, which in practice means *only* the operatorList of the *first* page where the Type3-font is being used.
However, there's one thing that's slightly unfortunate with that approach: Since fonts are global to the PDF document, we really ought to ensure that any Type3 dependencies are appended to the operatorList of *all* pages where the Type3-font is being used. Otherwise there's a theoretical risk that, if one page has its rendering paused, another page may try to use a Type3-font whose dependencies are not yet fully resolved. In that case there would be errors, since Type3 operatorLists are executed synchronously.
Hence this patch, which ensures that all relevant pages will have Type3 dependencies appended to the main operatorList. (Note here that the `OperatorList.addDependencies` method, via `OperatorList.addDependency`, ensures that a dependency is only added *once* to any operatorList.)
Finally, these changes also remove the need for the "waiting for the main-thread"-hack that was added to `PartialEvaluator.buildPaintImageXObject` as part of fixing issue 10717.
2020-07-26 19:23:28 +09:00
|
|
|
|
this.type3Dependencies = font.isType3Font ? new Set() : null;
|
2014-05-20 06:27:54 +09:00
|
|
|
|
this.sent = false;
|
|
|
|
|
}
|
Fallback to the built-in font renderer when font loading fails
After PR 9340 all glyphs are now re-mapped to a Private Use Area (PUA) which means that if a font fails to load, for whatever reason[1], all glyphs in the font will now render as Unicode glyph outlines.
This obviously doesn't look good, to say the least, and might be seen as a "regression" since previously many glyphs were left in their original positions which provided a slightly better fallback[2].
Hence this patch, which implements a *general* fallback to the PDF.js built-in font renderer for fonts that fail to load (i.e. are rejected by the sanitizer). One caveat here is that this only works for the Font Loading API, since it's easy to handle errors in that case[3].
The solution implemented in this patch does *not* in any way delay the loading of valid fonts, which was the problem with my previous attempt at a solution, and will only require a bit of extra work/waiting for those fonts that actually fail to load.
*Please note:* This patch doesn't fix any of the underlying PDF.js font conversion bugs that's responsible for creating corrupt font files, however it does *improve* rendering in a number of cases; refer to this possibly incomplete list:
[Bug 1524888](https://bugzilla.mozilla.org/show_bug.cgi?id=1524888)
Issue 10175
Issue 10232
---
[1] Usually because the PDF.js font conversion code wasn't able to parse the font file correctly.
[2] Glyphs fell back to some default font, which while not accurate was more useful than the current state.
[3] Furthermore I'm not sure how to implement this generally, assuming that's even possible, and don't really have time/interest to look into it either.
2019-02-11 08:47:56 +09:00
|
|
|
|
|
2020-04-03 17:19:02 +09:00
|
|
|
|
send(handler) {
|
|
|
|
|
if (this.sent) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
this.sent = true;
|
Fallback to the built-in font renderer when font loading fails
After PR 9340 all glyphs are now re-mapped to a Private Use Area (PUA) which means that if a font fails to load, for whatever reason[1], all glyphs in the font will now render as Unicode glyph outlines.
This obviously doesn't look good, to say the least, and might be seen as a "regression" since previously many glyphs were left in their original positions which provided a slightly better fallback[2].
Hence this patch, which implements a *general* fallback to the PDF.js built-in font renderer for fonts that fail to load (i.e. are rejected by the sanitizer). One caveat here is that this only works for the Font Loading API, since it's easy to handle errors in that case[3].
The solution implemented in this patch does *not* in any way delay the loading of valid fonts, which was the problem with my previous attempt at a solution, and will only require a bit of extra work/waiting for those fonts that actually fail to load.
*Please note:* This patch doesn't fix any of the underlying PDF.js font conversion bugs that's responsible for creating corrupt font files, however it does *improve* rendering in a number of cases; refer to this possibly incomplete list:
[Bug 1524888](https://bugzilla.mozilla.org/show_bug.cgi?id=1524888)
Issue 10175
Issue 10232
---
[1] Usually because the PDF.js font conversion code wasn't able to parse the font file correctly.
[2] Glyphs fell back to some default font, which while not accurate was more useful than the current state.
[3] Furthermore I'm not sure how to implement this generally, assuming that's even possible, and don't really have time/interest to look into it either.
2019-02-11 08:47:56 +09:00
|
|
|
|
|
2020-04-03 17:19:02 +09:00
|
|
|
|
handler.send("commonobj", [
|
|
|
|
|
this.loadedName,
|
|
|
|
|
"Font",
|
[api-minor] Change `Font.exportData` to, by default, stop exporting properties which are completely unused on the main-thread and/or in the API (PR 11773 follow-up)
For years now, the `Font.exportData` method has (because of its previous implementation) been exporting many properties despite them being completely unused on the main-thread and/or in the API.
This is unfortunate, since among those properties there's a number of potentially very large data-structures, containing e.g. Arrays and Objects, which thus have to be first structured cloned and then stored on the main-thread.
With the changes in this patch, we'll thus by default save memory for *every* `Font` instance created (there can be a lot in longer documents). The memory savings obviously depends a lot on the actual font data, but some approximate figures are: For non-embedded fonts it can save a couple of kilobytes, for simple embedded fonts a handful of kilobytes, and for composite fonts the size of this auxiliary can even be larger than the actual font program itself.
All-in-all, there's no good reason to keep exporting these properties by default when they're unused. However, since we cannot be sure that every property is unused in custom implementations of the PDF.js library, this patch adds a new `getDocument` option (named `fontExtraProperties`) that still allows access to the following properties:
- "cMap": An internal data structure, only used with composite fonts and never really intended to be exposed on the main-thread and/or in the API.
Note also that the `CMap`/`IdentityCMap` classes are a lot more complex than simple Objects, but only their "internal" properties survive the structured cloning used to send data to the main-thread. Given that CMaps can often be *very* large, not exporting them can also save a fair bit of memory.
- "defaultEncoding": An internal property used with simple fonts, and used when building the glyph mapping on the worker-thread. Considering how complex that topic is, and given that not all font types are handled identically, exposing this on the main-thread and/or in the API most likely isn't useful.
- "differences": An internal property used with simple fonts, and used when building the glyph mapping on the worker-thread. Considering how complex that topic is, and given that not all font types are handled identically, exposing this on the main-thread and/or in the API most likely isn't useful.
- "isSymbolicFont": An internal property, used during font parsing and building of the glyph mapping on the worker-thread.
- "seacMap": An internal map, only potentially used with *some* Type1/CFF fonts and never intended to be exposed in the API. The existing `Font.{charToGlyph, charToGlyphs}` functionality already takes this data into account when handling text.
- "toFontChar": The glyph map, necessary for mapping characters to glyphs in the font, which is built upon the various encoding information contained in the font dictionary and/or font program. This is not directly used on the main-thread and/or in the API.
- "toUnicode": The unicode map, necessary for text-extraction to work correctly, which is built upon the ToUnicode/CMap information contained in the font dictionary, but not directly used on the main-thread and/or in the API.
- "vmetrics": An array of width data used with fonts which are composite *and* vertical, but not directly used on the main-thread and/or in the API.
- "widths": An array of width data used with most fonts, but not directly used on the main-thread and/or in the API.
2020-04-03 18:51:46 +09:00
|
|
|
|
this.font.exportData(this._extraProperties),
|
2020-04-03 17:19:02 +09:00
|
|
|
|
]);
|
|
|
|
|
}
|
Fallback to the built-in font renderer when font loading fails
After PR 9340 all glyphs are now re-mapped to a Private Use Area (PUA) which means that if a font fails to load, for whatever reason[1], all glyphs in the font will now render as Unicode glyph outlines.
This obviously doesn't look good, to say the least, and might be seen as a "regression" since previously many glyphs were left in their original positions which provided a slightly better fallback[2].
Hence this patch, which implements a *general* fallback to the PDF.js built-in font renderer for fonts that fail to load (i.e. are rejected by the sanitizer). One caveat here is that this only works for the Font Loading API, since it's easy to handle errors in that case[3].
The solution implemented in this patch does *not* in any way delay the loading of valid fonts, which was the problem with my previous attempt at a solution, and will only require a bit of extra work/waiting for those fonts that actually fail to load.
*Please note:* This patch doesn't fix any of the underlying PDF.js font conversion bugs that's responsible for creating corrupt font files, however it does *improve* rendering in a number of cases; refer to this possibly incomplete list:
[Bug 1524888](https://bugzilla.mozilla.org/show_bug.cgi?id=1524888)
Issue 10175
Issue 10232
---
[1] Usually because the PDF.js font conversion code wasn't able to parse the font file correctly.
[2] Glyphs fell back to some default font, which while not accurate was more useful than the current state.
[3] Furthermore I'm not sure how to implement this generally, assuming that's even possible, and don't really have time/interest to look into it either.
2019-02-11 08:47:56 +09:00
|
|
|
|
|
2020-04-03 17:19:02 +09:00
|
|
|
|
fallback(handler) {
|
|
|
|
|
if (!this.font.data) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
// When font loading failed, fall back to the built-in font renderer.
|
|
|
|
|
this.font.disableFontFace = true;
|
|
|
|
|
// An arbitrary number of text rendering operators could have been
|
|
|
|
|
// encountered between the point in time when the 'Font' message was sent
|
|
|
|
|
// to the main-thread, and the point in time when the 'FontFallback'
|
|
|
|
|
// message was received on the worker-thread.
|
|
|
|
|
// To ensure that all 'FontPath's are available on the main-thread, when
|
|
|
|
|
// font loading failed, attempt to resend *all* previously parsed glyphs.
|
|
|
|
|
const glyphs = this.font.glyphCacheValues;
|
|
|
|
|
PartialEvaluator.buildFontPaths(this.font, glyphs, handler);
|
|
|
|
|
}
|
2014-05-20 06:27:54 +09:00
|
|
|
|
|
Improve how Type3-fonts with dependencies are handled
While the `CharProcs` streams of Type3-fonts *usually* don't rely on dependencies, such as e.g. images, it does happen in some cases.
Currently any dependencies are simply appended to the parent operatorList, which in practice means *only* the operatorList of the *first* page where the Type3-font is being used.
However, there's one thing that's slightly unfortunate with that approach: Since fonts are global to the PDF document, we really ought to ensure that any Type3 dependencies are appended to the operatorList of *all* pages where the Type3-font is being used. Otherwise there's a theoretical risk that, if one page has its rendering paused, another page may try to use a Type3-font whose dependencies are not yet fully resolved. In that case there would be errors, since Type3 operatorLists are executed synchronously.
Hence this patch, which ensures that all relevant pages will have Type3 dependencies appended to the main operatorList. (Note here that the `OperatorList.addDependencies` method, via `OperatorList.addDependency`, ensures that a dependency is only added *once* to any operatorList.)
Finally, these changes also remove the need for the "waiting for the main-thread"-hack that was added to `PartialEvaluator.buildPaintImageXObject` as part of fixing issue 10717.
2020-07-26 19:23:28 +09:00
|
|
|
|
loadType3Data(evaluator, resources, task) {
|
2020-04-03 17:19:02 +09:00
|
|
|
|
if (this.type3Loaded) {
|
2014-05-20 06:27:54 +09:00
|
|
|
|
return this.type3Loaded;
|
2020-04-03 17:19:02 +09:00
|
|
|
|
}
|
2020-07-27 20:00:24 +09:00
|
|
|
|
if (!this.font.isType3Font) {
|
|
|
|
|
throw new Error("Must be a Type3 font.");
|
|
|
|
|
}
|
2020-04-03 17:19:02 +09:00
|
|
|
|
// When parsing Type3 glyphs, always ignore them if there are errors.
|
|
|
|
|
// Compared to the parsing of e.g. an entire page, it doesn't really
|
|
|
|
|
// make sense to only be able to render a Type3 glyph partially.
|
|
|
|
|
var type3Options = Object.create(evaluator.options);
|
|
|
|
|
type3Options.ignoreErrors = false;
|
|
|
|
|
var type3Evaluator = evaluator.clone(type3Options);
|
|
|
|
|
type3Evaluator.parsingType3Font = true;
|
|
|
|
|
|
Improve how Type3-fonts with dependencies are handled
While the `CharProcs` streams of Type3-fonts *usually* don't rely on dependencies, such as e.g. images, it does happen in some cases.
Currently any dependencies are simply appended to the parent operatorList, which in practice means *only* the operatorList of the *first* page where the Type3-font is being used.
However, there's one thing that's slightly unfortunate with that approach: Since fonts are global to the PDF document, we really ought to ensure that any Type3 dependencies are appended to the operatorList of *all* pages where the Type3-font is being used. Otherwise there's a theoretical risk that, if one page has its rendering paused, another page may try to use a Type3-font whose dependencies are not yet fully resolved. In that case there would be errors, since Type3 operatorLists are executed synchronously.
Hence this patch, which ensures that all relevant pages will have Type3 dependencies appended to the main operatorList. (Note here that the `OperatorList.addDependencies` method, via `OperatorList.addDependency`, ensures that a dependency is only added *once* to any operatorList.)
Finally, these changes also remove the need for the "waiting for the main-thread"-hack that was added to `PartialEvaluator.buildPaintImageXObject` as part of fixing issue 10717.
2020-07-26 19:23:28 +09:00
|
|
|
|
const translatedFont = this.font,
|
|
|
|
|
type3Dependencies = this.type3Dependencies;
|
2020-04-03 17:19:02 +09:00
|
|
|
|
var loadCharProcsPromise = Promise.resolve();
|
|
|
|
|
var charProcs = this.dict.get("CharProcs");
|
|
|
|
|
var fontResources = this.dict.get("Resources") || resources;
|
|
|
|
|
var charProcOperatorList = Object.create(null);
|
|
|
|
|
|
Improve how Type3-fonts with dependencies are handled
While the `CharProcs` streams of Type3-fonts *usually* don't rely on dependencies, such as e.g. images, it does happen in some cases.
Currently any dependencies are simply appended to the parent operatorList, which in practice means *only* the operatorList of the *first* page where the Type3-font is being used.
However, there's one thing that's slightly unfortunate with that approach: Since fonts are global to the PDF document, we really ought to ensure that any Type3 dependencies are appended to the operatorList of *all* pages where the Type3-font is being used. Otherwise there's a theoretical risk that, if one page has its rendering paused, another page may try to use a Type3-font whose dependencies are not yet fully resolved. In that case there would be errors, since Type3 operatorLists are executed synchronously.
Hence this patch, which ensures that all relevant pages will have Type3 dependencies appended to the main operatorList. (Note here that the `OperatorList.addDependencies` method, via `OperatorList.addDependency`, ensures that a dependency is only added *once* to any operatorList.)
Finally, these changes also remove the need for the "waiting for the main-thread"-hack that was added to `PartialEvaluator.buildPaintImageXObject` as part of fixing issue 10717.
2020-07-26 19:23:28 +09:00
|
|
|
|
for (const key of charProcs.getKeys()) {
|
2020-04-14 19:28:14 +09:00
|
|
|
|
loadCharProcsPromise = loadCharProcsPromise.then(function () {
|
2020-04-03 17:19:02 +09:00
|
|
|
|
var glyphStream = charProcs.get(key);
|
|
|
|
|
var operatorList = new OperatorList();
|
|
|
|
|
return type3Evaluator
|
|
|
|
|
.getOperatorList({
|
|
|
|
|
stream: glyphStream,
|
|
|
|
|
task,
|
|
|
|
|
resources: fontResources,
|
|
|
|
|
operatorList,
|
|
|
|
|
})
|
2020-04-14 19:28:14 +09:00
|
|
|
|
.then(function () {
|
2020-04-03 17:19:02 +09:00
|
|
|
|
charProcOperatorList[key] = operatorList.getIR();
|
|
|
|
|
|
Improve how Type3-fonts with dependencies are handled
While the `CharProcs` streams of Type3-fonts *usually* don't rely on dependencies, such as e.g. images, it does happen in some cases.
Currently any dependencies are simply appended to the parent operatorList, which in practice means *only* the operatorList of the *first* page where the Type3-font is being used.
However, there's one thing that's slightly unfortunate with that approach: Since fonts are global to the PDF document, we really ought to ensure that any Type3 dependencies are appended to the operatorList of *all* pages where the Type3-font is being used. Otherwise there's a theoretical risk that, if one page has its rendering paused, another page may try to use a Type3-font whose dependencies are not yet fully resolved. In that case there would be errors, since Type3 operatorLists are executed synchronously.
Hence this patch, which ensures that all relevant pages will have Type3 dependencies appended to the main operatorList. (Note here that the `OperatorList.addDependencies` method, via `OperatorList.addDependency`, ensures that a dependency is only added *once* to any operatorList.)
Finally, these changes also remove the need for the "waiting for the main-thread"-hack that was added to `PartialEvaluator.buildPaintImageXObject` as part of fixing issue 10717.
2020-07-26 19:23:28 +09:00
|
|
|
|
for (const dependency of operatorList.dependencies) {
|
|
|
|
|
type3Dependencies.add(dependency);
|
|
|
|
|
}
|
2020-04-03 17:19:02 +09:00
|
|
|
|
})
|
2020-04-14 19:28:14 +09:00
|
|
|
|
.catch(function (reason) {
|
2020-04-03 17:19:02 +09:00
|
|
|
|
warn(`Type3 font resource "${key}" is not available.`);
|
|
|
|
|
const dummyOperatorList = new OperatorList();
|
|
|
|
|
charProcOperatorList[key] = dummyOperatorList.getIR();
|
|
|
|
|
});
|
|
|
|
|
});
|
|
|
|
|
}
|
2020-04-14 19:28:14 +09:00
|
|
|
|
this.type3Loaded = loadCharProcsPromise.then(function () {
|
2020-04-03 17:19:02 +09:00
|
|
|
|
translatedFont.charProcOperatorList = charProcOperatorList;
|
|
|
|
|
});
|
|
|
|
|
return this.type3Loaded;
|
|
|
|
|
}
|
|
|
|
|
}
|
2014-05-20 06:27:54 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
class StateManager {
|
|
|
|
|
constructor(initialState) {
|
2014-04-10 08:44:07 +09:00
|
|
|
|
this.state = initialState;
|
|
|
|
|
this.stateStack = [];
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
|
|
|
|
|
save() {
|
|
|
|
|
var old = this.state;
|
|
|
|
|
this.stateStack.push(this.state);
|
|
|
|
|
this.state = old.clone();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
restore() {
|
|
|
|
|
var prev = this.stateStack.pop();
|
|
|
|
|
if (prev) {
|
|
|
|
|
this.state = prev;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
transform(args) {
|
|
|
|
|
this.state.ctm = Util.transform(this.state.ctm, args);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
class TextState {
|
|
|
|
|
constructor() {
|
2014-04-10 08:44:07 +09:00
|
|
|
|
this.ctm = new Float32Array(IDENTITY_MATRIX);
|
2016-06-01 06:01:35 +09:00
|
|
|
|
this.fontName = null;
|
2013-09-15 02:58:58 +09:00
|
|
|
|
this.fontSize = 0;
|
2014-04-10 08:44:07 +09:00
|
|
|
|
this.font = null;
|
|
|
|
|
this.fontMatrix = FONT_IDENTITY_MATRIX;
|
|
|
|
|
this.textMatrix = IDENTITY_MATRIX.slice();
|
|
|
|
|
this.textLineMatrix = IDENTITY_MATRIX.slice();
|
|
|
|
|
this.charSpacing = 0;
|
|
|
|
|
this.wordSpacing = 0;
|
2013-09-15 02:58:58 +09:00
|
|
|
|
this.leading = 0;
|
|
|
|
|
this.textHScale = 1;
|
|
|
|
|
this.textRise = 0;
|
|
|
|
|
}
|
2014-03-23 03:15:51 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
setTextMatrix(a, b, c, d, e, f) {
|
|
|
|
|
var m = this.textMatrix;
|
|
|
|
|
m[0] = a;
|
|
|
|
|
m[1] = b;
|
|
|
|
|
m[2] = c;
|
|
|
|
|
m[3] = d;
|
|
|
|
|
m[4] = e;
|
|
|
|
|
m[5] = f;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
setTextLineMatrix(a, b, c, d, e, f) {
|
|
|
|
|
var m = this.textLineMatrix;
|
|
|
|
|
m[0] = a;
|
|
|
|
|
m[1] = b;
|
|
|
|
|
m[2] = c;
|
|
|
|
|
m[3] = d;
|
|
|
|
|
m[4] = e;
|
|
|
|
|
m[5] = f;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
translateTextMatrix(x, y) {
|
|
|
|
|
var m = this.textMatrix;
|
|
|
|
|
m[4] = m[0] * x + m[2] * y + m[4];
|
|
|
|
|
m[5] = m[1] * x + m[3] * y + m[5];
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
translateTextLineMatrix(x, y) {
|
|
|
|
|
var m = this.textLineMatrix;
|
|
|
|
|
m[4] = m[0] * x + m[2] * y + m[4];
|
|
|
|
|
m[5] = m[1] * x + m[3] * y + m[5];
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
calcTextLineMatrixAdvance(a, b, c, d, e, f) {
|
|
|
|
|
var font = this.font;
|
|
|
|
|
if (!font) {
|
|
|
|
|
return null;
|
|
|
|
|
}
|
|
|
|
|
var m = this.textLineMatrix;
|
|
|
|
|
if (!(a === m[0] && b === m[1] && c === m[2] && d === m[3])) {
|
|
|
|
|
return null;
|
|
|
|
|
}
|
|
|
|
|
var txDiff = e - m[4],
|
|
|
|
|
tyDiff = f - m[5];
|
|
|
|
|
if ((font.vertical && txDiff !== 0) || (!font.vertical && tyDiff !== 0)) {
|
|
|
|
|
return null;
|
|
|
|
|
}
|
|
|
|
|
var tx,
|
|
|
|
|
ty,
|
|
|
|
|
denominator = a * d - b * c;
|
|
|
|
|
if (font.vertical) {
|
|
|
|
|
tx = (-tyDiff * c) / denominator;
|
|
|
|
|
ty = (tyDiff * a) / denominator;
|
|
|
|
|
} else {
|
|
|
|
|
tx = (txDiff * d) / denominator;
|
|
|
|
|
ty = (-txDiff * b) / denominator;
|
|
|
|
|
}
|
|
|
|
|
return { width: tx, height: ty, value: font.vertical ? ty : tx };
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
calcRenderMatrix(ctm) {
|
|
|
|
|
// 9.4.4 Text Space Details
|
|
|
|
|
var tsm = [
|
|
|
|
|
this.fontSize * this.textHScale,
|
|
|
|
|
0,
|
|
|
|
|
0,
|
|
|
|
|
this.fontSize,
|
|
|
|
|
0,
|
|
|
|
|
this.textRise,
|
|
|
|
|
];
|
|
|
|
|
return Util.transform(ctm, Util.transform(this.textMatrix, tsm));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
carriageReturn() {
|
|
|
|
|
this.translateTextLineMatrix(0, -this.leading);
|
|
|
|
|
this.textMatrix = this.textLineMatrix.slice();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
clone() {
|
|
|
|
|
var clone = Object.create(this);
|
|
|
|
|
clone.textMatrix = this.textMatrix.slice();
|
|
|
|
|
clone.textLineMatrix = this.textLineMatrix.slice();
|
|
|
|
|
clone.fontMatrix = this.fontMatrix.slice();
|
|
|
|
|
return clone;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
class EvalState {
|
|
|
|
|
constructor() {
|
2014-04-10 08:44:07 +09:00
|
|
|
|
this.ctm = new Float32Array(IDENTITY_MATRIX);
|
2013-08-01 06:01:55 +09:00
|
|
|
|
this.font = null;
|
2013-08-20 08:33:20 +09:00
|
|
|
|
this.textRenderingMode = TextRenderingMode.FILL;
|
2014-05-22 02:47:42 +09:00
|
|
|
|
this.fillColorSpace = ColorSpace.singletons.gray;
|
|
|
|
|
this.strokeColorSpace = ColorSpace.singletons.gray;
|
2011-10-25 08:55:23 +09:00
|
|
|
|
}
|
2011-10-28 03:51:10 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
clone() {
|
|
|
|
|
return Object.create(this);
|
|
|
|
|
}
|
|
|
|
|
}
|
2014-01-17 22:16:52 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
class EvaluatorPreprocessor {
|
|
|
|
|
static get opMap() {
|
|
|
|
|
// Specifies properties for each command
|
|
|
|
|
//
|
|
|
|
|
// If variableArgs === true: [0, `numArgs`] expected
|
|
|
|
|
// If variableArgs === false: exactly `numArgs` expected
|
|
|
|
|
const getOPMap = getLookupTableFactory(function (t) {
|
|
|
|
|
// Graphic state
|
|
|
|
|
t.w = { id: OPS.setLineWidth, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.J = { id: OPS.setLineCap, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.j = { id: OPS.setLineJoin, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.M = { id: OPS.setMiterLimit, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.d = { id: OPS.setDash, numArgs: 2, variableArgs: false };
|
|
|
|
|
t.ri = { id: OPS.setRenderingIntent, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.i = { id: OPS.setFlatness, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.gs = { id: OPS.setGState, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.q = { id: OPS.save, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.Q = { id: OPS.restore, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.cm = { id: OPS.transform, numArgs: 6, variableArgs: false };
|
|
|
|
|
|
|
|
|
|
// Path
|
|
|
|
|
t.m = { id: OPS.moveTo, numArgs: 2, variableArgs: false };
|
|
|
|
|
t.l = { id: OPS.lineTo, numArgs: 2, variableArgs: false };
|
|
|
|
|
t.c = { id: OPS.curveTo, numArgs: 6, variableArgs: false };
|
|
|
|
|
t.v = { id: OPS.curveTo2, numArgs: 4, variableArgs: false };
|
|
|
|
|
t.y = { id: OPS.curveTo3, numArgs: 4, variableArgs: false };
|
|
|
|
|
t.h = { id: OPS.closePath, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.re = { id: OPS.rectangle, numArgs: 4, variableArgs: false };
|
|
|
|
|
t.S = { id: OPS.stroke, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.s = { id: OPS.closeStroke, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.f = { id: OPS.fill, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.F = { id: OPS.fill, numArgs: 0, variableArgs: false };
|
|
|
|
|
t["f*"] = { id: OPS.eoFill, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.B = { id: OPS.fillStroke, numArgs: 0, variableArgs: false };
|
|
|
|
|
t["B*"] = { id: OPS.eoFillStroke, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.b = { id: OPS.closeFillStroke, numArgs: 0, variableArgs: false };
|
|
|
|
|
t["b*"] = { id: OPS.closeEOFillStroke, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.n = { id: OPS.endPath, numArgs: 0, variableArgs: false };
|
|
|
|
|
|
|
|
|
|
// Clipping
|
|
|
|
|
t.W = { id: OPS.clip, numArgs: 0, variableArgs: false };
|
|
|
|
|
t["W*"] = { id: OPS.eoClip, numArgs: 0, variableArgs: false };
|
|
|
|
|
|
|
|
|
|
// Text
|
|
|
|
|
t.BT = { id: OPS.beginText, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.ET = { id: OPS.endText, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.Tc = { id: OPS.setCharSpacing, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.Tw = { id: OPS.setWordSpacing, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.Tz = { id: OPS.setHScale, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.TL = { id: OPS.setLeading, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.Tf = { id: OPS.setFont, numArgs: 2, variableArgs: false };
|
|
|
|
|
t.Tr = { id: OPS.setTextRenderingMode, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.Ts = { id: OPS.setTextRise, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.Td = { id: OPS.moveText, numArgs: 2, variableArgs: false };
|
|
|
|
|
t.TD = { id: OPS.setLeadingMoveText, numArgs: 2, variableArgs: false };
|
|
|
|
|
t.Tm = { id: OPS.setTextMatrix, numArgs: 6, variableArgs: false };
|
|
|
|
|
t["T*"] = { id: OPS.nextLine, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.Tj = { id: OPS.showText, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.TJ = { id: OPS.showSpacedText, numArgs: 1, variableArgs: false };
|
|
|
|
|
t["'"] = { id: OPS.nextLineShowText, numArgs: 1, variableArgs: false };
|
|
|
|
|
t['"'] = {
|
|
|
|
|
id: OPS.nextLineSetSpacingShowText,
|
|
|
|
|
numArgs: 3,
|
|
|
|
|
variableArgs: false,
|
|
|
|
|
};
|
2014-01-17 22:16:52 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// Type3 fonts
|
|
|
|
|
t.d0 = { id: OPS.setCharWidth, numArgs: 2, variableArgs: false };
|
|
|
|
|
t.d1 = {
|
|
|
|
|
id: OPS.setCharWidthAndBounds,
|
|
|
|
|
numArgs: 6,
|
|
|
|
|
variableArgs: false,
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
// Color
|
|
|
|
|
t.CS = { id: OPS.setStrokeColorSpace, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.cs = { id: OPS.setFillColorSpace, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.SC = { id: OPS.setStrokeColor, numArgs: 4, variableArgs: true };
|
|
|
|
|
t.SCN = { id: OPS.setStrokeColorN, numArgs: 33, variableArgs: true };
|
|
|
|
|
t.sc = { id: OPS.setFillColor, numArgs: 4, variableArgs: true };
|
|
|
|
|
t.scn = { id: OPS.setFillColorN, numArgs: 33, variableArgs: true };
|
|
|
|
|
t.G = { id: OPS.setStrokeGray, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.g = { id: OPS.setFillGray, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.RG = { id: OPS.setStrokeRGBColor, numArgs: 3, variableArgs: false };
|
|
|
|
|
t.rg = { id: OPS.setFillRGBColor, numArgs: 3, variableArgs: false };
|
|
|
|
|
t.K = { id: OPS.setStrokeCMYKColor, numArgs: 4, variableArgs: false };
|
|
|
|
|
t.k = { id: OPS.setFillCMYKColor, numArgs: 4, variableArgs: false };
|
|
|
|
|
|
|
|
|
|
// Shading
|
|
|
|
|
t.sh = { id: OPS.shadingFill, numArgs: 1, variableArgs: false };
|
|
|
|
|
|
|
|
|
|
// Images
|
|
|
|
|
t.BI = { id: OPS.beginInlineImage, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.ID = { id: OPS.beginImageData, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.EI = { id: OPS.endInlineImage, numArgs: 1, variableArgs: false };
|
|
|
|
|
|
|
|
|
|
// XObjects
|
|
|
|
|
t.Do = { id: OPS.paintXObject, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.MP = { id: OPS.markPoint, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.DP = { id: OPS.markPointProps, numArgs: 2, variableArgs: false };
|
|
|
|
|
t.BMC = { id: OPS.beginMarkedContent, numArgs: 1, variableArgs: false };
|
|
|
|
|
t.BDC = {
|
|
|
|
|
id: OPS.beginMarkedContentProps,
|
|
|
|
|
numArgs: 2,
|
|
|
|
|
variableArgs: false,
|
|
|
|
|
};
|
|
|
|
|
t.EMC = { id: OPS.endMarkedContent, numArgs: 0, variableArgs: false };
|
|
|
|
|
|
|
|
|
|
// Compatibility
|
|
|
|
|
t.BX = { id: OPS.beginCompat, numArgs: 0, variableArgs: false };
|
|
|
|
|
t.EX = { id: OPS.endCompat, numArgs: 0, variableArgs: false };
|
|
|
|
|
|
|
|
|
|
// (reserved partial commands for the lexer)
|
|
|
|
|
t.BM = null;
|
|
|
|
|
t.BD = null;
|
|
|
|
|
t.true = null;
|
|
|
|
|
t.fa = null;
|
|
|
|
|
t.fal = null;
|
|
|
|
|
t.fals = null;
|
|
|
|
|
t.false = null;
|
|
|
|
|
t.nu = null;
|
|
|
|
|
t.nul = null;
|
|
|
|
|
t.null = null;
|
|
|
|
|
});
|
|
|
|
|
|
|
|
|
|
return shadow(this, "opMap", getOPMap());
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static get MAX_INVALID_PATH_OPS() {
|
|
|
|
|
return shadow(this, "MAX_INVALID_PATH_OPS", 20);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
constructor(stream, xref, stateManager) {
|
2016-01-22 07:43:27 +09:00
|
|
|
|
// TODO(mduan): pass array of knownCommands rather than this.opMap
|
2014-01-17 22:16:52 +09:00
|
|
|
|
// dictionary
|
2019-06-23 23:01:45 +09:00
|
|
|
|
this.parser = new Parser({
|
2020-07-05 19:20:10 +09:00
|
|
|
|
lexer: new Lexer(stream, EvaluatorPreprocessor.opMap),
|
2019-06-23 23:01:45 +09:00
|
|
|
|
xref,
|
|
|
|
|
});
|
2014-04-10 08:44:07 +09:00
|
|
|
|
this.stateManager = stateManager;
|
2014-05-15 15:07:43 +09:00
|
|
|
|
this.nonProcessedArgs = [];
|
Error, rather than warn, once a number of invalid path operators are encountered in `EvaluatorPreprocessor.read` (bug 1443140)
Incomplete path operators, in particular, can result in fairly chaotic rendering artifacts, as can be observed on page four of the referenced PDF file.
The initial (naive) solution that was attempted, was to simply throw a `FormatError` as soon as any invalid (i.e. too short) operator was found and rely on the existing `ignoreErrors` code-paths. However, doing so would have caused regressions in some files; see the existing `issue2391-1` test-case, which was promoted to an `eq` test to help prevent future bugs.
Hence this patch, which adds special handling for invalid path operators since those may cause quite bad rendering artifacts.
You could, in all fairness, argue that the patch is a handwavy solution and I wouldn't object. However, given that this only concerns *corrupt* PDF files, the way that PDF viewers (PDF.js included) try to gracefully deal with those could probably be described as a best-effort solution anyway.
This patch also adjusts the existing `warn`/`info` messages to print the command name according to the PDF specification, rather than an internal PDF.js enumeration value. The former should be much more useful for debugging purposes.
Fixes https://bugzilla.mozilla.org/show_bug.cgi?id=1443140.
2018-06-24 16:53:32 +09:00
|
|
|
|
this._numInvalidPathOPS = 0;
|
2014-01-17 22:16:52 +09:00
|
|
|
|
}
|
2014-03-23 03:15:51 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
get savedStatesDepth() {
|
|
|
|
|
return this.stateManager.stateStack.length;
|
|
|
|
|
}
|
2014-03-23 03:15:51 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// |operation| is an object with two fields:
|
|
|
|
|
//
|
|
|
|
|
// - |fn| is an out param.
|
|
|
|
|
//
|
|
|
|
|
// - |args| is an inout param. On entry, it should have one of two values.
|
|
|
|
|
//
|
|
|
|
|
// - An empty array. This indicates that the caller is providing the
|
|
|
|
|
// array in which the args will be stored in. The caller should use
|
|
|
|
|
// this value if it can reuse a single array for each call to read().
|
|
|
|
|
//
|
|
|
|
|
// - |null|. This indicates that the caller needs this function to create
|
|
|
|
|
// the array in which any args are stored in. If there are zero args,
|
|
|
|
|
// this function will leave |operation.args| as |null| (thus avoiding
|
|
|
|
|
// allocations that would occur if we used an empty array to represent
|
|
|
|
|
// zero arguments). Otherwise, it will replace |null| with a new array
|
|
|
|
|
// containing the arguments. The caller should use this value if it
|
|
|
|
|
// cannot reuse an array for each call to read().
|
|
|
|
|
//
|
|
|
|
|
// These two modes are present because this function is very hot and so
|
|
|
|
|
// avoiding allocations where possible is worthwhile.
|
|
|
|
|
//
|
|
|
|
|
read(operation) {
|
|
|
|
|
var args = operation.args;
|
|
|
|
|
while (true) {
|
|
|
|
|
var obj = this.parser.getObj();
|
|
|
|
|
if (obj instanceof Cmd) {
|
|
|
|
|
var cmd = obj.cmd;
|
|
|
|
|
// Check that the command is valid
|
|
|
|
|
var opSpec = EvaluatorPreprocessor.opMap[cmd];
|
|
|
|
|
if (!opSpec) {
|
|
|
|
|
warn(`Unknown command "${cmd}".`);
|
|
|
|
|
continue;
|
|
|
|
|
}
|
2014-06-19 19:47:00 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
var fn = opSpec.id;
|
|
|
|
|
var numArgs = opSpec.numArgs;
|
|
|
|
|
var argsLength = args !== null ? args.length : 0;
|
|
|
|
|
|
|
|
|
|
if (!opSpec.variableArgs) {
|
|
|
|
|
// Postscript commands can be nested, e.g. /F2 /GS2 gs 5.711 Tf
|
|
|
|
|
if (argsLength !== numArgs) {
|
|
|
|
|
var nonProcessedArgs = this.nonProcessedArgs;
|
|
|
|
|
while (argsLength > numArgs) {
|
|
|
|
|
nonProcessedArgs.push(args.shift());
|
|
|
|
|
argsLength--;
|
|
|
|
|
}
|
|
|
|
|
while (argsLength < numArgs && nonProcessedArgs.length !== 0) {
|
|
|
|
|
if (args === null) {
|
|
|
|
|
args = [];
|
2014-06-19 19:47:00 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
args.unshift(nonProcessedArgs.pop());
|
|
|
|
|
argsLength++;
|
2014-06-19 19:47:00 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
2014-06-19 19:47:00 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
if (argsLength < numArgs) {
|
|
|
|
|
const partialMsg =
|
|
|
|
|
`command ${cmd}: expected ${numArgs} args, ` +
|
|
|
|
|
`but received ${argsLength} args.`;
|
|
|
|
|
|
|
|
|
|
// Incomplete path operators, in particular, can result in fairly
|
|
|
|
|
// chaotic rendering artifacts. Hence the following heuristics is
|
|
|
|
|
// used to error, rather than just warn, once a number of invalid
|
|
|
|
|
// path operators have been encountered (fixes bug1443140.pdf).
|
|
|
|
|
if (
|
|
|
|
|
fn >= OPS.moveTo &&
|
|
|
|
|
fn <= OPS.endPath && // Path operator
|
|
|
|
|
++this._numInvalidPathOPS >
|
|
|
|
|
EvaluatorPreprocessor.MAX_INVALID_PATH_OPS
|
|
|
|
|
) {
|
|
|
|
|
throw new FormatError(`Invalid ${partialMsg}`);
|
2014-06-19 19:47:00 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// If we receive too few arguments, it's not possible to execute
|
|
|
|
|
// the command, hence we skip the command.
|
|
|
|
|
warn(`Skipping ${partialMsg}`);
|
|
|
|
|
if (args !== null) {
|
|
|
|
|
args.length = 0;
|
|
|
|
|
}
|
|
|
|
|
continue;
|
2014-06-19 19:47:00 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
} else if (argsLength > numArgs) {
|
|
|
|
|
info(
|
|
|
|
|
`Command ${cmd}: expected [0, ${numArgs}] args, ` +
|
|
|
|
|
`but received ${argsLength} args.`
|
|
|
|
|
);
|
|
|
|
|
}
|
2014-06-19 19:47:00 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
// TODO figure out how to type-check vararg functions
|
|
|
|
|
this.preprocessCommand(fn, args);
|
2014-06-19 19:47:00 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
operation.fn = fn;
|
|
|
|
|
operation.args = args;
|
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
if (obj === EOF) {
|
|
|
|
|
return false; // no more commands
|
|
|
|
|
}
|
|
|
|
|
// argument
|
|
|
|
|
if (obj !== null) {
|
|
|
|
|
if (args === null) {
|
|
|
|
|
args = [];
|
2016-12-16 21:05:33 +09:00
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
args.push(obj);
|
|
|
|
|
if (args.length > 33) {
|
|
|
|
|
throw new FormatError("Too many arguments");
|
2014-01-17 22:16:52 +09:00
|
|
|
|
}
|
|
|
|
|
}
|
2020-07-05 19:20:10 +09:00
|
|
|
|
}
|
|
|
|
|
}
|
2014-03-23 03:15:51 +09:00
|
|
|
|
|
2020-07-05 19:20:10 +09:00
|
|
|
|
preprocessCommand(fn, args) {
|
|
|
|
|
switch (fn | 0) {
|
|
|
|
|
case OPS.save:
|
|
|
|
|
this.stateManager.save();
|
|
|
|
|
break;
|
|
|
|
|
case OPS.restore:
|
|
|
|
|
this.stateManager.restore();
|
|
|
|
|
break;
|
|
|
|
|
case OPS.transform:
|
|
|
|
|
this.stateManager.transform(args);
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
2014-02-24 11:42:54 +09:00
|
|
|
|
|
Enable auto-formatting of the entire code-base using Prettier (issue 11444)
Note that Prettier, purposely, has only limited [configuration options](https://prettier.io/docs/en/options.html). The configuration file is based on [the one in `mozilla central`](https://searchfox.org/mozilla-central/source/.prettierrc) with just a few additions (to avoid future breakage if the defaults ever changes).
Prettier is being used for a couple of reasons:
- To be consistent with `mozilla-central`, where Prettier is already in use across the tree.
- To ensure a *consistent* coding style everywhere, which is automatically enforced during linting (since Prettier is used as an ESLint plugin). This thus ends "all" formatting disussions once and for all, removing the need for review comments on most stylistic matters.
Many ESLint options are now redundant, and I've tried my best to remove all the now unnecessary options (but I may have missed some).
Note also that since Prettier considers the `printWidth` option as a guide, rather than a hard rule, this patch resorts to a small hack in the ESLint config to ensure that *comments* won't become too long.
*Please note:* This patch is generated automatically, by appending the `--fix` argument to the ESLint call used in the `gulp lint` task. It will thus require some additional clean-up, which will be done in a *separate* commit.
(On a more personal note, I'll readily admit that some of the changes Prettier makes are *extremely* ugly. However, in the name of consistency we'll probably have to live with that.)
2019-12-25 23:59:37 +09:00
|
|
|
|
export { PartialEvaluator };
|