| // Copyright (c) 2013, the Dart project authors. Please see the AUTHORS file |
| // for details. All rights reserved. Use of this source code is governed by a |
| // BSD-style license that can be found in the LICENSE file. |
| |
| library _fe_analyzer_shared.scanner.utf8_bytes_scanner; |
| |
| import 'dart:convert' show unicodeBomCharacterRune, utf8; |
| |
| import 'token.dart' show SyntheticStringToken, TokenType; |
| |
| import 'token.dart' as analyzer show StringToken; |
| |
| import 'scanner.dart' show unicodeReplacementCharacter; |
| |
| import 'abstract_scanner.dart' |
| show AbstractScanner, LanguageVersionChanged, ScannerConfiguration; |
| |
| import 'token_impl.dart' |
| show CommentToken, DartDocToken, LanguageVersionToken, StringToken; |
| |
| /** |
| * Scanner that reads from a UTF-8 encoded list of bytes and creates tokens |
| * that points to substrings. |
| */ |
| class Utf8BytesScanner extends AbstractScanner { |
| /** |
| * The file content. |
| * |
| * The content is zero-terminated. |
| */ |
| List<int> bytes; |
| |
| /** |
| * Points to the offset of the last byte returned by [advance]. |
| * |
| * After invoking [currentAsUnicode], the [byteOffset] points to the last |
| * byte that is part of the (unicode or ASCII) character. That way, [advance] |
| * can always increase the byte offset by 1. |
| */ |
| int byteOffset = -1; |
| |
| /** |
| * The getter [scanOffset] is expected to return the index where the current |
| * character *starts*. In case of a non-ascii character, after invoking |
| * [currentAsUnicode], the byte offset points to the *last* byte. |
| * |
| * This field keeps track of the number of bytes for the current unicode |
| * character. For example, if bytes 7,8,9 encode one unicode character, the |
| * [byteOffset] is 9 (after invoking [currentAsUnicode]). The [scanSlack] |
| * will be 2, so that [scanOffset] returns 7. |
| */ |
| int scanSlack = 0; |
| |
| /** |
| * Holds the [byteOffset] value for which the current [scanSlack] is valid. |
| */ |
| int scanSlackOffset = -1; |
| |
| /** |
| * Returns the byte offset of the first byte that belongs to the current |
| * character. |
| */ |
| int get scanOffset { |
| if (byteOffset == scanSlackOffset) { |
| return byteOffset - scanSlack; |
| } else { |
| return byteOffset; |
| } |
| } |
| |
| /** |
| * The difference between the number of bytes and the number of corresponding |
| * string characters, up to the current [byteOffset]. |
| */ |
| int utf8Slack = 0; |
| |
| /** |
| * Creates a new Utf8BytesScanner. The source file is expected to be a |
| * [Utf8BytesSourceFile] that holds a list of UTF-8 bytes. Otherwise the |
| * string text of the source file is decoded. |
| * |
| * The list of UTF-8 bytes [file.slowUtf8Bytes()] is expected to return an |
| * array whose last element is '0' to signal the end of the file. If this |
| * is not the case, the entire array is copied before scanning. |
| */ |
| Utf8BytesScanner(this.bytes, |
| {ScannerConfiguration configuration, |
| bool includeComments: false, |
| LanguageVersionChanged languageVersionChanged}) |
| : super(configuration, includeComments, languageVersionChanged, |
| numberOfBytesHint: bytes.length) { |
| assert(bytes.last == 0); |
| // Skip a leading BOM. |
| if (containsBomAt(/* offset = */ 0)) { |
| byteOffset += 3; |
| utf8Slack += 3; |
| } |
| } |
| |
| Utf8BytesScanner.createRecoveryOptionScanner(Utf8BytesScanner copyFrom) |
| : super.recoveryOptionScanner(copyFrom) { |
| this.bytes = copyFrom.bytes; |
| this.byteOffset = copyFrom.byteOffset; |
| this.scanSlack = copyFrom.scanSlack; |
| this.scanSlackOffset = copyFrom.scanSlackOffset; |
| this.utf8Slack = copyFrom.utf8Slack; |
| } |
| |
| Utf8BytesScanner createRecoveryOptionScanner() { |
| return new Utf8BytesScanner.createRecoveryOptionScanner(this); |
| } |
| |
| bool containsBomAt(int offset) { |
| const List<int> BOM_UTF8 = const [0xEF, 0xBB, 0xBF]; |
| |
| return offset + 3 < bytes.length && |
| bytes[offset] == BOM_UTF8[0] && |
| bytes[offset + 1] == BOM_UTF8[1] && |
| bytes[offset + 2] == BOM_UTF8[2]; |
| } |
| |
| int advance() => bytes[++byteOffset]; |
| |
| int peek() => bytes[byteOffset + 1]; |
| |
| /// Returns the unicode code point starting at the byte offset [startOffset] |
| /// with the byte [nextByte]. |
| int nextCodePoint(int startOffset, int nextByte) { |
| int expectedHighBytes; |
| if (nextByte < 0xC2) { |
| expectedHighBytes = 1; // Bad code unit. |
| } else if (nextByte < 0xE0) { |
| expectedHighBytes = 2; |
| } else if (nextByte < 0xF0) { |
| expectedHighBytes = 3; |
| } else if (nextByte < 0xF5) { |
| expectedHighBytes = 4; |
| } else { |
| expectedHighBytes = 1; // Bad code unit. |
| } |
| int numBytes = 0; |
| for (int i = 0; i < expectedHighBytes; i++) { |
| if (bytes[byteOffset + i] < 0x80) { |
| break; |
| } |
| numBytes++; |
| } |
| int end = startOffset + numBytes; |
| byteOffset = end - 1; |
| if (expectedHighBytes == 1 || numBytes != expectedHighBytes) { |
| return unicodeReplacementCharacter; |
| } |
| // TODO(lry): measurably slow, decode creates first a Utf8Decoder and a |
| // _Utf8Decoder instance. Also the sublist is eagerly allocated. |
| String codePoint = |
| utf8.decode(bytes.sublist(startOffset, end), allowMalformed: true); |
| if (codePoint.length == 0) { |
| // The UTF-8 decoder discards leading BOM characters. |
| // TODO(floitsch): don't just assume that removed characters were the |
| // BOM. |
| assert(containsBomAt(startOffset)); |
| codePoint = new String.fromCharCode(unicodeBomCharacterRune); |
| } |
| if (codePoint.length == 1) { |
| utf8Slack += (numBytes - 1); |
| scanSlack = numBytes - 1; |
| scanSlackOffset = byteOffset; |
| return codePoint.codeUnitAt(/* index = */ 0); |
| } else if (codePoint.length == 2) { |
| utf8Slack += (numBytes - 2); |
| scanSlack = numBytes - 1; |
| scanSlackOffset = byteOffset; |
| stringOffsetSlackOffset = byteOffset; |
| // In case of a surrogate pair, return a single code point. |
| // Gracefully degrade given invalid UTF-8. |
| RuneIterator runes = codePoint.runes.iterator; |
| if (!runes.moveNext()) return unicodeReplacementCharacter; |
| int codeUnit = runes.current; |
| return !runes.moveNext() ? codeUnit : unicodeReplacementCharacter; |
| } else { |
| return unicodeReplacementCharacter; |
| } |
| } |
| |
| int lastUnicodeOffset = -1; |
| int currentAsUnicode(int next) { |
| if (next < 128) return next; |
| // Check if currentAsUnicode was already invoked. |
| if (byteOffset == lastUnicodeOffset) return next; |
| int res = nextCodePoint(byteOffset, next); |
| lastUnicodeOffset = byteOffset; |
| return res; |
| } |
| |
| void handleUnicode(int startScanOffset) { |
| int end = byteOffset; |
| // TODO(lry): this measurably slows down the scanner for files with unicode. |
| String s = |
| utf8.decode(bytes.sublist(startScanOffset, end), allowMalformed: true); |
| utf8Slack += (end - startScanOffset) - s.length; |
| } |
| |
| /** |
| * This field remembers the byte offset of the last character decoded with |
| * [nextCodePoint] that used two code units in UTF-16. |
| * |
| * [nextCodePoint] returns a single code point for each unicode character, |
| * even if it needs two code units in UTF-16. |
| * |
| * For example, '\u{1d11e}' uses 4 bytes in UTF-8, and two code units in |
| * UTF-16. The [utf8Slack] is therefore 2. After invoking [nextCodePoint], the |
| * [byteOffset] points to the last (of 4) bytes. The [stringOffset] should |
| * return the offset of the first one, which is one position more left than |
| * the [utf8Slack]. |
| */ |
| int stringOffsetSlackOffset = -1; |
| |
| int get stringOffset { |
| if (stringOffsetSlackOffset == byteOffset) { |
| return byteOffset - utf8Slack - 1; |
| } else { |
| return byteOffset - utf8Slack; |
| } |
| } |
| |
| @override |
| analyzer.StringToken createSubstringToken( |
| TokenType type, int start, bool asciiOnly, |
| [int extraOffset = 0]) { |
| return new StringToken.fromUtf8Bytes( |
| type, bytes, start, byteOffset + extraOffset, asciiOnly, tokenStart, |
| precedingComments: comments); |
| } |
| |
| @override |
| analyzer.StringToken createSyntheticSubstringToken( |
| TokenType type, int start, bool asciiOnly, String syntheticChars) { |
| String source = StringToken.decodeUtf8(bytes, start, byteOffset, asciiOnly); |
| return new SyntheticStringToken( |
| type, source + syntheticChars, tokenStart, source.length); |
| } |
| |
| @override |
| CommentToken createCommentToken(TokenType type, int start, bool asciiOnly, |
| [int extraOffset = 0]) { |
| return new CommentToken.fromUtf8Bytes( |
| type, bytes, start, byteOffset + extraOffset, asciiOnly, tokenStart); |
| } |
| |
| @override |
| DartDocToken createDartDocToken(TokenType type, int start, bool asciiOnly, |
| [int extraOffset = 0]) { |
| return new DartDocToken.fromUtf8Bytes( |
| type, bytes, start, byteOffset + extraOffset, asciiOnly, tokenStart); |
| } |
| |
| @override |
| LanguageVersionToken createLanguageVersionToken( |
| int start, int major, int minor) { |
| return new LanguageVersionToken.fromUtf8Bytes( |
| bytes, start, byteOffset, tokenStart, major, minor); |
| } |
| |
| bool atEndOfFile() => byteOffset >= bytes.length - 1; |
| } |