192 lines
5.4 KiB
JavaScript
192 lines
5.4 KiB
JavaScript
|
"use strict";
|
||
|
|
||
|
Object.defineProperty(exports, "__esModule", {
|
||
|
value: true
|
||
|
});
|
||
|
exports.TokenMap = void 0;
|
||
|
var _t = require("@babel/types");
|
||
|
const {
|
||
|
traverseFast,
|
||
|
VISITOR_KEYS
|
||
|
} = _t;
|
||
|
class TokenMap {
|
||
|
constructor(ast, tokens, source) {
|
||
|
this._tokens = void 0;
|
||
|
this._source = void 0;
|
||
|
this._nodesToTokenIndexes = new Map();
|
||
|
this._nodesOccurrencesCountCache = new Map();
|
||
|
this._tokensCache = new Map();
|
||
|
this._tokens = tokens;
|
||
|
this._source = source;
|
||
|
traverseFast(ast, node => {
|
||
|
const indexes = this._getTokensIndexesOfNode(node);
|
||
|
if (indexes.length > 0) this._nodesToTokenIndexes.set(node, indexes);
|
||
|
});
|
||
|
this._tokensCache = null;
|
||
|
}
|
||
|
has(node) {
|
||
|
return this._nodesToTokenIndexes.has(node);
|
||
|
}
|
||
|
getIndexes(node) {
|
||
|
return this._nodesToTokenIndexes.get(node);
|
||
|
}
|
||
|
find(node, condition) {
|
||
|
const indexes = this._nodesToTokenIndexes.get(node);
|
||
|
if (indexes) {
|
||
|
for (let k = 0; k < indexes.length; k++) {
|
||
|
const index = indexes[k];
|
||
|
const tok = this._tokens[index];
|
||
|
if (condition(tok, index)) return tok;
|
||
|
}
|
||
|
}
|
||
|
return null;
|
||
|
}
|
||
|
findLastIndex(node, condition) {
|
||
|
const indexes = this._nodesToTokenIndexes.get(node);
|
||
|
if (indexes) {
|
||
|
for (let k = indexes.length - 1; k >= 0; k--) {
|
||
|
const index = indexes[k];
|
||
|
const tok = this._tokens[index];
|
||
|
if (condition(tok, index)) return index;
|
||
|
}
|
||
|
}
|
||
|
return -1;
|
||
|
}
|
||
|
findMatching(node, test, occurrenceCount = 0) {
|
||
|
const indexes = this._nodesToTokenIndexes.get(node);
|
||
|
if (indexes) {
|
||
|
let i = 0;
|
||
|
const count = occurrenceCount;
|
||
|
if (count > 1) {
|
||
|
const cache = this._nodesOccurrencesCountCache.get(node);
|
||
|
if (cache && cache.test === test && cache.count < count) {
|
||
|
i = cache.i + 1;
|
||
|
occurrenceCount -= cache.count + 1;
|
||
|
}
|
||
|
}
|
||
|
for (; i < indexes.length; i++) {
|
||
|
const tok = this._tokens[indexes[i]];
|
||
|
if (this.matchesOriginal(tok, test)) {
|
||
|
if (occurrenceCount === 0) {
|
||
|
if (count > 0) {
|
||
|
this._nodesOccurrencesCountCache.set(node, {
|
||
|
test,
|
||
|
count,
|
||
|
i
|
||
|
});
|
||
|
}
|
||
|
return tok;
|
||
|
}
|
||
|
occurrenceCount--;
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
return null;
|
||
|
}
|
||
|
matchesOriginal(token, test) {
|
||
|
if (token.end - token.start !== test.length) return false;
|
||
|
if (token.value != null) return token.value === test;
|
||
|
return this._source.startsWith(test, token.start);
|
||
|
}
|
||
|
startMatches(node, test) {
|
||
|
const indexes = this._nodesToTokenIndexes.get(node);
|
||
|
if (!indexes) return false;
|
||
|
const tok = this._tokens[indexes[0]];
|
||
|
if (tok.start !== node.start) return false;
|
||
|
return this.matchesOriginal(tok, test);
|
||
|
}
|
||
|
endMatches(node, test) {
|
||
|
const indexes = this._nodesToTokenIndexes.get(node);
|
||
|
if (!indexes) return false;
|
||
|
const tok = this._tokens[indexes[indexes.length - 1]];
|
||
|
if (tok.end !== node.end) return false;
|
||
|
return this.matchesOriginal(tok, test);
|
||
|
}
|
||
|
_getTokensIndexesOfNode(node) {
|
||
|
if (node.start == null || node.end == null) return [];
|
||
|
const {
|
||
|
first,
|
||
|
last
|
||
|
} = this._findTokensOfNode(node, 0, this._tokens.length - 1);
|
||
|
let low = first;
|
||
|
const children = childrenIterator(node);
|
||
|
if ((node.type === "ExportNamedDeclaration" || node.type === "ExportDefaultDeclaration") && node.declaration && node.declaration.type === "ClassDeclaration") {
|
||
|
children.next();
|
||
|
}
|
||
|
const indexes = [];
|
||
|
for (const child of children) {
|
||
|
if (child == null) continue;
|
||
|
if (child.start == null || child.end == null) continue;
|
||
|
const childTok = this._findTokensOfNode(child, low, last);
|
||
|
const high = childTok.first;
|
||
|
for (let k = low; k < high; k++) indexes.push(k);
|
||
|
low = childTok.last + 1;
|
||
|
}
|
||
|
for (let k = low; k <= last; k++) indexes.push(k);
|
||
|
return indexes;
|
||
|
}
|
||
|
_findTokensOfNode(node, low, high) {
|
||
|
const cached = this._tokensCache.get(node);
|
||
|
if (cached) return cached;
|
||
|
const first = this._findFirstTokenOfNode(node.start, low, high);
|
||
|
const last = this._findLastTokenOfNode(node.end, first, high);
|
||
|
this._tokensCache.set(node, {
|
||
|
first,
|
||
|
last
|
||
|
});
|
||
|
return {
|
||
|
first,
|
||
|
last
|
||
|
};
|
||
|
}
|
||
|
_findFirstTokenOfNode(start, low, high) {
|
||
|
while (low <= high) {
|
||
|
const mid = high + low >> 1;
|
||
|
if (start < this._tokens[mid].start) {
|
||
|
high = mid - 1;
|
||
|
} else if (start > this._tokens[mid].start) {
|
||
|
low = mid + 1;
|
||
|
} else {
|
||
|
return mid;
|
||
|
}
|
||
|
}
|
||
|
return low;
|
||
|
}
|
||
|
_findLastTokenOfNode(end, low, high) {
|
||
|
while (low <= high) {
|
||
|
const mid = high + low >> 1;
|
||
|
if (end < this._tokens[mid].end) {
|
||
|
high = mid - 1;
|
||
|
} else if (end > this._tokens[mid].end) {
|
||
|
low = mid + 1;
|
||
|
} else {
|
||
|
return mid;
|
||
|
}
|
||
|
}
|
||
|
return high;
|
||
|
}
|
||
|
}
|
||
|
exports.TokenMap = TokenMap;
|
||
|
function* childrenIterator(node) {
|
||
|
if (node.type === "TemplateLiteral") {
|
||
|
yield node.quasis[0];
|
||
|
for (let i = 1; i < node.quasis.length; i++) {
|
||
|
yield node.expressions[i - 1];
|
||
|
yield node.quasis[i];
|
||
|
}
|
||
|
return;
|
||
|
}
|
||
|
const keys = VISITOR_KEYS[node.type];
|
||
|
for (const key of keys) {
|
||
|
const child = node[key];
|
||
|
if (!child) continue;
|
||
|
if (Array.isArray(child)) {
|
||
|
yield* child;
|
||
|
} else {
|
||
|
yield child;
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
//# sourceMappingURL=token-map.js.map
|