blob: 126d403fe417d9d6bc83fd4abf1c69858aed7929 [file] [log] [blame]
/* Copyright (c) 2012-2017 The ANTLR Project. All rights reserved.
* Use of this file is governed by the BSD 3-clause license that
* can be found in the LICENSE.txt file in the project root.
*/
#include "atn/EmptyPredictionContext.h"
#include "misc/MurmurHash.h"
#include "atn/ArrayPredictionContext.h"
#include "RuleContext.h"
#include "ParserRuleContext.h"
#include "atn/RuleTransition.h"
#include "support/Arrays.h"
#include "support/CPPUtils.h"
#include "atn/PredictionContext.h"
using namespace antlr4;
using namespace antlr4::misc;
using namespace antlr4::atn;
using namespace antlrcpp;
std::atomic<size_t> PredictionContext::globalNodeCount(0);
const Ref<PredictionContext> PredictionContext::EMPTY = std::make_shared<EmptyPredictionContext>();
//----------------- PredictionContext ----------------------------------------------------------------------------------
PredictionContext::PredictionContext(size_t cachedHashCode) : id(globalNodeCount.fetch_add(1, std::memory_order_relaxed)), cachedHashCode(cachedHashCode) {
}
PredictionContext::~PredictionContext() {
}
Ref<PredictionContext> PredictionContext::fromRuleContext(const ATN &atn, RuleContext *outerContext) {
if (outerContext == nullptr) {
return PredictionContext::EMPTY;
}
// if we are in RuleContext of start rule, s, then PredictionContext
// is EMPTY. Nobody called us. (if we are empty, return empty)
if (outerContext->parent == nullptr || outerContext == &ParserRuleContext::EMPTY) {
return PredictionContext::EMPTY;
}
// If we have a parent, convert it to a PredictionContext graph
Ref<PredictionContext> parent = PredictionContext::fromRuleContext(atn, dynamic_cast<RuleContext *>(outerContext->parent));
ATNState *state = atn.states.at(outerContext->invokingState);
RuleTransition *transition = (RuleTransition *)state->transitions[0];
return SingletonPredictionContext::create(parent, transition->followState->stateNumber);
}
bool PredictionContext::isEmpty() const {
return this == EMPTY.get();
}
bool PredictionContext::hasEmptyPath() const {
// since EMPTY_RETURN_STATE can only appear in the last position, we check last one
return getReturnState(size() - 1) == EMPTY_RETURN_STATE;
}
size_t PredictionContext::hashCode() const {
return cachedHashCode;
}
size_t PredictionContext::calculateEmptyHashCode() {
size_t hash = MurmurHash::initialize(INITIAL_HASH);
hash = MurmurHash::finish(hash, 0);
return hash;
}
size_t PredictionContext::calculateHashCode(Ref<PredictionContext> parent, size_t returnState) {
size_t hash = MurmurHash::initialize(INITIAL_HASH);
hash = MurmurHash::update(hash, parent);
hash = MurmurHash::update(hash, returnState);
hash = MurmurHash::finish(hash, 2);
return hash;
}
size_t PredictionContext::calculateHashCode(const std::vector<Ref<PredictionContext>> &parents,
const std::vector<size_t> &returnStates) {
size_t hash = MurmurHash::initialize(INITIAL_HASH);
for (auto parent : parents) {
hash = MurmurHash::update(hash, parent);
}
for (auto returnState : returnStates) {
hash = MurmurHash::update(hash, returnState);
}
return MurmurHash::finish(hash, parents.size() + returnStates.size());
}
Ref<PredictionContext> PredictionContext::merge(const Ref<PredictionContext> &a,
const Ref<PredictionContext> &b, bool rootIsWildcard, PredictionContextMergeCache *mergeCache) {
assert(a && b);
// share same graph if both same
if (a == b || *a == *b) {
return a;
}
if (is<SingletonPredictionContext>(a) && is<SingletonPredictionContext>(b)) {
return mergeSingletons(std::dynamic_pointer_cast<SingletonPredictionContext>(a),
std::dynamic_pointer_cast<SingletonPredictionContext>(b), rootIsWildcard, mergeCache);
}
// At least one of a or b is array.
// If one is $ and rootIsWildcard, return $ as * wildcard.
if (rootIsWildcard) {
if (is<EmptyPredictionContext>(a)) {
return a;
}
if (is<EmptyPredictionContext>(b)) {
return b;
}
}
// convert singleton so both are arrays to normalize
Ref<ArrayPredictionContext> left;
if (is<SingletonPredictionContext>(a)) {
left = std::make_shared<ArrayPredictionContext>(std::dynamic_pointer_cast<SingletonPredictionContext>(a));
} else {
left = std::dynamic_pointer_cast<ArrayPredictionContext>(a);
}
Ref<ArrayPredictionContext> right;
if (is<SingletonPredictionContext>(b)) {
right = std::make_shared<ArrayPredictionContext>(std::dynamic_pointer_cast<SingletonPredictionContext>(b));
} else {
right = std::dynamic_pointer_cast<ArrayPredictionContext>(b);
}
return mergeArrays(left, right, rootIsWildcard, mergeCache);
}
Ref<PredictionContext> PredictionContext::mergeSingletons(const Ref<SingletonPredictionContext> &a,
const Ref<SingletonPredictionContext> &b, bool rootIsWildcard, PredictionContextMergeCache *mergeCache) {
if (mergeCache != nullptr) { // Can be null if not given to the ATNState from which this call originates.
auto existing = mergeCache->get(a, b);
if (existing) {
return existing;
}
existing = mergeCache->get(b, a);
if (existing) {
return existing;
}
}
Ref<PredictionContext> rootMerge = mergeRoot(a, b, rootIsWildcard);
if (rootMerge) {
if (mergeCache != nullptr) {
mergeCache->put(a, b, rootMerge);
}
return rootMerge;
}
Ref<PredictionContext> parentA = a->parent;
Ref<PredictionContext> parentB = b->parent;
if (a->returnState == b->returnState) { // a == b
Ref<PredictionContext> parent = merge(parentA, parentB, rootIsWildcard, mergeCache);
// If parent is same as existing a or b parent or reduced to a parent, return it.
if (parent == parentA) { // ax + bx = ax, if a=b
return a;
}
if (parent == parentB) { // ax + bx = bx, if a=b
return b;
}
// else: ax + ay = a'[x,y]
// merge parents x and y, giving array node with x,y then remainders
// of those graphs. dup a, a' points at merged array
// new joined parent so create new singleton pointing to it, a'
Ref<PredictionContext> a_ = SingletonPredictionContext::create(parent, a->returnState);
if (mergeCache != nullptr) {
mergeCache->put(a, b, a_);
}
return a_;
} else {
// a != b payloads differ
// see if we can collapse parents due to $+x parents if local ctx
Ref<PredictionContext> singleParent;
if (a == b || (*parentA == *parentB)) { // ax + bx = [a,b]x
singleParent = parentA;
}
if (singleParent) { // parents are same, sort payloads and use same parent
std::vector<size_t> payloads = { a->returnState, b->returnState };
if (a->returnState > b->returnState) {
payloads[0] = b->returnState;
payloads[1] = a->returnState;
}
std::vector<Ref<PredictionContext>> parents = { singleParent, singleParent };
Ref<PredictionContext> a_ = std::make_shared<ArrayPredictionContext>(parents, payloads);
if (mergeCache != nullptr) {
mergeCache->put(a, b, a_);
}
return a_;
}
// parents differ and can't merge them. Just pack together
// into array; can't merge.
// ax + by = [ax,by]
Ref<PredictionContext> a_;
if (a->returnState > b->returnState) { // sort by payload
std::vector<size_t> payloads = { b->returnState, a->returnState };
std::vector<Ref<PredictionContext>> parents = { b->parent, a->parent };
a_ = std::make_shared<ArrayPredictionContext>(parents, payloads);
} else {
std::vector<size_t> payloads = {a->returnState, b->returnState};
std::vector<Ref<PredictionContext>> parents = { a->parent, b->parent };
a_ = std::make_shared<ArrayPredictionContext>(parents, payloads);
}
if (mergeCache != nullptr) {
mergeCache->put(a, b, a_);
}
return a_;
}
}
Ref<PredictionContext> PredictionContext::mergeRoot(const Ref<SingletonPredictionContext> &a,
const Ref<SingletonPredictionContext> &b, bool rootIsWildcard) {
if (rootIsWildcard) {
if (a == EMPTY) { // * + b = *
return EMPTY;
}
if (b == EMPTY) { // a + * = *
return EMPTY;
}
} else {
if (a == EMPTY && b == EMPTY) { // $ + $ = $
return EMPTY;
}
if (a == EMPTY) { // $ + x = [$,x]
std::vector<size_t> payloads = { b->returnState, EMPTY_RETURN_STATE };
std::vector<Ref<PredictionContext>> parents = { b->parent, nullptr };
Ref<PredictionContext> joined = std::make_shared<ArrayPredictionContext>(parents, payloads);
return joined;
}
if (b == EMPTY) { // x + $ = [$,x] ($ is always first if present)
std::vector<size_t> payloads = { a->returnState, EMPTY_RETURN_STATE };
std::vector<Ref<PredictionContext>> parents = { a->parent, nullptr };
Ref<PredictionContext> joined = std::make_shared<ArrayPredictionContext>(parents, payloads);
return joined;
}
}
return nullptr;
}
Ref<PredictionContext> PredictionContext::mergeArrays(const Ref<ArrayPredictionContext> &a,
const Ref<ArrayPredictionContext> &b, bool rootIsWildcard, PredictionContextMergeCache *mergeCache) {
if (mergeCache != nullptr) {
auto existing = mergeCache->get(a, b);
if (existing) {
return existing;
}
existing = mergeCache->get(b, a);
if (existing) {
return existing;
}
}
// merge sorted payloads a + b => M
size_t i = 0; // walks a
size_t j = 0; // walks b
size_t k = 0; // walks target M array
std::vector<size_t> mergedReturnStates(a->returnStates.size() + b->returnStates.size());
std::vector<Ref<PredictionContext>> mergedParents(a->returnStates.size() + b->returnStates.size());
// walk and merge to yield mergedParents, mergedReturnStates
while (i < a->returnStates.size() && j < b->returnStates.size()) {
Ref<PredictionContext> a_parent = a->parents[i];
Ref<PredictionContext> b_parent = b->parents[j];
if (a->returnStates[i] == b->returnStates[j]) {
// same payload (stack tops are equal), must yield merged singleton
size_t payload = a->returnStates[i];
// $+$ = $
bool both$ = payload == EMPTY_RETURN_STATE && !a_parent && !b_parent;
bool ax_ax = (a_parent && b_parent) && *a_parent == *b_parent; // ax+ax -> ax
if (both$ || ax_ax) {
mergedParents[k] = a_parent; // choose left
mergedReturnStates[k] = payload;
}
else { // ax+ay -> a'[x,y]
Ref<PredictionContext> mergedParent = merge(a_parent, b_parent, rootIsWildcard, mergeCache);
mergedParents[k] = mergedParent;
mergedReturnStates[k] = payload;
}
i++; // hop over left one as usual
j++; // but also skip one in right side since we merge
} else if (a->returnStates[i] < b->returnStates[j]) { // copy a[i] to M
mergedParents[k] = a_parent;
mergedReturnStates[k] = a->returnStates[i];
i++;
}
else { // b > a, copy b[j] to M
mergedParents[k] = b_parent;
mergedReturnStates[k] = b->returnStates[j];
j++;
}
k++;
}
// copy over any payloads remaining in either array
if (i < a->returnStates.size()) {
for (std::vector<int>::size_type p = i; p < a->returnStates.size(); p++) {
mergedParents[k] = a->parents[p];
mergedReturnStates[k] = a->returnStates[p];
k++;
}
} else {
for (std::vector<int>::size_type p = j; p < b->returnStates.size(); p++) {
mergedParents[k] = b->parents[p];
mergedReturnStates[k] = b->returnStates[p];
k++;
}
}
// trim merged if we combined a few that had same stack tops
if (k < mergedParents.size()) { // write index < last position; trim
if (k == 1) { // for just one merged element, return singleton top
Ref<PredictionContext> a_ = SingletonPredictionContext::create(mergedParents[0], mergedReturnStates[0]);
if (mergeCache != nullptr) {
mergeCache->put(a, b, a_);
}
return a_;
}
mergedParents.resize(k);
mergedReturnStates.resize(k);
}
Ref<ArrayPredictionContext> M = std::make_shared<ArrayPredictionContext>(mergedParents, mergedReturnStates);
// if we created same array as a or b, return that instead
// TODO: track whether this is possible above during merge sort for speed
if (*M == *a) {
if (mergeCache != nullptr) {
mergeCache->put(a, b, a);
}
return a;
}
if (*M == *b) {
if (mergeCache != nullptr) {
mergeCache->put(a, b, b);
}
return b;
}
// ml: this part differs from Java code. We have to recreate the context as the parents array is copied on creation.
if (combineCommonParents(mergedParents)) {
mergedReturnStates.resize(mergedParents.size());
M = std::make_shared<ArrayPredictionContext>(mergedParents, mergedReturnStates);
}
if (mergeCache != nullptr) {
mergeCache->put(a, b, M);
}
return M;
}
bool PredictionContext::combineCommonParents(std::vector<Ref<PredictionContext>> &parents) {
std::set<Ref<PredictionContext>> uniqueParents;
for (size_t p = 0; p < parents.size(); ++p) {
Ref<PredictionContext> parent = parents[p];
if (uniqueParents.find(parent) == uniqueParents.end()) { // don't replace
uniqueParents.insert(parent);
}
}
for (size_t p = 0; p < parents.size(); ++p) {
parents[p] = *uniqueParents.find(parents[p]);
}
return true;
}
std::string PredictionContext::toDOTString(const Ref<PredictionContext> &context) {
if (context == nullptr) {
return "";
}
std::stringstream ss;
ss << "digraph G {\n" << "rankdir=LR;\n";
std::vector<Ref<PredictionContext>> nodes = getAllContextNodes(context);
std::sort(nodes.begin(), nodes.end(), [](const Ref<PredictionContext> &o1, const Ref<PredictionContext> &o2) {
return o1->id - o2->id;
});
for (auto current : nodes) {
if (is<SingletonPredictionContext>(current)) {
std::string s = std::to_string(current->id);
ss << " s" << s;
std::string returnState = std::to_string(current->getReturnState(0));
if (is<EmptyPredictionContext>(current)) {
returnState = "$";
}
ss << " [label=\"" << returnState << "\"];\n";
continue;
}
Ref<ArrayPredictionContext> arr = std::static_pointer_cast<ArrayPredictionContext>(current);
ss << " s" << arr->id << " [shape=box, label=\"" << "[";
bool first = true;
for (auto inv : arr->returnStates) {
if (!first) {
ss << ", ";
}
if (inv == EMPTY_RETURN_STATE) {
ss << "$";
} else {
ss << inv;
}
first = false;
}
ss << "]";
ss << "\"];\n";
}
for (auto current : nodes) {
if (current == EMPTY) {
continue;
}
for (size_t i = 0; i < current->size(); i++) {
if (!current->getParent(i)) {
continue;
}
ss << " s" << current->id << "->" << "s" << current->getParent(i)->id;
if (current->size() > 1) {
ss << " [label=\"parent[" << i << "]\"];\n";
} else {
ss << ";\n";
}
}
}
ss << "}\n";
return ss.str();
}
// The "visited" map is just a temporary structure to control the retrieval process (which is recursive).
Ref<PredictionContext> PredictionContext::getCachedContext(const Ref<PredictionContext> &context,
PredictionContextCache &contextCache, std::map<Ref<PredictionContext>, Ref<PredictionContext>> &visited) {
if (context->isEmpty()) {
return context;
}
{
auto iterator = visited.find(context);
if (iterator != visited.end())
return iterator->second; // Not necessarly the same as context.
}
auto iterator = contextCache.find(context);
if (iterator != contextCache.end()) {
visited[context] = *iterator;
return *iterator;
}
bool changed = false;
std::vector<Ref<PredictionContext>> parents(context->size());
for (size_t i = 0; i < parents.size(); i++) {
Ref<PredictionContext> parent = getCachedContext(context->getParent(i), contextCache, visited);
if (changed || parent != context->getParent(i)) {
if (!changed) {
parents.clear();
for (size_t j = 0; j < context->size(); j++) {
parents.push_back(context->getParent(j));
}
changed = true;
}
parents[i] = parent;
}
}
if (!changed) {
contextCache.insert(context);
visited[context] = context;
return context;
}
Ref<PredictionContext> updated;
if (parents.empty()) {
updated = EMPTY;
} else if (parents.size() == 1) {
updated = SingletonPredictionContext::create(parents[0], context->getReturnState(0));
contextCache.insert(updated);
} else {
updated = std::make_shared<ArrayPredictionContext>(parents, std::dynamic_pointer_cast<ArrayPredictionContext>(context)->returnStates);
contextCache.insert(updated);
}
visited[updated] = updated;
visited[context] = updated;
return updated;
}
std::vector<Ref<PredictionContext>> PredictionContext::getAllContextNodes(const Ref<PredictionContext> &context) {
std::vector<Ref<PredictionContext>> nodes;
std::set<PredictionContext *> visited;
getAllContextNodes_(context, nodes, visited);
return nodes;
}
void PredictionContext::getAllContextNodes_(const Ref<PredictionContext> &context, std::vector<Ref<PredictionContext>> &nodes,
std::set<PredictionContext *> &visited) {
if (visited.find(context.get()) != visited.end()) {
return; // Already done.
}
visited.insert(context.get());
nodes.push_back(context);
for (size_t i = 0; i < context->size(); i++) {
getAllContextNodes_(context->getParent(i), nodes, visited);
}
}
std::string PredictionContext::toString() const {
return antlrcpp::toString(this);
}
std::string PredictionContext::toString(Recognizer * /*recog*/) const {
return toString();
}
std::vector<std::string> PredictionContext::toStrings(Recognizer *recognizer, int currentState) {
return toStrings(recognizer, EMPTY, currentState);
}
std::vector<std::string> PredictionContext::toStrings(Recognizer *recognizer, const Ref<PredictionContext> &stop, int currentState) {
std::vector<std::string> result;
for (size_t perm = 0; ; perm++) {
size_t offset = 0;
bool last = true;
PredictionContext *p = this;
size_t stateNumber = currentState;
std::stringstream ss;
ss << "[";
bool outerContinue = false;
while (!p->isEmpty() && p != stop.get()) {
size_t index = 0;
if (p->size() > 0) {
size_t bits = 1;
while ((1ULL << bits) < p->size()) {
bits++;
}
size_t mask = (1 << bits) - 1;
index = (perm >> offset) & mask;
last &= index >= p->size() - 1;
if (index >= p->size()) {
outerContinue = true;
break;
}
offset += bits;
}
if (recognizer != nullptr) {
if (ss.tellp() > 1) {
// first char is '[', if more than that this isn't the first rule
ss << ' ';
}
const ATN &atn = recognizer->getATN();
ATNState *s = atn.states[stateNumber];
std::string ruleName = recognizer->getRuleNames()[s->ruleIndex];
ss << ruleName;
} else if (p->getReturnState(index) != EMPTY_RETURN_STATE) {
if (!p->isEmpty()) {
if (ss.tellp() > 1) {
// first char is '[', if more than that this isn't the first rule
ss << ' ';
}
ss << p->getReturnState(index);
}
}
stateNumber = p->getReturnState(index);
p = p->getParent(index).get();
}
if (outerContinue)
continue;
ss << "]";
result.push_back(ss.str());
if (last) {
break;
}
}
return result;
}
//----------------- PredictionContextMergeCache ------------------------------------------------------------------------
Ref<PredictionContext> PredictionContextMergeCache::put(Ref<PredictionContext> const& key1, Ref<PredictionContext> const& key2,
Ref<PredictionContext> const& value) {
Ref<PredictionContext> previous;
auto iterator = _data.find(key1);
if (iterator == _data.end())
_data[key1][key2] = value;
else {
auto iterator2 = iterator->second.find(key2);
if (iterator2 != iterator->second.end())
previous = iterator2->second;
iterator->second[key2] = value;
}
return previous;
}
Ref<PredictionContext> PredictionContextMergeCache::get(Ref<PredictionContext> const& key1, Ref<PredictionContext> const& key2) {
auto iterator = _data.find(key1);
if (iterator == _data.end())
return nullptr;
auto iterator2 = iterator->second.find(key2);
if (iterator2 == iterator->second.end())
return nullptr;
return iterator2->second;
}
void PredictionContextMergeCache::clear() {
_data.clear();
}
std::string PredictionContextMergeCache::toString() const {
std::string result;
for (auto pair : _data)
for (auto pair2 : pair.second)
result += pair2.second->toString() + "\n";
return result;
}
size_t PredictionContextMergeCache::count() const {
size_t result = 0;
for (auto entry : _data)
result += entry.second.size();
return result;
}