| //===--- PPCaching.cpp - Handle caching lexed tokens ----------------------===// |
| // |
| // The LLVM Compiler Infrastructure |
| // |
| // This file is distributed under the University of Illinois Open Source |
| // License. See LICENSE.TXT for details. |
| // |
| //===----------------------------------------------------------------------===// |
| // |
| // This file implements pieces of the Preprocessor interface that manage the |
| // caching of lexed tokens. |
| // |
| //===----------------------------------------------------------------------===// |
| |
| #include "clang/Lex/Preprocessor.h" |
| using namespace clang; |
| |
| // EnableBacktrackAtThisPos - From the point that this method is called, and |
| // until CommitBacktrackedTokens() or Backtrack() is called, the Preprocessor |
| // keeps track of the lexed tokens so that a subsequent Backtrack() call will |
| // make the Preprocessor re-lex the same tokens. |
| // |
| // Nested backtracks are allowed, meaning that EnableBacktrackAtThisPos can |
| // be called multiple times and CommitBacktrackedTokens/Backtrack calls will |
| // be combined with the EnableBacktrackAtThisPos calls in reverse order. |
| void Preprocessor::EnableBacktrackAtThisPos() { |
| BacktrackPositions.push_back(CachedLexPos); |
| EnterCachingLexMode(); |
| } |
| |
| // Disable the last EnableBacktrackAtThisPos call. |
| void Preprocessor::CommitBacktrackedTokens() { |
| assert(!BacktrackPositions.empty() |
| && "EnableBacktrackAtThisPos was not called!"); |
| BacktrackPositions.pop_back(); |
| } |
| |
| Preprocessor::CachedTokensRange Preprocessor::LastCachedTokenRange() { |
| assert(isBacktrackEnabled()); |
| auto PrevCachedLexPos = BacktrackPositions.back(); |
| return CachedTokensRange{PrevCachedLexPos, CachedLexPos}; |
| } |
| |
| void Preprocessor::EraseCachedTokens(CachedTokensRange TokenRange) { |
| assert(TokenRange.Begin <= TokenRange.End); |
| if (CachedLexPos == TokenRange.Begin && TokenRange.Begin != TokenRange.End) { |
| // We have backtracked to the start of the token range as we want to consume |
| // them again. Erase the tokens only after consuming then. |
| assert(!CachedTokenRangeToErase); |
| CachedTokenRangeToErase = TokenRange; |
| return; |
| } |
| // The cached tokens were committed, so they should be erased now. |
| assert(TokenRange.End == CachedLexPos); |
| CachedTokens.erase(CachedTokens.begin() + TokenRange.Begin, |
| CachedTokens.begin() + TokenRange.End); |
| CachedLexPos = TokenRange.Begin; |
| ExitCachingLexMode(); |
| } |
| |
| // Make Preprocessor re-lex the tokens that were lexed since |
| // EnableBacktrackAtThisPos() was previously called. |
| void Preprocessor::Backtrack() { |
| assert(!BacktrackPositions.empty() |
| && "EnableBacktrackAtThisPos was not called!"); |
| CachedLexPos = BacktrackPositions.back(); |
| BacktrackPositions.pop_back(); |
| recomputeCurLexerKind(); |
| } |
| |
| void Preprocessor::CachingLex(Token &Result) { |
| if (!InCachingLexMode()) |
| return; |
| |
| if (CachedLexPos < CachedTokens.size()) { |
| Result = CachedTokens[CachedLexPos++]; |
| // Erase the some of the cached tokens after they are consumed when |
| // asked to do so. |
| if (CachedTokenRangeToErase && |
| CachedTokenRangeToErase->End == CachedLexPos) { |
| EraseCachedTokens(*CachedTokenRangeToErase); |
| CachedTokenRangeToErase = None; |
| } |
| return; |
| } |
| |
| ExitCachingLexMode(); |
| Lex(Result); |
| |
| if (isBacktrackEnabled()) { |
| // Cache the lexed token. |
| EnterCachingLexMode(); |
| CachedTokens.push_back(Result); |
| ++CachedLexPos; |
| return; |
| } |
| |
| if (CachedLexPos < CachedTokens.size()) { |
| EnterCachingLexMode(); |
| } else { |
| // All cached tokens were consumed. |
| CachedTokens.clear(); |
| CachedLexPos = 0; |
| } |
| } |
| |
| void Preprocessor::EnterCachingLexMode() { |
| if (InCachingLexMode()) { |
| assert(CurLexerKind == CLK_CachingLexer && "Unexpected lexer kind"); |
| return; |
| } |
| |
| PushIncludeMacroStack(); |
| CurLexerKind = CLK_CachingLexer; |
| } |
| |
| |
| const Token &Preprocessor::PeekAhead(unsigned N) { |
| assert(CachedLexPos + N > CachedTokens.size() && "Confused caching."); |
| ExitCachingLexMode(); |
| for (size_t C = CachedLexPos + N - CachedTokens.size(); C > 0; --C) { |
| CachedTokens.push_back(Token()); |
| Lex(CachedTokens.back()); |
| } |
| EnterCachingLexMode(); |
| return CachedTokens.back(); |
| } |
| |
| void Preprocessor::AnnotatePreviousCachedTokens(const Token &Tok) { |
| assert(Tok.isAnnotation() && "Expected annotation token"); |
| assert(CachedLexPos != 0 && "Expected to have some cached tokens"); |
| assert(CachedTokens[CachedLexPos-1].getLastLoc() == Tok.getAnnotationEndLoc() |
| && "The annotation should be until the most recent cached token"); |
| |
| // Start from the end of the cached tokens list and look for the token |
| // that is the beginning of the annotation token. |
| for (CachedTokensTy::size_type i = CachedLexPos; i != 0; --i) { |
| CachedTokensTy::iterator AnnotBegin = CachedTokens.begin() + i-1; |
| if (AnnotBegin->getLocation() == Tok.getLocation()) { |
| assert((BacktrackPositions.empty() || BacktrackPositions.back() <= i) && |
| "The backtrack pos points inside the annotated tokens!"); |
| // Replace the cached tokens with the single annotation token. |
| if (i < CachedLexPos) |
| CachedTokens.erase(AnnotBegin + 1, CachedTokens.begin() + CachedLexPos); |
| *AnnotBegin = Tok; |
| CachedLexPos = i; |
| return; |
| } |
| } |
| } |
| |
| bool Preprocessor::IsPreviousCachedToken(const Token &Tok) const { |
| // There's currently no cached token... |
| if (!CachedLexPos) |
| return false; |
| |
| const Token LastCachedTok = CachedTokens[CachedLexPos - 1]; |
| if (LastCachedTok.getKind() != Tok.getKind()) |
| return false; |
| |
| int RelOffset = 0; |
| if ((!getSourceManager().isInSameSLocAddrSpace( |
| Tok.getLocation(), getLastCachedTokenLocation(), &RelOffset)) || |
| RelOffset) |
| return false; |
| |
| return true; |
| } |
| |
| void Preprocessor::ReplacePreviousCachedToken(ArrayRef<Token> NewToks) { |
| assert(CachedLexPos != 0 && "Expected to have some cached tokens"); |
| CachedTokens.insert(CachedTokens.begin() + CachedLexPos - 1, NewToks.begin(), |
| NewToks.end()); |
| CachedTokens.erase(CachedTokens.begin() + CachedLexPos - 1 + NewToks.size()); |
| CachedLexPos += NewToks.size() - 1; |
| } |