|  | //===--- PPCaching.cpp - Handle caching lexed tokens ----------------------===// | 
|  | // | 
|  | //                     The LLVM Compiler Infrastructure | 
|  | // | 
|  | // This file is distributed under the University of Illinois Open Source | 
|  | // License. See LICENSE.TXT for details. | 
|  | // | 
|  | //===----------------------------------------------------------------------===// | 
|  | // | 
|  | // This file implements pieces of the Preprocessor interface that manage the | 
|  | // caching of lexed tokens. | 
|  | // | 
|  | //===----------------------------------------------------------------------===// | 
|  |  | 
|  | #include "clang/Lex/Preprocessor.h" | 
|  | using namespace clang; | 
|  |  | 
|  | // EnableBacktrackAtThisPos - From the point that this method is called, and | 
|  | // until CommitBacktrackedTokens() or Backtrack() is called, the Preprocessor | 
|  | // keeps track of the lexed tokens so that a subsequent Backtrack() call will | 
|  | // make the Preprocessor re-lex the same tokens. | 
|  | // | 
|  | // Nested backtracks are allowed, meaning that EnableBacktrackAtThisPos can | 
|  | // be called multiple times and CommitBacktrackedTokens/Backtrack calls will | 
|  | // be combined with the EnableBacktrackAtThisPos calls in reverse order. | 
|  | void Preprocessor::EnableBacktrackAtThisPos() { | 
|  | BacktrackPositions.push_back(CachedLexPos); | 
|  | EnterCachingLexMode(); | 
|  | } | 
|  |  | 
|  | // Disable the last EnableBacktrackAtThisPos call. | 
|  | void Preprocessor::CommitBacktrackedTokens() { | 
|  | assert(!BacktrackPositions.empty() | 
|  | && "EnableBacktrackAtThisPos was not called!"); | 
|  | BacktrackPositions.pop_back(); | 
|  | } | 
|  |  | 
|  | Preprocessor::CachedTokensRange Preprocessor::LastCachedTokenRange() { | 
|  | assert(isBacktrackEnabled()); | 
|  | auto PrevCachedLexPos = BacktrackPositions.back(); | 
|  | return CachedTokensRange{PrevCachedLexPos, CachedLexPos}; | 
|  | } | 
|  |  | 
|  | void Preprocessor::EraseCachedTokens(CachedTokensRange TokenRange) { | 
|  | assert(TokenRange.Begin <= TokenRange.End); | 
|  | if (CachedLexPos == TokenRange.Begin && TokenRange.Begin != TokenRange.End) { | 
|  | // We have backtracked to the start of the token range as we want to consume | 
|  | // them again. Erase the tokens only after consuming then. | 
|  | assert(!CachedTokenRangeToErase); | 
|  | CachedTokenRangeToErase = TokenRange; | 
|  | return; | 
|  | } | 
|  | // The cached tokens were committed, so they should be erased now. | 
|  | assert(TokenRange.End == CachedLexPos); | 
|  | CachedTokens.erase(CachedTokens.begin() + TokenRange.Begin, | 
|  | CachedTokens.begin() + TokenRange.End); | 
|  | CachedLexPos = TokenRange.Begin; | 
|  | ExitCachingLexMode(); | 
|  | } | 
|  |  | 
|  | // Make Preprocessor re-lex the tokens that were lexed since | 
|  | // EnableBacktrackAtThisPos() was previously called. | 
|  | void Preprocessor::Backtrack() { | 
|  | assert(!BacktrackPositions.empty() | 
|  | && "EnableBacktrackAtThisPos was not called!"); | 
|  | CachedLexPos = BacktrackPositions.back(); | 
|  | BacktrackPositions.pop_back(); | 
|  | recomputeCurLexerKind(); | 
|  | } | 
|  |  | 
|  | void Preprocessor::CachingLex(Token &Result) { | 
|  | if (!InCachingLexMode()) | 
|  | return; | 
|  |  | 
|  | if (CachedLexPos < CachedTokens.size()) { | 
|  | Result = CachedTokens[CachedLexPos++]; | 
|  | // Erase the some of the cached tokens after they are consumed when | 
|  | // asked to do so. | 
|  | if (CachedTokenRangeToErase && | 
|  | CachedTokenRangeToErase->End == CachedLexPos) { | 
|  | EraseCachedTokens(*CachedTokenRangeToErase); | 
|  | CachedTokenRangeToErase = None; | 
|  | } | 
|  | return; | 
|  | } | 
|  |  | 
|  | ExitCachingLexMode(); | 
|  | Lex(Result); | 
|  |  | 
|  | if (isBacktrackEnabled()) { | 
|  | // Cache the lexed token. | 
|  | EnterCachingLexMode(); | 
|  | CachedTokens.push_back(Result); | 
|  | ++CachedLexPos; | 
|  | return; | 
|  | } | 
|  |  | 
|  | if (CachedLexPos < CachedTokens.size()) { | 
|  | EnterCachingLexMode(); | 
|  | } else { | 
|  | // All cached tokens were consumed. | 
|  | CachedTokens.clear(); | 
|  | CachedLexPos = 0; | 
|  | } | 
|  | } | 
|  |  | 
|  | void Preprocessor::EnterCachingLexMode() { | 
|  | if (InCachingLexMode()) { | 
|  | assert(CurLexerKind == CLK_CachingLexer && "Unexpected lexer kind"); | 
|  | return; | 
|  | } | 
|  |  | 
|  | PushIncludeMacroStack(); | 
|  | CurLexerKind = CLK_CachingLexer; | 
|  | } | 
|  |  | 
|  |  | 
|  | const Token &Preprocessor::PeekAhead(unsigned N) { | 
|  | assert(CachedLexPos + N > CachedTokens.size() && "Confused caching."); | 
|  | ExitCachingLexMode(); | 
|  | for (size_t C = CachedLexPos + N - CachedTokens.size(); C > 0; --C) { | 
|  | CachedTokens.push_back(Token()); | 
|  | Lex(CachedTokens.back()); | 
|  | } | 
|  | EnterCachingLexMode(); | 
|  | return CachedTokens.back(); | 
|  | } | 
|  |  | 
|  | void Preprocessor::AnnotatePreviousCachedTokens(const Token &Tok) { | 
|  | assert(Tok.isAnnotation() && "Expected annotation token"); | 
|  | assert(CachedLexPos != 0 && "Expected to have some cached tokens"); | 
|  | assert(CachedTokens[CachedLexPos-1].getLastLoc() == Tok.getAnnotationEndLoc() | 
|  | && "The annotation should be until the most recent cached token"); | 
|  |  | 
|  | // Start from the end of the cached tokens list and look for the token | 
|  | // that is the beginning of the annotation token. | 
|  | for (CachedTokensTy::size_type i = CachedLexPos; i != 0; --i) { | 
|  | CachedTokensTy::iterator AnnotBegin = CachedTokens.begin() + i-1; | 
|  | if (AnnotBegin->getLocation() == Tok.getLocation()) { | 
|  | assert((BacktrackPositions.empty() || BacktrackPositions.back() <= i) && | 
|  | "The backtrack pos points inside the annotated tokens!"); | 
|  | // Replace the cached tokens with the single annotation token. | 
|  | if (i < CachedLexPos) | 
|  | CachedTokens.erase(AnnotBegin + 1, CachedTokens.begin() + CachedLexPos); | 
|  | *AnnotBegin = Tok; | 
|  | CachedLexPos = i; | 
|  | return; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | bool Preprocessor::IsPreviousCachedToken(const Token &Tok) const { | 
|  | // There's currently no cached token... | 
|  | if (!CachedLexPos) | 
|  | return false; | 
|  |  | 
|  | const Token LastCachedTok = CachedTokens[CachedLexPos - 1]; | 
|  | if (LastCachedTok.getKind() != Tok.getKind()) | 
|  | return false; | 
|  |  | 
|  | int RelOffset = 0; | 
|  | if ((!getSourceManager().isInSameSLocAddrSpace( | 
|  | Tok.getLocation(), getLastCachedTokenLocation(), &RelOffset)) || | 
|  | RelOffset) | 
|  | return false; | 
|  |  | 
|  | return true; | 
|  | } | 
|  |  | 
|  | void Preprocessor::ReplacePreviousCachedToken(ArrayRef<Token> NewToks) { | 
|  | assert(CachedLexPos != 0 && "Expected to have some cached tokens"); | 
|  | CachedTokens.insert(CachedTokens.begin() + CachedLexPos - 1, NewToks.begin(), | 
|  | NewToks.end()); | 
|  | CachedTokens.erase(CachedTokens.begin() + CachedLexPos - 1 + NewToks.size()); | 
|  | CachedLexPos += NewToks.size() - 1; | 
|  | } |