/* * cppcheck - c/c++ syntax checking * Copyright (C) 2007-2008 Daniel Marjamäki, Reijo Tomperi, Nicolas Le Cam * * This program is free software: you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program. If not, see #include #include #include #include #include #include #include #include // <- strtoul #include #ifdef __BORLANDC__ #include #include #endif //--------------------------------------------------------------------------- Tokenizer::Tokenizer() { _tokens = 0; _tokensBack = 0; _dsymlist = 0; } Tokenizer::~Tokenizer() { DeallocateTokens(); } //--------------------------------------------------------------------------- // Helper functions.. TOKEN *Tokenizer::_gettok(TOKEN *tok, int index) { while (tok && index>0) { tok = tok->next(); index--; } return tok; } //--------------------------------------------------------------------------- const TOKEN *Tokenizer::tokens() const { return _tokens; } //--------------------------------------------------------------------------- // Defined symbols. // "#define abc 123" will create a defined symbol "abc" with the value 123 //--------------------------------------------------------------------------- const std::vector *Tokenizer::getFiles() const { return &_files; } void Tokenizer::Define(const char Name[], const char Value[]) { if (!(Name && Name[0])) return; if (!(Value && Value[0])) return; // Is 'Value' a decimal value.. bool dec = true, hex = true; for (int i = 0; Value[i]; i++) { if ( ! isdigit(Value[i]) ) dec = false; if ( ! isxdigit(Value[i]) && (!(i==1 && Value[i]=='x'))) hex = false; } if (!dec && !hex) return; char *strValue = strdup(Value); if (!dec && hex) { // Convert Value from hexadecimal to decimal unsigned long value; std::istringstream istr(Value+2); istr >> std::hex >> value; std::ostringstream ostr; ostr << value; free(strValue); strValue = strdup(ostr.str().c_str()); } DefineSymbol *NewSym = new DefineSymbol; memset(NewSym, 0, sizeof(DefineSymbol)); NewSym->name = strdup(Name); NewSym->value = strValue; NewSym->next = _dsymlist; _dsymlist = NewSym; } //--------------------------------------------------------------------------- //--------------------------------------------------------------------------- // addtoken // add a token. Used by 'Tokenizer' //--------------------------------------------------------------------------- void Tokenizer::addtoken(const char str[], const unsigned int lineno, const unsigned int fileno) { if (str[0] == 0) return; // Replace hexadecimal value with decimal std::ostringstream str2; if (strncmp(str,"0x",2)==0) { str2 << strtoul(str+2, NULL, 16); } else { str2 << str; } if (_tokensBack) { _tokensBack->insertToken( str2.str().c_str() ); _tokensBack = _tokensBack->next(); } else { _tokens = new TOKEN; _tokensBack = _tokens; _tokensBack->setstr( str2.str().c_str() ); } _tokensBack->linenr( lineno ); _tokensBack->fileIndex( fileno ); // Check if str is defined.. for (DefineSymbol *sym = _dsymlist; sym; sym = sym->next) { if (strcmp(str,sym->name)==0) { _tokensBack->setstr(sym->value); break; } } } //--------------------------------------------------------------------------- //--------------------------------------------------------------------------- // SizeOfType - gives the size of a type //--------------------------------------------------------------------------- int Tokenizer::SizeOfType(const char type[]) const { if (!type) return 0; std::map::const_iterator it = _typeSize.find(type); if ( it == _typeSize.end() ) return 0; return it->second; } //--------------------------------------------------------------------------- //--------------------------------------------------------------------------- // InsertTokens - Copy and insert tokens //--------------------------------------------------------------------------- void Tokenizer::InsertTokens(TOKEN *dest, TOKEN *src, unsigned int n) { while (n > 0) { dest->insertToken( src->aaaa() ); dest = dest->next(); dest->fileIndex( src->fileIndex() ); dest->linenr( src->linenr() ); dest->varId( src->varId() ); src = src->next(); n--; } } //--------------------------------------------------------------------------- //--------------------------------------------------------------------------- // Tokenize - tokenizes a given file. //--------------------------------------------------------------------------- void Tokenizer::tokenize(std::istream &code, const char FileName[]) { // Has this file been tokenized already? for (unsigned int i = 0; i < _files.size(); i++) { if ( SameFileName( _files[i].c_str(), FileName ) ) return; } // The "_files" vector remembers what files have been tokenized.. _files.push_back( FileLister::simplifyPath( FileName ) ); // Tokenize the file.. tokenizeCode( code, (unsigned int)(_files.size() - 1) ); } //--------------------------------------------------------------------------- //--------------------------------------------------------------------------- // Tokenize - tokenizes input stream //--------------------------------------------------------------------------- void Tokenizer::tokenizeCode(std::istream &code, const unsigned int FileIndex) { // Tokenize the file. unsigned int lineno = 1; std::string CurrentToken; for (char ch = (char)code.get(); code.good(); ch = (char)code.get()) { // We are not handling UTF and stuff like that. Code is supposed to plain simple text. if ( ch < 0 ) continue; // Preprocessor stuff? if (ch == '#' && CurrentToken.empty()) { std::string line("#"); { char chPrev = '#'; while ( code.good() ) { ch = (char)code.get(); if (chPrev!='\\' && ch=='\n') break; if (ch!=' ') chPrev = ch; if (ch!='\\' && ch!='\n') line += ch; if (ch=='\n') ++lineno; } } if (strncmp(line.c_str(),"#include",8)==0 && line.find("\"") != std::string::npos) { // Extract the filename line.erase(0, line.find("\"")+1); line.erase(line.find("\"")); // Relative path.. if (_files.back().find_first_of("\\/") != std::string::npos) { std::string path = _files.back(); path.erase( 1 + path.find_last_of("\\/") ); line = path + line; } addtoken("#include", lineno, FileIndex); addtoken(line.c_str(), lineno, FileIndex); std::ifstream fin( line.c_str() ); tokenize(fin, line.c_str()); } else if (strncmp(line.c_str(), "#define", 7) == 0) { std::string strId; enum {Space1, Id, Space2, Value} State; State = Space1; for (unsigned int i = 8; i < line.length(); i++) { if (State==Space1 || State==Space2) { if (isspace(line[i])) continue; State = (State==Space1) ? Id : Value; } else if (State==Id) { if ( isspace( line[i] ) ) { strId = CurrentToken; CurrentToken.clear(); State = Space2; continue; } else if ( ! isalnum(line[i]) ) { break; } } CurrentToken += line[i]; } if (State==Value) { addtoken("def", lineno, FileIndex); addtoken(strId.c_str(), lineno, FileIndex); addtoken(";", lineno, FileIndex); Define(strId.c_str(), CurrentToken.c_str()); } CurrentToken.clear(); } else { addtoken("#", lineno, FileIndex); addtoken(";", lineno, FileIndex); } lineno++; continue; } if (ch == '\n') { // Add current token.. addtoken(CurrentToken.c_str(), lineno++, FileIndex); CurrentToken.clear(); continue; } // Comments.. if (ch == '/' && code.good()) { bool newstatement = bool( strchr(";{}", CurrentToken.empty() ? '\0' : CurrentToken[0]) != NULL ); // Add current token.. addtoken(CurrentToken.c_str(), lineno, FileIndex); CurrentToken.clear(); // Read next character.. ch = (char)code.get(); // If '//'.. if (ch == '/') { std::string comment; getline( code, comment ); // Parse in the whole comment // If the comment says something like "fred is deleted" then generate appropriate tokens for that comment = comment + " "; if ( newstatement && comment.find(" deleted ")!=std::string::npos ) { // delete addtoken( "delete", lineno, FileIndex ); // fred std::string::size_type pos1 = comment.find_first_not_of(" \t"); std::string::size_type pos2 = comment.find(" ", pos1); std::string firstWord = comment.substr( pos1, pos2-pos1 ); addtoken( firstWord.c_str(), lineno, FileIndex ); // ; addtoken( ";", lineno, FileIndex ); } lineno++; continue; } // If '/*'.. if (ch == '*') { char chPrev; ch = chPrev = 'A'; while (code.good() && (chPrev!='*' || ch!='/')) { chPrev = ch; ch = (char)code.get(); if (ch == '\n') lineno++; } continue; } // Not a comment.. add token.. addtoken("/", lineno, FileIndex); } // char.. if (ch == '\'') { // Add previous token addtoken(CurrentToken.c_str(), lineno, FileIndex); CurrentToken.clear(); // Read this .. CurrentToken += ch; CurrentToken += (char)code.get(); CurrentToken += (char)code.get(); if (CurrentToken[1] == '\\') CurrentToken += (char)code.get(); // Add token and start on next.. addtoken(CurrentToken.c_str(), lineno, FileIndex); CurrentToken.clear(); continue; } // String.. if (ch == '\"') { addtoken(CurrentToken.c_str(), lineno, FileIndex); CurrentToken.clear(); bool special = false; char c = ch; do { // Append token.. CurrentToken += c; if ( c == '\n' ) ++lineno; // Special sequence '\.' if (special) special = false; else special = (c == '\\'); // Get next character c = (char)code.get(); } while (code.good() && (special || c != '\"')); CurrentToken += '\"'; addtoken(CurrentToken.c_str(), lineno, FileIndex); CurrentToken.clear(); continue; } if (strchr("+-*/%&|^?!=<>[](){};:,.~",ch)) { addtoken(CurrentToken.c_str(), lineno, FileIndex); CurrentToken.clear(); CurrentToken += ch; addtoken(CurrentToken.c_str(), lineno, FileIndex); CurrentToken.clear(); continue; } if (isspace(ch) || iscntrl(ch)) { addtoken(CurrentToken.c_str(), lineno, FileIndex); CurrentToken.clear(); continue; } CurrentToken += ch; } addtoken( CurrentToken.c_str(), lineno, FileIndex ); // Combine tokens.. for (TOKEN *tok = _tokens; tok && tok->next(); tok = tok->next()) { static const char* combineWithNext[][3] = { { "<", "<", "<<" }, { ">", ">", ">>" }, { "&", "&", "&&" }, { "|", "|", "||" }, { "+", "=", "+=" }, { "-", "=", "-=" }, { "*", "=", "*=" }, { "/", "=", "/=" }, { "&", "=", "&=" }, { "|", "=", "|=" }, { "=", "=", "==" }, { "!", "=", "!=" }, { "<", "=", "<=" }, { ">", "=", ">=" }, { ":", ":", "::" }, { "-", ">", "." }, // Replace "->" with "." { "private", ":", "private:" }, { "protected", ":", "protected:" }, { "public", ":", "public:" } }; for (unsigned ui = 0; ui < sizeof(combineWithNext) / sizeof(combineWithNext[0]); ui++) { if ( tok->str() == combineWithNext[ui][0] && tok->next()->str() == combineWithNext[ui][1] ) { tok->setstr(combineWithNext[ui][2]); tok->deleteNext(); } } } // typedef.. for ( TOKEN *tok = _tokens; tok; ) { if ( TOKEN::Match(tok, "typedef %type% %type% ;") ) { const char *type1 = tok->strAt(1); const char *type2 = tok->strAt(2); tok = const_cast(tok->tokAt(4)); for ( TOKEN *tok2 = tok; tok2; tok2 = tok2->next() ) { if ( tok2->str() == type2 ) tok2->setstr(type1); } continue; } else if ( TOKEN::Match(tok, "typedef %type% %type% %type% ;") ) { const char *type1 = tok->strAt(1); const char *type2 = tok->strAt(2); const char *type3 = tok->strAt(3); tok = const_cast(tok->tokAt(5)); for ( TOKEN *tok2 = tok; tok2; tok2 = tok2->next() ) { if ( tok2->str() == type3 ) { tok2->setstr(type1); tok2->insertToken(type2); tok2 = tok2->next(); } } continue; } tok = tok->next(); } // Remove __asm.. for ( TOKEN *tok = _tokens; tok; tok = tok->next() ) { if ( TOKEN::simpleMatch(tok->next(), "__asm {") ) { while ( tok->next() ) { bool last = TOKEN::simpleMatch( tok->next(), "}" ); // Unlink and delete tok->next() tok->deleteNext(); // break if this was the last token to delete.. if (last) break; } } } // Remove "volatile" while ( TOKEN::simpleMatch(_tokens, "volatile") ) { TOKEN *tok = _tokens; _tokens = _tokens->next(); delete tok; } for ( TOKEN *tok = _tokens; tok; tok = tok->next() ) { while ( TOKEN::simpleMatch(tok->next(), "volatile") ) { tok->deleteNext(); } } } //--------------------------------------------------------------------------- void Tokenizer::setVarId() { // Clear all variable ids for ( TOKEN *tok = _tokens; tok; tok = tok->next() ) tok->varId( 0 ); // Set variable ids.. unsigned int _varId = 0; for ( TOKEN *tok = _tokens; tok; tok = tok->next() ) { if ( ! TOKEN::Match(tok, "[;{}(] %type% %var%") ) continue; // Determine name of declared variable.. const char *varname = 0; TOKEN *tok2 = tok->next(); while ( tok2 && ! TOKEN::Match( tok2, "[;[=(]" ) ) { if ( tok2->isName() ) varname = tok2->strAt(0); else if ( tok2->str() != "*" ) break; tok2 = tok2->next(); } // Variable declaration found => Set variable ids if ( TOKEN::Match(tok2, "[;[=]") && varname ) { ++_varId; int indentlevel = 0; int parlevel = 0; for ( tok2 = tok->next(); tok2; tok2 = tok2->next() ) { if ( tok2->str() == varname ) tok2->varId( _varId ); else if ( tok2->str() == "{" ) ++indentlevel; else if ( tok2->str() == "}" ) { --indentlevel; if ( indentlevel < 0 ) break; } else if ( tok2->str() == "(" ) ++parlevel; else if ( tok2->str() == ")" ) --parlevel; else if ( parlevel < 0 && tok2->str() == ";" ) break; } } } // Struct/Class members for ( TOKEN *tok = _tokens; tok; tok = tok->next() ) { if ( tok->varId() != 0 && TOKEN::Match(tok, "%var% . %var%") && tok->tokAt(2)->varId() == 0 ) { ++_varId; const std::string pattern(std::string("%varid% . ") + tok->strAt(2)); for ( TOKEN *tok2 = tok; tok2; tok2 = tok2->next() ) { if ( TOKEN::Match(tok2, pattern.c_str(), 0, tok->varId()) ) tok2->next()->next()->varId( _varId ); } } } } //--------------------------------------------------------------------------- // Simplify token list //--------------------------------------------------------------------------- void Tokenizer::simplifyTokenList() { // Remove the keyword 'unsigned' for ( TOKEN *tok = _tokens; tok; tok = tok->next() ) { if (tok->next() && (tok->next()->str() == "unsigned")) { tok->deleteNext(); } } // Replace constants.. for (TOKEN *tok = _tokens; tok; tok = tok->next()) { if (TOKEN::Match(tok,"const %type% %var% = %num% ;")) { const char *sym = tok->strAt(2); const char *num = tok->strAt(4); for (TOKEN *tok2 = _gettok(tok,6); tok2; tok2 = tok2->next()) { if (tok2->str() == sym) { tok2->setstr(num); } } } } // Fill the map _typeSize.. _typeSize.clear(); _typeSize["char"] = sizeof(char); _typeSize["short"] = sizeof(short); _typeSize["int"] = sizeof(int); _typeSize["long"] = sizeof(long); _typeSize["float"] = sizeof(float); _typeSize["double"] = sizeof(double); for (TOKEN *tok = _tokens; tok; tok = tok->next()) { if (TOKEN::Match(tok,"class %var%")) { _typeSize[tok->strAt(1)] = 11; } else if (TOKEN::Match(tok, "struct %var%")) { _typeSize[tok->strAt(1)] = 13; } } // Replace 'sizeof(type)'.. for (TOKEN *tok = _tokens; tok; tok = tok->next()) { if (tok->str() != "sizeof") continue; if (TOKEN::Match(tok, "sizeof ( %type% * )")) { std::ostringstream str; // 'sizeof(type *)' has the same size as 'sizeof(char *)' str << sizeof(char *); tok->setstr( str.str().c_str() ); for (int i = 0; i < 4; i++) { tok->deleteNext(); } } else if (TOKEN::Match(tok, "sizeof ( %type% )")) { const char *type = tok->strAt( 2); int size = SizeOfType(type); if (size > 0) { std::ostringstream str; str << size; tok->setstr( str.str().c_str() ); for (int i = 0; i < 3; i++) { tok->deleteNext(); } } } else if (TOKEN::Match(tok, "sizeof ( * %var% )")) { tok->setstr("100"); for ( int i = 0; i < 4; ++i ) tok->deleteNext(); } } // Replace 'sizeof(var)' for (TOKEN *tok = _tokens; tok; tok = tok->next()) { // type array [ num ] ; if ( ! TOKEN::Match(tok, "%type% %var% [ %num% ] ;") ) continue; int size = SizeOfType(tok->aaaa()); if (size <= 0) continue; const char *varname = tok->strAt( 1); int total_size = size * atoi( tok->strAt( 3) ); // Replace 'sizeof(var)' with number int indentlevel = 0; for ( TOKEN *tok2 = _gettok(tok,5); tok2; tok2 = tok2->next() ) { if (tok2->str() == "{") { ++indentlevel; } else if (tok2->str() == "}") { --indentlevel; if (indentlevel < 0) break; } // Todo: TOKEN::Match varname directly else if (TOKEN::Match(tok2, "sizeof ( %var% )")) { if (strcmp(tok2->strAt(2), varname) == 0) { std::ostringstream str; str << total_size; tok2->setstr(str.str().c_str()); // Delete the other tokens.. for (int i = 0; i < 3; i++) { tok2->deleteNext(); } } } } } // Simple calculations.. for ( bool done = false; !done; done = true ) { for (TOKEN *tok = _tokens; tok; tok = tok->next()) { if (TOKEN::simpleMatch(tok->next(), "* 1") || TOKEN::simpleMatch(tok->next(), "1 *")) { for (int i = 0; i < 2; i++) tok->deleteNext(); done = false; } // (1-2) if (TOKEN::Match(tok, "[[,(=<>] %num% [+-*/] %num% [],);=<>]")) { int i1 = atoi(tok->strAt(1)); int i2 = atoi(tok->strAt(3)); if ( i2 == 0 && *(tok->strAt(2)) == '/' ) { continue; } switch (*(tok->strAt(2))) { case '+': i1 += i2; break; case '-': i1 -= i2; break; case '*': i1 *= i2; break; case '/': i1 /= i2; break; } tok = tok->next(); std::ostringstream str; str << i1; tok->setstr(str.str().c_str()); for (int i = 0; i < 2; i++) { tok->deleteNext(); } done = false; } } } // Replace "*(str + num)" => "str[num]" for (TOKEN *tok = _tokens; tok; tok = tok->next()) { if ( ! strchr(";{}(=<>", tok->aaaa0()) ) continue; TOKEN *next = tok->next(); if ( ! next ) break; if (TOKEN::Match(next, "* ( %var% + %num% )")) { const char *str[4] = {"var","[","num","]"}; str[0] = tok->strAt(3); str[2] = tok->strAt(5); for (int i = 0; i < 4; i++) { tok = tok->next(); tok->setstr(str[i]); } tok->deleteNext(); tok->deleteNext(); } } // Split up variable declarations if possible.. for (TOKEN *tok = _tokens; tok; tok = tok->next()) { if ( ! TOKEN::Match(tok, "[{};]") ) continue; TOKEN *type0 = tok->next(); if (!TOKEN::Match(type0, "%type%")) continue; if (TOKEN::Match(type0, "else|return")) continue; TOKEN *tok2 = NULL; unsigned int typelen = 0; if ( TOKEN::Match(type0, "%type% %var% ,|=") ) { if ( type0->next()->str() != "operator" ) { tok2 = _gettok(type0, 2); // The ',' or '=' token typelen = 1; } } else if ( TOKEN::Match(type0, "%type% * %var% ,|=") ) { if ( type0->next()->next()->str() != "operator" ) { tok2 = _gettok(type0, 3); // The ',' token typelen = 1; } } else if ( TOKEN::Match(type0, "%type% %var% [ %num% ] ,|=") ) { tok2 = _gettok(type0, 5); // The ',' token typelen = 1; } else if ( TOKEN::Match(type0, "%type% * %var% [ %num% ] ,|=") ) { tok2 = _gettok(type0, 6); // The ',' token typelen = 1; } else if ( TOKEN::Match(type0, "struct %type% %var% ,|=") ) { tok2 = _gettok(type0, 3); typelen = 2; } else if ( TOKEN::Match(type0, "struct %type% * %var% ,|=") ) { tok2 = _gettok(type0, 4); typelen = 2; } if (tok2) { if (tok2->str() == ",") { tok2->setstr(";"); InsertTokens(tok2, type0, typelen); } else { TOKEN *eq = tok2; int parlevel = 0; while (tok2) { if ( strchr("{(", tok2->aaaa0()) ) { parlevel++; } else if ( strchr("})", tok2->aaaa0()) ) { if (parlevel<0) break; parlevel--; } else if ( parlevel==0 && strchr(";,",tok2->aaaa0()) ) { // "type var =" => "type var; var =" TOKEN *VarTok = _gettok(type0,typelen); if (VarTok->aaaa0()=='*') VarTok = VarTok->next(); InsertTokens(eq, VarTok, 2); eq->setstr(";"); // "= x, " => "= x; type " if (tok2->str() == ",") { tok2->setstr(";"); InsertTokens( tok2, type0, typelen ); } break; } tok2 = tok2->next(); } } } } // Replace NULL with 0.. for ( TOKEN *tok = _tokens; tok; tok = tok->next() ) { if ( tok->str() == "NULL" ) tok->setstr("0"); } // Replace pointer casts of 0.. "(char *)0" => "0" for ( TOKEN *tok = _tokens; tok; tok = tok->next() ) { if ( TOKEN::Match(tok->next(), "( %type% * ) 0") || TOKEN::Match(tok->next(),"( %type% %type% * ) 0") ) { while (!TOKEN::simpleMatch(tok->next(),"0")) tok->deleteNext(); } } simplifyIfAddBraces(); bool modified = true; while ( modified ) { modified = false; modified |= simplifyConditions(); modified |= simplifyCasts(); modified |= simplifyFunctionReturn(); modified |= simplifyKnownVariables(); modified |= removeReduntantConditions(); } } //--------------------------------------------------------------------------- const TOKEN *Tokenizer::findClosing( const TOKEN *tok, const char *start, const char *end ) { if( !tok ) return 0; // Find the closing "}" int indentLevel = 0; for ( const TOKEN *closing = tok->next(); closing; closing = closing->next() ) { if( closing->str() == start ) { indentLevel++; continue; } if( closing->str() == end ) indentLevel--; if( indentLevel >= 0 ) continue; // Closing } is found. return closing; } return 0; } bool Tokenizer::removeReduntantConditions() { bool ret = false; for ( TOKEN *tok = _tokens; tok; tok = tok->next() ) { if (!TOKEN::simpleMatch(tok, "if")) continue; if (!TOKEN::Match(tok->tokAt(1), "( %bool% ) {")) continue; // Find matching else const TOKEN *elseTag = 0; // Find the closing "}" elseTag = Tokenizer::findClosing( tok->tokAt( 4 ), "{", "}" ); if( elseTag ) elseTag = elseTag->next(); bool boolValue = false; if( tok->tokAt( 2 )->str() == "true" ) boolValue = true; // Handle if with else if( elseTag && elseTag->str()=="else" ) { if( TOKEN::simpleMatch( elseTag->next(), "if" ) ) { // Handle "else if" if( boolValue == false ) { // Convert "if( false ) {aaa;} else if() {bbb;}" => "if() {bbb;}" TOKEN::eraseTokens( tok, elseTag->tokAt( 2 ) ); ret = true; } else { // Keep first if, remove every else if and else after it const TOKEN *lastTagInIf = elseTag->tokAt( 2 ); while( lastTagInIf ) { if( lastTagInIf->str() == "(" ) { lastTagInIf = Tokenizer::findClosing( lastTagInIf, "(", ")" ); lastTagInIf = lastTagInIf->next(); } lastTagInIf = Tokenizer::findClosing( lastTagInIf, "{", "}" ); lastTagInIf = lastTagInIf->next(); if( !TOKEN::simpleMatch( lastTagInIf, "else" ) ) break; lastTagInIf = lastTagInIf->next(); if( TOKEN::simpleMatch( lastTagInIf, "if" ) ) lastTagInIf = lastTagInIf->next(); } TOKEN::eraseTokens( elseTag->previous(), lastTagInIf ); ret = true; } } else { // Handle else if( boolValue == false ) { // Convert "if( false ) {aaa;} else {bbb;}" => "{bbb;}" or ";{bbb;}" if( tok->previous() ) tok = tok->previous(); else tok->setstr( ";" ); TOKEN::eraseTokens( tok, elseTag->tokAt( 1 ) ); } else { if( TOKEN::simpleMatch( elseTag->tokAt( 1 ), "{" ) ) { // Convert "if( true ) {aaa;} else {bbb;}" => "{aaa;}" const TOKEN *end = Tokenizer::findClosing( elseTag->tokAt( 1 ), "{", "}" ); if( !end ) { // Possibly syntax error in code return false; } // Remove the "else { aaa; }" TOKEN::eraseTokens( elseTag->previous(), end->tokAt( 1 ) ); } // Remove "if( true )" if( tok->previous() ) tok = tok->previous(); else tok->setstr( ";" ); TOKEN::eraseTokens( tok, tok->tokAt(5) ); } ret = true; } } // Handle if without else else { if( boolValue == false ) { // Remove if and its content if( tok->previous() ) tok = tok->previous(); else tok->setstr( ";" ); TOKEN::eraseTokens( tok, elseTag ); } else { // convert "if( true ) {aaa;}" => "{aaa;}" if( tok->previous() ) tok = tok->previous(); else tok->setstr( ";" ); TOKEN::eraseTokens( tok, tok->tokAt( 5 ) ); } ret = true; } } return ret; } bool Tokenizer::simplifyIfAddBraces() { bool ret = false; for ( TOKEN *tok = _tokens; tok; tok = tok ? tok->next() : NULL ) { if ( TOKEN::Match(tok, "if|for|while (") ) { // Goto the ending ')' int parlevel = 1; tok = tok->next(); while ( parlevel >= 1 && (tok = tok->next()) ) { if ( tok->str() == "(" ) ++parlevel; else if ( tok->str() == ")" ) --parlevel; } // ')' should be followed by '{' if (!tok || TOKEN::simpleMatch(tok, ") {")) continue; } else if ( tok->str() == "else" ) { // An else followed by an if or brace don't need to be processed further if ( TOKEN::Match( tok, "else if|{" ) ) continue; } else { continue; } // insert open brace.. tok->insertToken("{"); tok = tok->next(); // insert close brace.. // In most cases it would work to just search for the next ';' and insert a closing brace after it. // But here are special cases.. // * if (cond) for (;;) break; // * if (cond1) if (cond2) { } int parlevel = 0; int indentlevel = 0; while ( (tok = tok->next()) != NULL ) { if ( tok->str() == "{" ) ++indentlevel; else if ( tok->str() == "}" ) { --indentlevel; if ( indentlevel == 0 ) break; } else if ( tok->str() == "(" ) ++parlevel; else if ( tok->str() == ")" ) --parlevel; else if ( indentlevel == 0 && parlevel == 0 && tok->str() == ";" ) break; } if ( tok ) { tok->insertToken("}"); ret = true; } } return ret; } bool Tokenizer::simplifyConditions() { bool ret = false; for ( TOKEN *tok = _tokens; tok; tok = tok->next() ) { if (TOKEN::simpleMatch(tok, "( true &&") || TOKEN::simpleMatch(tok, "&& true &&") || TOKEN::simpleMatch(tok->next(), "&& true )")) { tok->deleteNext(); tok->deleteNext(); ret = true; } else if (TOKEN::simpleMatch(tok, "( false ||") || TOKEN::simpleMatch(tok, "|| false ||") || TOKEN::simpleMatch(tok->next(), "|| false )")) { tok->deleteNext(); tok->deleteNext(); ret = true; } // Change numeric constant in condition to "true" or "false" const TOKEN *tok2 = tok->tokAt(2); if ((tok->str()=="(" || tok->str()=="&&" || tok->str()=="||") && TOKEN::Match(tok->next(), "%num%") && tok2 && (tok2->str()==")" || tok2->str()=="&&" || tok2->str()=="||")) { tok->next()->setstr((tok->next()->str() != "0") ? "true" : "false"); ret = true; } // Reduce "(%num% == %num%)" => "(true)"/"(false)" const TOKEN *tok4 = tok->tokAt(4); if ( ! tok4 ) break; if ( (tok->str()=="&&" || tok->str()=="||" || tok->str()=="(") && TOKEN::Match(tok->tokAt(1), "%num% %any% %num%") && (tok4->str()=="&&" || tok4->str()=="||" || tok4->str()==")") ) { double op1 = (strstr(tok->strAt(1), "0x")) ? strtol(tok->strAt(1),0,16) : atof( tok->strAt(1) ); double op2 = (strstr(tok->strAt(3), "0x")) ? strtol(tok->strAt(3),0,16) : atof( tok->strAt(3) ); std::string cmp = tok->strAt(2); bool result = false; if ( cmp == "==" ) result = (op1 == op2); else if ( cmp == "!=" ) result = (op1 != op2); else if ( cmp == ">=" ) result = (op1 >= op2); else if ( cmp == ">" ) result = (op1 > op2); else if ( cmp == "<=" ) result = (op1 <= op2); else if ( cmp == "<" ) result = (op1 < op2); else cmp = ""; if ( ! cmp.empty() ) { tok = tok->next(); tok->deleteNext(); tok->deleteNext(); tok->setstr( result ? "true" : "false" ); ret = true; } } } return ret; } bool Tokenizer::simplifyCasts() { bool ret = false; for ( TOKEN *tok = _tokens; tok; tok = tok->next() ) { if ( TOKEN::Match(tok->next(), "( %type% * )") ) { tok->deleteNext(); tok->deleteNext(); tok->deleteNext(); tok->deleteNext(); ret = true; } else if ( TOKEN::Match(tok->next(), "dynamic_cast|reinterpret_cast|const_cast|static_cast <" ) ) { while ( tok->next() && tok->next()->str() != ">" ) tok->deleteNext(); tok->deleteNext(); tok->deleteNext(); TOKEN *tok2 = tok; int parlevel = 0; while ( tok2->next() && parlevel >= 0 ) { tok2 = tok2->next(); if ( TOKEN::simpleMatch(tok2->next(), "(") ) ++parlevel; else if ( TOKEN::simpleMatch(tok2->next(), ")") ) --parlevel; } if (tok2->next()) tok2->deleteNext(); ret = true; } } return ret; } bool Tokenizer::simplifyFunctionReturn() { bool ret = false; int indentlevel = 0; for ( const TOKEN *tok = tokens(); tok; tok = tok->next() ) { if ( tok->str() == "{" ) ++indentlevel; else if ( tok->str() == "}" ) --indentlevel; else if ( indentlevel == 0 && TOKEN::Match(tok, "%var% ( ) { return %num% ; }") ) { std::ostringstream pattern; pattern << "[(=+-*/] " << tok->str() << " ( ) [;)+-*/]"; for ( TOKEN *tok2 = _tokens; tok2; tok2 = tok2->next() ) { if ( TOKEN::Match(tok2, pattern.str().c_str()) ) { tok2 = tok2->next(); tok2->setstr( tok->strAt(5) ); tok2->deleteNext(); tok2->deleteNext(); ret = true; } } } } return ret; } bool Tokenizer::simplifyKnownVariables() { bool ret = false; for ( TOKEN *tok = _tokens; tok; tok = tok->next() ) { // Search for a block of code if ( ! TOKEN::Match(tok, ") const| {") ) continue; // parse the block of code.. int indentlevel = 0; for ( TOKEN *tok2 = tok; tok2; tok2 = tok2->next() ) { if ( tok2->str() == "{" ) ++indentlevel; else if ( tok2->str() == "}" ) { --indentlevel; if ( indentlevel <= 0 ) break; } else if ( TOKEN::Match(tok2, "%var% = %num% ;") || TOKEN::Match(tok2, "%var% = %bool% ;")) { unsigned int varid = tok2->varId(); if( varid == 0 ) continue; for ( TOKEN *tok3 = tok2->next(); tok3; tok3 = tok3->next() ) { // Perhaps it's a loop => bail out if ( TOKEN::Match(tok3, "[{}]") ) break; // Variable is used somehow in a non-defined pattern => bail out if ( tok3->varId() == varid ) break; // Replace variable with numeric constant.. if ( TOKEN::Match(tok3, "if ( %varid% )", 0, varid) ) { tok3 = tok3->next()->next(); tok3->setstr( tok2->strAt(2) ); ret = true; } } } } } return ret; } //--------------------------------------------------------------------------- // Helper functions for handling the tokens list //--------------------------------------------------------------------------- //--------------------------------------------------------------------------- const TOKEN *Tokenizer::GetFunctionTokenByName( const char funcname[] ) const { for ( unsigned int i = 0; i < _functionList.size(); ++i ) { if ( _functionList[i]->str() == funcname ) { return _functionList[i]; } } return NULL; } void Tokenizer::fillFunctionList() { _functionList.clear(); bool classfunc = false; int indentlevel = 0; for ( const TOKEN *tok = _tokens; tok; tok = tok->next() ) { if ( tok->str() == "{" ) ++indentlevel; else if ( tok->str() == "}" ) --indentlevel; if (indentlevel > 0) { continue; } if (strchr("};", tok->aaaa0())) classfunc = false; else if ( tok->str() == "::" ) classfunc = true; else if (TOKEN::Match(tok, "%var% (")) { // Check if this is the first token of a function implementation.. for ( const TOKEN *tok2 = tok->tokAt(2); tok2; tok2 = tok2->next() ) { if ( tok2->str() == ";" ) { tok = tok2; break; } else if ( tok2->str() == "{" ) { break; } else if ( tok2->str() == ")" ) { if ( TOKEN::Match(tok2, ") const| {") ) { _functionList.push_back( tok ); tok = tok2; } else { tok = tok2; while (tok->next() && !strchr(";{", tok->next()->aaaa0())) tok = tok->next(); } break; } } } } // If the _functionList functions with duplicate names, remove them // TODO this will need some better handling for ( unsigned int func1 = 0; func1 < _functionList.size(); ) { bool hasDuplicates = false; for ( unsigned int func2 = func1 + 1; func2 < _functionList.size(); ) { if ( _functionList[func1]->str() == _functionList[func2]->str() ) { hasDuplicates = true; _functionList.erase( _functionList.begin() + func2 ); } else { ++func2; } } if ( ! hasDuplicates ) { ++func1; } else { _functionList.erase( _functionList.begin() + func1 ); } } } //--------------------------------------------------------------------------- // Deallocate lists.. void Tokenizer::DeallocateTokens() { deleteTokens( _tokens ); _tokens = 0; _tokensBack = 0; while (_dsymlist) { struct DefineSymbol *next = _dsymlist->next; free(_dsymlist->name); free(_dsymlist->value); delete _dsymlist; _dsymlist = next; } _files.clear(); } void Tokenizer::deleteTokens(TOKEN *tok) { while (tok) { TOKEN *next = tok->next(); delete tok; tok = next; } } //--------------------------------------------------------------------------- const char *Tokenizer::getParameterName( const TOKEN *ftok, int par ) { int _par = 1; for ( ; ftok; ftok = ftok->next()) { if ( ftok->str()=="," ) ++_par; if ( par==_par && TOKEN::Match(ftok, "%var% [,)]") ) return ftok->aaaa(); } return NULL; } //--------------------------------------------------------------------------- std::string Tokenizer::fileLine( const TOKEN *tok ) const { std::ostringstream ostr; ostr << "[" << _files.at(tok->fileIndex()) << ":" << tok->linenr() << "]"; return ostr.str(); } //--------------------------------------------------------------------------- bool Tokenizer::SameFileName( const char fname1[], const char fname2[] ) { #ifdef __linux__ return bool( strcmp(fname1, fname2) == 0 ); #endif #ifdef __GNUC__ return bool( strcasecmp(fname1, fname2) == 0 ); #endif #ifdef __BORLANDC__ return bool( stricmp(fname1, fname2) == 0 ); #endif #ifdef _MSC_VER return bool( _stricmp(fname1, fname2) == 0 ); #endif } //---------------------------------------------------------------------------