diff --git a/src/tokenize.cpp b/src/tokenize.cpp index 1a9e9fd3b..00f65ae49 100644 --- a/src/tokenize.cpp +++ b/src/tokenize.cpp @@ -368,7 +368,7 @@ bool Tokenizer::tokenize(std::istream &code, const char FileName[]) // Combine tokens.. for (Token *tok = _tokens; tok && tok->next(); tok = tok->next()) { - static const char* combineWithNext[][3] = + static const char * const combineWithNext[][3] = { { "<", "<", "<<" },