0
|
1 |
/****************************************************************************
|
|
2 |
**
|
|
3 |
** Copyright (C) 2001-2004 Roberto Raggi
|
|
4 |
** Copyright (C) 2009 Nokia Corporation and/or its subsidiary(-ies).
|
|
5 |
** All rights reserved.
|
|
6 |
** Contact: Nokia Corporation (qt-info@nokia.com)
|
|
7 |
**
|
|
8 |
** This file is part of the qt3to4 porting application of the Qt Toolkit.
|
|
9 |
**
|
|
10 |
** $QT_BEGIN_LICENSE:LGPL$
|
|
11 |
** No Commercial Usage
|
|
12 |
** This file contains pre-release code and may not be distributed.
|
|
13 |
** You may use this file in accordance with the terms and conditions
|
|
14 |
** contained in the Technology Preview License Agreement accompanying
|
|
15 |
** this package.
|
|
16 |
**
|
|
17 |
** GNU Lesser General Public License Usage
|
|
18 |
** Alternatively, this file may be used under the terms of the GNU Lesser
|
|
19 |
** General Public License version 2.1 as published by the Free Software
|
|
20 |
** Foundation and appearing in the file LICENSE.LGPL included in the
|
|
21 |
** packaging of this file. Please review the following information to
|
|
22 |
** ensure the GNU Lesser General Public License version 2.1 requirements
|
|
23 |
** will be met: http://www.gnu.org/licenses/old-licenses/lgpl-2.1.html.
|
|
24 |
**
|
|
25 |
** In addition, as a special exception, Nokia gives you certain additional
|
|
26 |
** rights. These rights are described in the Nokia Qt LGPL Exception
|
|
27 |
** version 1.1, included in the file LGPL_EXCEPTION.txt in this package.
|
|
28 |
**
|
|
29 |
** If you have questions regarding the use of this file, please contact
|
|
30 |
** Nokia at qt-info@nokia.com.
|
|
31 |
**
|
|
32 |
**
|
|
33 |
**
|
|
34 |
**
|
|
35 |
**
|
|
36 |
**
|
|
37 |
**
|
|
38 |
**
|
|
39 |
** $QT_END_LICENSE$
|
|
40 |
**
|
|
41 |
****************************************************************************/
|
|
42 |
|
|
43 |
#include "rpp.h"
|
|
44 |
#include "rppexpressionbuilder.h"
|
|
45 |
|
|
46 |
QT_BEGIN_NAMESPACE
|
|
47 |
|
|
48 |
using namespace TokenEngine;
|
|
49 |
|
|
50 |
namespace Rpp
|
|
51 |
{
|
|
52 |
|
|
53 |
Preprocessor::Preprocessor()
|
|
54 |
: lexerTokenIndex(0), numTokens(0)
|
|
55 |
{
|
|
56 |
|
|
57 |
}
|
|
58 |
|
|
59 |
Source *Preprocessor::parse(const TokenEngine::TokenContainer &tokenContainer,
|
|
60 |
const QVector<Type> &tokenTypeList, TypedPool<Item> *memoryPool)
|
|
61 |
{
|
|
62 |
m_memoryPool = memoryPool;
|
|
63 |
Source *m_source = createNode<Source>(m_memoryPool); //node whith no parent
|
|
64 |
m_tokenContainer = tokenContainer;
|
|
65 |
m_tokenTypeList = tokenTypeList;
|
|
66 |
lexerTokenIndex = 0;
|
|
67 |
numTokens = m_tokenContainer.count();
|
|
68 |
|
|
69 |
if(m_tokenContainer.count() != tokenTypeList.count()) {
|
|
70 |
emit error(QLatin1String("Error"), QLatin1String("Internal error in preprocessor: Number of tokens is not equal to number of types in type list"));
|
|
71 |
return m_source;
|
|
72 |
}
|
|
73 |
|
|
74 |
if(tokenTypeList.isEmpty()) {
|
|
75 |
// emit error("Warning:", "Trying to parse empty source file");
|
|
76 |
return m_source;
|
|
77 |
}
|
|
78 |
Q_ASSERT(m_source->toItemComposite());
|
|
79 |
parseGroup(m_source);
|
|
80 |
|
|
81 |
return m_source;
|
|
82 |
}
|
|
83 |
|
|
84 |
// group-part
|
|
85 |
// group group-part
|
|
86 |
bool Preprocessor::parseGroup(Item *group)
|
|
87 |
{
|
|
88 |
Q_ASSERT(group->toItemComposite());
|
|
89 |
bool gotGroup = false;
|
|
90 |
while(lexerTokenIndex < numTokens) {
|
|
91 |
if (!parseGroupPart(group))
|
|
92 |
break;
|
|
93 |
gotGroup = true;
|
|
94 |
}
|
|
95 |
return gotGroup;
|
|
96 |
}
|
|
97 |
|
|
98 |
//if-section (# if / # ifdef / #ifndef )
|
|
99 |
//control-line ( #include / etc )
|
|
100 |
//# non-directive ( # text newline
|
|
101 |
//text-line (text newline )
|
|
102 |
bool Preprocessor::parseGroupPart(Item *group)
|
|
103 |
{
|
|
104 |
//cout << "parse group part" << endl;
|
|
105 |
Q_ASSERT(group->toItemComposite());
|
|
106 |
|
|
107 |
//look up first significant token
|
|
108 |
Type token = lookAhead();
|
|
109 |
if(token == Token_eof)
|
|
110 |
return false;
|
|
111 |
|
|
112 |
//look for '#'
|
|
113 |
if(token != Token_preproc)
|
|
114 |
return parseTextLine(group);
|
|
115 |
|
|
116 |
//look up first significant token after the '#'
|
|
117 |
token = lookAheadSkipHash();
|
|
118 |
if(token == Token_eof)
|
|
119 |
return false;
|
|
120 |
|
|
121 |
// Check if we are at the end of a group. This is not an neccesarely an
|
|
122 |
// error, it happens when we reach an #endif for example.
|
|
123 |
if (token == Token_directive_elif || token == Token_directive_else ||
|
|
124 |
token == Token_directive_endif)
|
|
125 |
return false;
|
|
126 |
|
|
127 |
// if-section?
|
|
128 |
if(token == Token_directive_if || token == Token_directive_ifdef ||
|
|
129 |
token == Token_directive_ifndef)
|
|
130 |
return parseIfSection(group);
|
|
131 |
|
|
132 |
// control-line?
|
|
133 |
if (token == Token_directive_define)
|
|
134 |
return parseDefineDirective(group);
|
|
135 |
if (token == Token_directive_undef)
|
|
136 |
return parseUndefDirective(group);
|
|
137 |
if (token == Token_directive_include)
|
|
138 |
return parseIncludeDirective(group);
|
|
139 |
if (token == Token_directive_error)
|
|
140 |
return parseErrorDirective(group);
|
|
141 |
if (token == Token_directive_pragma)
|
|
142 |
return parsePragmaDirective(group);
|
|
143 |
|
|
144 |
return parseNonDirective(group);
|
|
145 |
}
|
|
146 |
|
|
147 |
// if-section -> if-group elif-groups[opt] else-group[opt] endif-line
|
|
148 |
bool Preprocessor::parseIfSection(Item *group)
|
|
149 |
{
|
|
150 |
// cout << "parse if section" << endl ;
|
|
151 |
Q_ASSERT(group->toItemComposite());
|
|
152 |
IfSection *ifSection = createNode<IfSection>(m_memoryPool, group);
|
|
153 |
group->toItemComposite()->add(ifSection);
|
|
154 |
|
|
155 |
if (!parseIfGroup(ifSection))
|
|
156 |
return false;
|
|
157 |
|
|
158 |
Type type = lookAheadSkipHash();
|
|
159 |
if(type == Token_directive_elif)
|
|
160 |
if(!parseElifGroups(ifSection))
|
|
161 |
return false;
|
|
162 |
|
|
163 |
type = lookAheadSkipHash();
|
|
164 |
if(type == Token_directive_else)
|
|
165 |
if(!parseElseGroup(ifSection))
|
|
166 |
return false;
|
|
167 |
|
|
168 |
return parseEndifLine(ifSection);
|
|
169 |
}
|
|
170 |
|
|
171 |
bool Preprocessor::parseNonDirective(Item *group)
|
|
172 |
{
|
|
173 |
// cout << "parsenondirective" << endl;
|
|
174 |
Q_ASSERT(group->toItemComposite());
|
|
175 |
TokenSection tokenSection = readLine();
|
|
176 |
if(tokenSection.count() == 0)
|
|
177 |
return false;
|
|
178 |
|
|
179 |
NonDirective *nonDirective = createNode<NonDirective>(m_memoryPool, group);
|
|
180 |
group->toItemComposite()->add(nonDirective);
|
|
181 |
nonDirective->setTokenSection(tokenSection);
|
|
182 |
return true;
|
|
183 |
}
|
|
184 |
|
|
185 |
|
|
186 |
bool Preprocessor::parseTextLine(Item *group)
|
|
187 |
{
|
|
188 |
//cout << "parsetextline" << endl;
|
|
189 |
Q_ASSERT(group->toItemComposite());
|
|
190 |
const TokenSection tokenSection = readLine();
|
|
191 |
// cout << tokenSection.fullText().constData() << endl;
|
|
192 |
|
|
193 |
if(tokenSection.count() == 0)
|
|
194 |
return false;
|
|
195 |
|
|
196 |
Text *text = createNode<Text>(m_memoryPool, group);
|
|
197 |
group->toItemComposite()->add(text);
|
|
198 |
text->setTokenSection(tokenSection);
|
|
199 |
|
|
200 |
// Create Token-derived nodes and atach to text
|
|
201 |
QVector<Token *> tokens;
|
|
202 |
tokens.reserve(tokenSection.count());
|
|
203 |
for (int t = 0; t < tokenSection.count(); ++t) {
|
|
204 |
Token *node = 0;
|
|
205 |
const int containerIndex = tokenSection.containerIndex(t);
|
|
206 |
switch(m_tokenTypeList.at(containerIndex)) {
|
|
207 |
case Token_identifier:
|
|
208 |
case Token_defined:
|
|
209 |
case Token_directive_if:
|
|
210 |
case Token_directive_elif:
|
|
211 |
case Token_directive_else:
|
|
212 |
case Token_directive_undef:
|
|
213 |
case Token_directive_endif:
|
|
214 |
case Token_directive_ifdef:
|
|
215 |
case Token_directive_ifndef:
|
|
216 |
case Token_directive_define:
|
|
217 |
case Token_directive_include:
|
|
218 |
node = createNode<IdToken>(m_memoryPool, text);
|
|
219 |
break;
|
|
220 |
case Token_line_comment:
|
|
221 |
node = createNode<LineComment>(m_memoryPool, text);
|
|
222 |
break;
|
|
223 |
case Token_multiline_comment:
|
|
224 |
node = createNode<MultiLineComment>(m_memoryPool, text);
|
|
225 |
break;
|
|
226 |
case Token_whitespaces:
|
|
227 |
case Token_char_literal:
|
|
228 |
case Token_string_literal:
|
|
229 |
default:
|
|
230 |
node = createNode<NonIdToken>(m_memoryPool, text);
|
|
231 |
break;
|
|
232 |
}
|
|
233 |
Q_ASSERT(node);
|
|
234 |
node->setToken(containerIndex);
|
|
235 |
tokens.append(node);
|
|
236 |
}
|
|
237 |
|
|
238 |
text->setTokens(tokens);
|
|
239 |
|
|
240 |
return true;
|
|
241 |
}
|
|
242 |
|
|
243 |
// if-group -> ifDirective
|
|
244 |
// if-group -> ifdefDirevtive
|
|
245 |
// if-group -> ifndefDirevtive
|
|
246 |
bool Preprocessor::parseIfGroup(IfSection *ifSection)
|
|
247 |
{
|
|
248 |
// cout << "parse if group" << endl;
|
|
249 |
Q_ASSERT(ifSection->toItemComposite());
|
|
250 |
bool result;
|
|
251 |
const Type type = lookAheadSkipHash();
|
|
252 |
if (type == Token_directive_ifdef) {
|
|
253 |
IfdefDirective *d = createNode<IfdefDirective>(m_memoryPool, ifSection);
|
|
254 |
result = parseIfdefLikeDirective(d);
|
|
255 |
ifSection->setIfGroup(d);
|
|
256 |
} else if (type == Token_directive_ifndef) {
|
|
257 |
IfndefDirective *d = createNode<IfndefDirective>(m_memoryPool, ifSection);
|
|
258 |
result = parseIfdefLikeDirective(d);
|
|
259 |
ifSection->setIfGroup(d);
|
|
260 |
} else if (type == Token_directive_if) {
|
|
261 |
IfDirective *d = createNode<IfDirective>(m_memoryPool, ifSection);
|
|
262 |
result = parseIfLikeDirective(d);
|
|
263 |
ifSection->setIfGroup(d);
|
|
264 |
} else {
|
|
265 |
result = false;
|
|
266 |
}
|
|
267 |
return result;
|
|
268 |
}
|
|
269 |
|
|
270 |
bool Preprocessor::parseElifGroups(IfSection *ifSection)
|
|
271 |
{
|
|
272 |
//cout << "parse ElifGroups" << endl;
|
|
273 |
bool gotElif = false;
|
|
274 |
while(lookAheadSkipHash() == Token_directive_elif ) {
|
|
275 |
if (!parseElifGroup(ifSection))
|
|
276 |
break;
|
|
277 |
gotElif = true;
|
|
278 |
}
|
|
279 |
return gotElif;
|
|
280 |
}
|
|
281 |
|
|
282 |
bool Preprocessor::parseElifGroup(IfSection *ifSection)
|
|
283 |
{
|
|
284 |
//cout << "parse ElifGroup" << endl;
|
|
285 |
ElifDirective *elifDirective = createNode<ElifDirective>(m_memoryPool, ifSection);
|
|
286 |
ifSection->addElifGroup(elifDirective);
|
|
287 |
return parseIfLikeDirective(elifDirective);
|
|
288 |
}
|
|
289 |
|
|
290 |
bool Preprocessor::parseElseGroup(IfSection *ifSection)
|
|
291 |
{
|
|
292 |
//cout << "parse else group" << endl;
|
|
293 |
TokenSection tokenSection = readLine();
|
|
294 |
if(tokenSection.count() == 0)
|
|
295 |
return false;
|
|
296 |
|
|
297 |
ElseDirective *elseDirective = createNode<ElseDirective>(m_memoryPool, ifSection);
|
|
298 |
ifSection->setElseGroup(elseDirective);
|
|
299 |
elseDirective->setTokenSection(tokenSection);
|
|
300 |
parseGroup(elseDirective);
|
|
301 |
return true;
|
|
302 |
}
|
|
303 |
|
|
304 |
//# endif newline
|
|
305 |
bool Preprocessor::parseEndifLine(IfSection *ifSection)
|
|
306 |
{
|
|
307 |
//cout << "parse endifline" << endl;
|
|
308 |
TokenSection tokenSection = readLine();
|
|
309 |
if(tokenSection.count() == 0)
|
|
310 |
return false;
|
|
311 |
|
|
312 |
EndifDirective *endifDirective = createNode<EndifDirective>(m_memoryPool, ifSection);
|
|
313 |
ifSection->setEndifLine(endifDirective);
|
|
314 |
endifDirective->setTokenSection(tokenSection);
|
|
315 |
|
|
316 |
return true;
|
|
317 |
}
|
|
318 |
|
|
319 |
//parses an "ifdef-like" directive, like #ifdef and #ifndef :)
|
|
320 |
//# ifdef identifier newline group[opt]
|
|
321 |
bool Preprocessor::parseIfdefLikeDirective(IfdefLikeDirective *node)
|
|
322 |
{
|
|
323 |
Q_ASSERT(node->toItemComposite());
|
|
324 |
const TokenSection tokenSection = readLine();
|
|
325 |
const QVector<int> cleanedLine = cleanTokenRange(tokenSection);
|
|
326 |
|
|
327 |
if(cleanedLine.count() < 3)
|
|
328 |
return false;
|
|
329 |
|
|
330 |
node->setTokenSection(tokenSection);
|
|
331 |
node->setIdentifier(TokenList(m_tokenContainer, QVector<int>() << cleanedLine.at(2)));
|
|
332 |
parseGroup(node);
|
|
333 |
|
|
334 |
return true;
|
|
335 |
}
|
|
336 |
|
|
337 |
//# if constant-expression newline group[opt]
|
|
338 |
bool Preprocessor::parseIfLikeDirective(IfLikeDirective *node)
|
|
339 |
{
|
|
340 |
//cout << "parse if-like directive" << endl;
|
|
341 |
Q_ASSERT(node->toItemComposite());
|
|
342 |
TokenSection tokenSection = readLine();
|
|
343 |
QVector<int> cleanedSection = cleanTokenRange(tokenSection);
|
|
344 |
if(cleanedSection.count() < 3)
|
|
345 |
return false;
|
|
346 |
|
|
347 |
cleanedSection.erase(cleanedSection.begin(), cleanedSection.begin() + 2); //remove # and if
|
|
348 |
cleanedSection.pop_back(); //remove endl;
|
|
349 |
|
|
350 |
const TokenList sectionList(m_tokenContainer, cleanedSection);
|
|
351 |
ExpressionBuilder expressionBuilder(sectionList, m_tokenTypeList, m_memoryPool);
|
|
352 |
Expression *expr = expressionBuilder.parse();
|
|
353 |
node->setTokenSection(tokenSection);
|
|
354 |
node->setExpression(expr);
|
|
355 |
|
|
356 |
parseGroup(node);
|
|
357 |
return true;
|
|
358 |
}
|
|
359 |
|
|
360 |
/*
|
|
361 |
# define identifier replacement-list new-line
|
|
362 |
# define identifier lparen identifier-list[opt] ) replacement-list new-line
|
|
363 |
# define identifier lparen ... ) replacement-list new-line
|
|
364 |
# define identifier lparen identifier-list, ... ) replacement-list new-line
|
|
365 |
*/
|
|
366 |
bool Preprocessor::parseDefineDirective(Item *group)
|
|
367 |
{
|
|
368 |
Q_ASSERT(group->toItemComposite());
|
|
369 |
const TokenSection line = readLine();
|
|
370 |
const QVector<int> cleanedLine = cleanTokenRange(line);
|
|
371 |
if(cleanedLine.count() < 3)
|
|
372 |
return false;
|
|
373 |
|
|
374 |
// get identifier
|
|
375 |
const int identifier = cleanedLine.at(2); //skip "#" and "define"
|
|
376 |
DefineDirective *defineDirective = 0;
|
|
377 |
int replacementListStart;
|
|
378 |
|
|
379 |
// check if this is a macro function
|
|
380 |
if (cleanedLine.count() >= 4
|
|
381 |
&& m_tokenContainer.text(cleanedLine.at(3)) == "("
|
|
382 |
&& !isWhiteSpace(cleanedLine.at(3) - 1)) {
|
|
383 |
MacroFunctionDefinition *macro;
|
|
384 |
macro = createNode<MacroFunctionDefinition>(m_memoryPool, group);
|
|
385 |
|
|
386 |
int tokenIndex = 4; //point to first argument or ')'
|
|
387 |
QVector<int> macroParameterList;
|
|
388 |
while(tokenIndex < cleanedLine.count()) {
|
|
389 |
QByteArray currentText = m_tokenContainer.text(cleanedLine.at(tokenIndex));
|
|
390 |
++tokenIndex;
|
|
391 |
if(currentText == ")")
|
|
392 |
break;
|
|
393 |
if(currentText == ",")
|
|
394 |
continue;
|
|
395 |
macroParameterList.append(cleanedLine.at(tokenIndex - 1));
|
|
396 |
}
|
|
397 |
macro->setParameters(TokenList(m_tokenContainer, macroParameterList));
|
|
398 |
defineDirective = macro;
|
|
399 |
replacementListStart = tokenIndex;
|
|
400 |
} else {
|
|
401 |
MacroDefinition *macro;
|
|
402 |
macro = createNode<MacroDefinition>(m_memoryPool, group);
|
|
403 |
defineDirective = macro;
|
|
404 |
replacementListStart = 3;
|
|
405 |
}
|
|
406 |
Q_ASSERT(defineDirective);
|
|
407 |
|
|
408 |
// This is a bit hackish.. we want the replacement list with whitepspace
|
|
409 |
// tokens, but cleanedLine() has already removed those. And we can't use
|
|
410 |
// the original line, because that may contain escaped newline tokens.
|
|
411 |
// So we remove the esacped newlines and search for the token number
|
|
412 |
// given by cleanedLine.at(replacementListStart)
|
|
413 |
QVector<int> replacementList;
|
|
414 |
const QVector<int> noEscNewline = cleanEscapedNewLines(line);
|
|
415 |
if (replacementListStart < cleanedLine.count()) {
|
|
416 |
const int cleanedLineReplacementListStart = cleanedLine.at(replacementListStart);
|
|
417 |
const int rListStart = noEscNewline.indexOf(cleanedLineReplacementListStart);
|
|
418 |
if (rListStart != -1) {
|
|
419 |
const int skipNewLineToken = 1;
|
|
420 |
for (int i = rListStart; i < noEscNewline.count() - skipNewLineToken; ++i) {
|
|
421 |
const int tokenContainerIndex = noEscNewline.at(i);
|
|
422 |
const Type type = m_tokenTypeList.at(tokenContainerIndex);
|
|
423 |
// Don't append comment tokens.
|
|
424 |
if (type != Token_line_comment && type != Token_multiline_comment) {
|
|
425 |
replacementList.append(tokenContainerIndex);
|
|
426 |
|
|
427 |
}
|
|
428 |
}
|
|
429 |
}
|
|
430 |
}
|
|
431 |
|
|
432 |
defineDirective->setTokenSection(line);
|
|
433 |
defineDirective->setIdentifier(TokenList(m_tokenContainer, QVector<int>() << identifier));
|
|
434 |
defineDirective->setReplacementList(TokenList(m_tokenContainer, replacementList));
|
|
435 |
group->toItemComposite()->add(defineDirective);
|
|
436 |
return true;
|
|
437 |
}
|
|
438 |
|
|
439 |
|
|
440 |
// # undef identifier newline
|
|
441 |
bool Preprocessor::parseUndefDirective(Item *group)
|
|
442 |
{
|
|
443 |
Q_ASSERT(group->toItemComposite());
|
|
444 |
const TokenSection tokenSection = readLine();
|
|
445 |
const QVector<int> cleanedLine = cleanTokenRange(tokenSection);
|
|
446 |
|
|
447 |
if(cleanedLine.count() < 3)
|
|
448 |
return false;
|
|
449 |
|
|
450 |
UndefDirective *undefDirective = createNode<UndefDirective>(m_memoryPool, group);
|
|
451 |
group->toItemComposite()->add(undefDirective);
|
|
452 |
undefDirective->setIdentifier(TokenList(m_tokenContainer, QVector<int>() << cleanedLine.at(2)));
|
|
453 |
undefDirective->setTokenSection(tokenSection);
|
|
454 |
return true;
|
|
455 |
}
|
|
456 |
|
|
457 |
//include pp-tokens new-line
|
|
458 |
bool Preprocessor::parseIncludeDirective(Item *group)
|
|
459 |
{
|
|
460 |
// cout << "parseIncludeDirective" << endl;
|
|
461 |
Q_ASSERT(group->toItemComposite());
|
|
462 |
TokenSection tokenSection = readLine();
|
|
463 |
if(tokenSection.count() == 0)
|
|
464 |
return false;
|
|
465 |
|
|
466 |
const TokenEngine::TokenContainer tokenContainer = tokenSection.tokenContainer(0);
|
|
467 |
IncludeDirective *includeDirective = createNode<IncludeDirective>(m_memoryPool, group);
|
|
468 |
group->toItemComposite()->add(includeDirective);
|
|
469 |
includeDirective->setTokenSection(tokenSection);
|
|
470 |
|
|
471 |
//remove whitepspace and comment tokens
|
|
472 |
TokenList tokenList(m_tokenContainer, cleanTokenRange(tokenSection));
|
|
473 |
|
|
474 |
//iterate through the tokens, look for a string literal or a '<'.
|
|
475 |
int tokenIndex = 0;
|
|
476 |
const int endIndex = tokenList.count();
|
|
477 |
while (tokenIndex < endIndex) {
|
|
478 |
const int containerTokenIndex = tokenList.containerIndex(tokenIndex);
|
|
479 |
if(m_tokenTypeList.at(containerTokenIndex) == Token_string_literal) {
|
|
480 |
QByteArray tokenText = tokenList.text(tokenIndex);
|
|
481 |
includeDirective->setFilename(tokenText.mid(1, tokenText.size() -2)); //remove quotes
|
|
482 |
includeDirective->setFilenameTokens(TokenEngine::TokenList(tokenContainer, QVector<int>() << containerTokenIndex));
|
|
483 |
includeDirective->setIncludeType(IncludeDirective::QuoteInclude);
|
|
484 |
break;
|
|
485 |
} else if(tokenList.text(tokenIndex) == "<") {
|
|
486 |
// We found a <, all following tokens until we read a
|
|
487 |
// > is a part of the file anme
|
|
488 |
QByteArray filename;
|
|
489 |
++tokenIndex; //skip '<'
|
|
490 |
QVector<int> filenameTokens;
|
|
491 |
while(tokenIndex < endIndex) {
|
|
492 |
const QByteArray tokenText = tokenList.text(tokenIndex);
|
|
493 |
if(tokenText == ">")
|
|
494 |
break;
|
|
495 |
filenameTokens.append(tokenList.containerIndex(tokenIndex));
|
|
496 |
filename += tokenText;
|
|
497 |
++tokenIndex;
|
|
498 |
}
|
|
499 |
if(tokenIndex < endIndex) {
|
|
500 |
includeDirective->setFilename(filename);
|
|
501 |
includeDirective->setFilenameTokens(TokenEngine::TokenList(tokenContainer, filenameTokens));
|
|
502 |
includeDirective->setIncludeType(IncludeDirective::AngleBracketInclude);
|
|
503 |
}
|
|
504 |
break;
|
|
505 |
}
|
|
506 |
++tokenIndex;
|
|
507 |
}
|
|
508 |
|
|
509 |
return true;
|
|
510 |
}
|
|
511 |
|
|
512 |
//# error pp-tokens[opt] new-line
|
|
513 |
bool Preprocessor::parseErrorDirective(Item *group)
|
|
514 |
{
|
|
515 |
Q_ASSERT(group->toItemComposite());
|
|
516 |
TokenSection tokenSection = readLine();
|
|
517 |
if(tokenSection.count() == 0)
|
|
518 |
return false;
|
|
519 |
|
|
520 |
ErrorDirective *errorDirective = createNode<ErrorDirective>(m_memoryPool, group);
|
|
521 |
group->toItemComposite()->add(errorDirective);
|
|
522 |
errorDirective->setTokenSection(tokenSection);
|
|
523 |
return true;
|
|
524 |
}
|
|
525 |
|
|
526 |
//# pragma pp-tokens[opt] new-line
|
|
527 |
bool Preprocessor::parsePragmaDirective(Item *group)
|
|
528 |
{
|
|
529 |
Q_ASSERT(group->toItemComposite());
|
|
530 |
TokenSection tokenSection = readLine();
|
|
531 |
if(tokenSection.count() == 0)
|
|
532 |
return false;
|
|
533 |
|
|
534 |
PragmaDirective *pragmaDirective = createNode<PragmaDirective>(m_memoryPool, group);
|
|
535 |
group->toItemComposite()->add(pragmaDirective);
|
|
536 |
pragmaDirective->setTokenSection(tokenSection);
|
|
537 |
return true;
|
|
538 |
}
|
|
539 |
/*
|
|
540 |
Reads a preprocessor line from the source by advancing lexerTokenIndex and
|
|
541 |
returing a TokenSection containg the read line. Text lines separated by
|
|
542 |
an escaped newline are joined.
|
|
543 |
*/
|
|
544 |
TokenSection Preprocessor::readLine()
|
|
545 |
{
|
|
546 |
const int startIndex = lexerTokenIndex;
|
|
547 |
bool gotNewline = false;
|
|
548 |
|
|
549 |
while(isValidIndex(lexerTokenIndex) && !gotNewline) {
|
|
550 |
if(m_tokenTypeList.at(lexerTokenIndex) == Token_newline) {
|
|
551 |
if (lexerTokenIndex == 0 || m_tokenTypeList.at(lexerTokenIndex-1) != '\\') {
|
|
552 |
gotNewline = true;
|
|
553 |
break;
|
|
554 |
}
|
|
555 |
}
|
|
556 |
++lexerTokenIndex;
|
|
557 |
}
|
|
558 |
|
|
559 |
if(gotNewline)
|
|
560 |
++lexerTokenIndex; //include newline
|
|
561 |
else
|
|
562 |
emit error(QLatin1String("Error"), QLatin1String("Unexpected end of source"));
|
|
563 |
|
|
564 |
return TokenSection(m_tokenContainer, startIndex, lexerTokenIndex - startIndex);
|
|
565 |
}
|
|
566 |
|
|
567 |
/*
|
|
568 |
Returns false if index is past the end of m_tokenContainer.
|
|
569 |
*/
|
|
570 |
inline bool Preprocessor::isValidIndex(const int index) const
|
|
571 |
{
|
|
572 |
return (index < m_tokenContainer.count());
|
|
573 |
}
|
|
574 |
|
|
575 |
/*
|
|
576 |
Returns true if the token at index is a whitepsace token.
|
|
577 |
*/
|
|
578 |
inline bool Preprocessor::isWhiteSpace(const int index) const
|
|
579 |
{
|
|
580 |
return (m_tokenTypeList.at(index) == Token_whitespaces);
|
|
581 |
}
|
|
582 |
|
|
583 |
/*
|
|
584 |
Looks ahead from lexerTokenIndex, returns the token type found at the first
|
|
585 |
token that is not a comment or whitespace token.
|
|
586 |
*/
|
|
587 |
Type Preprocessor::lookAhead() const
|
|
588 |
{
|
|
589 |
const int index = skipWhiteSpaceAndComments();
|
|
590 |
if (index == -1)
|
|
591 |
return Token_eof;
|
|
592 |
return m_tokenTypeList.at(index);
|
|
593 |
}
|
|
594 |
/*
|
|
595 |
Looks ahead from lexerTokenIndex, returns the token type found at the first
|
|
596 |
token that is not a comment, whitespace or '#' token.
|
|
597 |
*/
|
|
598 |
Type Preprocessor::lookAheadSkipHash() const
|
|
599 |
{
|
|
600 |
const int index = skipWhiteSpaceCommentsHash();
|
|
601 |
if (index == -1)
|
|
602 |
return Token_eof;
|
|
603 |
return m_tokenTypeList.at(index);
|
|
604 |
}
|
|
605 |
|
|
606 |
/*
|
|
607 |
Returns the index for the first token after lexerTokenIndex that is not a
|
|
608 |
whitespace or comment token.
|
|
609 |
*/
|
|
610 |
inline int Preprocessor::skipWhiteSpaceAndComments() const
|
|
611 |
{
|
|
612 |
int index = lexerTokenIndex;
|
|
613 |
if(!isValidIndex(index))
|
|
614 |
return -1;
|
|
615 |
while(m_tokenTypeList.at(index) == Token_whitespaces
|
|
616 |
|| m_tokenTypeList.at(index) == Token_comment
|
|
617 |
|| m_tokenTypeList.at(index) == Token_line_comment
|
|
618 |
|| m_tokenTypeList.at(index) == Token_multiline_comment ) {
|
|
619 |
++index;
|
|
620 |
if(!isValidIndex(index))
|
|
621 |
return -1;
|
|
622 |
}
|
|
623 |
return index;
|
|
624 |
}
|
|
625 |
|
|
626 |
/*
|
|
627 |
Returns the index for the first token after lexerTokenIndex that is not a
|
|
628 |
whitespace, comment or '#' token.
|
|
629 |
*/
|
|
630 |
inline int Preprocessor::skipWhiteSpaceCommentsHash() const
|
|
631 |
{
|
|
632 |
int index = lexerTokenIndex;
|
|
633 |
if(!isValidIndex(index))
|
|
634 |
return -1;
|
|
635 |
while(m_tokenTypeList.at(index) == Token_whitespaces
|
|
636 |
|| m_tokenTypeList.at(index) == Token_comment
|
|
637 |
|| m_tokenTypeList.at(index) == Token_line_comment
|
|
638 |
|| m_tokenTypeList.at(index) == Token_multiline_comment
|
|
639 |
|| m_tokenTypeList.at(index) == Token_preproc ) {
|
|
640 |
++index;
|
|
641 |
if(!isValidIndex(index))
|
|
642 |
return -1;
|
|
643 |
}
|
|
644 |
return index;
|
|
645 |
}
|
|
646 |
|
|
647 |
/*
|
|
648 |
Removes escaped newlines from tokenSection. Both the escape token ('\')
|
|
649 |
and the newline token ('\n') are removed.
|
|
650 |
*/
|
|
651 |
QVector<int> Preprocessor::cleanEscapedNewLines(const TokenSection &tokenSection) const
|
|
652 |
{
|
|
653 |
QVector<int> indexList;
|
|
654 |
|
|
655 |
int t = 0;
|
|
656 |
const int numTokens = tokenSection.count();
|
|
657 |
while (t < numTokens) {
|
|
658 |
const int containerIndex = tokenSection.containerIndex(t);
|
|
659 |
const int currentToken = t;
|
|
660 |
++t;
|
|
661 |
|
|
662 |
//handle escaped newlines
|
|
663 |
if (tokenSection.text(currentToken) == "\\"
|
|
664 |
&& currentToken + 1 < numTokens
|
|
665 |
&& m_tokenTypeList.at(containerIndex + 1) == Token_newline)
|
|
666 |
continue;
|
|
667 |
|
|
668 |
indexList.append(containerIndex);
|
|
669 |
}
|
|
670 |
return indexList;
|
|
671 |
}
|
|
672 |
|
|
673 |
/*
|
|
674 |
Removes escaped newlines, whitespace and comment tokens from tokenSection
|
|
675 |
*/
|
|
676 |
QVector<int> Preprocessor::cleanTokenRange(const TokenSection &tokenSection) const
|
|
677 |
{
|
|
678 |
QVector<int> indexList;
|
|
679 |
|
|
680 |
int t = 0;
|
|
681 |
const int numTokens = tokenSection.count();
|
|
682 |
while (t < numTokens) {
|
|
683 |
const int containerIndex = tokenSection.containerIndex(t);
|
|
684 |
const Type tokenType = m_tokenTypeList.at(containerIndex);
|
|
685 |
const int currentToken = t;
|
|
686 |
++t;
|
|
687 |
|
|
688 |
if(tokenType == Token_whitespaces ||
|
|
689 |
tokenType == Token_line_comment ||
|
|
690 |
tokenType == Token_multiline_comment )
|
|
691 |
continue;
|
|
692 |
|
|
693 |
//handle escaped newlines
|
|
694 |
if(tokenSection.text(currentToken) == "\\" &&
|
|
695 |
currentToken + 1 < numTokens &&
|
|
696 |
m_tokenTypeList.at(containerIndex + 1) == Token_newline)
|
|
697 |
continue;
|
|
698 |
|
|
699 |
indexList.append(containerIndex);
|
|
700 |
}
|
|
701 |
return indexList;
|
|
702 |
}
|
|
703 |
/*
|
|
704 |
Returns the text for an Item node and all its children.
|
|
705 |
*/
|
|
706 |
QByteArray visitGetText(Item *item)
|
|
707 |
{
|
|
708 |
QByteArray text;
|
|
709 |
|
|
710 |
text += item->text().fullText();
|
|
711 |
|
|
712 |
if(item->toItemComposite()) {
|
|
713 |
ItemComposite *composite = item->toItemComposite();
|
|
714 |
for (int i=0; i <composite->count(); ++i)
|
|
715 |
text += visitGetText(composite->item(i));
|
|
716 |
}
|
|
717 |
|
|
718 |
return text;
|
|
719 |
}
|
|
720 |
|
|
721 |
void Source::setFileName(const QString &fileName)
|
|
722 |
{
|
|
723 |
m_fileName = fileName;
|
|
724 |
}
|
|
725 |
|
|
726 |
} // namespace Rpp
|
|
727 |
|
|
728 |
QT_END_NAMESPACE
|