OpenCores
URL https://opencores.org/ocsvn/openrisc_me/openrisc_me/trunk

Subversion Repositories openrisc_me

[/] [openrisc/] [trunk/] [gnu-src/] [gcc-4.2.2/] [gcc/] [doc/] [cppinternals.texi] - Blame information for rev 154

Details | Compare with Previous | View Log

Line No. Rev Author Line
1 38 julius
\input texinfo
2
@setfilename cppinternals.info
3
@settitle The GNU C Preprocessor Internals
4
 
5
@include gcc-common.texi
6
 
7
@ifinfo
8
@dircategory Software development
9
@direntry
10
* Cpplib: (cppinternals).      Cpplib internals.
11
@end direntry
12
@end ifinfo
13
 
14
@c @smallbook
15
@c @cropmarks
16
@c @finalout
17
@setchapternewpage odd
18
@ifinfo
19
This file documents the internals of the GNU C Preprocessor.
20
 
21
Copyright 2000, 2001, 2002, 2004, 2005 Free Software Foundation, Inc.
22
 
23
Permission is granted to make and distribute verbatim copies of
24
this manual provided the copyright notice and this permission notice
25
are preserved on all copies.
26
 
27
@ignore
28
Permission is granted to process this file through Tex and print the
29
results, provided the printed document carries copying permission
30
notice identical to this one except for the removal of this paragraph
31
(this paragraph not being relevant to the printed manual).
32
 
33
@end ignore
34
Permission is granted to copy and distribute modified versions of this
35
manual under the conditions for verbatim copying, provided also that
36
the entire resulting derived work is distributed under the terms of a
37
permission notice identical to this one.
38
 
39
Permission is granted to copy and distribute translations of this manual
40
into another language, under the above conditions for modified versions.
41
@end ifinfo
42
 
43
@titlepage
44
@title Cpplib Internals
45
@versionsubtitle
46
@author Neil Booth
47
@page
48
@vskip 0pt plus 1filll
49
@c man begin COPYRIGHT
50
Copyright @copyright{} 2000, 2001, 2002, 2004, 2005
51
Free Software Foundation, Inc.
52
 
53
Permission is granted to make and distribute verbatim copies of
54
this manual provided the copyright notice and this permission notice
55
are preserved on all copies.
56
 
57
Permission is granted to copy and distribute modified versions of this
58
manual under the conditions for verbatim copying, provided also that
59
the entire resulting derived work is distributed under the terms of a
60
permission notice identical to this one.
61
 
62
Permission is granted to copy and distribute translations of this manual
63
into another language, under the above conditions for modified versions.
64
@c man end
65
@end titlepage
66
@contents
67
@page
68
 
69
@node Top
70
@top
71
@chapter Cpplib---the GNU C Preprocessor
72
 
73
The GNU C preprocessor is
74
implemented as a library, @dfn{cpplib}, so it can be easily shared between
75
a stand-alone preprocessor, and a preprocessor integrated with the C,
76
C++ and Objective-C front ends.  It is also available for use by other
77
programs, though this is not recommended as its exposed interface has
78
not yet reached a point of reasonable stability.
79
 
80
The library has been written to be re-entrant, so that it can be used
81
to preprocess many files simultaneously if necessary.  It has also been
82
written with the preprocessing token as the fundamental unit; the
83
preprocessor in previous versions of GCC would operate on text strings
84
as the fundamental unit.
85
 
86
This brief manual documents the internals of cpplib, and explains some
87
of the tricky issues.  It is intended that, along with the comments in
88
the source code, a reasonably competent C programmer should be able to
89
figure out what the code is doing, and why things have been implemented
90
the way they have.
91
 
92
@menu
93
* Conventions::         Conventions used in the code.
94
* Lexer::               The combined C, C++ and Objective-C Lexer.
95
* Hash Nodes::          All identifiers are entered into a hash table.
96
* Macro Expansion::     Macro expansion algorithm.
97
* Token Spacing::       Spacing and paste avoidance issues.
98
* Line Numbering::      Tracking location within files.
99
* Guard Macros::        Optimizing header files with guard macros.
100
* Files::               File handling.
101
* Concept Index::       Index.
102
@end menu
103
 
104
@node Conventions
105
@unnumbered Conventions
106
@cindex interface
107
@cindex header files
108
 
109
cpplib has two interfaces---one is exposed internally only, and the
110
other is for both internal and external use.
111
 
112
The convention is that functions and types that are exposed to multiple
113
files internally are prefixed with @samp{_cpp_}, and are to be found in
114
the file @file{internal.h}.  Functions and types exposed to external
115
clients are in @file{cpplib.h}, and prefixed with @samp{cpp_}.  For
116
historical reasons this is no longer quite true, but we should strive to
117
stick to it.
118
 
119
We are striving to reduce the information exposed in @file{cpplib.h} to the
120
bare minimum necessary, and then to keep it there.  This makes clear
121
exactly what external clients are entitled to assume, and allows us to
122
change internals in the future without worrying whether library clients
123
are perhaps relying on some kind of undocumented implementation-specific
124
behavior.
125
 
126
@node Lexer
127
@unnumbered The Lexer
128
@cindex lexer
129
@cindex newlines
130
@cindex escaped newlines
131
 
132
@section Overview
133
The lexer is contained in the file @file{lex.c}.  It is a hand-coded
134
lexer, and not implemented as a state machine.  It can understand C, C++
135
and Objective-C source code, and has been extended to allow reasonably
136
successful preprocessing of assembly language.  The lexer does not make
137
an initial pass to strip out trigraphs and escaped newlines, but handles
138
them as they are encountered in a single pass of the input file.  It
139
returns preprocessing tokens individually, not a line at a time.
140
 
141
It is mostly transparent to users of the library, since the library's
142
interface for obtaining the next token, @code{cpp_get_token}, takes care
143
of lexing new tokens, handling directives, and expanding macros as
144
necessary.  However, the lexer does expose some functionality so that
145
clients of the library can easily spell a given token, such as
146
@code{cpp_spell_token} and @code{cpp_token_len}.  These functions are
147
useful when generating diagnostics, and for emitting the preprocessed
148
output.
149
 
150
@section Lexing a token
151
Lexing of an individual token is handled by @code{_cpp_lex_direct} and
152
its subroutines.  In its current form the code is quite complicated,
153
with read ahead characters and such-like, since it strives to not step
154
back in the character stream in preparation for handling non-ASCII file
155
encodings.  The current plan is to convert any such files to UTF-8
156
before processing them.  This complexity is therefore unnecessary and
157
will be removed, so I'll not discuss it further here.
158
 
159
The job of @code{_cpp_lex_direct} is simply to lex a token.  It is not
160
responsible for issues like directive handling, returning lookahead
161
tokens directly, multiple-include optimization, or conditional block
162
skipping.  It necessarily has a minor r@^ole to play in memory
163
management of lexed lines.  I discuss these issues in a separate section
164
(@pxref{Lexing a line}).
165
 
166
The lexer places the token it lexes into storage pointed to by the
167
variable @code{cur_token}, and then increments it.  This variable is
168
important for correct diagnostic positioning.  Unless a specific line
169
and column are passed to the diagnostic routines, they will examine the
170
@code{line} and @code{col} values of the token just before the location
171
that @code{cur_token} points to, and use that location to report the
172
diagnostic.
173
 
174
The lexer does not consider whitespace to be a token in its own right.
175
If whitespace (other than a new line) precedes a token, it sets the
176
@code{PREV_WHITE} bit in the token's flags.  Each token has its
177
@code{line} and @code{col} variables set to the line and column of the
178
first character of the token.  This line number is the line number in
179
the translation unit, and can be converted to a source (file, line) pair
180
using the line map code.
181
 
182
The first token on a logical, i.e.@: unescaped, line has the flag
183
@code{BOL} set for beginning-of-line.  This flag is intended for
184
internal use, both to distinguish a @samp{#} that begins a directive
185
from one that doesn't, and to generate a call-back to clients that want
186
to be notified about the start of every non-directive line with tokens
187
on it.  Clients cannot reliably determine this for themselves: the first
188
token might be a macro, and the tokens of a macro expansion do not have
189
the @code{BOL} flag set.  The macro expansion may even be empty, and the
190
next token on the line certainly won't have the @code{BOL} flag set.
191
 
192
New lines are treated specially; exactly how the lexer handles them is
193
context-dependent.  The C standard mandates that directives are
194
terminated by the first unescaped newline character, even if it appears
195
in the middle of a macro expansion.  Therefore, if the state variable
196
@code{in_directive} is set, the lexer returns a @code{CPP_EOF} token,
197
which is normally used to indicate end-of-file, to indicate
198
end-of-directive.  In a directive a @code{CPP_EOF} token never means
199
end-of-file.  Conveniently, if the caller was @code{collect_args}, it
200
already handles @code{CPP_EOF} as if it were end-of-file, and reports an
201
error about an unterminated macro argument list.
202
 
203
The C standard also specifies that a new line in the middle of the
204
arguments to a macro is treated as whitespace.  This white space is
205
important in case the macro argument is stringified.  The state variable
206
@code{parsing_args} is nonzero when the preprocessor is collecting the
207
arguments to a macro call.  It is set to 1 when looking for the opening
208
parenthesis to a function-like macro, and 2 when collecting the actual
209
arguments up to the closing parenthesis, since these two cases need to
210
be distinguished sometimes.  One such time is here: the lexer sets the
211
@code{PREV_WHITE} flag of a token if it meets a new line when
212
@code{parsing_args} is set to 2.  It doesn't set it if it meets a new
213
line when @code{parsing_args} is 1, since then code like
214
 
215
@smallexample
216
#define foo() bar
217
foo
218
baz
219
@end smallexample
220
 
221
@noindent would be output with an erroneous space before @samp{baz}:
222
 
223
@smallexample
224
foo
225
 baz
226
@end smallexample
227
 
228
This is a good example of the subtlety of getting token spacing correct
229
in the preprocessor; there are plenty of tests in the testsuite for
230
corner cases like this.
231
 
232
The lexer is written to treat each of @samp{\r}, @samp{\n}, @samp{\r\n}
233
and @samp{\n\r} as a single new line indicator.  This allows it to
234
transparently preprocess MS-DOS, Macintosh and Unix files without their
235
needing to pass through a special filter beforehand.
236
 
237
We also decided to treat a backslash, either @samp{\} or the trigraph
238
@samp{??/}, separated from one of the above newline indicators by
239
non-comment whitespace only, as intending to escape the newline.  It
240
tends to be a typing mistake, and cannot reasonably be mistaken for
241
anything else in any of the C-family grammars.  Since handling it this
242
way is not strictly conforming to the ISO standard, the library issues a
243
warning wherever it encounters it.
244
 
245
Handling newlines like this is made simpler by doing it in one place
246
only.  The function @code{handle_newline} takes care of all newline
247
characters, and @code{skip_escaped_newlines} takes care of arbitrarily
248
long sequences of escaped newlines, deferring to @code{handle_newline}
249
to handle the newlines themselves.
250
 
251
The most painful aspect of lexing ISO-standard C and C++ is handling
252
trigraphs and backlash-escaped newlines.  Trigraphs are processed before
253
any interpretation of the meaning of a character is made, and unfortunately
254
there is a trigraph representation for a backslash, so it is possible for
255
the trigraph @samp{??/} to introduce an escaped newline.
256
 
257
Escaped newlines are tedious because theoretically they can occur
258
anywhere---between the @samp{+} and @samp{=} of the @samp{+=} token,
259
within the characters of an identifier, and even between the @samp{*}
260
and @samp{/} that terminates a comment.  Moreover, you cannot be sure
261
there is just one---there might be an arbitrarily long sequence of them.
262
 
263
So, for example, the routine that lexes a number, @code{parse_number},
264
cannot assume that it can scan forwards until the first non-number
265
character and be done with it, because this could be the @samp{\}
266
introducing an escaped newline, or the @samp{?} introducing the trigraph
267
sequence that represents the @samp{\} of an escaped newline.  If it
268
encounters a @samp{?} or @samp{\}, it calls @code{skip_escaped_newlines}
269
to skip over any potential escaped newlines before checking whether the
270
number has been finished.
271
 
272
Similarly code in the main body of @code{_cpp_lex_direct} cannot simply
273
check for a @samp{=} after a @samp{+} character to determine whether it
274
has a @samp{+=} token; it needs to be prepared for an escaped newline of
275
some sort.  Such cases use the function @code{get_effective_char}, which
276
returns the first character after any intervening escaped newlines.
277
 
278
The lexer needs to keep track of the correct column position, including
279
counting tabs as specified by the @option{-ftabstop=} option.  This
280
should be done even within C-style comments; they can appear in the
281
middle of a line, and we want to report diagnostics in the correct
282
position for text appearing after the end of the comment.
283
 
284
@anchor{Invalid identifiers}
285
Some identifiers, such as @code{__VA_ARGS__} and poisoned identifiers,
286
may be invalid and require a diagnostic.  However, if they appear in a
287
macro expansion we don't want to complain with each use of the macro.
288
It is therefore best to catch them during the lexing stage, in
289
@code{parse_identifier}.  In both cases, whether a diagnostic is needed
290
or not is dependent upon the lexer's state.  For example, we don't want
291
to issue a diagnostic for re-poisoning a poisoned identifier, or for
292
using @code{__VA_ARGS__} in the expansion of a variable-argument macro.
293
Therefore @code{parse_identifier} makes use of state flags to determine
294
whether a diagnostic is appropriate.  Since we change state on a
295
per-token basis, and don't lex whole lines at a time, this is not a
296
problem.
297
 
298
Another place where state flags are used to change behavior is whilst
299
lexing header names.  Normally, a @samp{<} would be lexed as a single
300
token.  After a @code{#include} directive, though, it should be lexed as
301
a single token as far as the nearest @samp{>} character.  Note that we
302
don't allow the terminators of header names to be escaped; the first
303
@samp{"} or @samp{>} terminates the header name.
304
 
305
Interpretation of some character sequences depends upon whether we are
306
lexing C, C++ or Objective-C, and on the revision of the standard in
307
force.  For example, @samp{::} is a single token in C++, but in C it is
308
two separate @samp{:} tokens and almost certainly a syntax error.  Such
309
cases are handled by @code{_cpp_lex_direct} based upon command-line
310
flags stored in the @code{cpp_options} structure.
311
 
312
Once a token has been lexed, it leads an independent existence.  The
313
spelling of numbers, identifiers and strings is copied to permanent
314
storage from the original input buffer, so a token remains valid and
315
correct even if its source buffer is freed with @code{_cpp_pop_buffer}.
316
The storage holding the spellings of such tokens remains until the
317
client program calls cpp_destroy, probably at the end of the translation
318
unit.
319
 
320
@anchor{Lexing a line}
321
@section Lexing a line
322
@cindex token run
323
 
324
When the preprocessor was changed to return pointers to tokens, one
325
feature I wanted was some sort of guarantee regarding how long a
326
returned pointer remains valid.  This is important to the stand-alone
327
preprocessor, the future direction of the C family front ends, and even
328
to cpplib itself internally.
329
 
330
Occasionally the preprocessor wants to be able to peek ahead in the
331
token stream.  For example, after the name of a function-like macro, it
332
wants to check the next token to see if it is an opening parenthesis.
333
Another example is that, after reading the first few tokens of a
334
@code{#pragma} directive and not recognizing it as a registered pragma,
335
it wants to backtrack and allow the user-defined handler for unknown
336
pragmas to access the full @code{#pragma} token stream.  The stand-alone
337
preprocessor wants to be able to test the current token with the
338
previous one to see if a space needs to be inserted to preserve their
339
separate tokenization upon re-lexing (paste avoidance), so it needs to
340
be sure the pointer to the previous token is still valid.  The
341
recursive-descent C++ parser wants to be able to perform tentative
342
parsing arbitrarily far ahead in the token stream, and then to be able
343
to jump back to a prior position in that stream if necessary.
344
 
345
The rule I chose, which is fairly natural, is to arrange that the
346
preprocessor lex all tokens on a line consecutively into a token buffer,
347
which I call a @dfn{token run}, and when meeting an unescaped new line
348
(newlines within comments do not count either), to start lexing back at
349
the beginning of the run.  Note that we do @emph{not} lex a line of
350
tokens at once; if we did that @code{parse_identifier} would not have
351
state flags available to warn about invalid identifiers (@pxref{Invalid
352
identifiers}).
353
 
354
In other words, accessing tokens that appeared earlier in the current
355
line is valid, but since each logical line overwrites the tokens of the
356
previous line, tokens from prior lines are unavailable.  In particular,
357
since a directive only occupies a single logical line, this means that
358
the directive handlers like the @code{#pragma} handler can jump around
359
in the directive's tokens if necessary.
360
 
361
Two issues remain: what about tokens that arise from macro expansions,
362
and what happens when we have a long line that overflows the token run?
363
 
364
Since we promise clients that we preserve the validity of pointers that
365
we have already returned for tokens that appeared earlier in the line,
366
we cannot reallocate the run.  Instead, on overflow it is expanded by
367
chaining a new token run on to the end of the existing one.
368
 
369
The tokens forming a macro's replacement list are collected by the
370
@code{#define} handler, and placed in storage that is only freed by
371
@code{cpp_destroy}.  So if a macro is expanded in the line of tokens,
372
the pointers to the tokens of its expansion that are returned will always
373
remain valid.  However, macros are a little trickier than that, since
374
they give rise to three sources of fresh tokens.  They are the built-in
375
macros like @code{__LINE__}, and the @samp{#} and @samp{##} operators
376
for stringification and token pasting.  I handled this by allocating
377
space for these tokens from the lexer's token run chain.  This means
378
they automatically receive the same lifetime guarantees as lexed tokens,
379
and we don't need to concern ourselves with freeing them.
380
 
381
Lexing into a line of tokens solves some of the token memory management
382
issues, but not all.  The opening parenthesis after a function-like
383
macro name might lie on a different line, and the front ends definitely
384
want the ability to look ahead past the end of the current line.  So
385
cpplib only moves back to the start of the token run at the end of a
386
line if the variable @code{keep_tokens} is zero.  Line-buffering is
387
quite natural for the preprocessor, and as a result the only time cpplib
388
needs to increment this variable is whilst looking for the opening
389
parenthesis to, and reading the arguments of, a function-like macro.  In
390
the near future cpplib will export an interface to increment and
391
decrement this variable, so that clients can share full control over the
392
lifetime of token pointers too.
393
 
394
The routine @code{_cpp_lex_token} handles moving to new token runs,
395
calling @code{_cpp_lex_direct} to lex new tokens, or returning
396
previously-lexed tokens if we stepped back in the token stream.  It also
397
checks each token for the @code{BOL} flag, which might indicate a
398
directive that needs to be handled, or require a start-of-line call-back
399
to be made.  @code{_cpp_lex_token} also handles skipping over tokens in
400
failed conditional blocks, and invalidates the control macro of the
401
multiple-include optimization if a token was successfully lexed outside
402
a directive.  In other words, its callers do not need to concern
403
themselves with such issues.
404
 
405
@node Hash Nodes
406
@unnumbered Hash Nodes
407
@cindex hash table
408
@cindex identifiers
409
@cindex macros
410
@cindex assertions
411
@cindex named operators
412
 
413
When cpplib encounters an ``identifier'', it generates a hash code for
414
it and stores it in the hash table.  By ``identifier'' we mean tokens
415
with type @code{CPP_NAME}; this includes identifiers in the usual C
416
sense, as well as keywords, directive names, macro names and so on.  For
417
example, all of @code{pragma}, @code{int}, @code{foo} and
418
@code{__GNUC__} are identifiers and hashed when lexed.
419
 
420
Each node in the hash table contain various information about the
421
identifier it represents.  For example, its length and type.  At any one
422
time, each identifier falls into exactly one of three categories:
423
 
424
@itemize @bullet
425
@item Macros
426
 
427
These have been declared to be macros, either on the command line or
428
with @code{#define}.  A few, such as @code{__TIME__} are built-ins
429
entered in the hash table during initialization.  The hash node for a
430
normal macro points to a structure with more information about the
431
macro, such as whether it is function-like, how many arguments it takes,
432
and its expansion.  Built-in macros are flagged as special, and instead
433
contain an enum indicating which of the various built-in macros it is.
434
 
435
@item Assertions
436
 
437
Assertions are in a separate namespace to macros.  To enforce this, cpp
438
actually prepends a @code{#} character before hashing and entering it in
439
the hash table.  An assertion's node points to a chain of answers to
440
that assertion.
441
 
442
@item Void
443
 
444
Everything else falls into this category---an identifier that is not
445
currently a macro, or a macro that has since been undefined with
446
@code{#undef}.
447
 
448
When preprocessing C++, this category also includes the named operators,
449
such as @code{xor}.  In expressions these behave like the operators they
450
represent, but in contexts where the spelling of a token matters they
451
are spelt differently.  This spelling distinction is relevant when they
452
are operands of the stringizing and pasting macro operators @code{#} and
453
@code{##}.  Named operator hash nodes are flagged, both to catch the
454
spelling distinction and to prevent them from being defined as macros.
455
@end itemize
456
 
457
The same identifiers share the same hash node.  Since each identifier
458
token, after lexing, contains a pointer to its hash node, this is used
459
to provide rapid lookup of various information.  For example, when
460
parsing a @code{#define} statement, CPP flags each argument's identifier
461
hash node with the index of that argument.  This makes duplicated
462
argument checking an O(1) operation for each argument.  Similarly, for
463
each identifier in the macro's expansion, lookup to see if it is an
464
argument, and which argument it is, is also an O(1) operation.  Further,
465
each directive name, such as @code{endif}, has an associated directive
466
enum stored in its hash node, so that directive lookup is also O(1).
467
 
468
@node Macro Expansion
469
@unnumbered Macro Expansion Algorithm
470
@cindex macro expansion
471
 
472
Macro expansion is a tricky operation, fraught with nasty corner cases
473
and situations that render what you thought was a nifty way to
474
optimize the preprocessor's expansion algorithm wrong in quite subtle
475
ways.
476
 
477
I strongly recommend you have a good grasp of how the C and C++
478
standards require macros to be expanded before diving into this
479
section, let alone the code!.  If you don't have a clear mental
480
picture of how things like nested macro expansion, stringification and
481
token pasting are supposed to work, damage to your sanity can quickly
482
result.
483
 
484
@section Internal representation of macros
485
@cindex macro representation (internal)
486
 
487
The preprocessor stores macro expansions in tokenized form.  This
488
saves repeated lexing passes during expansion, at the cost of a small
489
increase in memory consumption on average.  The tokens are stored
490
contiguously in memory, so a pointer to the first one and a token
491
count is all you need to get the replacement list of a macro.
492
 
493
If the macro is a function-like macro the preprocessor also stores its
494
parameters, in the form of an ordered list of pointers to the hash
495
table entry of each parameter's identifier.  Further, in the macro's
496
stored expansion each occurrence of a parameter is replaced with a
497
special token of type @code{CPP_MACRO_ARG}.  Each such token holds the
498
index of the parameter it represents in the parameter list, which
499
allows rapid replacement of parameters with their arguments during
500
expansion.  Despite this optimization it is still necessary to store
501
the original parameters to the macro, both for dumping with e.g.,
502
@option{-dD}, and to warn about non-trivial macro redefinitions when
503
the parameter names have changed.
504
 
505
@section Macro expansion overview
506
The preprocessor maintains a @dfn{context stack}, implemented as a
507
linked list of @code{cpp_context} structures, which together represent
508
the macro expansion state at any one time.  The @code{struct
509
cpp_reader} member variable @code{context} points to the current top
510
of this stack.  The top normally holds the unexpanded replacement list
511
of the innermost macro under expansion, except when cpplib is about to
512
pre-expand an argument, in which case it holds that argument's
513
unexpanded tokens.
514
 
515
When there are no macros under expansion, cpplib is in @dfn{base
516
context}.  All contexts other than the base context contain a
517
contiguous list of tokens delimited by a starting and ending token.
518
When not in base context, cpplib obtains the next token from the list
519
of the top context.  If there are no tokens left in the list, it pops
520
that context off the stack, and subsequent ones if necessary, until an
521
unexhausted context is found or it returns to base context.  In base
522
context, cpplib reads tokens directly from the lexer.
523
 
524
If it encounters an identifier that is both a macro and enabled for
525
expansion, cpplib prepares to push a new context for that macro on the
526
stack by calling the routine @code{enter_macro_context}.  When this
527
routine returns, the new context will contain the unexpanded tokens of
528
the replacement list of that macro.  In the case of function-like
529
macros, @code{enter_macro_context} also replaces any parameters in the
530
replacement list, stored as @code{CPP_MACRO_ARG} tokens, with the
531
appropriate macro argument.  If the standard requires that the
532
parameter be replaced with its expanded argument, the argument will
533
have been fully macro expanded first.
534
 
535
@code{enter_macro_context} also handles special macros like
536
@code{__LINE__}.  Although these macros expand to a single token which
537
cannot contain any further macros, for reasons of token spacing
538
(@pxref{Token Spacing}) and simplicity of implementation, cpplib
539
handles these special macros by pushing a context containing just that
540
one token.
541
 
542
The final thing that @code{enter_macro_context} does before returning
543
is to mark the macro disabled for expansion (except for special macros
544
like @code{__TIME__}).  The macro is re-enabled when its context is
545
later popped from the context stack, as described above.  This strict
546
ordering ensures that a macro is disabled whilst its expansion is
547
being scanned, but that it is @emph{not} disabled whilst any arguments
548
to it are being expanded.
549
 
550
@section Scanning the replacement list for macros to expand
551
The C standard states that, after any parameters have been replaced
552
with their possibly-expanded arguments, the replacement list is
553
scanned for nested macros.  Further, any identifiers in the
554
replacement list that are not expanded during this scan are never
555
again eligible for expansion in the future, if the reason they were
556
not expanded is that the macro in question was disabled.
557
 
558
Clearly this latter condition can only apply to tokens resulting from
559
argument pre-expansion.  Other tokens never have an opportunity to be
560
re-tested for expansion.  It is possible for identifiers that are
561
function-like macros to not expand initially but to expand during a
562
later scan.  This occurs when the identifier is the last token of an
563
argument (and therefore originally followed by a comma or a closing
564
parenthesis in its macro's argument list), and when it replaces its
565
parameter in the macro's replacement list, the subsequent token
566
happens to be an opening parenthesis (itself possibly the first token
567
of an argument).
568
 
569
It is important to note that when cpplib reads the last token of a
570
given context, that context still remains on the stack.  Only when
571
looking for the @emph{next} token do we pop it off the stack and drop
572
to a lower context.  This makes backing up by one token easy, but more
573
importantly ensures that the macro corresponding to the current
574
context is still disabled when we are considering the last token of
575
its replacement list for expansion (or indeed expanding it).  As an
576
example, which illustrates many of the points above, consider
577
 
578
@smallexample
579
#define foo(x) bar x
580
foo(foo) (2)
581
@end smallexample
582
 
583
@noindent which fully expands to @samp{bar foo (2)}.  During pre-expansion
584
of the argument, @samp{foo} does not expand even though the macro is
585
enabled, since it has no following parenthesis [pre-expansion of an
586
argument only uses tokens from that argument; it cannot take tokens
587
from whatever follows the macro invocation].  This still leaves the
588
argument token @samp{foo} eligible for future expansion.  Then, when
589
re-scanning after argument replacement, the token @samp{foo} is
590
rejected for expansion, and marked ineligible for future expansion,
591
since the macro is now disabled.  It is disabled because the
592
replacement list @samp{bar foo} of the macro is still on the context
593
stack.
594
 
595
If instead the algorithm looked for an opening parenthesis first and
596
then tested whether the macro were disabled it would be subtly wrong.
597
In the example above, the replacement list of @samp{foo} would be
598
popped in the process of finding the parenthesis, re-enabling
599
@samp{foo} and expanding it a second time.
600
 
601
@section Looking for a function-like macro's opening parenthesis
602
Function-like macros only expand when immediately followed by a
603
parenthesis.  To do this cpplib needs to temporarily disable macros
604
and read the next token.  Unfortunately, because of spacing issues
605
(@pxref{Token Spacing}), there can be fake padding tokens in-between,
606
and if the next real token is not a parenthesis cpplib needs to be
607
able to back up that one token as well as retain the information in
608
any intervening padding tokens.
609
 
610
Backing up more than one token when macros are involved is not
611
permitted by cpplib, because in general it might involve issues like
612
restoring popped contexts onto the context stack, which are too hard.
613
Instead, searching for the parenthesis is handled by a special
614
function, @code{funlike_invocation_p}, which remembers padding
615
information as it reads tokens.  If the next real token is not an
616
opening parenthesis, it backs up that one token, and then pushes an
617
extra context just containing the padding information if necessary.
618
 
619
@section Marking tokens ineligible for future expansion
620
As discussed above, cpplib needs a way of marking tokens as
621
unexpandable.  Since the tokens cpplib handles are read-only once they
622
have been lexed, it instead makes a copy of the token and adds the
623
flag @code{NO_EXPAND} to the copy.
624
 
625
For efficiency and to simplify memory management by avoiding having to
626
remember to free these tokens, they are allocated as temporary tokens
627
from the lexer's current token run (@pxref{Lexing a line}) using the
628
function @code{_cpp_temp_token}.  The tokens are then re-used once the
629
current line of tokens has been read in.
630
 
631
This might sound unsafe.  However, tokens runs are not re-used at the
632
end of a line if it happens to be in the middle of a macro argument
633
list, and cpplib only wants to back-up more than one lexer token in
634
situations where no macro expansion is involved, so the optimization
635
is safe.
636
 
637
@node Token Spacing
638
@unnumbered Token Spacing
639
@cindex paste avoidance
640
@cindex spacing
641
@cindex token spacing
642
 
643
First, consider an issue that only concerns the stand-alone
644
preprocessor: there needs to be a guarantee that re-reading its preprocessed
645
output results in an identical token stream.  Without taking special
646
measures, this might not be the case because of macro substitution.
647
For example:
648
 
649
@smallexample
650
#define PLUS +
651
#define EMPTY
652
#define f(x) =x=
653
+PLUS -EMPTY- PLUS+ f(=)
654
        @expansion{} + + - - + + = = =
655
@emph{not}
656
        @expansion{} ++ -- ++ ===
657
@end smallexample
658
 
659
One solution would be to simply insert a space between all adjacent
660
tokens.  However, we would like to keep space insertion to a minimum,
661
both for aesthetic reasons and because it causes problems for people who
662
still try to abuse the preprocessor for things like Fortran source and
663
Makefiles.
664
 
665
For now, just notice that when tokens are added (or removed, as shown by
666
the @code{EMPTY} example) from the original lexed token stream, we need
667
to check for accidental token pasting.  We call this @dfn{paste
668
avoidance}.  Token addition and removal can only occur because of macro
669
expansion, but accidental pasting can occur in many places: both before
670
and after each macro replacement, each argument replacement, and
671
additionally each token created by the @samp{#} and @samp{##} operators.
672
 
673
Look at how the preprocessor gets whitespace output correct
674
normally.  The @code{cpp_token} structure contains a flags byte, and one
675
of those flags is @code{PREV_WHITE}.  This is flagged by the lexer, and
676
indicates that the token was preceded by whitespace of some form other
677
than a new line.  The stand-alone preprocessor can use this flag to
678
decide whether to insert a space between tokens in the output.
679
 
680
Now consider the result of the following macro expansion:
681
 
682
@smallexample
683
#define add(x, y, z) x + y +z;
684
sum = add (1,2, 3);
685
        @expansion{} sum = 1 + 2 +3;
686
@end smallexample
687
 
688
The interesting thing here is that the tokens @samp{1} and @samp{2} are
689
output with a preceding space, and @samp{3} is output without a
690
preceding space, but when lexed none of these tokens had that property.
691
Careful consideration reveals that @samp{1} gets its preceding
692
whitespace from the space preceding @samp{add} in the macro invocation,
693
@emph{not} replacement list.  @samp{2} gets its whitespace from the
694
space preceding the parameter @samp{y} in the macro replacement list,
695
and @samp{3} has no preceding space because parameter @samp{z} has none
696
in the replacement list.
697
 
698
Once lexed, tokens are effectively fixed and cannot be altered, since
699
pointers to them might be held in many places, in particular by
700
in-progress macro expansions.  So instead of modifying the two tokens
701
above, the preprocessor inserts a special token, which I call a
702
@dfn{padding token}, into the token stream to indicate that spacing of
703
the subsequent token is special.  The preprocessor inserts padding
704
tokens in front of every macro expansion and expanded macro argument.
705
These point to a @dfn{source token} from which the subsequent real token
706
should inherit its spacing.  In the above example, the source tokens are
707
@samp{add} in the macro invocation, and @samp{y} and @samp{z} in the
708
macro replacement list, respectively.
709
 
710
It is quite easy to get multiple padding tokens in a row, for example if
711
a macro's first replacement token expands straight into another macro.
712
 
713
@smallexample
714
#define foo bar
715
#define bar baz
716
[foo]
717
        @expansion{} [baz]
718
@end smallexample
719
 
720
Here, two padding tokens are generated with sources the @samp{foo} token
721
between the brackets, and the @samp{bar} token from foo's replacement
722
list, respectively.  Clearly the first padding token is the one to
723
use, so the output code should contain a rule that the first
724
padding token in a sequence is the one that matters.
725
 
726
But what if a macro expansion is left?  Adjusting the above
727
example slightly:
728
 
729
@smallexample
730
#define foo bar
731
#define bar EMPTY baz
732
#define EMPTY
733
[foo] EMPTY;
734
        @expansion{} [ baz] ;
735
@end smallexample
736
 
737
As shown, now there should be a space before @samp{baz} and the
738
semicolon in the output.
739
 
740
The rules we decided above fail for @samp{baz}: we generate three
741
padding tokens, one per macro invocation, before the token @samp{baz}.
742
We would then have it take its spacing from the first of these, which
743
carries source token @samp{foo} with no leading space.
744
 
745
It is vital that cpplib get spacing correct in these examples since any
746
of these macro expansions could be stringified, where spacing matters.
747
 
748
So, this demonstrates that not just entering macro and argument
749
expansions, but leaving them requires special handling too.  I made
750
cpplib insert a padding token with a @code{NULL} source token when
751
leaving macro expansions, as well as after each replaced argument in a
752
macro's replacement list.  It also inserts appropriate padding tokens on
753
either side of tokens created by the @samp{#} and @samp{##} operators.
754
I expanded the rule so that, if we see a padding token with a
755
@code{NULL} source token, @emph{and} that source token has no leading
756
space, then we behave as if we have seen no padding tokens at all.  A
757
quick check shows this rule will then get the above example correct as
758
well.
759
 
760
Now a relationship with paste avoidance is apparent: we have to be
761
careful about paste avoidance in exactly the same locations we have
762
padding tokens in order to get white space correct.  This makes
763
implementation of paste avoidance easy: wherever the stand-alone
764
preprocessor is fixing up spacing because of padding tokens, and it
765
turns out that no space is needed, it has to take the extra step to
766
check that a space is not needed after all to avoid an accidental paste.
767
The function @code{cpp_avoid_paste} advises whether a space is required
768
between two consecutive tokens.  To avoid excessive spacing, it tries
769
hard to only require a space if one is likely to be necessary, but for
770
reasons of efficiency it is slightly conservative and might recommend a
771
space where one is not strictly needed.
772
 
773
@node Line Numbering
774
@unnumbered Line numbering
775
@cindex line numbers
776
 
777
@section Just which line number anyway?
778
 
779
There are three reasonable requirements a cpplib client might have for
780
the line number of a token passed to it:
781
 
782
@itemize @bullet
783
@item
784
The source line it was lexed on.
785
@item
786
The line it is output on.  This can be different to the line it was
787
lexed on if, for example, there are intervening escaped newlines or
788
C-style comments.  For example:
789
 
790
@smallexample
791
foo /* @r{A long
792
comment} */ bar \
793
baz
794
@result{}
795
foo bar baz
796
@end smallexample
797
 
798
@item
799
If the token results from a macro expansion, the line of the macro name,
800
or possibly the line of the closing parenthesis in the case of
801
function-like macro expansion.
802
@end itemize
803
 
804
The @code{cpp_token} structure contains @code{line} and @code{col}
805
members.  The lexer fills these in with the line and column of the first
806
character of the token.  Consequently, but maybe unexpectedly, a token
807
from the replacement list of a macro expansion carries the location of
808
the token within the @code{#define} directive, because cpplib expands a
809
macro by returning pointers to the tokens in its replacement list.  The
810
current implementation of cpplib assigns tokens created from built-in
811
macros and the @samp{#} and @samp{##} operators the location of the most
812
recently lexed token.  This is a because they are allocated from the
813
lexer's token runs, and because of the way the diagnostic routines infer
814
the appropriate location to report.
815
 
816
The diagnostic routines in cpplib display the location of the most
817
recently @emph{lexed} token, unless they are passed a specific line and
818
column to report.  For diagnostics regarding tokens that arise from
819
macro expansions, it might also be helpful for the user to see the
820
original location in the macro definition that the token came from.
821
Since that is exactly the information each token carries, such an
822
enhancement could be made relatively easily in future.
823
 
824
The stand-alone preprocessor faces a similar problem when determining
825
the correct line to output the token on: the position attached to a
826
token is fairly useless if the token came from a macro expansion.  All
827
tokens on a logical line should be output on its first physical line, so
828
the token's reported location is also wrong if it is part of a physical
829
line other than the first.
830
 
831
To solve these issues, cpplib provides a callback that is generated
832
whenever it lexes a preprocessing token that starts a new logical line
833
other than a directive.  It passes this token (which may be a
834
@code{CPP_EOF} token indicating the end of the translation unit) to the
835
callback routine, which can then use the line and column of this token
836
to produce correct output.
837
 
838
@section Representation of line numbers
839
 
840
As mentioned above, cpplib stores with each token the line number that
841
it was lexed on.  In fact, this number is not the number of the line in
842
the source file, but instead bears more resemblance to the number of the
843
line in the translation unit.
844
 
845
The preprocessor maintains a monotonic increasing line count, which is
846
incremented at every new line character (and also at the end of any
847
buffer that does not end in a new line).  Since a line number of zero is
848
useful to indicate certain special states and conditions, this variable
849
starts counting from one.
850
 
851
This variable therefore uniquely enumerates each line in the translation
852
unit.  With some simple infrastructure, it is straight forward to map
853
from this to the original source file and line number pair, saving space
854
whenever line number information needs to be saved.  The code the
855
implements this mapping lies in the files @file{line-map.c} and
856
@file{line-map.h}.
857
 
858
Command-line macros and assertions are implemented by pushing a buffer
859
containing the right hand side of an equivalent @code{#define} or
860
@code{#assert} directive.  Some built-in macros are handled similarly.
861
Since these are all processed before the first line of the main input
862
file, it will typically have an assigned line closer to twenty than to
863
one.
864
 
865
@node Guard Macros
866
@unnumbered The Multiple-Include Optimization
867
@cindex guard macros
868
@cindex controlling macros
869
@cindex multiple-include optimization
870
 
871
Header files are often of the form
872
 
873
@smallexample
874
#ifndef FOO
875
#define FOO
876
@dots{}
877
#endif
878
@end smallexample
879
 
880
@noindent
881
to prevent the compiler from processing them more than once.  The
882
preprocessor notices such header files, so that if the header file
883
appears in a subsequent @code{#include} directive and @code{FOO} is
884
defined, then it is ignored and it doesn't preprocess or even re-open
885
the file a second time.  This is referred to as the @dfn{multiple
886
include optimization}.
887
 
888
Under what circumstances is such an optimization valid?  If the file
889
were included a second time, it can only be optimized away if that
890
inclusion would result in no tokens to return, and no relevant
891
directives to process.  Therefore the current implementation imposes
892
requirements and makes some allowances as follows:
893
 
894
@enumerate
895
@item
896
There must be no tokens outside the controlling @code{#if}-@code{#endif}
897
pair, but whitespace and comments are permitted.
898
 
899
@item
900
There must be no directives outside the controlling directive pair, but
901
the @dfn{null directive} (a line containing nothing other than a single
902
@samp{#} and possibly whitespace) is permitted.
903
 
904
@item
905
The opening directive must be of the form
906
 
907
@smallexample
908
#ifndef FOO
909
@end smallexample
910
 
911
or
912
 
913
@smallexample
914
#if !defined FOO     [equivalently, #if !defined(FOO)]
915
@end smallexample
916
 
917
@item
918
In the second form above, the tokens forming the @code{#if} expression
919
must have come directly from the source file---no macro expansion must
920
have been involved.  This is because macro definitions can change, and
921
tracking whether or not a relevant change has been made is not worth the
922
implementation cost.
923
 
924
@item
925
There can be no @code{#else} or @code{#elif} directives at the outer
926
conditional block level, because they would probably contain something
927
of interest to a subsequent pass.
928
@end enumerate
929
 
930
First, when pushing a new file on the buffer stack,
931
@code{_stack_include_file} sets the controlling macro @code{mi_cmacro} to
932
@code{NULL}, and sets @code{mi_valid} to @code{true}.  This indicates
933
that the preprocessor has not yet encountered anything that would
934
invalidate the multiple-include optimization.  As described in the next
935
few paragraphs, these two variables having these values effectively
936
indicates top-of-file.
937
 
938
When about to return a token that is not part of a directive,
939
@code{_cpp_lex_token} sets @code{mi_valid} to @code{false}.  This
940
enforces the constraint that tokens outside the controlling conditional
941
block invalidate the optimization.
942
 
943
The @code{do_if}, when appropriate, and @code{do_ifndef} directive
944
handlers pass the controlling macro to the function
945
@code{push_conditional}.  cpplib maintains a stack of nested conditional
946
blocks, and after processing every opening conditional this function
947
pushes an @code{if_stack} structure onto the stack.  In this structure
948
it records the controlling macro for the block, provided there is one
949
and we're at top-of-file (as described above).  If an @code{#elif} or
950
@code{#else} directive is encountered, the controlling macro for that
951
block is cleared to @code{NULL}.  Otherwise, it survives until the
952
@code{#endif} closing the block, upon which @code{do_endif} sets
953
@code{mi_valid} to true and stores the controlling macro in
954
@code{mi_cmacro}.
955
 
956
@code{_cpp_handle_directive} clears @code{mi_valid} when processing any
957
directive other than an opening conditional and the null directive.
958
With this, and requiring top-of-file to record a controlling macro, and
959
no @code{#else} or @code{#elif} for it to survive and be copied to
960
@code{mi_cmacro} by @code{do_endif}, we have enforced the absence of
961
directives outside the main conditional block for the optimization to be
962
on.
963
 
964
Note that whilst we are inside the conditional block, @code{mi_valid} is
965
likely to be reset to @code{false}, but this does not matter since
966
the closing @code{#endif} restores it to @code{true} if appropriate.
967
 
968
Finally, since @code{_cpp_lex_direct} pops the file off the buffer stack
969
at @code{EOF} without returning a token, if the @code{#endif} directive
970
was not followed by any tokens, @code{mi_valid} is @code{true} and
971
@code{_cpp_pop_file_buffer} remembers the controlling macro associated
972
with the file.  Subsequent calls to @code{stack_include_file} result in
973
no buffer being pushed if the controlling macro is defined, effecting
974
the optimization.
975
 
976
A quick word on how we handle the
977
 
978
@smallexample
979
#if !defined FOO
980
@end smallexample
981
 
982
@noindent
983
case.  @code{_cpp_parse_expr} and @code{parse_defined} take steps to see
984
whether the three stages @samp{!}, @samp{defined-expression} and
985
@samp{end-of-directive} occur in order in a @code{#if} expression.  If
986
so, they return the guard macro to @code{do_if} in the variable
987
@code{mi_ind_cmacro}, and otherwise set it to @code{NULL}.
988
@code{enter_macro_context} sets @code{mi_valid} to false, so if a macro
989
was expanded whilst parsing any part of the expression, then the
990
top-of-file test in @code{push_conditional} fails and the optimization
991
is turned off.
992
 
993
@node Files
994
@unnumbered File Handling
995
@cindex files
996
 
997
Fairly obviously, the file handling code of cpplib resides in the file
998
@file{files.c}.  It takes care of the details of file searching,
999
opening, reading and caching, for both the main source file and all the
1000
headers it recursively includes.
1001
 
1002
The basic strategy is to minimize the number of system calls.  On many
1003
systems, the basic @code{open ()} and @code{fstat ()} system calls can
1004
be quite expensive.  For every @code{#include}-d file, we need to try
1005
all the directories in the search path until we find a match.  Some
1006
projects, such as glibc, pass twenty or thirty include paths on the
1007
command line, so this can rapidly become time consuming.
1008
 
1009
For a header file we have not encountered before we have little choice
1010
but to do this.  However, it is often the case that the same headers are
1011
repeatedly included, and in these cases we try to avoid repeating the
1012
filesystem queries whilst searching for the correct file.
1013
 
1014
For each file we try to open, we store the constructed path in a splay
1015
tree.  This path first undergoes simplification by the function
1016
@code{_cpp_simplify_pathname}.  For example,
1017
@file{/usr/include/bits/../foo.h} is simplified to
1018
@file{/usr/include/foo.h} before we enter it in the splay tree and try
1019
to @code{open ()} the file.  CPP will then find subsequent uses of
1020
@file{foo.h}, even as @file{/usr/include/foo.h}, in the splay tree and
1021
save system calls.
1022
 
1023
Further, it is likely the file contents have also been cached, saving a
1024
@code{read ()} system call.  We don't bother caching the contents of
1025
header files that are re-inclusion protected, and whose re-inclusion
1026
macro is defined when we leave the header file for the first time.  If
1027
the host supports it, we try to map suitably large files into memory,
1028
rather than reading them in directly.
1029
 
1030
The include paths are internally stored on a null-terminated
1031
singly-linked list, starting with the @code{"header.h"} directory search
1032
chain, which then links into the @code{<header.h>} directory chain.
1033
 
1034
Files included with the @code{<foo.h>} syntax start the lookup directly
1035
in the second half of this chain.  However, files included with the
1036
@code{"foo.h"} syntax start at the beginning of the chain, but with one
1037
extra directory prepended.  This is the directory of the current file;
1038
the one containing the @code{#include} directive.  Prepending this
1039
directory on a per-file basis is handled by the function
1040
@code{search_from}.
1041
 
1042
Note that a header included with a directory component, such as
1043
@code{#include "mydir/foo.h"} and opened as
1044
@file{/usr/local/include/mydir/foo.h}, will have the complete path minus
1045
the basename @samp{foo.h} as the current directory.
1046
 
1047
Enough information is stored in the splay tree that CPP can immediately
1048
tell whether it can skip the header file because of the multiple include
1049
optimization, whether the file didn't exist or couldn't be opened for
1050
some reason, or whether the header was flagged not to be re-used, as it
1051
is with the obsolete @code{#import} directive.
1052
 
1053
For the benefit of MS-DOS filesystems with an 8.3 filename limitation,
1054
CPP offers the ability to treat various include file names as aliases
1055
for the real header files with shorter names.  The map from one to the
1056
other is found in a special file called @samp{header.gcc}, stored in the
1057
command line (or system) include directories to which the mapping
1058
applies.  This may be higher up the directory tree than the full path to
1059
the file minus the base name.
1060
 
1061
@node Concept Index
1062
@unnumbered Concept Index
1063
@printindex cp
1064
 
1065
@bye

powered by: WebSVN 2.1.0

© copyright 1999-2024 OpenCores.org, equivalent to Oliscience, all rights reserved. OpenCores®, registered trademark.