The C Preprocessor Versus D
Back when C was invented, compiler technology was primitive. Installing a text macro preprocessor onto the front end was a straightforward and easy way to add many powerful features. The increasing size & complexity of programs have illustrated that these features come with many inherent problems. D doesn't have a preprocessor; but D provides a more scalable means to solve the same problems.
- Header Files
- #pragma once
- #pragma pack
- Macros
- Conditional Compilation
- Code Factoring
- #error and Static Asserts
- Template Mixins
Header Files
The C Preprocessor Way
C and C++ rely heavily on textual inclusion of header files. This frequently results in the compiler having to recompile tens of thousands of lines of code over and over again for every source file, an obvious source of slow compile times. What header files are normally used for is more appropriately done doing a symbolic, rather than textual, insertion. This is done with the import statement. Symbolic inclusion means the compiler just loads an already compiled symbol table. The needs for macro "wrappers" to prevent multiple #inclusion, funky #pragma once syntax, and incomprehensible fragile syntax for precompiled headers are simply unnecessary and irrelevant to D.
#include <stdio.h>
The D Way
D uses symbolic imports:
import std.c.stdio;
#pragma once
The C Preprocessor Way
C header files frequently need to be protected against being #include'd multiple times. To do it, a header file will contain the line:
#pragma once
or the more portable:
#ifndef __STDIO_INCLUDE
#define __STDIO_INCLUDE
... header file contents
#endif
The D Way
Completely unnecessary since D does a symbolic include of import files; they only get imported once no matter how many times the import declaration appears.
#pragma pack
The C Preprocessor Way
This is used in C to adjust the alignment for structs.
The D Way
For D classes, there is no need to adjust the alignment (in fact, the compiler is free to rearrange the data fields to get the optimum layout, much as the compiler will rearrange local variables on the stack frame). For D structs that get mapped onto externally defined data structures, there is a need, and it is handled with:
struct Foo
{
align (4): // use 4 byte alignment
...
}
Macros
Preprocessor macros add powerful features and flexibility to C. But they have a downside:
- Macros have no concept of scope; they are valid from the point of definition to the end of the source. They cut a swath across .h files, nested code, etc. When #include'ing tens of thousands of lines of macro definitions, it becomes problematical to avoid inadvertent macro expansions.
- Macros are unknown to the debugger. Trying to debug a program with symbolic data is undermined by the debugger only knowing about macro expansions, not the macros themselves.
- Macros make it impossible to tokenize source code, as an earlier macro change can arbitrarily redo tokens.
- The purely textual basis of macros leads to arbitrary and inconsistent usage, making code using macros error prone. (Some attempt to resolve this was introduced with templates in C++.)
- Macros are still used to make up for deficits in the language's expressive capability, such as for "wrappers" around header files.
Here's an enumeration of the common uses for macros, and the corresponding feature in D:
- Defining literal constants:
The C Preprocessor Way
#define VALUE 5
The D Way
const int VALUE = 5;
- Creating a list of values or flags:
The C Preprocessor Way
int flags: #define FLAG_X 0x1 #define FLAG_Y 0x2 #define FLAG_Z 0x4 ... flags |= FLAG_X;
The D Way
enum FLAGS { X = 0x1, Y = 0x2, Z = 0x4 }; FLAGS flags; ... flags |= FLAGS.X;
- Distinguishing between ascii chars and wchar chars:
The C Preprocessor Way
#if UNICODE #define dchar wchar_t #define TEXT(s) L##s #else #define dchar char #define TEXT(s) s #endif ... dchar h[] = TEXT("hello");
The D Way
dchar[] h = "hello";
D's optimizer will inline the function, and will do the conversion of the string constant at compile time. - Supporting legacy compilers:
The C Preprocessor Way
#if PROTOTYPES #define P(p) p #else #define P(p) () #endif int func P((int x, int y));
The D Way
By making the D compiler open source, it will largely avoid the problem of syntactical backwards compatibility. - Type aliasing:
The C Preprocessor Way
#define INT int
The D Way
alias int INT;
- Using one header file for both declaration and definition:
The C Preprocessor Way
#define EXTERN extern #include "declarations.h" #undef EXTERN #define EXTERN #include "declarations.h"
In declarations.h:EXTERN int foo;
The D Way
The declaration and the definition are the same, so there is no need to muck with the storage class to generate both a declaration and a definition from the same source. - Lightweight inline functions:
The C Preprocessor Way
#define X(i) ((i) = (i) / 3)
The D Way
int X(ref int i) { return i = i / 3; }
The compiler optimizer will inline it; no efficiency is lost. - Assert function file and line number information:
The C Preprocessor Way
#define assert(e) ((e) || _assert(__LINE__, __FILE__))
The D Way
assert() is a built-in expression primitive. Giving the compiler such knowledge of assert() also enables the optimizer to know about things like the _assert() function never returns. - Setting function calling conventions:
The C Preprocessor Way
#ifndef _CRTAPI1 #define _CRTAPI1 __cdecl #endif #ifndef _CRTAPI2 #define _CRTAPI2 __cdecl #endif int _CRTAPI2 func();
The D Way
Calling conventions can be specified in blocks, so there's no need to change it for every function:extern (Windows) { int onefunc(); int anotherfunc(); }
- Hiding __near or __far pointer weirdness:
The C Preprocessor Way
#define LPSTR char FAR *
The D Way
D doesn't support 16 bit code, mixed pointer sizes, and different kinds of pointers, and so the problem is just irrelevant. - Simple generic programming:
The C Preprocessor Way
Selecting which function to use based on text substitution:#ifdef UNICODE int getValueW(wchar_t *p); #define getValue getValueW #else int getValueA(char *p); #define getValue getValueA #endif
The D Way
D enables declarations of symbols that are aliases of other symbols:version (UNICODE) { int getValueW(wchar[] p); alias getValueW getValue; } else { int getValueA(char[] p); alias getValueA getValue; }
Conditional Compilation
The C Preprocessor Way
Conditional compilation is a powerful feature of the C preprocessor, but it has its downside:
- The preprocessor has no concept of scope. #if/#endif can be interleaved with code in a completely unstructured and disorganized fashion, making things difficult to follow.
- Conditional compilation triggers off of macros - macros that can conflict with identifiers used in the program.
- #if expressions are evaluated in subtly different ways than C expressions are.
- The preprocessor language is fundamentally different in concept than C, for example, whitespace and line terminators mean things to the preprocessor that they do not in C.
The D Way
D supports conditional compilation:
- Separating version specific functionality into separate modules.
- The debug statement for enabling/disabling debug harnesses, extra printing, etc.
- The version statement for dealing with multiple versions of the program generated from a single set of sources.
- The if (0) statement.
- The /+ +/ nesting comment can be used to comment out blocks of code.
Code Factoring
The C Preprocessor Way
It's common in a function to have a repetitive sequence of code to be executed in multiple places. Performance considerations preclude factoring it out into a separate function, so it is implemented as a macro. For example, consider this fragment from a byte code interpreter:
unsigned char *ip; // byte code instruction pointer
int *stack;
int spi; // stack pointer
...
#define pop() (stack[--spi])
#define push(i) (stack[spi++] = (i))
while (1)
{
switch (*ip++)
{
case ADD:
op1 = pop();
op2 = pop();
result = op1 + op2;
push(result);
break;
case SUB:
...
}
}
This suffers from numerous problems:
- The macros must evaluate to expressions and cannot declare any variables. Consider the difficulty of extending them to check for stack overflow/underflow.
- The macros exist outside of the semantic symbol table, so remain in scope even outside of the function they are declared in.
- Parameters to macros are passed textually, not by value, meaning that the macro implementation needs to be careful to not use the parameter more than once, and must protect it with ().
- Macros are invisible to the debugger, which sees only the expanded expressions.
The D Way
D neatly addresses this with nested functions:
ubyte* ip; // byte code instruction pointer
int[] stack; // operand stack
int spi; // stack pointer
...
int pop() { return stack[--spi]; }
void push(int i) { stack[spi++] = i; }
while (1)
{
switch (*ip++)
{
case ADD:
op1 = pop();
op2 = pop();
push(op1 + op2);
break;
case SUB:
...
}
}
The problems addressed are:
- The nested functions have available the full expressive power of D functions. The array accesses already are bounds checked (adjustable by compile time switch).
- Nested function names are scoped just like any other name.
- Parameters are passed by value, so need to worry about side effects in the parameter expressions.
- Nested functions are visible to the debugger.
Additionally, nested functions can be inlined by the implementation resulting in the same high performance that the C macro version exhibits.
#error and Static Asserts
Static asserts are user defined checks made at compile time; if the check fails the compile issues an error and fails.
The C Preprocessor Way
The first way is to use the #error preprocessing directive:
#if FOO || BAR
... code to compile ...
#else
#error "there must be either FOO or BAR"
#endif
This has the limitations inherent in preprocessor expressions (i.e. integer constant expressions only, no casts, no sizeof, no symbolic constants, etc.).
These problems can be circumvented to some extent by defining a static_assert macro (thanks to M. Wilson):
#define static_assert(_x) do { typedef int ai[(_x) ? 1 : 0]; } while(0)
and using it like:
void foo(T t)
{
static_assert(sizeof(T) < 4);
...
}
This works by causing a compile time semantic error if the condition evaluates to false. The limitations of this technique are a sometimes very confusing error message from the compiler, along with an inability to use a static_assert outside of a function body.
The D Way
D has the static assert, which can be used anywhere a declaration or a statement can be used. For example:
version (FOO)
{
class Bar
{
const int x = 5;
static assert(Bar.x == 5 || Bar.x == 6);
void foo(T t)
{
static assert(T.sizeof < 4);
...
}
}
}
else version (BAR)
{
...
}
else
{
static assert(0); // unsupported version
}
Template Mixins
D template mixins superficially look just like using C's preprocessor to insert blocks of code and parse them in the scope of where they are instantiated. But the advantages of mixins over macros are:
- Mixins substitute in parsed declaration trees that pass muster with the language syntax, macros substitute in arbitrary preprocessor tokens that have no organization.
- Mixins are in the same language. Macros are a separate and distinct language layered on top of C++, with its own expression rules, its own types, its distinct symbol table, its own scoping rules, etc.
- Mixins are selected based on partial specialization rules, macros have no overloading.
- Mixins create a scope, macros do not.
- Mixins are compatible with syntax parsing tools, macros are not.
- Mixin semantic information and symbol tables are passed through to the debugger, macros are lost in translation.
- Mixins have override conflict resolution rules, macros just collide.
- Mixins automatically create unique identifiers as required using a standard algorithm, macros have to do it manually with kludgy token pasting.
- Mixin value arguments with side effects are evaluated once, macro value arguments get evaluated each time they are used in the expansion (leading to weird bugs).
- Mixin argument replacements don't need to be 'protected' with parentheses to avoid operator precedence regrouping.
- Mixins can be typed as normal D code of arbitrary length, multiline macros have to be backslash line-spliced, can't use // to end of line comments, etc.
- Mixins can define other mixins. Macros cannot create other macros.