So in natural languages syntax carries with it some general indications of meaning.
The same is true of the grammar of a programming language.
Consider a simple arithmetic expression
As long as you know the rules of arithmetic you will realize that you have to do the multiplication first. Arithmetic notation is a remarkably sophisticated mini-language which is why it takes some time to learn in school and why beginners make mistakes.
Implementing this arithmetic expression as a program is difficult because you can't simply read it from left to right and implement each operation as you meet it. That is 3+2*5 isn't (3+2)*5 but 3+(2*5) the multiplication has a higher priority.
A simple grammar for this type of expression, leaving out the obvious detail, might be
This means that this particular grammar only gives the syntax tree that corresponds to the correct grouping of the arithmetic operators and their operands.
In this case we have a grammar that reflects the semantics or meaning of the language and this is vital if the grammar is going to help with the translation.
There may be many grammars that generate a language and any one of these is fine for generating an expression or proving an expression legal but when it comes to parsing we need a grammar that means something.
Travelling the tree
Now that we have said the deeply unfashionable thing that syntax is not isolated from semantics we can now see why we bother to use a grammar analyser within a compiler.
Put simply a syntax tree or its equivalent can be used to generate the machine code or intermediate code that the expression corresponds to.
The syntax tree can be considered as a program that tells you how to evaluate the expression.
For example, a common method of generating code from a tree is to walk all its nodes using a “depth first” algorithm.
That is, visit the deepest nodes first and generate an instruction corresponding to the value or operator stored at each node. The details of how to do this vary but you can see the general idea in this diagram.
So now you know. We use grammar to parse expressions, to make syntax trees, to generate the code. Now find out about different types of grammar and parsing methods – they are important.
Exactly how does computer memory work? What is surprising is that it still works in more or less the same way as when Babbage designed his Analytical Engine or the IBM 360 accessed core memory. [ ... ]
Performing a computation sounds simple like a simple enough task and it is easy to suppose that everything is computable. In fact there are a range of different types of non-computability that we need [ ... ]