Ken Thompson did many wonderful things. Probably more than any programmer, he influenced the programming style which we have right now. While he did not invent C (that was done by his friend Dennis Ritchie), he invented the programming style which underlies C. Dealing with pointers, knowing how pointers are subtracted, and stuff like that, all comes from Ken Thompson.
Believe it or not, the best and brightest at that time were heavily on the march to get rid of pointers. Absolutely brilliant people who would give Turing award speeches. Tony Hoare is the case in point saying that pointers have to be abolished1. The idea was that pointer is illegitimate thing. Language designers were fully convinced that even if you provide pointers nobody should call them pointers. They would call them something like access types. They have to prohibit iterations, like you could never subtract pointers. Niklaus Wirth, the designer of Pascal, was very certain that you should never allow subtraction between pointers. We still have it at least in some languages. I know that our Java friends do not believe in subtracting pointers and subtraction of pointers. But, pointers are still at least partially with us.
So Ken is an absolutely great man in many respects. His career started, not with UNIX, but when he was freshly out of school coming up with the brilliant practical algorithm for matching regular expressions2. Every time you write grep or something like that you’re most likely exercising code written by Ken in the late 60s.
There is so much practical reliance on regular expressions. But they were invented believe it or not by theoreticians, specifically Stephen Kleene II who was a logician. So Ken made them practical3.
Then he did UNIX, this totally brilliant operating system on which we all rely. All of our livelihoods come from Ken, in one shape or form. Do you remember the fortune cookie program? One of the fortune cookies was: “it is all Ken’s fault”4. We have to remember that.
After he does that, just to demonstrate that he can do anything, he decides to show people how to play computer chess. Mind it, at that point as far as I know he barely knew the moves. There was total dominance by Russian chess playing program Kaissa and two or three years he builds this specialized hardware, totally revolutionizes the approach to just playing. Of course nobody remembers what Kaissa was. He is a guy who could enter in a totally new field and do something beyond, not just world class, but beyond world class in a short period of time.
In 1984 he was given the Turing award. It was shared with Dennis and he gave a beautiful speech which I highly recommend that you read called “Reflections on Trusting Trust” which talks about many things. But, I’ll use just one little episode in the beginning which is very important from my point of view. He says that he was very blessed with collaborators specifically with Dennis5. Then he continues that during many years of their collaboration not once they would attempt to do the same thing. They had this good sense to work on complementary things. Except once. Both of them needed to write an assembly program with about 20 lines of code and continues Ken, “I checked, it was character by character identical”. It sounds like mystical thing right, two people coming together, this is a Vulcan mind-meld (joke).
But, there is something profound there. I actually think that it’s less amazing than Ken makes it be. That is a central point of what I am trying to teach here. I actually had such mind-meld with several of my colleagues after we worked together for a while. For example, it’s literally true that when Dave Musser and I were working together long-distance, we would write the same thing. It was invariably, character by character, identical (even though we were not as good as Ken and Dennis).
I claim that if we are really turning programming into professional discipline then the story of Ken and Dennis will not be a miracle. We should be writing basic code, character by character, identical. Imagine how wonderful it would be if you could understand what someone else wrote.
Whomever really wants to learn, will learn, and that is a challenge because it is my experience that if we take the intersection of what a room of programmers know it’s going to be an empty set. That doesn’t mean that you as an individual person doesn’t know things, but intersection is going to be relatively small.
Unfortunately we’ve got to a point where nobody teaches programming. Because there’s no professor of computer science who has any idea how to program6. Part of what I am trying to do here is to start discussing little things like, how do we name variables? Why is it important? We want to get to the point where everybody writes a consistent code, recognizable code. This is why I want to go slow and develop so that we all agree7.
We have to start with a very very simple program. Most of you recognize it, right? Hello World.
#include <iostream>
int main() {
std::cout << "Hello, world!" << std::endl;
}
There is no return value.
Do you think it will compile?
It does.
C++ treats main
as a special function; meaning when the control goes
through the last brace, zero is returned.
Because the standard UNIX convention, which became Universal
convention, that on success you return zero.
The language actually allows you to do things like that.
One warning, cout
stuff works pretty much like how you think it works.
However a friend of mine wrote something like:
std::cout << f() << g();
Lo and behold depending on the compiler, different results were printed.
The order of evaluation of arguments to a function is not defined
and the operator <<
is just a function call8.
Never depend on the arguments to a function
being evaluated in a given order.
The person in question actually immediately declared that he found the compiler bug. Compilers do have bugs but they are not as common as people believe.
This whole talk about performance started me and some coworkers investigating problems and involved very long queries. When we narrowed things down we always found something pertaining to incorrect use of STL9. It’s usually a one liner and the most egregious.
The most amazing thing is the following one liner which I will start in the beginning of the course, because it was just so important. We could weave the whole fabric of the course around this one little example. There was a range of integers. For example:
int a[] = { 1, 3, 1, 4, 1, 5 };
The person wanted to find out how many of these integers are unique. The person decided to write the following code10:
#include <iostream>
#include <set>
int main() {
std::set<int> set_of_ints(a, a + 6);
std::cout << set_of_ints.size() << std::endl;
}
This is perfectly good. It computed the thing. But it is actually very slow.
The algorithm for set
11 is red black tree which is
universally stated in every book to be one of the best implementation of sets.
Do we have to believe textbooks? We will see.
In any case, we’ll learn to count number of operations.
We will observe that indeed it is going to do O(n log(n))
number of comparisons.
But, it is awfully slow.
Why?
Well, it has to do sort, but without actually sorting.
Otherwise, you really cannot find out how many unique elements are in a sequence of integers.
This is a paradoxical thing since finding unique elements is much
easier operation12.
It seems that finding unique elements does not require ordering, it just requires equality. But, what we are actually doing is optimizing a search or find. Equality gives us linear search while sorting gives us binary search so we can find much much faster. One of the amazing things which we will discover is that ordering is very important. Things which we could do with ordering cannot be effectively done just with equality13.
If we are going by the book we will say sorting is good as long it does approximately O(n log(n))
operations.
This is good for the books.
It’s actually not good for our code, because Big O
could have an arbitrary coefficient.
It could be 50(n log(n))
14,
The effective coefficient, not just in terms of number of operations,
but in terms of time for this is very very
large. How large? We will be doing measurements together of this very problem.
But, let me tell you, it slows down between factor of 22 and a factor of 50 on large data sets.
That is a lot.
How should we solve the problem? Could we replace it with a one-liner?
There is an STL function just for this.
It is called std::unique
15.
What does it do?
It goes through a range and removes all non-unique elements,
giving back a range of unique elements.
Of course the presupposition is that the input range is sorted.
It will work if the range is not sorted either.
It will just not eliminate all equal elements.
It returns a pointer to the element past the last element of the range.
This is a standard STL convention, that we will study.
For example given:
1 2 2 2
^
The returned pointer is indicated. So let’s use it
std::sort(a, a + 6);
std::cout << std::unique(a, a + 6) - a << std::endl;
Why am I not going to use std::distance
here (instead
of subtracting the pointer a
from the result of std::unique
)?
std::distance(a, std::unique(a, a + 6));
There is no need. Never use a function which is intended to be used in a general case, when you sit deeper inside some specific case.
Whenever people have problems with STL, they ask me questions. They show me their examples of using STL. The most common thing is something like map of maps:
std::map<..., std::map<...> >
You sit there and you wonder. Before STL, nobody in his right mind would dare to say “I’m going to have a self-balancing tree which at every node has a self-balancing tree”, because it will take them five years to implement it.
It’s maybe one of STL’s strong points, that such a construction is so trivial. You could not just have map of maps, you could have maps of vectors of list of maps of sets. And in one line. Will it work? Yes, for some definition of work. Will it be fast? No. It cannot be fast because you have this enormous complex stuff. Its complexity is hidden but it’s still there.
Miracles are not possible. If you ask somebody to create a very complex data structure that’s what you’re going to get. You’re going to get problems with node allocation. You are going to get problems with rebalancing. You’re going to get problem with whatever these advanced data structures do. These problems are going to get worse and worse. You’re going to get a cache miss on every step through the set.
As our computers become faster and faster and faster they’re getting slower and slower and slower16. Meaning that going to the main memory is very slow. You want to have locality of reference. You want all your data that you are working on in the local cache. If you have a huge set or map it’s not going to be like that.
Let me tell you a very simple piece of advice. If you remember just one thing out of this course it should be this:
std::vector
.Avoid any data structures except arrays. “Well aren’t there exceptions?” No, not for you. Because typically advanced data structures are slower than simple data structures. Data structures which appear to be alright when textbook writers wrote their books, are no longer all right now. Computers change so much. Things no longer work.
If set
is bad? Why is it in the standard?
Think about it.
If you do all of your insertions first, and then start searching,
you do not need a set
.
The set is a very specific data structure which is needed only
when you have a very specific workload.
If you have a thing which grows
and shrinks constantly dynamically.
Then you have to have a set
.
You need a set
only if you need a thing which does not move things around.
As long as something gets into a set
and it is not erased the pointer, it is fixed.
For example, him sitting in this chair is in the set
.
As long as he’s in this set he will not move from his chair.
You could find him in constant time.
It’s a very useful thing except most people do not use set for that.
There is a great book all of you should have and all of you should aspire to read (but actual reading that we don’t know…) The book I am referring to (there’s only one book like that in computer science) Art of Computer Programming, written by Donald Knuth. It’s a very imperfect book but it’s the greatest book we got.
One of the things which I want to convince you of is that we want to be professionals. As professionals we want to have certain things on our bookshelves even if we don’t read them. Lawyers have all this legal stuff in their office, they don’t know anything about what’s inside. But, they know they have to put it on the shelves. We at least have to aspire to the same level of professionalism. We put Knuth on the shelf.
When we started reading Knuth we observed that even a very long time ago, Knuth started doing measuring operations. That is, when he would describe an algorithm he would not just say Big O, that came later (in spite of the fact that he introduced Big O into analysis of algorithms) but in his book he would carefully annotate how many operations.
I want to get us to a point where we start learning how to measure things and when we measure algorithms there are two ways we could measure:
Both of them are important.
Because if you count operations you could have
analysis of algorithm in terms of abstract operations not depending on the
specific type.
This is very good, we have to have that.
But then we also need to have times for specific commonly used types.
So these are two things which we need to do.
Our goal is to be able to measure this problem with set
and with sort.
But, first we need to learn more about C++.
<<
is not some special form (like the notion in Lisp), it is just a function call.↩STL (standard template library) was of course originally written by the teacher Alex Stepanov. The original library was written for SGI, but then later integrated into the standard library. However, the version available in most implementations is much more complex and less consistent than Alex’s original work. I have archived the SGI STL code which probably contains more of Alex' original code than other that is available.
The SGI STL documentation has also been archived and is a very useful reference alongside modern references, especially for understanding concepts.
↩#include <set>
when I first did STL.
I had one gigantic file called stl.h
At that time people said it’s utterly unacceptable, you have twenty thousand lines of code.
If you think about it nowadays it appears to be very funny. Twenty thousand is nothing.
That’s why stl.h
was replaced with a gazillion little headers.
Sometimes I have no idea what they are.↩2^32
entries and count how many times integers are found only once.↩remove-duplicates
and it relied on equality.
It actually would go through N
times removing things.
They did not sort.
They use equality.
It did work and it worked exceptionally well for most of the applications.
In Lisp you usually have a list with 5 elements so it works just fine.
The moment you go to larger lists things stop working.
Quadratic algorithms are slow.↩To say a function f
is O(g)
it’s only required that f
be bounded
by a constant multiple for all inputs beyond some point.
Formally, if f(x) <= Ag(x)
for all x > M
, where A
and M
are constant real numbers
(see “The Art of Computer Programming” 1.2.11).
Alex will tend to say “order of” to distinguish O(f)
from exactly f
.
In this section, Alex is observing that there is no requirement for A
to be small.
This performance description is based on several trends in computer architecture. The first is that memory, especially main memory hasn’t increased in read/write speed relative to the increase in processor speed. So reading and writing to memory is relatively expensive.
Second, and related is that to compensate CPUs now have many layers of caches with layers increasing in speed and decreasing in size as they get closer to the CPU. The CPU does a lot of guessing to anticipate what data needs to be in the cache for a given section of code. Data structures and algorithms which jump over the address space tend to mess this up so the CPU must pause and load data in and out of cache.
↩