RedBlack Tree
A redblack tree is a type of binary search tree. It is self balancing like the AVL tree, though it uses different properties to maintain the invariant of being balanced. Balanced binary search trees are much more efficient at search than unbalanced binary search trees, so the complexity needed to maintain balance is often worth it. They are called redblack trees because each node in the tree is labeled as red or black.
Redblack trees maintain a slightly looser height invariant than AVL trees. Because the height of the redblack tree is slightly larger, lookup will be slower in a redblack tree. However, the looser height invariant makes insertion and deletion faster. Also, redblack trees are popular due to the relative ease of implementation.
This image shows a representation of a redblack tree. Notice how each leaf is actually a black, null value. These properties will be important in proving the tree's height invariant.
Contents
Overview
The redblack tree is similar to the binary search tree in that it is made up of nodes and each node has at most two children. However, there are new properties that are specific to the redblack tree.
 Each node is either red or black, this can be saved in memory as a single bit (e.g. 'red' = 1, 'black' = 0).
 The root of the tree is always black.
 All leaves are null and they are black.
 If a node is red, then its parent is black.
 Any path from a given node to any of its descendant leaves contains the same amount of black nodes. This is sometimes known as the blackdepth.
 The height of the redblack tree is at most \(2 \cdot \log_2(n + 1)\) ; this property will be proven later.
When certain nodes are inserted that upset the height invariant of the tree, the tree is then rearranged using the current coloring scheme of its nodes. Once the tree is rearranged, it is repainted to ensure that the coloring properties are maintained.
RedBlack Tree Height
Proving the height of the redblack tree is especially important because the height of the redblack tree is what allows us to calculate it asymptotic complexity and performance. This is one method of doing so.
First imagine a redblack tree with height \(h\). Now, we merge all red nodes into their black parents. A given black node can either have:
 2 black children, in which case the black parent still has 2 children.
 1 black child and 1 red child, in which case the black parent now has 3 children.
 2 red children, in which case the black parent now has 4 children.
Here is a graphical example of that merging process (assume any stray arrow points to a black node).
Now, we merge all of the red nodes into their parent nodes. This means that any black node will now have \(2 \cdot r + 2\) pointers coming out of it, where \(r\) is the number of red children they have.
As you can see, every black node has either 2, 3, or 4 children.
This new tree has a height, \(h_1\). Because any given path in the original redblack tree had at most half its nodes red, we know that this new height is at least half the original height. So,
\[h_1 \ge h/2\]
The number of leaves in a tree is exactly equal to \(n+1\), so
\[n+1 \ge 2^{h_1}\] \[\log_2(n+1) \ge h_1 \ge h/2\] \[h \le 2\log_2(n+1).\]
Operations
In a redblack tree, there are two operations that can change the structure of the tree, insert and delete. These changes might involve the addition or subtraction of nodes, the changing of a node's color, or the reorganization of nodes via a rotation. By showing the various cases and algorithms for the insertion process, though, we can infer the same things about deletion.
Inserting a node involves first searching through the tree to find its rightful spot. Then the node is inserted as a red node. This might violate the property that a red node's parent is black, though. So, now we have three potential cases that we might have to deal with.
1) Case 1
In the first case, the node we've inserted is red, it's parent is red, and it's parent sibling is red. We know that the inserted node's grandparent will be black, so all we need to do is switch the coloring of the inserted node's grandparent with the coloring of the inserted node's parent and its parent's sibling. This case might need to continue to be fixed up through the root of the tree, though, because the inserted node's grandparent may have a parent who is red.
This following graphic shows this case. The node that was inserted is labeled as "INSERT". The left tree shows the case, and the right tree shows it rectified.
2) Case 2
Case 2 is a case that will be transformed into case 3. Case 2 is the first case in which we will perform a rotation. There are two kinds of rotations, a left rotation and a right rotation. Case 2 uses a left rotation because the nodes involved are rotated counterclockwise. The following image shows the rotation in case 2. Remember, this rotation does not fully fix the problem, but it sets it up to be solved in case 3.
As you can see, a left rotation was performed on the node labeled "ROTATE".
3) Case 3
Case 3 involves a right rotation on the grandparent. In the following graphic, the node to be rotated about it labeled "ROTATE", the inserted node is labeled "INSERT", and a third node "MIDDLE" has been labeled to show where it ends up after the rotation.
Deletion, as stated above, works in the exact same way. Once a node has been taken out of the tree, the tree can be in any of the above 3 cases. Then the issue is resolved in the same way.
Asymptotic Complexity
We know what needs to be done in the events of an insertion or a deletion. So what's the runtime of each operation?
As we proved earlier, the height is at most \(h \le 2 \cdot \log_2(n+1)\). This means that the process of findind an index at which we can insert or delete will be an \(O(\log_2(n))\) operation. At that point we can fall into any of our three cases.
In case 1, we need to back out of the tree, recoloring nodes as we go. This process also takes \(O\log_2(n)\), so it doesn't increase our complexity at all.
In cases 2 and 3 we perform 1 or 2 rotations, respectively. At that point, we terminate. These cases have constant time operations. So, we know that insertion and deletion takes \(O\log_2(n)\) time.
Search in a redblack tree is the same as any balanaced binary search tree, \(O\log_2(n)\) time. Traversal is a \(O(n)\) amortized operation because to search through the entire tree, you simply have to enter and exit each node.
See also: big O notation.
Average  
Space  \(O(n)\) 
Search  \(O(\log_2(n))\) 
Traversal  *\(O(n)\) 
Insert  \(O(\log_2(n))\) 
Delete  \(O(\log_2(n))\) 
*amortized analysis
Python Implementation
A python implementation might look something like this. Note that this implementation is for academic purposes only and does not guarantee any functionality.
The basic node and tree class can look something like this. Each node needs to keep track of its children, its parent, its color and its key. The RBTree
class will start out with nothing as its leaves and nothing as its root. Only the root will change, though.
1 2 3 4 5 6 7 8 9 10 

A search function can be easily implemented using binary search. Note that this implementation assumes that no two keys will have the same value.
1 2 3 4 5 6 7 8 

The insert function will behave similarly to the search in that it will find that place in the tree that the new node should go. However, it will need to call functions that will make sure the tree is fixed afterwards.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 

The fixTree
method is where most of the heavy lifting takes place. It will decide what kind of case we are in and act appropriately.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 

And here is the left_rotate
method which will rotate about the given node. The right_rotate
method is the exact same thing in the other direction.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 

References
 burnett, C. Wikipedia RedBlack Trees. Retrieved April 25, 2016, from https://en.wikipedia.org/wiki/Red–black_tree