-- Secret sauce
A lot of the magic of F1 comes from Spanner, the distributed storage system. The name "F1" itself is an allusion to "inheriting" some of the properties of Spanner.
-- Hierarchical tables
What they call hierarchical tables, I would think best be viewed as one-to-many relationships. In guess they've privileged this model in their storage because that's what a lot of their AdWords schema looks like.
-- Change History
I like the observation that keeping full histories is relatively straightforward with atomic, granular timestamping and indeed that it should be baked in. Every database schema I've ever worked with always goes through a similar evolutionary cycle:
1. We only need to capture the current state of the model.
2. Wait, we do need to capture historical states of the model.
3. Wait, the model is changed, we need to capture historical states and the models that were current.
(You can think of this as taking progressive differentials of incoming transactions).
The F1 designers have baked that right into the database, where it belongs. Weak temporal support has long been the sore point in SQL.
-- Remote data
I was struck by their observation that most database storage engines are built around the concept of seeks and reads, whereas theirs is necessarily built around batching and pipelining over a network. If I am reading them correctly, their engine takes advantage of having multiple copies of data by sending reads to multiple disk nodes and then working from the first copy that is returned.
From what I understand of Datomic temporal support is baked in, but it's about the only database I can think of that does that, and I haven't heard stories about it being used in production for large volumes of data.
http://www.cs.arizona.edu/people/rts/tdbbook.pdf
More generally, you can usually build a model that handles (2), the changing state, by generalising your original model. I've done in both of classical ways, having either validity period fields (thus pushing complexity into every query), or by having audit tables (thus relying on triggers).
But building a changing model is hard, because at the schema level, SQL only supports (1). That is, it provides primitives to change the current state of the model.
So you wind up having to build a meta-model. The deficiencies of SQL (and I'm an RDBMS bigot) mean that we wind up building inner platforms. Or outer platforms -- witness migration toolkits.
Yet support for changing models is essential. No model is ever constant. Most of my work is for a public-sector client whose legal reporting requirements are constantly changing. You cannot throw away old reports which were made under old regulations; you must be able to recreate the world at any point in time. Which requires either a lot of bookkeeping code or taking periodic snapshots.
Datomic looks neat. I see that I'm not the only one who made the leap from transactional memory to transactional models.
Edit: I just found notes I wrote in 2009 on database languages while looking for something else. Spooky: http://chester.id.au/2013/08/28/notes-towards-a-set-objectiv...
This is actually the model they've used in things like AppEngine as well - nested entities allow them to take advantage of locality in transactions.
I think recursive structures like lists and trees are one weakness of the relational model, and haven't found a fully satisfying relational answer to the limitations. It seems to me like the assumption of atomic column types is a major weakness inherent in the relational model itself when it comes to recursion. Any thoughts/comments?
If I am reading them correctly, it's a storage strategy, not a specific "feature" per se. The closest analogy to hstore is that they provide native support for storing and querying protobuf blobs.
> I think recursive structures like lists and trees are one weakness of the relational model, and haven't found a fully satisfying relational answer to the limitations.
It depends on why you're using trees or graphs.
If it's inherent in the data, then modern SQL has recursive queries that make it much easier than the old methods.
If it's inherent in the model, you will find it harder. You might need to pick a non-standard approach, such as PostgreSQL's inherited tables. I'd think long and hard before saying it's inherent in the model, by the way. Strictly speaking you can represent the same thing as sets of relations or as a graph; it's better to utilise the strengths of the tool in front of you.
If you are interested in how to do it take a look here: https://blog.dsl-platform.com/postgres-bridge-between-worlds...
<http://stackoverflow.com/questions/18384883/why-is-googles-t...
However, I think this HN thread seems like it might be a good place to get comments on why my line of reasoning may be incorrect. Does anyone have any thoughts on why building something like Spanner on top of basic Paxos quorums and NTP would be a bad idea?
They do. They have a range of commercial, including enterprise, offerings (many of which are based on originally-internal technologies, including their MySQL-based distributed database that preceded F1.)
> why wouldn't google commercialize that technology themselves?
They probably will, just as they have commercialized their previous internal storage technologies.