Your SlideShare is downloading. ×
  • Like
Putting freebase in a star schema
Upcoming SlideShare
Loading in...5

Thanks for flagging this SlideShare!

Oops! An error has occurred.


Now you can save presentations on your phone or tablet

Available for both IPhone and Android

Text the download link to your phone

Standard text messaging rates apply

Putting freebase in a star schema


Freebase is a open database of things that exist in the world: things like people, places, songs and television shows. As of the January 2009 dump, Freebase contained about 241 million facts, …

Freebase is a open database of things that exist in the world: things like people, places, songs and television shows. As of the January 2009 dump, Freebase contained about 241 million facts, and it’s growing all the time. You can browse it via the web and even edit it, much like Wikipedia. Freebase also has an API that lets programs add data and make queries using a language called MQL. Freebase is complementary to DBpedia and other sources of information. Although it takes a different approach to the semantic web than systems based on RDF standards, it interoperates with them via linked data.

The January 2009 Freebase dump is about 500 MB in size. Insi

Published in Technology
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Be the first to comment
    Be the first to like this
No Downloads


Total Views
On SlideShare
From Embeds
Number of Embeds



Embeds 0

No embeds

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

    No notes for slide


  • 1. Generation 5 » Putting Freebase in a Star Schema  Subscribe to our RSS Feed | About Us Putting Freebase in a Star Schema What’s Freebase? Freebase is a open database of things that exist in the world:  things like people,  places,  songs and television shows.   As of the January 2009 dump,  Freebase contained about 241 million facts,  and it’s growing all the time.  You can browse it via the web and even edit it,  much like Wikipedia.  Freebase also has an API that lets programs add data and make queries using a language called MQL.  Freebase is complementary to DBpedia and other sources of information.  Although it takes a different approach to the semantic web than systems based on RDF standards,  it interoperates with them via  linked data. The January 2009 Freebase dump is about 500 MB in size.  Inside a bzip-compressed files,  you’ll find something that’s similar in spirit to a Turtle RDF file,  but is in a simpler format and represents facts as a collection of four values rather than just three. Your Own Personal Freebase To start exploring and extracting from Freebase,  I wanted to load the database into a star schema in a mysql database — an architecture similar to some RDF stores,  such as ARC.  The project took about a week of time on a modern x86 server with 4 cores and 4 GB of RAM and resulted in a 18 GB collection of database files and indexes. This is sufficient for my immediate purposes,  but future versions of Freebase promise to be much larger:  this article examines the means that could be used to improve performance and scalability using parallelism as well as improved data structures and algorithms. I’m interested in using generic databases such as Freebase and Dbpedia as a data source for building web sites.  It’s possible to access generic databases through  APIs,  but there are advantages to having your own copy:  you don’t need to worry about API limits and network latency,  and you can ask questions that cover the entire universe of discourse. Many RDF stores use variations of a format known as a Star Schema for representing RDF triples;  the Star Schema is commonly used in data warehousing application because it can efficiently represent repetitive data.   Freebase is similar to,  but not quite an RDF system.  Although there are multiple ways of thinking about Freebase,  the quarterly dump files provided by Metaweb are presented as quads:  groups of four related terms in tab-delimited terms.  To have a platform for exploring freebase,  I began a project of loading Freebase into a Star Schema in a relational database. A Disclaimer Timings reported in this article are approximate.  This work was done on a server that was doing other things; little effort was made to control sources of variation such as foreign workload,  software configuration and hardware characteristics.  I think it’s orders of magnitude that matter here:  with much larger data sets becoming available,  we need tools that can handle databases 10-100 times as big,  and quibbling about 20% here or there isn’t so important.  I’ve gotten similar results with the ARC triple store.  Some products do about an order of magnitude better:  the Virtuoso server can load DBpedia,  a larger database,   in about 16 to 22 hours on a 16 GB computer:  several papers on RDF store performance are available [1] [2] [3].  Although the system described in this paper isn’t quite an RDF store,  it’s performance is comprable to a relatively untuned RDF store.[1/12/2014 8:17:56 PM] Search for: Search Archives June 2012 (1) August 2010 (1) May 2010 (1) June 2009 (2) April 2009 (1) March 2009 (1) February 2009 (3) January 2009 (3) November 2008 (1) August 2008 (2) July 2008 (5) June 2008 (5) May 2008 (2) April 2008 (6) March 2008 (8) June 2006 (1) February 2006 (1) Categories AJAX (2) Asynchronous Communications (16) Biology (1) Books (1) Design (1) Distributed (1) Exceptions (2) Functional Programming (1) GIS (1) Ithaca (1) Japan (1) Math (1) Media (3) Nature (1) Semantic Web (3) Tools (28) CRUD (1) Dot Net (17) Freebase (2) GWT (9) Java (7) Linq (2) PHP (6) Server Frameworks (1) Silverlight (12) SQL (5) Uncategorized (1) Web (2) Analytics (1)
  • 2. Generation 5 » Putting Freebase in a Star Schema It took about a week of calendar time to load the 241 million quads in the January 2009 Freebase into a Star Schema using a modern 4-core web server with 4GB of RAM;  this time could certainly be improved with microoptimizations,  but it’s in the same range that people are observing that it takes to load 10^8 triples into other RDF stores.  (One product is claimed to load DBPedia,  which contains about 100 million triples,  in about 16 hours with “heavy-duty hardware”.)   Data sets exceeding 10^9 triples are becoming rapidly available — these will soon exceed what can be handled with simple hardware and software and will require new techniques:  both the use of parallel computers and optimized data structures. The Star Schema In a star schema,  data is represented in separate fact and dimension tables, all of the rows in the fact table ( quad ) contain integer keys — the values associated with the keys are defined in dimension tables ( cN_value ).  This efficiently compresses the data and indexes for the fact table,  particularly when the values are highly repetitive. I loaded data into the following schema: create table c1_value (    id             integer primary key auto_increment,    value          text,    key(value(255)) ) type=myisam; ... identical c2_value, c3_value and c4_value tables ... create table quad    id                c1                c2                c3                c4             ) type=myisam; ( integer integer integer integer integer primary key auto_increment, not null, not null, not null, not null Although I later created indexes on c1, c2, c3, and c4 in the quad table,  I left unnecessary indexes off of the tables during the loading process because it’s more efficient to create indexes after loading data in a table.  The keys on the value fields of the dimension tables are important,  because the loading process does frequent queries to see if values already exist in the dimension table.  The sequentially assigned id in the quad field isn’t necessary for many applications,  but it gives each a fact a unique identity and makes the system aware of the order of facts in the dump file. The Loading Process The loading script was written in PHP and used a naive method to build the index incrementally.  In pseudo code it looked something like this: function insert_quad($q1,$q2,$q3,$q4) {     $c1=get_dimension_key(1,$q1);     $c2=get_dimension_key(2,$q2);     $c3=get_dimension_key(3,$q3);     $c4=get_dimension_key(4,$q4);     $conn->insert_row("quad",null,$c1,$c2,$c3,$c4) } function get_dimension_key($index,$value) { $cached_value=check_cache($value); if ($cached_value) return $cached_value; $table="$c{$index}_value"; $row=$conn->fetch_row_by_value($table,$value); if ($row) return $row->id; $conn->insert_row($table,$value); return $conn->last_insert_id };[1/12/2014 8:17:56 PM]
  • 3. Generation 5 » Putting Freebase in a Star Schema Caching frequently used dimension values improves performance by a factor of five or so,  at least in the early stages of loading.  A simple cache management algorithm,  clearing the cache every 500,000 facts,  controls memory use.  Timing data shows that a larger cache or better replacement algorithm would make at most an increment improvement in performance.  (Unless a complete dimension table index can be held in RAM,  in which case all read queries can be eliminated.) I performed two steps after the initial load: 1. Created indexes on quad(c1), quad(c2), quad(c3) and quad(c4) 2. Used myisam table compression to reduce database size and improve performance Loading Performance It took about 140 hours (nearly 6 days) to do the initial load.  Here’s a graph of facts loaded vs elapsed time: The important thing Iabout this graph is that it’s convex upward:  the loading process slows down as the number of facts increases.  The first 50 quads are loaded at a rate of about 6 million per hour;  the last 50 are loaded at a rate of about 1 million per hour.  An explanation of the details of the curve would be complex,  but log N search performance of B-tree indexes and the ability of the database to answer queries out of the computer’s RAM cache would be significant.  Generically,  all databases will perform the same way,  becoming progressively slower as the size of the database increases:  you’ll eventually reach a database size where the time to load the database becomes unacceptable. The process of constructing b-tree indexes on the mysql tables took most of a day.  On average it took about four hours to construct a b-tree index on one column of quad : mysql> create index quad_c4 on quad(c4); Query OK, 243098077 rows affected (3 hours 40 min 50.03 sec) Records: 243098077  Duplicates: 0  Warnings: 0 It took about an hour to compress the tables and rebuild indexes,  at which point the data directory looks like: -rw-r-----rw-r-----rw-r-----rw-r-----rw-r-----rw-r-----rw-r-----rw-r-----rw-r-----rw-r-----rw-r-----rw-r-----rw-r-----rw-rw---- 1 1 1 1 1 1 1 1 1 1 1 1 1 1 mysql mysql mysql mysql mysql mysql mysql mysql mysql mysql mysql mysql mysql mysql root        8588 root   713598307 root   557990912 root        8588 root      485254 root      961536 root        8588 root   472636380 root   370497536 root        8588 root  1365899624 root  1849223168 root          65 mysql       8660 Feb Feb Feb Feb Feb Feb Feb Feb Feb Feb Feb Feb Feb Feb 22 22 24 22 22 24 22 22 24 22 22 24 22 23 18:42 18:48 10:48 18:56 18:46 10:48 18:56 18:51 10:51 18:56 18:44 11:01 18:42 17:16[1/12/2014 8:17:56 PM] c1_value.frm c1_value.MYD c1_value.MYI c2_value.frm c2_value.MYD c2_value.MYI c3_value.frm c3_value.MYD c3_value.MYI c4_value.frm c4_value.MYD c4_value.MYI db.opt quad.frm
  • 4. Generation 5 » Putting Freebase in a Star Schema -rw-rw---- 1 mysql mysql 3378855902 Feb 23 20:08 quad.MYD -rw-rw---- 1 mysql mysql 9927788544 Feb 24 11:42 quad.MYI At this point it’s clear that the indexes are larger than the actual databases:  note that c2_value is much smaller than the other tables because it holds a relatively small number of predicate types: mysql> select count(*) from c2_value; +----------+ | count(*) | +----------+ | 14771 | +----------+ 1 row in set (0.04 sec) mysql> select * from c2_value limit 10; +----+-------------------------------------------------------+ | id | value | +----+-------------------------------------------------------+ | 1 | /type/type/expected_by | | 2 | reverse_of:/community/discussion_thread/topic | | 3 | reverse_of:/freebase/user_profile/watched_discussions | | 4 | reverse_of:/freebase/type_hints/included_types | | 5 | /type/object/name | | 6 | /freebase/documented_object/tip | | 7 | /type/type/default_property | | 8 | /type/type/extends | | 9 | /type/type/domain | | 10 | /type/object/type | +----+-------------------------------------------------------+ 10 rows in set (0.00 sec) The total size of the mysql tablespace comes to about 18GB,  anexpansion of about 40 times relative to the bzip2 compressed dump file. Query Performance After all of this trouble,  how does it perform?  Not too bad if we’re asking a simple question,  such as pulling up the facts associated with a particular object mysql> select * from quad where c1=34493; +---------+-------+------+---------+--------+ | id      | c1    | c2   | c3      | c4     | +---------+-------+------+---------+--------+ | 2125876 | 34493 |   11 |      69 | 148106 | | 2125877 | 34493 |   12 | 1821399 |      1 | | 2125878 | 34493 |   13 | 1176303 | 148107 | | 2125879 | 34493 | 1577 |      69 | 148108 | | 2125880 | 34493 |   13 | 1176301 | 148109 | | 2125881 | 34493 |   10 | 1713782 |      1 | | 2125882 | 34493 |    5 | 1174826 | 148110 | | 2125883 | 34493 | 1369 | 1826183 |      1 | | 2125884 | 34493 | 1578 | 1826184 |      1 | | 2125885 | 34493 |    5 |      66 | 148110 | | 2125886 | 34493 | 1579 | 1826185 |      1 | +---------+-------+------+---------+--------+ 11 rows in set (0.05 sec) Certain sorts of aggregate queries are reasonably efficient,  if you don’t need to do them too often:  we can look up the most common 20 predicates in about a minute: select (select value from c2_value as v where as predicate,count(*) from quad as q group by c2 order by count(*) desc limit 20; +-----------------------------------------+----------+ | predicate                               | count(*) | +-----------------------------------------+----------+ | /type/object/type                       | 27911090 | | /type/type/instance                     | 27911090 | | /type/object/key                        | 23540311 | | /type/object/timestamp                  | 19462011 | | /type/object/creator                    | 19462011 | | /type/permission/controls               | 19462010 | | /type/object/name                       | 14200072 | | master:9202a8c04000641f800000000000012e |  5541319 | | master:9202a8c04000641f800000000000012b |  4732113 | | /music/release/track                    |  4260825 | | reverse_of:/music/release/track         |  4260825 | | /music/track/length                     |  4104120 | | /music/album/track                      |  4056938 | | /music/track/album                      |  4056938 | | /common/document/source_uri             |  3411482 | | /common/topic/article                   |  3369110 | | reverse_of:/common/topic/article        |  3369110 | | /type/content/blob_id                   |  1174046 | | /type/content/media_type                |  1174044 | | reverse_of:/type/content/media_type     |  1174044 | +-----------------------------------------+----------+ 20 rows in set (43.47 sec) You’ve got to be careful how you write your queries:  the above query with the subselect is efficient,  but I found it took 5 hours to run when I joined c2_value with[1/12/2014 8:17:56 PM]
  • 5. Generation 5 » Putting Freebase in a Star Schema quad and grouped on value .  A person who wishes to do frequent aggregate queries would find it most efficient to create a materialized views of the aggregates. Faster And Large It’s obvious that the Jan 2009 Freebase is pretty big to handle with the techniques I’m using.  One thing I’m sure of is that that Freebase will be much bigger next quarter — I’m not going to do it the same way again.  What can I do to speed the process up? Don’t Screw Up This kind of process involves a number of lengthy steps.  Mistakes,  particularly if repeated,  can waste days or weeks.  Although services such as EC2 are a good way to provision servers to do this kind of work,  the use of automation and careful procedures is key to saving time and money. Partition it Remember how the loading rate of a data set decreases as the size of the set increase?  If I could split the data set into 5 partitions of 50 M quads each,  I could increase the loading rate by a factor of 3 or so.  If I can build those 5 partitions in parallel (which is trivial),  I can reduce wallclock time by a factor of 15. Eliminate Random Access I/O This loading process is slow because of the involvement of random access disk I/O.  All of Freebase canbe loaded into mysql with the following statement, LOAD DATA INFILE ‘/tmp/freebase.dat’ INTO TABLE q FIELDS TERMINATED  BY ‘t’; which took me about 40 minutes to run.   Processes that do a “full table scan” on the raw Freebase table with a grep or awk -type pipeline take about 20-30 minutes to complete.  Dimension tables can be built quickly if they can be indexed by a RAM  hasthable.   The process that builds the dimension table can emit a list of key values for the associated quads:  this output can be sequentially merged to produce the fact table. Bottle It Once a data source has been loaded into a database,  a physical copy of the database can be made and copied to another machine.  Copies can be made in the fraction of the time that it takes to construct the database.  A good example is the Amazon EC2 AMI that contains a preinstalled and preloaded Virtuoso database loaded with billions of triples from DBPedia,  MusicBrainz,  NeuroCommons and a number of other databases.  Although the process of creating the image is complex,  a new instance can be provisioned in 1.5 hours at the click of a button. Compress Data Values Unique object identifiers in freebase are coded in an inefficient ASCII representation: mysql> select * from c1_value limit 10; +----+----------------------------------------+ | id | value                                  | +----+----------------------------------------+ |  1 | /guid/9202a8c04000641f800000000000003b | |  2 | /guid/9202a8c04000641f80000000000000ba | |  3 | /guid/9202a8c04000641f8000000000000528 | |  4 | /guid/9202a8c04000641f8000000000000836 | |  5 | /guid/9202a8c04000641f8000000000000df3 | |  6 | /guid/9202a8c04000641f800000000000116f | |  7 | /guid/9202a8c04000641f8000000000001207 | |  8 | /guid/9202a8c04000641f80000000000015f0 | |  9 | /guid/9202a8c04000641f80000000000017dc | | 10 | /guid/9202a8c04000641f80000000000018a2 | +----+----------------------------------------+ 10 rows in set (0.00 sec) These are 38 bytes apiece.  The hexadecimal part of the guid could be represented in 16 bytes in a binary format,  and it appears that about half of the guid is a constant prefix that could be further excised. A similar efficiency can be gained in the construction of in-memory dimension tables: md5 or sha1 hashes could be used as proxies for values. The freebase dump is littered with “reverse_of:” properties which are superfluous if the correct index structures exist to do forward and backward searches. Parallelize it[1/12/2014 8:17:56 PM]
  • 6. Generation 5 » Putting Freebase in a Star Schema Loading can be parallelized in many ways:  for instance,  the four dimension tables can be built in parallel.  Dimension tables can also be built by a sorting process that can be performed on a computer cluster using map/reduce techniques.  A cluster of computers can also store a knowledge base in RAM,  trading sequential disk I/O for communication costs.  Since the availability of data is going to grow faster than the speed of storage systems,  parallelism is going to become essential for handling large knowledge bases — an issue identified by Japanese AI workers in the early 1980′s. Cube it? Some queries  benefit from indexes built on combinations of tables,  such as CREATE INDEX quad_c1_c2 ON quad(c1,c2); there are 40 combinations of columns on which an index could be useful — however,  the cost in time and storage involved in creating those indexes would be excessively expensive.  If such indexes were indeed necessary, a Multidimensional database can create a cube index that is less expensive than a complete set of B-tree indexes. Break it up into separate tables? It might be anathema to many semweb enthusiasts,  but I think that Freebase (and parts of Freebase) could be efficiently mapped to conventional relational tables.  That’s because facts in Freebase are associated with types,  see,  for instance,  Composer from the Music Commons.  It seems reasonable to map types to relational tables and to create satellite tables to represent many-to-many relationships between types.  This scheme would automatically partition Freebase in a reasonable way and provide an efficient representation where many obvious questions (ex. “Find Female Composers Born In 1963 Who Are More Than 65 inches tall”) can be answered with a minimum number of joins. Conclusion Large knowledge bases are becoming available that cover large areas of human concern:  we’re finding many applications for them.  It’s possible to to handle databases such as Freebase and DBpedia on a single computer of moderate size,  however,  the size of generic databases and the hardware to store them on are going to grow larger than the ability of a singler computer to process them.  Fact stores that (i) use efficient data structures,  (ii) take advantage of parallelism,  and (iii) can be tuned to the requirements of particular applications,  are going to be essential for further progress in the Semantic Web. Credits Metaweb Technologies, Freebase Data Dumps, January 13, 2009 Kingsley Idehen,  for several links about RDF store performance. Stewart Butterfield for encyclopedia photo. Paul Houle on February 25th 2009 in Freebase, SQL, Semantic Web Comments (8) Comments (8) Login Sort by: Date Rating Last Activity L.G. · 254 weeks ago 0 Just use Sphinx: Reply 1 reply · active 254 weeks ago paul_houle 18p · 254 weeks ago +1 Sphinx looks like a nice product, but it looks like it addresses a different question: full-text search. Freebase is a semantic system where, instead of using an imprecise word like "jaguar", you can reference "jaguar the cat" or "jaguar the game console" by a guid. That said, full-text search can be a useful complement to this kind of system. Years ago I worked on a project called the Global Performing Arts Database where text about an media item was distributed in literally hundreds of different tables, since the system coded statements like[1/12/2014 8:17:56 PM]
  • 7. Generation 5 » Putting Freebase in a Star Schema "Picture A was taken during a production of Hamlet" "Hamlet was written by Shakespeare" in machine-readable (RDBMS) form. Of course we wanted Picture A to show up in a picture of Shakespeare, so we had to do a graph traversal of the RDBMS tables to collect any text that might be relevant to an item (careful not to follow paths that would lead to irrelevant results.) This way we'd build up a document vector which we'd index in a conventional full text system. Can Sphinx support that kind of thing easily? Reply John Sichi · 254 weeks ago 0 If you're going to the effort to transform to a star, it's worth looking into a column store for the DB as well. Besides query acceleration and automatic compression, you can also get much faster load times using bulk load facilities which avoid row-at-a-time index updates. I won't mention the names of any of the contenders here :) 2 replies · active 254 weeks ago Reply paul_houle 18p · 254 weeks ago +1 Got a specific one in mind? Reply John Sichi · 254 weeks ago 0 Well, yeah, LucidDB (, but since I work on that, I'm biased. :) Last OSCON, I actually loaded up a WEX dump into LucidDB in between sessions, but the bulk of it was semi-/un-structured text, which wasn't very interesting since LucidDB doesn't support lobs and fulltext/XML search yet. Also in the open source camp: Infobright is the lowest learning curve for MySQL users; MonetDB has XQuery support and is very fast as long as everything fits in memory. Reply Martin Gajdos · 254 weeks ago +1 It would be nice if you could post the scripts you have used for this. I'd love to take a look at those. Thanks Reply Josh Ribakoff · 254 weeks ago 0 Nice work, this will be very interesting to anyone coding knowledge based agents for reasoning about the real world Reply madtiger 17p · 252 weeks ago +1 I've been parsing through this data for a few days and I stumbled across this page. I agree, do you have a copy of these scripts, or is it something proprietary to your business? Reply Post a new comment Enter text right here! Comment as a Guest, or login: Name Email[1/12/2014 8:17:56 PM] Website (optional)
  • 8. Generation 5 » Putting Freebase in a Star Schema Displayed next to your comments. Not displayed publicly. If you have a website, link to it here. None Subscribe to None Submit Comment Copyright © 2013 Generation 5. WordPress Theme design.[1/12/2014 8:17:56 PM]