When I got it to work this morning, I was greeted by an email from 23andMe’s PR company, saying they have “built one of the world’s largest databases of individual genetic information.” Normally, I wouldn’t even bat an eye at a claim like that. I’m pretty sure it is a big database of variation… but I thought I should throw down the gauntlet and give 23andMe a run for their money. (-:
The timing for it couldn’t be better for me. My own database actually ran out of auto-increment IDs this week, as we surpassed 2^31 snps entered into the db and had to upgrade the key field to bigint from int. (Some variant calls have been deleted and replaced as variant callers have improved, so we actually have only 1.2 Billion variations recorded against the hg18 version of the human genome. A few hundred million more than that for hg19.) So, I thought I might have a bit of a claim to having one of the largest databases of human variation as well. Of course, comparing databases really is dependent on the metric being used, but hey, there’s some academic value in trying anyhow.
In the first corner, my database stores information from 2200+ samples (cancer and non-cancer tissue), genome wide (or transcriptome wide, depending on the source of the information.), giving us a wide sampling of data, including variations unique to individuals, as well as common polymorphisms. In the other corner, 23andMe has sampled a much greater number of individuals (100,000) using a SNP chip, meaning that they’re only able to sample a small amount of the variation in an individual – about 1/3rd of a single percent of the total amount of DNA in each individual.
(According to this page, they look at only 1 million possible SNPs, instead of the 3 Billion bases at which single nucleotide variations can be found – although arguments can be made about the importance of that specific fraction of a percent.)
The nature of the data being stored is pretty important, however. For many studies, the number of people sampled has a greater impact on the statistics than the number of sites studied and, since those are mainly the ones 23andMe are doing, clearly their database is more useful in that regard. In contrast, my database stores data from both cancer and non-cancer samples, which allows us to make sense of variations observed in specific types of cancers – and because cancer derived variations are less predictable (ie, not in the same 1M snps each time) than the run-of-the-mill-standard-human-variation-type snps, the same technology 23andMe used would have been entirely inappropriate for the cancer research we do.
Unfortunately, that means comparing the two databases is completely impossible – they have different purposes, different data and probably different designs. They have a database of 100k individuals, covering 1 million sites, whereas my database has 2k individuals, covering closer to 3 billion base pairs. So yeah, apples and oranges.
(In practice, however, we don’t see variations at all 3 Billion base pairs, so that metric is somewhat skewed itself. The number is closer to 100 Million bp – a fraction of the genome nearly 100 times larger than what 23andMe is actually sampling.)
But, I’d still be interested in knowing the absolute number of variations they’ve observed… a great prize upon which we could hold this epic battle of “largest database of human variations.” At best, 23andMe’s database holds 10^11 variations, (1×10^6 SNPs x 1×10^5 people), if every single variant was found in every single person – a rather unlikely case. With my database currently at 1.2×10^9 variations, I think we’ve got some pretty even odds here.
Really, despite the joking about comparing database sizes, the real deal would be the fantastic opportunity to learn something interesting by merging the two databases, which could teach use something both about cancer and about the frequencies of variations in the human population.
Alas, that is pretty much certain to never happen. I doubt 23andMe will make their database public – and our organization never will either. Beyond the ethical issues of making that type of information public, there are pretty good reasons why this data can only be shared with collaborators – and in measured doses at that. That’s another topic for another day, which I won’t go into here.
For now, 23andMe and I will just have to settle for both having “one of the world’s largest databases of individual genetic information.” The battle royale for the title will have to wait for another day… and who knows what other behemoths are lurking in other research labs around the world.
On the other hand, the irony of a graduate student challenging 23andMe for the title of largest database of human variation really does make my day. (=
[Note: I should mention that when I say that I have a database of human variation, the database was my creation but the data belongs to the Genome Sciences Centre – and credit should be given to all of those who did the biology and bench work, performed the sequencing, ran the bioinformatics pipelines and assisted in populating the database.]