My understanding of ML is that you need a huge volume of data to succeed. While each individual genome contains 25,000 genes, the actual stack of decoded Neanderthal genomes is pretty small. So, the data seems to have a large surface area but is shallow. I wonder if the former allows them to get around the latter in ML?