Apple acquires "dark data" specialist Lattice Data for $200M
In its latest buy, Apple has paid around $200 million to acquire Lattice Data Inc., a specialist in using machine learning to process "dark data" to efficiently build structured data sets that can be analyzed.

The buy was reported by Tech Crunch, and Apple responded with its usual boilerplate confirmation that it "buys smaller technology companies from time to time and we generally do not discuss our purpose or plans."
Lattice, based in Menlo Park, California, was commercializing a Stanford University research project known as "DeepDive," acting as a framework for statistical inference. The firm's website says "our mission is to unlock the value in dark data for critical real-world problems."
"Dark Data" pertains to the mountains of raw information collected by in various ways (such as logs or photos) but which remains difficult to analyze.
Lattice built a platform of proprietary software on top of Stanford's open source DeepDive technology, which had been developed over a six year period with $20 million of DARPA funding, aimed at rapidly make sense out of unstructured data.
DeepDive is aimed at "extracting value" from various "dark data" sources, serving, as the firm's site explains, as a "programming and execution framework for statistical inference, which allows us to solve data cleaning, extraction, and integration problems jointly. We model the known as features and the unknown as random variables connected in a factor graph."
Lattice chief scientist Christopher R won a MacArthur Genius Grantfor his work on DeepDive, and cofounded the firm with Michael Cafarella (who also co-founded Hadoop), Raphael Hoffmann andFeng Niu.
The company contrasts its approach to traditional machine learning, explaining "we do not require laborious manual annotations. Rather, we take advantage of domain knowledge and existing structured data to bootstrap learning via distant supervision. We solve data problems with data."
The firm also emphasizes the "machine scale" of its platform in being able to "push the envelope on machine learning speed and scale," resulting in "systems and applications that involve billions of webpages, thousands of machines, and terabytes of data" at what it describes as "human-level quality."
While Apple doesn't usually comment on why it buys up talent and technology, Lattice was reportedly shopping itself around as a solution to enhancing voice-based assistance, and had been in talks with Amazon's Alexa team as well as Samsung.
The applications that Lattice outlines on its website also suggest potential use in analyzing data for use in Maps and self driving vehicles; HealthKit and ResearchKit; camera logic and processing as well as Internet data document search.

The buy was reported by Tech Crunch, and Apple responded with its usual boilerplate confirmation that it "buys smaller technology companies from time to time and we generally do not discuss our purpose or plans."
Lattice, based in Menlo Park, California, was commercializing a Stanford University research project known as "DeepDive," acting as a framework for statistical inference. The firm's website says "our mission is to unlock the value in dark data for critical real-world problems."
"Dark Data" pertains to the mountains of raw information collected by in various ways (such as logs or photos) but which remains difficult to analyze.
Lattice built a platform of proprietary software on top of Stanford's open source DeepDive technology, which had been developed over a six year period with $20 million of DARPA funding, aimed at rapidly make sense out of unstructured data.
DeepDive is aimed at "extracting value" from various "dark data" sources, serving, as the firm's site explains, as a "programming and execution framework for statistical inference, which allows us to solve data cleaning, extraction, and integration problems jointly. We model the known as features and the unknown as random variables connected in a factor graph."
Lattice chief scientist Christopher R won a MacArthur Genius Grantfor his work on DeepDive, and cofounded the firm with Michael Cafarella (who also co-founded Hadoop), Raphael Hoffmann andFeng Niu.
The company contrasts its approach to traditional machine learning, explaining "we do not require laborious manual annotations. Rather, we take advantage of domain knowledge and existing structured data to bootstrap learning via distant supervision. We solve data problems with data."
The firm also emphasizes the "machine scale" of its platform in being able to "push the envelope on machine learning speed and scale," resulting in "systems and applications that involve billions of webpages, thousands of machines, and terabytes of data" at what it describes as "human-level quality."
While Apple doesn't usually comment on why it buys up talent and technology, Lattice was reportedly shopping itself around as a solution to enhancing voice-based assistance, and had been in talks with Amazon's Alexa team as well as Samsung.
The applications that Lattice outlines on its website also suggest potential use in analyzing data for use in Maps and self driving vehicles; HealthKit and ResearchKit; camera logic and processing as well as Internet data document search.
Comments
Dark Data though? Isn't that kind of racist?
Lots of what's counted as business analytics these days would have been considered "dark" decades ago.
AI can find correlations, cause and effects, in groups, or solo data to other groups or solo bits of data that would be very hard to surmize from just looking at it (probable that the effort just to find what to look for would be too large to even be profitable before).
Intent is what is important.
Can you think of any terms that weren't considered derogatory by society when you were a kid that are now demonized? I can think of plenty. It also works the other way, but to a lesser extent in the short term. For example, calling someone a scalawag isn't likely to offend anyone today, but not that long ago in American history that was one of the most offensive terms you could call certain groups.
We're wired poorly to accept things as being absolute when it's, at best, based on fundamental limitations of our species, or, at worst, actually random, but we all still perceive things about language that aren't actually real. Ideasthesia covers this phemonomen.
One popular psychological example are the names of these shapes:
Lots of sales related data is stored during the day, but not much of it is analysed in different contexts. A retail chain in the U.K. spent a lot money on one of these analysis package, and as a result the began to move items around the store depending on the time of day. For example, they moved beer and nappies to the front of the shop after 8pm and increased sales on both items.
These packages can can also show historical data across many regions that they can also link to events that may be missed by the marketing analysts.
Analyst: I need to know more about the sales spikes in condoms and toilet paper we are seeing in different regions at different times of the year. What's the connection?
Software: Music festivals.
improve Siri and search in general by deep diving the users data into a private encrypted database. Similar to photos image recognition system but using iCloud syncing to open that data and more across devices and app silos.
All the processing is done on the device (which is why the photo library on my phone is out of step with the photo library on my iPad), but if Apple doesn't want to put this data in the cloud, then what happens to all this learned stuff when I get a new iPhone? Does it have to learn it all again?
I assumed it was due to new file system coming soon. As in Multi-key encryption would allow them share this sort of data and still cut off devices as needed for security reason.