Google shared details Wednesday about Google Big Query, a cloud-based service that promises to bring the search giant's immense compute power and expertise with algorithms to bear on large data sets. The service is still in limited beta preview, but it promises to speed analysis of Google ad data while opening up ways to mash up and analyze huge data sets from external sources.
Google Big Query was described by Ju-Kay Kwek, product manager for Google Cloud Platform Team, as offering an array of SQL and graphical-user-interface-driven SQL analyses of tens of terabytes of data per customer, yet it doesn't require indexing or pre-caching. What's more, customers will get fine-grained analysis of all their data without summaries or aggregations.
"Fine-grained data is the key to the service because we don't know what questions customers are going to ask," said Kwek in an onstage interview at this week's GigaOm Structure Data conference in New York.
Some of Google's beta customers are uploading data to the service with batches and data streams and treating it as a cloud-based data warehouse, but Kwek said ad data would be the first priority, supporting a Google customer's need to understand massive global campaigns running in multiple languages.
"When an advertiser wants to understand the ROI or effectiveness of a keyword campaign running across the globe, that's a big-data problem," Kwek said. "They're currently extracting data using the Adwords API, building shared databases on-premises, doing all the indexing, and sometimes losing track of the questions they wanted to ask by the time they have the data available."
Thus, time to insight will be the biggest benefit of the service, Kwek said, with analyses taking a day or less, rather than days or weeks, when customers face extracting and structuring data on less robust and capable on-premises platforms.
For the full article, please click here.