Last ned artikkelen som pdf her
Big things can be intimidating. Research cannot allow Big Data to be one of them.
We stand on the edge of the most exciting and transformative period in our industry’s history: 90 percent of the data in the world today was created in the last two years and “data taps” such as mobile, social and POS will continue to pour out raw information for us to work with at an ever faster rate.
However, if our response to the new era of data is to retreat behind number-crunching technologies, then clients and indeed humanity as a whole, will be much the worse for it.
It may be tempting to conclude that human intuition must surely give way to computers and algorithms when it comes to keeping up with Big Data. But now, more than ever, we need to recognise the immense, unique power of our own minds when it comes to dealing with information – and deciding how to act on the basis of it.

So what do we mean by “Big” exactly?
Big Data wouldn’t be half as intimidating if it were just a question of having more numbers to deal with. But Big Data is bigger than that. It represents the coming together of several different themes, each of which would be fairly paradigm-shifting in its own right.
First of course, is the sheer scale of the data now being produced and stored. Walmart currently handles more than 1 million customer transactions every hour, in databases estimated to contain more than 2.5 petabytes. Such an organisation may soon have created more data every hour than research surveys have ever delivered. With data storage doubling every year, there appears no constraint on the amount of information that we are dealing with.
Connected to the size of data but equally significant is the fact that it now generates itself. Data no longer needs to be created through a questionnaire carefully crafted by a researcher, or painstakingly collected by a field agent; it is created and stored simply by virtue of things happening. It’s broken free of human control – and therefore isn’t limited as to how big it can get and how fast it comes at us. Data’s Velocity, the speed at which huge volumes of it can be generated, is every bit as breathtaking as its sheer size. And the speed with which it is available raises the opportunity and the demand to work with it in real-time.
Yet perhaps the most challenging shift of all is that this size and speed is combined with an explosion in variety of data forms. Big Data comes in all shapes and sizes. Researchers are leaping on new types of data source – and new types of source are leaping on us: from mobile activity to Twitter feeds, geo-location information, facial expression capture and much more. We are quickly moving from dealing in numerical scores to dealing in shapes, movement patterns, expressions – and human language. And such data does not come readily packaged for analysis; using it must involve translating it as well.
You created it: you deal with it
Faced with such challenges, it’s tempting to believe that computational power, which has taken the lead in creating this new world of information, must also take the lead in defining how we deal with it. In this view of the world, the researcher starts to look less like a person, more like supercomputer in a bunker: one where we simply have to feed in the right question or combination of questions, plug it into the river of Big Data – and wait for the answer to pop out. But there are significant dangers to this approach. If Big Data ends up becoming processed and commoditised data, then we are all in trouble.
Digesting really raw data
It’s a mistake to believe that data can ever speak for itself. Data always speaks with a human voice; it can’t say anything otherwise. Every statistic that we deal with is the result of subjective judgement about the problems that we should try to solve, what we think the answers should look like, and what data forms we can enlist to help provide those answers. And these judgements are human ones.
In the Big Data era, the human imagination continues to play an essential role in envisaging what our many different data sources can be made to do, and in aggregating, translating and coding them to enable them to do it. To take a very simple example, Google can predict a flu epidemic by spotting spikes in searches on cold and flu remedies. This is a tremendously cool thing, but it only works because somebody realised that this pattern is significant – and that it correlates to something meaningful and useful. Similarly, micro-location data gives TNS a powerful new tool for mapping movement around stores – but it is only powerful because we have established an understanding of what these movements mean.
In his book The Signal and The Noise, US election poll guru Nate Silver devotes a chapter to global warming and the fact that it would be impossible to find any evidence of this in the notoriously unstable climate record, were scientists not armed with a theory telling them exactly what to look for – and which data to prioritise. It’s an important reminder that...

From data creators to data curators
In the old days (of six months ago), the raw numbers that we sat down to analyse weren’t really raw at all; they were shaped by human hands even before they came into existence. The art of designing a questionnaire involves finely balanced judgements on which questions to ask and how to ask them. Whether to score preferences out of five, seven or ten can trigger some pretty serious debates with good reason – these things have a big influence on our ability to spot patterns, make connections and provide meaningful insight. And judging how to ask questions should, of course, be closely related to the challenge of what you are looking for.
In the Big Data era, we are no longer data creators, designing the structure of information from the outset; instead we are data curators, working with information that has been generated independently. As such, we will face many new challenges and require many new skillsets. However, as we evolve the role of research, we must continue to apply the same standards to independently generated Big Data that we would if we had created it ourselves. And this will require leveraging much hard-won experience about how data works. The skills that once went into the design of research instruments such as questionnaires will remain crucially important in aggregating and selecting data sources, and deciding exactly how they relate to one another. For now, this might involve incremental improvements such as linking spend and retention data to customer experience surveys, as we already do at TNS. In the future, we will find more and more scenarios where the data we aggregate does not include traditional surveys at all. In all of these contexts, it’s not just a question of being excited about what data can do. It’s equally important sometimes to step back, look at how complete and representative a given set of data is, and ask ourselves rigorous questions about what questions it is really qualified to answer.
The continuing evolution of analytics
At TNS, we’ve already evolved from the era of ad-hoc analysis, when researchers collected data with little reference to how it would eventually be used (and then looked through it in the hope it would reveal something useful). Today the design of the instruments for a particular piece of research is informed from the start by the challenge of how best to answer business questions.
The conceptual framework that we use for any type of analysis reflects how the human brain naturally makes sense of information. This framework consists of four different ways of looking at any set of data, whether it was generated through research or arrived, Big Data-style from independent sources. “Dimensions” and “Landscape” address the structure of information; the first seeking out common themes across a data set (the key themes defining a product category, for example), the second looking more closely at competitive relationships, owned and disputed territory and areas of opportunity. We then build on this structural understanding with more action-oriented means of addressing the data: “Groupings” to segment the subject matter and “Drivers” to reveal the variables that influence relevant results, including causal connections that can be far from immediately apparent.
This approach may be structured, but it retains grounds for flexibility. It provides a checklist or where and how to look for patterns and themes. In the Big Data era, we will learn to look for different types of patterns in vastly diverse forms of data, but human reason remains the key driving force in identifying them and drawing purposeful connections between them.
Computational muscle can give research the scale and speed that we will increasingly require in the Big Data era, but it is important to distinguish between automating processes and expecting machines to design them in the first place. We must not fool ourselves that Artificial Intelligence (AI) is ready to take on the task of formulating questions and crafting the algorithms to answer them. After all, even those that welcome the concept of a technological singularity in which human-designed AI surpasses that of humans themselves, don’t envisage it happening until at least 2045. That’s a long time to wait to take real advantage of Big Data.
Data and the human imagination
Imposing structure on Big Data will throw up some intriguing challenges – and these challenges will involve logical leaps and lateral thinking for which the human brain remains our best available tool. What is a meaningful means of scoring a positive tweet or Facebook rant? What aspect of somebody’s location is actually relevant to the client brief – and what other sources of information can be integrated or overlaid to give context to this information? The location of a car by itself is meaningless. If it’s a car unable to fit into the WalMart parking lot on Black Friday, it becomes a whole lot more interesting.
When we talk about deploying computational power in the Big Data era, we must therefore be pretty clear about what we are asking computers to do. We must continue to exercise our judgement as to which information is valid and valuable, and how its many varied forms can be coded in meaningful ways. As data curators, that’s our job. But by unleashing the power of today’s machines we can dramatically increase the scope of data that we can use, the range of questions that we can ask, and the speed with which we can answer them. Big Data can unleash the potential of human insight and human reason in ways never envisaged before.
The greater computational power that will enable us to make the most of Big Data must be harnessed to an expanded role for the human mind. Depending too much on non- human processing power creates two potential dangers: that we define in advance what it must look for and how it must look for it, leading to standardisation and blinkered, undifferentiated thinking, and that we confuse correlation with causation, failing to exercise human judgement about which results are meaningful and which are not. The challenges of the Big Data era will be challenges for the human imagination and human judgement as much as for IT infrastructure. We need to welcome them as such.
About Opinion Leaders
Opinion Leaders are a regular series of articles from TNS consultants, based on their expertise gathered through working on client assignments in over 80 markets globally, with additional insights gained through
TNS proprietary studies such as Digital Life, Mobile Life and the Commitment Economy.
About TNS
TNS advises clients on specific growth strategies around new market entry, innovation, brand switching and stakeholder management, based on long-established expertise and market-leading solutions. With a presence in over 80 countries, TNS has more conversations with the world’s consumers than anyone else and understands individual human behaviours and attitudes across every cultural, economic and political region of the world. TNS is part of Kantar, one of the world’s largest insight, information and consultancy groups.
Please visit www.tnsglobal.com for more information.
Get in touch
If you would like to talk to us about anything you have read in this report, please get in touch via info@tns-gallup.no or via Twitter @TNSGallupNorway
References
1. Nate Silver, The Signal and the Noise
2. Daniel-Kahneman, Thinking Fast and Slow