So where we have understood the overall innovation landscape and slowly increased our factor of magnifcation over the past five weeks, this week we continue in our highest magnification, but now we begin experimenting.
If we have created an innovation using the tools covered thus far, we have hopefully identified something which is in a fruitful vein, fits organizational strengths, and appears to have at least some potential to create value and impact. So we're off to a good start.
This week, we add to the toolset means by which we can quickly understand if we have created something with promise. The intent is to understand -- in a concise way and with minimal relative expense -- if the concept has 'legs.' Are we seeing the kind of spark of innovation in the market that we would expect? Are there weaknesses? Are there other veins of richness being identified in research which may be of even greater value than the initial concept?
We are now entering the phase of development where the concept is initially revealed to some sample of the public, and an initial view of what value the concept actually delivers into the market.
If we have done our due dilligence up to this point and indeed ended up inventing the next 'Corgi saddle,' now is exactly when we need to find out: before we have spend considerable time, effort, and resources on bringing the concept to fruition.
By the end of this lesson, you should be able to:
To Read | Chapters 11 and 12 (Keeley, et al.) Documents and assets as noted/linked in the Lesson (optional) |
---|---|
To Do |
Case Assignment: Structuring research
|
If you have any questions, please send them to my axj153@psu.edu [1] Faculty email. I will check daily to respond. If your question is one that is relevant to the entire class, I may respond to the entire class rather than individually.
Allow me to tell you a tale, one which has been disguised to protect the innocent.
I had the occasion to be a party to some research on "highly designed" vitamin bottles some years ago.
In this case, these vitamin bottles were not designed by some local or regional design firm, but were four designs crafted by literally a rock star of the design world. He has been on covers of major magazines, TED Talks, the whole bit. Let's call him Jacques.
These prototype bottles were 3D printed–before 3D printing was a thing–and were delivered to the test site, by hand, in four separate aluminum briefcases. They were unveiled with reverence usually reserved for ancient artifacts and the like, as they were extracted from their custom cut foam liners with white cotton-gloved hands. The cost just to build the prototypes could have bought a decent house at the time. (This is not an exaggeration, as I had momentary thoughts of absconding to Mexico with four aluminum briefcases and their precious cargos.)
They were indeed "highly designed."
So, these bottles were then sent through usability testing with the exact types of customers who would be using these "highly designed" bottles in the real world. People poked and prodded at them, which is terrifying if you are responsible for the prototypes' well-being (they indeed had a full-time handler), and rather entertaining as the observer.
There was one design called "Oval," which was a short, squat bottle with what looked like a blunted and flared 3" diameter "crown" on top. If you twisted the crown, a little port would raise from the center of the top of the crown, and this little port had a slot exactly the width of a vitamin, plus 1 mm. If I were to estimate the production cost of this tiny mechanical marvel, it would have been in the $15 range. Just for the bottle. By general retail packaging cost rule of thumb, that would have made a garden-variety bottle of drug-store vitamins around $150. (By the way, the bottle was single-use).
A funny thing happened as person after person used the vitamin bottles. Each thought they were on a hidden camera show.
Here's why: The design would dispense 10-15 vitamins at a time through that tiny slot as if they were shot out of a Lilliputian cannon. Many times they landed loudly on the glass tabletop as they were ejected.
From here, people did what they would naturally do, which is try to put the aforementioned vitamins back into the bottle. With any conventional bottle, this entails the usual 'hand cup and shake' maneuver. Not so with this "highly designed" bottle.
In this case, the participants would place the pills carefully back onto the tiny crown, daintily pushing them round and round in the hopes that one would fit into the tiny slot and drop back into the bottle. What was hilarious was that, for older participants, this typically entailed holding the bottle about 5 inches from one's eyes, and poking with all the gentle intent of trying to make a ladybug walk more quickly down a set of tiny stairs.
So, one might think that the design house would have received the message that the prototype *might* need a little refinement for usability.
Some weeks later, the research findings were presented, with Jacques in attendance. The researchers brought video and verbatim quotes of the encounters, as well as offering their own ratings, based on the participants' encounters, on a variety of facets. Without saying as much, it was clear the designs, in the presented iterations, scored between a C and an F- in the eyes of the participants. (The scores were actually presented on a soft scale, using descriptors instead of letter or percentage grades.)
Remember that by "participant" we are referring to a significant cross-section of the people who could conceivably be in the sweet spot to purchase this bottle. If you are prone to capitalistic dreaming, replace "participant" with "wallet with feet."
Jacques grew more and more agitated as the researchers presented, until, about 10 minutes in, he leapt up screaming, "Who are you? Who are you? You have no idea what you are talking about! You have no idea about design! These people (pointing at participant video frozen on the screen) are idiots! You chose them to insult me! I will not have this!" With this, he stormed out of the room, and his Senior Handler scurried after him. This left Junior Handler 1 and Junior Handler 2 in the room, who looked helpless for a couple minutes as the presentation continued, until they both decided to leave as well. Jacques was never heard from again.
Why do I tell you this story? Because in developing new offerings, there is a Jacques in all of us.
There is a natural tendency for us to protect what we create, and this is not a tendency that ever serves us well in innovation.
As we begin to test early-phase innovations, we indeed will have at least three personas or roles at play within ourselves, namely, the researcher, the creator, and the innovator. It is a conflict that really plays out no differently within each of us as it does between people, but when it happens internally, it is quiet and especially dangerous.
Here's what can happen, and we see it time and time again:
There is also a pervasive belief in the "creative genius" archetype, one that tells us anyone who is truly "creative" should be able to lock themselves in a room and just create. This is what we pay creative or innovative people to do, correct? To work in "genius isolation" and rely on only their flawed personal perspectives to create offerings which will be sold to hundreds of thousands, if not millions of individuals? We should believe in creative genius, but great research can make average people perform like creative geniuses.
I would suggest a shift in mindset as we embark into the joint acts of creating and trying to find flaws in our creations.
The following are a few mantras I have relied upon over the years and tend to come back to when in the thick of research and gathering insights.
There was a famous automotive engineer and designer named Carroll Smith, whose fastidious attention to detail has won his teams major victories at virtually every level. I do not use the term "fastidious attention to detail" lightly: he has written a 224-page book devoted solely to nut and bolt selection [9].
There is one phrase he is most famous for: "There is no such thing as material failure. All failures are human in origin." He argues that the role of the engineer is to account for everything from metal porosity to poor machining to the tendencies of the driver, and if a component fails in a crucial moment, the responsibility lies on a human–not an inert material.
While this is certainly a compelling platform for engineering, his statement may be adapted and applied to consumer research, as well. Research methodologies do not fail, research design is typically where the failure lies... and humans are responsible for this. As we will see, each methodology has strengths and weaknesses, and it is our responsibility to not only account for those, but to be smart about research design.
How do we tend to fail in consumer research design?
Note that none of the above are methodological failures, but failures on the part of the researcher. While professionals and professors alike devote their lives to the art and science of research design, in our case, we will be looking to perform quick–but "clean"–research which can be built upon. In essence, we will seek to perform research which will not be thrown away after a month, but which may be built upon to create a research narrative which accompanies our revisions to the offering over time.
What is freeing about research design is that if it is sound and done well, the results are the results. No explanation is needed for negative outcomes, no congratulations accepted for positive outcomes.
Our role is to think, structure, execute, and learn as much as possible. Only after, may we interpret results.
In the following pages, we will examine a few methodologies and techniques especially well-suited to testing early-phase ideas, messages, and offerings.
When discussing the realities and underpinnings of these research techniques, we will consider how each fits into a few different spectrums:
I will also share some experiences with the methodologies and applications of technologies you may find interesting. Some methodologies may be timeless, but there are ways you can deploy the methodologies which can also produce dividends for the research and generally make things more efficient.
Surveys are the most used, and unfortunately, the most abused research methodology. It is the least common denominator, in that everyone from 5th graders to college professors can use it, but it takes a significant amount of thought to structure a valid, scalable survey.
There was a time in the lifespan of this classic technique when participants felt "special" when asked to complete a survey that their opinion mattered. JD Power and Nielsen were masters of this, and Nielsen became famous for its technique of enclosing a crisp dollar bill in every survey mailed to create a sense of obligation on the part of the recipient.
With the advent of the internet, online surveys have become incredibly pervasive in our daily lives. To understand just how pervasive surveying is, I decided to count how many surveys I was presented with while going about my usual business on a Saturday. The final count? Fifteen, including three on shopping receipts. (I was fortunate in that CVS was an early errand that day, so I had an ample 24" receipt/scroll upon which to record the day's findings.)
Consider also that this survey overload is a reinforcing cycle: researchers receive fewer responses to surveys, so require higher number of survey sends (impressions) to reach the level of validity they require. Now multiply this escalation in survey sends by thousands upon thousands of companies, researchers, and media, all of which are in a similar situation. In essence, we have our own miniature Tragedy of The Commons in survey research, in that it is a race to deplete what is an openly-accessible, but finite resource: attention.
From USA Today, January 7, 2012, "For some consumers, surveys breed feedback fatigue [12]":
"Survey fatigue" has long been a concern among pollsters. Some social scientists fear a pushback on feedback could hamper important government data-gathering, as for the census or unemployment statistics.
If more people say no to those, "the data, possibly, become less trustworthy," said Judith Tanur, a retired Stony Brook University sociology professor specializing in survey methodology.
Response rates have been sinking fast in traditional public-opinion phone polls, including political ones, said Scott Keeter, the Pew Research Center's survey director and the president of the American Association for Public Opinion Research. Pew's response rates have fallen from about 36 percent in 1997 to 11 percent last year, he said. The rate includes households that weren't reachable, as well as those that said no.
The Associated Press conducts regular public opinion polling around the world and has seen similar trends in response rates. There's little consensus among researchers on whether lower response rates, in themselves, make results less reliable.
Keeter attributes the decline more to privacy concerns and an ever-busier population than to survey fatigue. But the flurry of customer-feedback requests "undoubtedly contributes to people putting up their guard," he said.
This has become a very real problem in consumer research, so we must get creative in how we first contact participants, continue the research relationship, and how we survey in the first place. We will discuss this a bit more in a moment, but in some cases, it can be helpful to take non-traditional approaches to find, approach, and continue relationships with your participant pool.
In regard to pure survey design, Duke University has an excellent set of condensed Tipsheets to help you create well-structured surveys. They have examples throughout to help explain not only the problem and solution, but also what the question looks like in practice.
These Tipsheets cover the major points you need to pay attention to in any survey design, and are a great resource to bookmark. If you happen to have staff helping to write or deploy surveys, it can be useful to point them to the Duke Tipsheets to not only help them do a better job in adding questions to the survey, but also in reviewing the other questions.
Beyond the prescriptions of the Duke Tipsheets, there are quite a few common survey mistakes that happen in the field, and anyone can have significantly negative outcomes for the survey results and completion. These mistakes are not limited to just the less experienced, they happen all the time to experienced research designers.
What can make mistakes especially damaging is if they happen early in a series of surveys. So, for example, let's imagine you seek to create a "survey narrative" over time, building a robust set of baseline data on a core set of questions. If you commit an especially egregious mistake, such as omitting what would be a very common response to one of the questions, it not only distorts the results of that survey, but every survey that contains that question. You may have built 3 years of baseline data, but the results may be in question because of one carried miscue from the first survey.
In a short-attention, low-engagement context, lengthy surveys are a liability. Participants will start to offer superficial, single word, or repetitive answers, or might abandon the survey altogether. Especially in the case of research pools you intend to survey over time, abandons can reduce the chance that your participants will opt in to the next round of survey. Furthermore, long surveys can skew results invisibly through attrition of participants, as we likely do not desire responses from only those who can devote 50 uninterrupted minutes in the middle of a workday.
Solutions: If you have a survey over 10 minutes, strongly consider carrying content to another survey or otherwise splitting the survey. Regardless of timing of the survey, one of the first things to do before the participant begins is to be clear about average completion time (i.e. "Average time required: 7 mins"). This creates not only an expectation on the part of the participant, but a type of implied contract. In the end, it is also common courtesy and good research practice. If you have stimuli integrated into the survey, such as animatics, videos, or prototypes, you can stretch the 10 minute boundary, but you still need to be conscious of (actual) time to complete.
Quite a few of the online survey tools have a feature you can enable to chart progress (as a percentage or bar chart) at the top of the page. This can allow someone to know where there are in the survey, generally allows better pacing, and can reduce opt-outs.
You may also have success gaining participants by using especially short survey length as a selling point. For example, a 'one-minute survey,' or even a 'one question survey.' There are a few mechanisms for continuance with a willing participant you can then use, such as an opt-in for future surveys in the future.
Multiple clause questions with multiple modifiers are confusing to participants, and can lead to dirty data without you ever knowing it. There is no mechanism for you to know your results were skewed.
Solutions: Even in highly-educated participant pools, simplicity and clarity are paramount. Chances are, multiple clause questions can be worded more simply, and if they can't be, you will need to split the question. Save the lawyerly questions for the courtroom, counselor.
It can also be a great practice for larger surveys to include a simple mechanism to allow the participant to "flag" questions. You can add this as a cell at the bottom of the survey page, and some survey software allows it as a pop-in tab from the side of the survey page. Either way, the goal is for a participant to let you know they found the question confusing and to perhaps put a sentence of why.
This is a classic issue that often goes unnoticed or unchallenged. These are surveys wherein each question is voiced in the third person, and using impersonal language and referring to "one" or "it" question constructions. For example: "It has been argued that one could tie a shoe with one hand. Agree or disagree." These types of constructions add a layer of interpretation for the participant, and may move their answers from their personal thoughts and feelings and into the hypothetical. For example, for the shoe tying, am I being asked if I can tie a shoe with one hand, if someone else conceivably could, of if I am aware of the argument.
Solutions: Be direct and personal. Your participants' responses are only valid in reference to themselves, not the thoughts of feelings of others or hypothetical "ones" hovering somewhere in the ether. Replace "one" and "it" with "you." If the more personal approach to these questions seems casual, it's because it is. If you prefer your surveys to sound clinical, yet be unclear and ineffective, that is entirely your prerogative.
This tends to be more of an overall issue, and can range from typos and awkward questions to a lack of logical flow through the survey.
Solutions: Use a group of peers or "pre-deploy" your survey to a small group of live participants and contact them immediately after. This is especially easy to do on the web, as you can watch survey completions come in as they happen, and call the participant to ask if all of the questions made sense and if there are any suggestions they would have. The goal here is to take just a handful of live, unprompted survey participants and intercept them immediately after taking the survey.
Surveys should feel like a good interview for participants, having a logical flow and seamlessly working from one question to the next. Choppy surveys tend to feel like being interviewed by a 4th grader: The questions in isolation may be valid, but it feels like you are being barraged with unlinked questions flowing from a stream of consciousness. This seems to be a more stylistic concern, but it can indeed confuse and detach participants from the survey.
Solutions: Don't be afraid to use "chapters" or breaker pages to allow your participants a bit of a break and to shift gears. So, if I were transitioning from a series of questions about demographics and into asking about experiences with the product category, I would insert a blank breaker page, and note something like, "Your Experiences With [Category]: In the following section of this survey, we would like to understand your thoughts, feelings and experiences with [category]. By [category], we are referring to products that [definition of the category to make sure all are on the same page]. It can be helpful to take a moment to think of specific times you have used (category) in the past to help you remember. Please restrict your answers to your experiences, and not the experiences of others."
Generally, the goal is to start with the more simple and straightforward background questions, and build on the questions until the most complex or emotionally charged are at about the 80% completion mark. The last 20% of the survey tends to act as a "cool-down" and reflection, capturing any closing remarks, feedback, or narrative responses. As a rule of thumb, if there is a chronological flow to the actual experience being examined (i.e., first impressions, use, disposal), the survey should mirror it.
There are few faster ways to lose a well-intentioned participant than to run them through two or three questions that do not allow them to answer as they intend. Essentially, the participant realizes that they will not be able to express their thoughts, and there is therefore no reason to complete the survey. They abandon the survey.
Solutions: The simplest way to resolve this problem is simple: always include open-ended "Other" as a selection. Not only will it allow you to capture the responses, but "Other" responses can be a hotbed for new thinking and unexpected answers.
Having an internal review and a limited "pre-deploy" of the survey will also help you avoid many incomplete answers.
Having thirty Agree/Disagree questions in a row is not a terribly engaging survey for participants.
Solutions: Before writing any questions, list the topics you seek to address in your survey. It may help to arrange them in outline format to create chapters, but the overall goal is to avoid general line-listing of topics that breed uninspired questions. While you do not need to balance the types of questions in your survey, it can be helpful to give it a read-through with an eye toward the question type to make sure you have some variation.
Having four of five questions with the scale arranged left to right, and the fifth with the scale reversed can lead to some unintended answers and the dirty data that comes with this confusion or omission.
Solutions: While it can occasionally be useful to add a "check question" to make sure that a participant is not running through the survey and clicking in the same place every time, varying question types does the same function. Having thirty Agree/Disagree questions in a row necessitates a check question or an axis reversal, but you shouldn't have thirty of the same question type in a row to begin with. One way or another, you shouldn't need to do axis reversals.
Too often, expediency in calculation overrules the thoroughness of results. People tend to lean toward questions that can be calculated and tallied for this reason, or that they do not know how to treat or score open-ended questions.
Solutions: Although there are a wide variety of ways you can compile and summarize open-ended responses, remember that just because you capture data does not mean that you have to instantly undertake calculations and manipulations. So, for example, if you have twenty questions ready to go and five open-ended questions you aren't sure how you're going to score, deploy the survey. As long as you are able to capture the open-ended responses, that information does not have a shelf life.
The researcher does not include filter questions and decision points in the survey (i.e., asking a group different questions based on their response to an earlier question). For this reason, the questions are either inaccurate for a large proportion of participants, or the logic and structure of the overly-broad questions are so convoluted it becomes difficult to read. Either way, the end result is not good.
Solutions: I would argue that one of the most useful functional benefits of online survey tools is the ability to introduce filter questions to create a branched survey. Use it. Not having branches and filter questions when they are needed is usually a sign of a poorly executed survey, an inattentive researcher, or generally sloppy research by thinking that you can send everyone every question, and they will answer them all.
If you want to check for the need for filters and branches in your survey, try taking the survey acting as a member of any different groups of participants. If the survey is all about understanding your thoughts about a product trial, and you have the potential that a handful of people receiving the survey have not yet used the product, include a filter question to separate that group and ask them questions specific to their experience, why they haven't used the product yet, etc.
When surveys are related to an experience the participant had with a prototype, for example. And there is a two-month lag because the researcher waited until everyone in the beta group had received their prototype. The participant has had the prototype on the shelf for one and a half of those two months and has forgotten all of their first impressions and reactions.
Solutions: Lay out the project timing before you begin. If there are going to be any significant lag times between the experience you seek to understand and survey deployment, either split the beta groups and survey them separately, or simply include the survey file with the prototype itself. I tend to be a heavy proponent of including the survey right along with the product/offering being tested, as it tends to make participants more attentive to the experience and remind them to think and record their thoughts and feelings.
It is also possible to send a survey too soon in regard to understanding experiences, as you can send the survey before the participant has even received the product.
While the best practices for survey design and question structure are the same regardless of how you deploy the survey, there are a few different approaches we may use to fit our research needs. Some offer almost instantaneous results by using massive pools of participants, while others allow you the flexibility to deploy surveys very quickly and effectively. For the sake of this discussion, I am going to assume we are already aware of our ability to survey via hard copy, phone, and other conventional means, especially for "nearfield" groups like customers.
In testing spaces and potential offerings, these tools can be used to deploy anything from surveys to a closed beta group of customers, to conducting wide-open public polling to gauge how many are familiar with a topic.
For example, I work with a charity benefiting children with Neurofibromatosis Type 1 (NF1), of which there are as many sufferers as multiple sclerosis worldwide (about 3MM). I wanted a quick gauge of awareness to test a hypothesis as I was setting up a messaging and branding platform, so I did a quick online awareness test with 500 American adults. The findings were that MS awareness was around 94% in American adults, while NF1 was less than 3%. Should I have needed ironclad results, I could have taken the next steps, but for my purposes, a sample of 500 was ample for my needs. Best of all, I had the responses within 30 minutes.
Here are a few tools and where they tend to fall on the survey spectrum:
Especially in the case of surveys, we want to not think of "a point in time" or a "snapshot," we want to think of a continuous line of research that needs to tell a story. If we want to understand shifts in perception over time in a closed group, for example, we need to pay attention to details like making sure we ask the same questions in the same order to help make sure our narrative is not skewed or interrupted.
The "snapshot" frame for research tends to lead to fractured efforts without an overarching structure or goal, and online surveys especially can worsen this condition with their instant feedback.
It may seem a bit esoteric at this point, but we will work on it in the Case this week a bit, as it is important to be able to build the entire narrative and understanding of the offering or topic. Isolated, unlinked efforts can be more confusing or distracting than they are worth.
Creating an early prototype of an offering or a concept to be tested is hard enough, as there are many variables at play. From messaging to features to pricing and more, much of the offering may still be yet to be determined. It is indeed a fun, creative endeavor, but the same possibilities which you identified in WGB analysis can be daunting to pare down.
What we should seek to do at this point is to try to not necessarily eliminate variables altogether, but to try to limit the variables to series of ranges. Gerald Zaltman, Professor Emeritus at Harvard University, used to refer to this type of thinking as, "Understanding the direction in which the wind is blowing, but not yet concerning ourselves with exact wind speed."
For practical purposes, we do not want to test 200 concepts or prototypes, but we we want to put our best thinking into perhaps five concepts.
Think of the internet as this seemingly infinite stream of customers and information, but in which we have a limited view of the individuals or their thought processes at any one time. What we seek to do with message or proposition testing is to place a sampling net into the water and see what ends up in the net. If eight of our ten sampling nets come up empty time after time again, we know we can stop sampling in that area of the stream for the time being and focus our efforts on the two sampling net regions which did show some promise.
So, how can we do this? By using the massive sample sizes of the internet and pay-per-click advertising to act as our sample nets.
First, a primer on Google AdWords/Pay-Per-Click advertising. Please watch the following 3:24 video.
This tutorial provides a view into what the AdWords interface looks like, and some of the practical concerns when starting out. It is really quite straightforward to get started with AdWords, and there are many tutorial resources, but there is an entire profession devoted to the art and science of PPC advertising. Our goal in this early phase research is not to engage in advanced e-commerce, but simply to conduct a small, highly focused and controlled market test to help us focus our research efforts. Please watch the following 12:00 video.
Have you ever wondered why so many great innovations and companies are born from garages? Innovation is messy and emergent. It's loud, it's chaotic, and it's definitely not tidy and clean and neat. Therefore, it isn't something people generally want taking place in the house. This also holds true when testing early-phase concepts.
Think of the core brand of your organization as "the house." We don't want to disrupt that when creating offerings, so we want to find a space away from the house in which we can work without disrupting the house. This is why we do our testing and creation in safe places which won't be a nuisance. Call it "going rogue" or "working in the garage," but we need to provide some isolation and insulation when creating and testing.
If you have thirty interesting messages or propositions from survey research or ideations, try to group them into five or so groups or topics at this point. For example, if you have some very powerful customer quotes from an earlier survey, group them into "Customer Quotes," and if there is a highly differentiated attribute that has been well-received in surveys, you could have a "Lead Attribute" group, and so on. Our goal here is to condense all of the different ideas as tightly as we can, so that we may then see which theme appears to show the most promise.
For each theme you have identified, create four or five different variations of each. These will be your AdWords ads. These will provide us with some very early indicators of what message or proposition may be the most interesting to customers in a live environment. While there is an entire science behind conversion and what makes customers buy, consider each click on your research ad as a "vote." We are still quite far away from selling anything, but a click is the first step a potential customer could take to show interest. If customers don't even show interest by clicking, they could never take the next step, no matter how appealing it may be.
This can get very, very complex. The easiest way is to use the same keywords for all of your ads to eliminate that variable, and Google will show suggested keywords directly on the page as you begin entering a few. We (i.e., a consultant) can worry about dialing in keywords in the live campaign, but at this point we want to keep things simple.
At this point, you don't have a product to sell, but yet you're testing in the live market. How do you get around this? Land any of the Adwords clicks on an information gathering page, because they were interested in the proposition, and we would want to potentially interview them. For this reason, our AdWords ad can link to a page that asks the visitor to "sign up to receive more information" or "sign up if they would like to be part of a beta test." This signup should be the first thing on the page, and it may indeed be the only thing.
Remember, we probably do not want to disclose the organization at this point, lest competition catch wind of our early phase projects. So our goal is for interested people to sign up for more information, and we may then fold them into the research.
By understanding the number of clicks each "mini-proposition" receives compared to the total number of impressions for the ad (known as Clickthrough Rate or CTR and expressed as a percentage), we may have some very early window into messages that may show more promise than others. If we see that Message A receives 80% more clicks than Message B over the span of two weeks, we may pencil in that Message A is what we will use as our headline in subsequent testing. This is a very simplistic view of analytics, but the nice part is that analytics are forever: they will be captured, and you can filter and refine them however you may prefer at a later date.
As our research and offering progresses, so too can our PPC advertising testing. As we will see later in the semester, PPC will be a cornerstone of our beta testing, as it will help to drive traffic into our microsite, from which we will refine the offering even further.
Some may consider the use of a commercial online advertising tool to be a bit removed from more "pure" research techniques. I would argue that "pure" research techniques are pure because they are theoretical, and that our goal as researchers at this point is to quickly understand if the offering has merit, and what customers find most attractive about the offering.
Here are a few ways PPC message testing can play a valuable part in your early-phase research mix:
Focus groups tend to be one of the "go to" choices for early-phase consumer research, many times because the methodology is common, and resources, such as facilities and moderators, are generally easy to locate. I would argue that focus groups are not well suited to the needs of early-phase innovation research, the type which we would be prone to conduct.
I've spent thousands of hours both "behind the glass" and "in the room" in research facilities conducting fieldwork of various types, and have had the occasion to both observe and participate as a member in focus groups. Even in the well-moderated sessions, I tended to come away thinking about how skewed the discussions tended to become, and how much I would have wanted to interview the participants one-on-one.
The effect of the focus group format on validity of findings as opposed to one-on-one depth interviews have received quite a bit of scholarly attention, and papers have been written exploring methodological issues with focus groups. This abstract from Boateng (2012) [24] summarizes the findings of the overall body of research:
The efficacy of Focus Group Discussion as a qualitative data collection methodology is put on the line by empirically comparing and contrasting data from two FGD sessions and one-on-one interviews to ascertain the consistency in terms of data retrieved from respondents using these two data collection methodologies. The study is guided by the hypothesis that data obtained by FGD may be influenced by groupthink rather than individual respondents' perspectives. A critical scrutiny of the data that emanated from the two organized focus groups discussion departed quite significantly from the data that was elicited from the one-on-one qualitative interviews. The difference in responses confirms that FGDs are not fully insulated from the shackles of groupthink. It is recommended, among others, that though FGD can stand unilaterally as a research methodology for nonsensitive topics with no direct personal implications for respondents; researchers should be encouraged to adopt FGD in league with other methodologies in a form of triangulation or mixed methodological approach for a more quality data, bearing in mind the central role occupied by data in the scientific research process.
Furthermore, in my experiences, the group discussion format of focus groups tends to elicit the following behaviors, each of which has its own way of biasing or eroding the findings. (If the group is composed of 18-35 year-old males and females, multiply the biasing factor by 3):
My overall point with this example is that, much like any social gathering with people who do not know each other, when you get 8-12 people in a room in a single conversation, people "overact" or take on personality traits they otherwise would not. Impromptu caucuses will form before your eyes, as people with similar thoughts will band together.
I offer the following as a humorous example of some of the exhibited traits you might see in a focus group. Interestingly, the morning after this Saturday Night Live sketch aired, the consumer research world exploded in agreement and story as to how realistic "Linda" was! Please watch the following 6:49 video.
If we break apart some of the research "jobs" we would likely need in evaluating ideas or early-phase concepts, the role of the focus group becomes more and more niche.
Surveys do a better job at understanding the overall space than focus groups, and are far less expensive. Furthermore, you are receiving "clean data" in a survey, unbiased by social pressures of the group or groupthink.
Individual interviews will give you far more depth than any focus group, while allowing the interviewer to explore topics and ideas of interest.
Observation and ethnography will tell you more about use phase in real application. Individual interviews will tell you more about initial impressions of a prototype in a controlled environment, free from group biases.
Message or proposition testing in a live environment will provide far more realistic, specific, and practicable information.
This leaves us with focus groups being used as ideation sessions to generate ideas and creative. Needless to say, in these applications, focus groups may have far more in common with the SNL skit than you might like.
Also referred to as individual interviews, IDI (in-depth interviews), or one-on-one interviews, the depth interview methodology may seem quite straightforward but can take years to do well. Luckily for our efforts, the depth interview is also one of the most accessible types of research: it is simply one person guiding the discussion and asking questions, and the interviewee responds. For this reason, many depth interviews tend to be quite conversational and natural in format, as this will also allow the participant to discuss the topic freely. The truly adept depth interviewers take it a level further, eliciting specific stories and experiences from the participant and using various probes to explore ideas further.
Dr. Kelly Page offers a nicely composed overview of depth interview techniques:
There are a few additions and points of emphasis I would make to the slides:
Observing the participant interacting with or reacting to a concept can be an invaluable way to gain unfiltered insights. While observation may be conducted as a methodology by itself or paired with many other research techniques, it is especially potent when paired with depth interviews. This potency comes from the fact that the interviewer may first depth interview the participant on the topic or offering category, learning about their experiences, thoughts, and feelings. After understanding the participant's thoughts and feelings about the general topic, they may then be exposed to the offering or prototype being tested to provide full feedback.
There are three major types of observation you will most commonly see when researching early-phase concepts or offerings:
Covert observation is exactly as it sounds: You are observing a behavior or interaction without the participant knowing it, either by camera or by blending into the surroundings in a public space. Only after you have observed the behavior fully do you reveal to the participants that you have been observing them.
In practice, covert observation tends to be limited to occasions when you can actually locate a camera somewhere to see how a group of participants might interact with a machine, for example. In a sense, proposition and message testing acts as a form of covert observation in that you are watching online behaviors, preferences, and analytics to create your findings.
Detached overt observation is similar to covert observation with the exception that you let the participant know you will be observing them, and then retreat to a detached vantage point to watch their behaviors. Overt detached tends to be a bit limiting in that participants may act a bit differently if they know they are being watched or recorded from afar, especially if they are the only participant.
Also sometimes referred to as "side-by-side" observation, this is when you not only tell the participant you will be watching them, but when you stay just behind them or by their side and ask them to narrate the experience. This is especially potent in usability or refinement studies, where participants may first try to interact with the product in an unguided way, then asking the researcher questions afterward. If they are operating from an "insight mindset," overt narrative observation is especially beneficial for designers and those most involved in the offering, as it allows them to see the experience through the eyes of the participant.
Much of the value of the pairing of depth interviews with observational research is that it provides the researcher with a frame of reference for the participants' experiences from which to better understand the observational portion. For example, if the participant talked about their electrical engineering degree and expertise in solar energy in the depth interview, and then had trouble understanding and interacting with the solar controller offering being observed, that should stand as a significant warning sign that an experienced lead user is having trouble understanding the product. If only using the observational research, one could think that the older gentleman was just not technologically savvy or otherwise try to fabricate some backstory to explain their trouble interacting with the offering.
I long so much to make beautiful things. But beautiful things require effort—and disappointment and perseverance."
In this Lesson, we have started to move from the theoreticals of strategy and into the realities of research and testing and learning things about the offering that no other organization in the world may know. It is these specific research efforts which can create meaningful organizational knowledge of the kind which is not found in textbooks or the boilerplate slide decks of consultants. The early-phase research you conduct on an offering is unique to your organization. At this point in time, it is likely no one holds more practical knowledge on this specific offering than your team.
But, many times, the first round of research will result in more unknowns than when you first started. As the unknowns expand, it can be easy to lose sight not only of the core proposition and strategy of the offering, but the research itself. In view of the overall project, the research phase can be both exciting and depressing for many of the same reasons: you are finally gaining real data on the offering, but rarely are all signs wildly positive. There are areas of further research, areas which are still unclear, and some of these areas may be crucial to the viability of the offering. It can indeed feel as if the research is being dragged down further and further into the unknown.
Take a breath.
You'll be fine.
Remember the Insight Mindset. Chant it if need be. But, remember that things shouldn't be clear at this point. Promising projects will have more questions than answers in early phases. Every result is a victory, and brings us closer to truth and innovation. The only way to continue to gain results and data is to continue to research and try to find the overall storyline of the offering.
Which brings us to the van Gogh at top. That portrait of a woman is not superimposed over the painting of a grassy meadow, van Gogh actually painted "Patch of Grass" over the portrait of the woman. This would not be the only time van Gogh would paint over complete works, in fact, it is estimated that up to a third of his known works are painted over his earlier works. [34]
Even as a frustrated and resource-constrained artist in need of more canvasses on which to paint, in his letters, van Gogh would go on to say:
"I've just kept on ceaselessly painting in order to learn painting."
He saw it not as a destruction of earlier works and efforts, but an act of continuous learning and evolving his craft. We should all aspire to be able to destroy our earlier works while embracing it as an opportunity to learn.
As we begin creating more and more advanced versions of the offering, it is important to know that frustration is inevitable, results will be unclear, and questions will multiply.
Like painting, take some solace in the fact that if it was easy, everyone could do it.
To refresh ourselves, our goals specifically for this Lesson are to:
To this end, this week's Case will pick up where we left off last week, and will take the next step by framing the types of insights we seek to gain in our initial research effort.
Links
[1] mailto:axj153@psu.edu
[2] https://depositphotos.com/photo/storage-and-protection-of-cash-and-valuable-goods-66127765.html
[3] https://depositphotos.com/portfolio-1030327.html
[4] https://depositphotos.com/content-license.html
[5] http://www.biography.com/people/igor-fyodorovich-stravinsky-9497118
[6] https://www.flickr.com/photos/sarabbit/5624898315/in/photolist-9z45ir-8SRmkj-5JTgw1-dpwWkE-kyGQks-kEDT2D-7YcHCq-7Y9swM-55N6Jy-itk6X-4wiv5r-5HEU9q-9a1T9M-dohr6M-n7Kq2-aPidv6-bxKM1R-ekcLk4-8QVsKk-oWhP6v-6H6MWB-gyHxXC-dMkuQd-dt1bNV-5f3NfP-7K63i2-9vvJHu-762DaY-q6WDMQ-bvAcqg-7xpmxA-pSYacS-bGxtug-edokZY-7Yfgpa-5JD2F3-d1MwTC-5PACyw-sk8WjC-dcLPs1-aaEgef-9stvCV-9te3Ux-rgTWmU-4dkJDN-4d7N5j-3gyHiS-225wnb-bfo1wn-4d3Nup
[7] https://www.flickr.com/photos/sarabbit/
[8] https://creativecommons.org/licenses/by-sa/2.0/
[9] http://www.amazon.com/Fasteners-Plumbing-Handbook-Motorbooks-Workshop/dp/0879384069
[10] https://www.shutterstock.com/ja/image-photo/closeup-survey-form-agerelated-question-216942508
[11] https://www.shutterstock.com/ja/g/zimmytws
[12] https://phys.org/news/2012-01-consumers-surveys-feedback-fatigue.html
[13] https://dism.duke.edu/
[14] https://www.surveymonkey.com/mp/take-a-tour/?ut_source=header
[15] http://www.questionpro.com/
[16] http://www.wufoo.com/
[17] https://www.mturk.com/mturk/welcome
[18] http://www.pickfu.com/
[19] https://www.survata.com/
[20] http://www.google.com/insights/consumersurveys/home
[21] https://www.fisheries.noaa.gov/feature-story/manage-fish-populations-scientists-study-entire-ecosystem
[22] https://www.youtube.com/watch?v=EYHegvg5kII
[23] https://www.youtube.com/watch?v=PnOlWtWBJqo
[24] http://ijbssnet.com/journals/Vol_3_No_7_April_2012/6.pdf
[25] https://www.youtube.com/watch?v=GQ4jPqU5r7k
[26] https://www.youtube.com/@SaturdayNightLive
[27] https://www.slideshare.net/drkellypage/depth-interviews-in-applied-marketing-research
[28] https://www.slideshare.net/drkellypage
[29] mailto:pagekl@cardiff.ac.uk
[30] mailto:kelly@caseinsights.com
[31] http://creativecommons.org/
[32] https://photon-science.desy.de/research/research_highlights/archive/visualizing_a_lost_painting_by_vincent_van_gogh/index_eng.html
[33] http://vangoghletters.org/vg/letters/let274/letter.html
[34] http://www.telegraph.co.uk/news/science/science-news/3352325/Van-Gogh-painting-uncovered-by-new-Xray-machine.html