Evaluating Qual Data Analysis abilities of AI powered platforms

Oct 21, 2024, Dean Stephens

 “AI is going to save you time and money!”

About five years ago, the first rumblings of AI were being mentioned in marketing research circles. Like most other people in my industry, I filed those rumblings away in my brain. Over time, those mentions of AI in marketing research collected dust on those proverbial bookshelves of my mind.

 

But roughly two years ago, I had to suddenly dust off those memories because AI came rushing in with a thunderous boom. I’d gone to an industry conference and was surprised to see the marketing research conference schedule included several sessions dedicated to AI. I was fascinated and interested, but also lost; I didn’t know how to interpret what I was hearing, nor imagine how AI could be applied in my line of work. There were whispers in the convention hallways about AI. “Did you see that session on AI?” “Wasn’t that cool?” There was a lot of buzz about the potential of AI in marketing research.

 

The following year, AI was unavoidable; it was everywhere. It came crashing into my world with the gale-force winds of a hurricane! It seemed like every vendor in our industry had launched a new AI-powered market research platform. And there were seemingly 100s of new AI entrants jumping into our industry that had never been there before.

 

So, it’s no surprise that many of these vendors started reaching out to me and courting me for my business. They were offering free or discounted trials of their new AI-powered tools. I began scheduling calls with some of the vendors so I could better understand the tools that were now being offered. And it was during my very first vendor call that the guy on the other end of the line said something akin to: “this AI tool is going to save you time by doing your analysis and reporting for you.” My ears perked up!  But what I heard in the back of my brain was “Artificial Intelligence is coming for my job.” This has been an unfounded refrain for decades in the marketing research industry; every new technology implies that “it’s the end of qualitative research.” And history has shown us that previous tech advances merely gave us more tools for qualitative researchers to do our jobs.

 

But this time seemed potentially different. If these AI-powered tools could really do what the vendors were saying they can do, then was I right to be truly worried?

 

Skeptical Optimism

As much as I enjoy sci-fi and have probably seen every movie Hollywood has ever produced about AI, there was something nagging at me. I’ve always been an optimistic skeptic or a skeptical optimist. I just couldn’t fathom that some fancy algorithm could replicate what millions of years of evolution had accomplished and culminated in the form of the human mind. In my estimation, the human brain is one neuron short of a miracle! It’s been estimated that the adult human brain has upwards of 100 billion neurons. And each neuron is connected to thousands of other neurons. Before synaptic pruning occurs in the adult brain, there are upwards of 150 quadrillion synaptic connections in a human child!  That’s a mind-numbingly large number. Yes, computers are faster than humans, but the human mind is immeasurably complex. In my humble opinion, it was hard for me to see how any computer could do what the human brain can do. Until science proves otherwise, the human brain is quite literally the most complex and powerful creation in the universe.

 

In my call with the vendor, he said his AI tool was going to revolutionize qualitative research and save me countless hours of analysis. I needed to see the receipts. I asked the vendor to show me how the AI tool worked and more importantly, I wanted to see the output—the AI-powered analysis. I listened to his pitch intently, took notes, and watched his demo of the AI-generated “reports.”  By the end of his pitch, he asked me for my thoughts.  I took a beat, collected myself, and calmly thanked him for the demo, told him that I was impressed with how far AI has come as a technological advancement. But more than anything else, I was truly relieved. I mentioned that I was relieved because based on the demo he had just shown me, AI was nowhere close to taking my job as a qualitative researcher. I explained that AI’s ability to summarize main themes is truly and amazingly impressive. I could see the benefit of using a tool like this to identify main themes of a research project. But his “report” contained zero synthesis of new ideas, or implications, or recommendations based on solid analysis. He agreed, that yes, perhaps his sales pitch was overpromising a bit.

 

But the drumbeat of the AI-powered market research platforms only got louder and louder. I was being contacted weekly, if not daily, by new vendors trying to tell me how their platform was truly different than the competition. They weren’t. 

 

I wasn’t only hearing that drumbeat from afar either—I was hearing the same drumbeat from my colleagues as well. And their evangelizing call for AI was getting louder and more widespread. I was starting to feel like an outsider—a naysayer of sorts.

 

The Experiment

So, I decided to run an experiment. I wanted to see if any of the market research platforms were living up to their promises. I had questions I wanted answered:

1.      Will these AI-powered tools really save me time?

2.      Can these AI-powered tools really create solid and accurate summaries of main themes?

3.      Can they synthesize the data into relevant implications and actionable insights?

4.      Are these AI-powered tools going to put me out of a job?

 

There were other very important questions that needed to be answered, but those questions are far more complex, more nuanced, and ethically sticky. Questions no one individual is qualified to answer—questions that the marketing research industry as a whole needs to answer. Questions like: 

1.      What do these AI tools do with the data we provide them?  And what should they do with our data?

2.      Should we divulge to our clients when we are using AI tools? 

3.      Who’s responsible if poor business decisions are made based on the results provided by these AI tools?

4.      How do we identify hallucinations?

5.      What do we do about hallucinations?

6.      Who’s responsible if a hallucination makes it through to the end client?

7.      Is good enough really good enough? 

8.      Do we want to make decisions or recommendations based on good enough?

9.      AI tools skew toward the mean.  Do we as professionals only want to focus on or report on the mean?  Is the mean what our clients really want? Or need?

 

To conduct my experiment, I reached out to six different AI-powered market research platforms. I purposely picked a range of brands from very well-known brands to lesser-known brands. The one criteria they each had to have:  they all had to claim that their AI platforms provided qualitative analysis and/or reporting with the promise that they would save researchers time.

 

I asked each vendor for free, unfettered access on a limited trial basis to their AI tools. I was upfront with the vendors about my intentions. And I explained that my experiment was not about pitting vendor against vendor, but rather solely about human vs. machine. I explained I was running an experiment to uncover two things:

1.      Are they living up to their product claims? (e.g., saving me time, saving me money, saving me cognitive resources)

2.      How does their AI-powered analysis and reporting stack-up against human-powered analysis and reporting?

 

I promised to share the unvarnished results of my experiment with them. And each vendor agreed to the terms of my experiment.

 

The Methodology

My goal was to compare machine-generated insights and analysis to human-generated insights and analysis. So, I decided to use a completed project of my own for one of my cybersecurity clients. I picked a report that I knew the client was very happy with the results. And then I set some parameters of the experiment:

1.      To avoid giving access to any non-relevant data that might confuse the AI or increase chances of the AI hallucinating, I opted to limit the experiment to a single, straightforward question the client wanted answered:

“What are the top 3 challenges you face in securing your organization?”

2.      Each transcript only included the answer to this single question.

3.      No PII was included except for the respondents’ industry.

4.      And all AI platforms were given the exact same queries or prompts.

 

The Results of My Experiment

After uploading the edited transcripts, I began running the queries on each platform.  The queries were not complex questions; they were very simple questions, which I’ve illustrated below.  The results of my experiment were mixed. Overall, I found that AI platforms are really, really good at one thing—summarizing main themes. And abysmal on most other analysis and reporting tasks.

 

1.      Generate a report of all pertinent and relevant insights.

 

For query #1 above, this is where AI truly shined. Nearly all of the AI-powered marketing research platforms excelled at summarizing main themes. Some were better than others, but overall they all did a good job; however, there was one disconcerting finding. A few of the platforms emphasized themes or insights that I had deemed less relevant, and more importantly, at least two of the six platforms actually hallucinated insights. They reported a theme that was never mentioned in the transcripts.

 

2.      Write a 1-2 sentence summary of each respondent's answer to the question: “What are the top 3 challenges you face in securing your organization?”

 

For query #2, it’s unsurprising that the AI-powered platforms did well on this task as well. They provided mostly accurate summaries of each respondent’s viewpoint. But like before, a few of the platforms hallucinated insights that were never said by any respondent.

 

3.      Write a single, short, concise, pithy headline that encapsulates the research findings.

 

For query #3, the results were almost laughable. In my human-generated report, I wrote a powerful, yet succinct, action-oriented, four-word headline:  “People are the problem.”  In my prompt that I gave to each AI tool, I requested a “single, short, concise, pithy headline that encapsulates the research findings.” The AI-generated responses were too long, unfocused, misleading, and sprinkled with baseless clichés. Here’s just a sample of a few of the AI-generated headlines:

Internal Employees and Budget Constraints: The Dual Achilles' Heels of Cybersecurity Across Industries

“Navigating Cybersecurity Challenges:  Staying Ahead of Threats, Selecting Suitable Solutions, and Empowering the Workforce”

“User Awareness and Legacy Systems: Key Cybersecurity Challenges for Organizations”

 

4.      Create a list of recommendations for the client if they asked what they should do to address the IT decision-makers' challenges in securing their organizations.

 

And finally for query #4, this is another area where the AI-powered platforms performed poorly. To illustrate how poorly the AI-generated responses were, I need to briefly summarize the challenges the IT decision-makers mentioned. First, their biggest problem was people—both internal and external users. Their other constraints were small IT budgets, an under-skilled talent pool, no money to train them, and exponential company growth which was stretching their IT staffs to the brink.

 

To understand the gravity of what the AI-platforms did, you have to recall that I only uploaded transcripts to a single question and that question asked what challenges IT decision-makers experience in securing their organizations. And keep in mind that the question or prompt I asked the AI to answer was to provide recommendations taking into account the challenges IT decision-makers faced.

 

In this instance, all of the platforms flat out ignored my instructions. Every platform provided recommendations that cost time, money, and lots of effort—the very three challenges the IT decision-makers expressly said were problematic for them! I will concede that all of the recommendations were very good recommendations—absent taking into account their challenges.

 

The Successes and Pitfalls of My Experiment

Granted this is only my one-man, small-scale, qualitative experiment.  Anecdotally, I’ve heard from several colleagues of quite similar experiences they’ve encountered with AI-powered analysis platforms.   But what successes and pitfalls can you glean from this experiment?

I am in favor of the use of AI-powered analysis tools to suss out main themes and insights quickly—but keep in mind, these are tools to be utilized by researchers; they are not final arbiters of research findings. In my view, we can use these powerful AI tools in two ways:  1) either we can use them as a starting point for further human analysis; 2) or alternatively, we can use them as a double-check of our own human-generated insights.  And what makes them especially useful is if we use tagging or coding of segments, then we can explore the differences and similarities in the main themes on a per segment basis. And that’s a true time-saver.

As for the pitfalls, there are many. As my experiment has shown, they can hallucinate and add in false insights—even when these AI agents were fed small, edited bits of data they still hallucinated. I’m left to imagine what would have happened if I had uploaded unedited transcripts?  My worry is that if we researchers cannot easily and quickly identify an AI-generated hallucination, then we become part of the problem. We become the facilitators of unreliable insights based on falsehoods. Additionally, it’s concerning that AI can actually ignore instructions in a query/prompt. And this leads me to an even bigger worry…these AI-powered tools are designed to provide answers. But we as qualitative researchers know, sometimes there is no answer.

 

For some valuable tips-n-tricks and best practices on using AI-powered marketing research platforms, head over to my blog post:  https://www.happytalkresearch.com/post/tips-n-tricks-for-using-ai-powered-qualitative-research-platforms


 

About US: 

flowresAI is an online qualitative research tool built on our proprietary AI technology.  Amongst several tools evaluated by Dean, flowresAI was one of them. While flowresAI has fared much better than our current competitors, We agree with Dean's evaluation of the overall usefulness of AI tools. We believe that the usage of AI tools is going to evolve continuously. We are developing flowresAI to ensure it augments the role of qualitative researcher and reduces grunt work from the qualitative research process.  To know more about flowresAI, please feel free to reach out to us at Maniish at [email protected] or schedule a Demo here https://calendly.com/maniish-flowres/flowres-demo?month=2024-10







Dean Stephens
Oct 21, 2024