top of page

PM Case Study: Oneshot. AI | Most AI Products Fail Because They Solve the Wrong Problem | Part 1 : Cohort Analysis

  • Writer: Gokul Rangarajan
    Gokul Rangarajan
  • 2 days ago
  • 11 min read

Updated: 1 day ago

Inside a Product Manager’s 12-Month Diagnosis of an AI Sales Tool


This blog is about the AI product management case of how Our GP and Group PM Gokul Rangarajan worked with OneShot AI, a B2B sales Chrome plugin, repositioned and reinvented its product's core from a tool to a platform using user research, customer research, secondary research, and strategic decisions. This blog breaks down Gokul's 9–12 months of working experience closely with Founder Oneshot.ai which helped them to raise 3M+ from 42CAP and eventually get to 2M ARR.


This is 3 part blog case study is based

  • how a promising AI product with users and revenue was not delivering real workflow value.

  • how user research and product analysis revealed the wrong problem being solved

  • how we rethought the product from a personalization tool → autonomous outreach platform When we started working with Oneshot.ai in November 2021 alongside Venki Pola and Gautham Rishi, the product already looked like what most people would call "working". The product had AI capabilities layered into the core experience. It had early traction, roughly fifty active users, and had already generated close to $20,000 in revenue and one paying enterprise customer custom-made to their workflow. From the outside, it checked the boxes most early-stage AI products aspire to reach: functional technology, initial adoption, and some signs of monetisation. But within the first few weeks of engaging with the product, team and leadership, something didn’t sit right.

Nov 2021

At that time, OneShot.ai was positioned as an AI-powered Chrome extension that helped sales teams generate personalised LinkedIn messages and cold emails.


The Product promise was straightforward: use AI to create highly tailored outreach at scale, something SDRs have historically struggled to do consistently. On paper, it made perfect sense. Personalisation has long been considered a key driver of outbound success. Response rates improve when messages feel relevant. Sales leaders talk about it constantly. The market, sales and SDR community reinforce it.

So what started as a simple engagement became a full product diagnosis:

  • Why users were not sticking

  • Why AI features were not translating to value

  • And what the product actually needed to become

So instead of assuming the premise was correct, I went deeper into three parallel tracks: Product usage analysis , User interviews, s, and market research.

Dec 2021-Product usage analysis


Before changing the product direction, we started with something more basic.

We looked at onboarding and Time to Value analysis . Because if users don’t understand value early, nothing else matters, including how users moved from:

installing the Chrome extension → actually using it inside LinkedIn or Eamil

On the surface, onboarding looked complete.

Users could:

  • sign up

  • connect accounts

  • land on LinkedIn

  • start using personalisation.

But when we tracked behaviour closely, the drop-offs were significant.


What We Actually Measured

We broke the onboarding into a step-by-step behavioural funnel.

Not vanity metrics like signups but product usage signals.

We started tracking:

  • How many users installed the extension

  • How many reached LinkedIn profile pages

  • How many triggered personalization

  • How many actually used the generated message

  • How many sent outreach

This was important because the product promise was:

“AI will help you personalize outreach at scale.”

So the real question became:

Are users actually reaching the moment where they experience this value? The user interviews

The “Aha Moment” We Expected

The expected “aha moment” in the product was simple:

A user opens a LinkedIn profile → OneShot analyses the profile →Generates a personalized intro →User sends it.

That moment should create a feeling of:

“This saves me time. I want to use this every day.”

But that wasn’t happening consistently. How we found ou this ? Looking at real usage patterns, a different picture emerged. A large majority of users were able to successfully reach LinkedIn after installing the extension, and nearly 65–70% triggered the AI at least once, with around 55% generating a personalised message. On the surface, this suggested strong feature adoption. But when we tracked what happened next, the funnel broke sharply. Less than 20–25% of users actually copied or used the generated message, and fewer than 10–12% went on to send outreach.


Even more concerning, onboarding completion itself was inconsistent—while users signed in, nearly 30–40% never fully completed onboarding flows, and a meaningful segment dropped off before experiencing the first real value moment. This meant users were reaching the feature but not crossing the threshold into meaningful usage.

Stage

Event Trigger

User Tier

Metric

How We Measured It

What We Learned

Install → LinkedIn Entry

Extension activated + LinkedIn page load

New Users

~70–75% reached LinkedIn

Chrome extension install event + domain tracking

Strong intent; users were curious and explored immediately

LinkedIn → AI Trigger

tracking and edit

Activated Users

~65–70% triggered AI

UI click events inside LinkedIn overlay

Feature discovery was high; users understood where value might be

AI Trigger → Message Generated

AI generation success event

Engaged Users

~50–55% generated message

Backend generation logs (API success events)

AI capability was working and being explored

Message Generated → Message Used

Copy/Edit interaction

Evaluating Users

~20–25% used output

Clipboard tracking + edit interactions

First major drop low trust in output or unclear usability

Message Used → Message Sent

“Copy to LinkedIn” / send action

High-Intent Users

~10–12% sent outreach

Proxy via copy + interview validation

Critical break product not driving real-world action

Onboarding Completion

Completed setup (company + context)

Signed-up Users

~60–70% completed onboarding

Funnel tracking across onboarding steps

Users signed in but didn’t fully activate

First Value Moment

Generated+ reviewed and stayed in flow

Activated Users

~30–35% reached aha moment

Multi-event condition (generate + dwell time + interaction)

Weak value realization; moment not strong enough

Repeat Usage

Returned to use feature again

Retained Users

8–9% repeat usage

Cohort retention (Day 1 / Day 7)

Product not part of daily workflow

Trial → Active

Integrated into outbound workflow

Power Users

Low (qualitative)

Combined usage and interview validation

Product explored, not operationalized


What became clear was that engagement was shallow. Many users tried the one-line personalisation once, but only ~8–9% returned to use it consistently, and trial-to-active conversion remained low.



Time-to-value was also longer than expected users weren’t immediately confident in using the output within their workflow, which reduced repeat behaviour. The product was being explored but not adopted. Users were onboarding but not internalising value. And even when they experienced the core feature, it wasn’t strong enough to change behaviour. The system was generating outputs but not driving action—and in a workflow product, that gap between generation and execution is where products fail.

This made us to dive deeper into other analsis to find pattern what we could get form users. Cohort Segmentation


By dividing users into groups with shared characteristics or experiences, we analysed this to understand whether users naturally progressed across value stages. We found out 50–60% of SDRs were generated once and dropped, while less than 10% moved to repeated usage.


  • The signal was clear: users were not transitioning from trial to usage, indicating broken value progression, not a top-of-funnel issue.

  • Value wasn’t progressive → users saw one-time benefit, not increasing returns

  • UX didn’t guide users from generate → act → repeat

  • Workflow mismatch → no natural next step after first use


Free → Paid Conversion Breakdown

We analysed this to diagnose conversion failure.






<5% of SDRs reached a stage where they could evaluate ROI meaningfully (multiple sends across prospects).

Users never accumulated enough usage depth to justify a buying decision—conversion was blocked before pricing even mattered.

  • Users didn’t reach outcome → no ROI signal formed

  • Trust insufficient → not enough successful sends to justify payment

  • Value not aggregated → benefit felt isolated, not compounding




Time-to-Value vs Time-to-Trust

Latency between text generation and action was analyzed <5 minutes to generate output, but >80% of SDRs never acted on it within the same session. Fast value exposure without immediate action indicates lack of trust, not lack of utility.



<5 min to generate · >80% never acted same session — trust gap, not utility gap
<5 min to generate · >80% never acted same session — trust gap, not utility gap

  • Generation fast (<1 min), but decision to send requires confidence in reply rate

  • SDRs mentally simulate: “Will this get ignored?” → high uncertainty → no send

  • No past success memory → each message judged from scratch

False Aha Moment Detection

We analyzed this gap to validate activation quality, In PLG SaaS, activation is typically measured by a user reaching a meaningful first action like creating a project, sending a campaign, or publishing something live. In outbound tools, this usually translates to a user completing and executing an outreach step, not just preparing for it. Across multiple PLG case studies (e.g., email tools, CRM workflows), the activation event is tied to a completed action inside the user’s workflow, not a preview or intermediate step.


In tools like Slack, value is realized when a team sends messages and gets responses—not when they just set up channels. In Notion, activation happens when users create and organize content they return to not when they browse templates. Similarly, in outbound tools like Apollo.io or Outreach.io, value is tied to sending sequences and engaging prospects, not drafting messages.


In the OneShot case, the product defined activation at the point of generating a personalized message. That aligns with feature usage, but not with how SDRs actually operate. For an SDR, generating a message is only a preparatory step the real job is reaching out and driving responses. Observationally, users were engaging with the generation flow, reading outputs, and sometimes refining them, but stopping short of integrating that output into their actual outreach workflow.

50% of SDRs generated personalization, but only ~10–12% acted on it. The insight: generation created a perceived aha, but real activation required action—our “aha definition” was incorrect.

  • SDRs react to output: “Looks good” → but next thought: “Will it work?” " do i really need it ?

  • Without outcome signal (reply/meeting), “aha” collapses within seconds

  • Generation ≠ quota impact → no behavioral shift



This creates what can be called a “false aha moment.” The product surfaces something that looks valuable in isolation, but it doesn’t map to the user’s definition of success. In PLG terms, the system is capturing intent signals without completion signals. The distinction matters because real activation requires the user to cross a behavioral threshold—moving from seeing value to relying on it. Without that shift, what appears as activation in product metrics is actually just structured exploration.

Activation Depth Analysis

~30% interacted beyond viewing · <half progressed to sending — engagement ≠ commitment
~30% interacted beyond viewing · <half progressed to sending — engagement ≠ commitment

Depth of interaction post-generation matters but ~30% of SDRs interacted beyond viewing, but less than half of them progressed to sending. Activation needs to include decision-making, not just interaction—users engaged but did not commit.



  • SDRs stopped at generation because editing cost ≈ writing from scratch

  • No reduction in thinking effort → cognitive load unchanged

  • If tool doesn’t save ≥30–40% effort, SDRs abandon

Retention Cohort (D1–D7)

~33% D0 → ~13% D3 → <10% D7 — product not tied to recurring workflow
~33% D0 → ~13% D3 → <10% D7 — product not tied to recurring workflow


~30–35% first-session usage dropped to ~10–15% within a few days and <10% by week one. We analyzed retention to check habit formation. Lack of repeat usage indicates the product is not tied to a recurring workflow.

Behavior Cohort Progression

~50% generated, ~20–25% used output, ~10–12% sent messages. We analyzed stepwise drop-offs across behavior.


the biggest break is between evaluation → action, not discovery → usage.

  • SDR flow: open profile → decide fast → send or skip

  • Tool inserted extra step → slowed decision loop

  • Anything breaking high-velocity loop gets dropped

Onboarding Drop-Off Analysis

We analyzed onboarding to check activation readiness.



~30–40% of SDRs dropped before completing onboarding, and an additional drop occurred before first meaningful use.


The signal: completing onboarding did not guarantee value understanding—onboarding was not aligned to outcome.

Value Depth Layer Analysis

 We analyzed depth of value engagement what we found is overall 50% reached shallow value (generation), ~20% mid-level (usage), ~10% deep value (sending). value was accessible but not compelling enough to drive deeper progression.


  • SDRs care only about message → reply → meeting

  • Generation alone doesn’t move pipeline → seen as low-value layer

  • Deeper value (reply outcomes) not visible → no progression

Free User Non-Conversion Analysis

80% of SDRs remained in exploratory usage without repeated action. We analyzed why free users didn’t convert. The takeaway: users didn’t see consistent outcomes, so they didn’t build conviction to pay.

Trust Formation Analysis


Trust formation analysis Seeing output does not equal believing it is good enough to use — trust degrades at each stage.
Trust formation analysis Seeing output does not equal believing it is good enough to use — trust degrades at each stage.

50% saw outputs, but only ~20–25% used them and ~10–12% acted. We analyzed the drop between output and usage. Trust degraded across stages seeing output ≠ believing it is good enough to use.

Exploration vs Habit Analysis

~60% of SDRs used the product once, while <10% formed repeat behavior. We analyzed session frequency over time.

Workflow Integration Analysis

Product was adjacent to the workflow, not embedded within it. <10% used it consistently.
Product was adjacent to the workflow, not embedded within it. <10% used it consistently.

<10% of SDRs used the product consistently within their outreach process. We analyzed integration into daily work. The product was adjacent to the workflow, not embedded within it.

Output vs Outcome Analysis

Output quality alone does not create value unless it leads to outcomes. 50% generated, only 11% acted.
Output quality alone does not create value unless it leads to outcomes. 50% generated, only 11% acted.


50% generated outputs, but only ~10–12% converted to outreach actions. We analyzed output-to-outcome conversion. The insight: output quality alone does not create value unless it leads to outcomes.

ICP Identification via Behavior

<10% of users demonstrated repeated usage patterns aligned with structured outbound workflows. We analyzed behavior of retained users. The product fit users with defined workflows, not ad-hoc usage patterns.


  • SDRs with structured outbound (lists, sequences) engaged more

  • Ad-hoc prospectors didn’t need tool consistently

  • Product required repeatable workflow to show value



Drop from ~50% generation to ~10–12% sending. We analyzed this gap specifically. The takeaway: the hardest problem is not generating content—it is enabling the decision to act on it.


Product-Market Fit Misalignment

Product sat between shallow and deep use cases, failing both segments. ~50% needed less, <10% needed more.
Product sat between shallow and deep use cases, failing both segments. ~50% needed less, <10% needed more.

50% of users didn’t need deep workflow automation, while <10% needed more than the product offered. We analyzed mismatch across cohorts. The insight: the product sat between shallow and deep use cases, failing both segments.



Final Takeaways (What We Actually Learned)

The core question wasn’t whether users found value—it was what kind of value they expected vs what we delivered. SDRs did find the one-line personalization interesting and occasionally useful, but not critical enough to change behavior. The value was incremental, not transformational. It helped improve a message, but it didn’t meaningfully improve what SDRs ultimately care about—getting replies and booking meetings. Because of that, the benefit didn’t compound over time, and without compounding value, there was no reason to build a habit or pay for it.



The second layer was a workflow issue. SDRs don’t operate in isolated steps—they work inside fast loops: find → write → send → move on. OneShot inserted itself into that loop but didn’t reduce enough friction across the full workflow. If the product had lived directly inside their existing flow (e.g., deeper LinkedIn automation or sequencing), it might have reduced effort meaningfully. Instead, it behaved like an assistive layer that required extra thinking at the decision moment, which is exactly where SDRs optimize for speed.



The third realization was around the nature of personalization itself. One-line personalization is helpful, but not a strong enough wedge. SDRs are comfortable writing one-liners themselves, and they don’t assign high monetary value to that task. More importantly, personalization alone doesn’t guarantee outcomes. The real value sits in who to reach out to, when to reach out, and why now. Without that context (signals, targeting, timing), improving the message itself felt insufficient. In short, we optimized the message, but SDRs optimize for pipeline.



Open Questions from the Analysis

Are SDRs actually not finding value, or is the value too small to change behavior?Users are engaging with message generation, but does that mean they believe it improves outcomes—or just that it’s interesting to try?

Is this a compounding value problem?If the benefit from one-line personalization is incremental, does it fail to build enough momentum across multiple sends to justify repeated usage?

Is this a workflow placement issue?If the product sits outside the core outbound loop, are SDRs dropping off simply because it interrupts their speed, not because it lacks utility?

Is one-line personalization itself a weak value wedge?If SDRs can already write one-liners quickly, is the product solving something they don’t perceive as worth outsourcing?

Is the problem actually upstream?Are SDRs struggling less with “what to write” and more with “who to target and why now,” making personalization feel secondary?

Does channel context affect behavior?If users are more cautious on LinkedIn, does the same output feel less usable compared to email workflows?



Hypothesis

% Range

Next action item

Users find value but don’t act on it

~55–65% → ~25–30%

User interview

Value doesn’t compound

~8–12%

User interview

Workflow friction causes drop-off

~55–65% → ~12–16%

Onboarding A/B experiment

One-line personalization is low-value

~60% gen vs ~10–15% send

User interview

Problem is upstream

~20–30% regenerate

Onboarding A/B experiment

Channel affects trust

~2–4% difference

Growth experiment

Did changing communication alone improve behavior or does reframing value from personalization to outcomes actually move SDR actions?

Do SDRs behave differently when context is added—like signals, timing, and “why now” on top of personalization?

And is personalization only valuable when it’s embedded directly into the outbound workflow, rather than sitting as a separate step?

In the next post, we break down what happened when we tested these assumptions in the product.

 
 
 

Comments


bottom of page