This is what your webinar data actually says.

A real, anonymized deep-dive report on three live runs of one host's sales webinar — generated from the raw Zoom + Brevo data in ManyMeet.

Every chart and number below comes from the actual production database. Names and brands are anonymized; the analysis is not.

Anyone using ManyMeet can produce a report like this by asking their AI assistant — Claude, ChatGPT, Gemini — to analyze their webinar via the MCP connection. No SQL, no exports, no spreadsheet wrangling.

Get reports like this for your own webinars

Free for your first webinar. No credit card.

Or scroll to read the full report below ↓

Take this report with you Download as image (.webp) Read inline ↓
ManyMeet · Custom Analysis
Report WBR-04 / 2026-05 · Rev. 2 Issued 2026-05-12 Confidential
Webinar deep-dive · 2026-03-03 – 2026-05-05

Live webinar for premium real estate investors.

Three runs on Tuesday, 19:00 Berlin time. 764 registrations, 313 attendees, median attendance 104 min against an advertised 90-min program. The run on April 14 was the outlier — and the most actionable lesson in this report. What works, where the pipeline bleeds, and which levers to pull next.

For
Anonymous Host
Anonymous Inc.
By
ManyMeet | Marketing
blue media labs GmbH
Engine
ManyMeet AI Engine
Proprietary webinar intelligence
Period
2026-03-03 – 2026-05-05
3 runs · all Tue 19:00 Berlin
764
Registrations
across 3 runs
313
Attendees
41 % attendance rate
104min
Median attendance
vs. 90-min program
73%
Punctuality · top run
≤ 5 min after start (Apr 14)
86%
Question-chatters · top run
Apr 14 · vs. 28 % on May 5
TL;DR

Six findings that carry this analysis.

06 Findings
01 · Outlier

April 14 is the blueprint.

~3× more day-0 registrations and 76 % same-day conversion (vs. 47 / 57 % in the other runs). Whatever triggered that day is the highest-leverage thing to replicate.

02 · Lead time

Registration recency beats everything.

Registrations in the last 24 h: 50 – 87 % attendance rate. ≥7 days out: 25 – 38 %. Day 0 is the highest-quality cohort but has no dedicated flow.

03 · Punctuality

The "drop-off" on Mar 3 was lateness.

Only 53 % were present by min 5 (vs. ~72 %). At the host's request, Mar 3 ran a reduced nurture and countdown sequence — the likely cause.

04 · Timing

Tue 17 – 19 Berlin is the sweet-spot window.

Registrations placed in that hour convert at 57 % — day of week and time of day match the event. Lifestyle fit by self-selection.

05 · Domain

Business emails attend more reliably.

45 % vs. 39 % for freemail. Stay duration is identical — the "pros leave before the pitch" hypothesis is not supported. Booking conversion remains untested.

06 · Reminder lift

Whoever clicks the T-1h reminder shows up.

In the May 5 run: T-1h clickers attended at 84 %, non-clickers at 13 %. The strongest single signal in the entire dataset — nearly identical to the attendee population itself.

— 01

The three runs at a glance.

Funnel · Retention · Engagement
Metric
Tue · Mar 3Run · 178 regs
Tue · Apr 14Outlier · 392 regs
Tue · May 5Run · 194 regs
Attended
66
170
77
Attendance rate
37.1% of regs
43.4% of regs
39.7% of regs
Avg attendance
84.1minutes
95.6minutes
92.6minutes
Punctual (≤ 5 min)
53.0% of attendees
72.9% of attendees
71.4% of attendees
Stays ≥ 90 min
66.7%
81.2%
74.0%
Chat messages
1021.55 / attendee
2711.59 / attendee
781.01 / attendee
Question density
38% of chatters
86% of chatters
28% of chatters

Retention rate across the webinar runtime

% of attendees still present at minute X. 90 min is the advertised program end — open Q&A follows.

Mar 3 Apr 14 May 5
100% 75% 50% 25% 0% 15 30 45 60 75 90 105 120 Minute · program ends at 90 offer / pitch
Insight

Once joined, the audience is highly engaged: median attendance 102 – 108 min against a 90-min program. The "early drop-off" on Mar 3 was a punctuality artifact — by min 60 all three runs are on the same plateau. A majority of attendees stay past program end for the Q&A: at min 105 (15 min into Q&A) 60 – 71 % are still present depending on the run.

— 02

Registration lead time is the strongest predictor of attendance.

Lead-time bucketing

Registrations per day before the webinar — attendees vs. no-shows

Stacked bars per day: green at bottom = attended, red on top = no-show. Day 0 = event day. Comparing Apr 14 (outlier) vs. May 5 reveals the volume shift.

Attended No-show Run's average attendance · if the line sits in the green, the day is above average
Run · 2026-04-14
50 40 30 20 10 0 Registrations 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21+ Days before the event
Total regs: 392 Attended: 170 (43 %) No-shows: 222
Run · 2026-05-05
50 40 30 20 10 0 Registrations 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21+ Days before the event
Total regs: 192 Attended: 77 (40 %) No-shows: 115

The black line marks each bar's run-average attendance (Apr 14 = 43 %, May 5 = 40 %). If the line sits within the green, that day converted above average — it made the webinar net better. If it sits in the red, the day made the webinar net worse.

Apr 14: Day 0 stands out clearly as a net plus; several mid-range days (2, 4, 13, 14, 15) also contribute above average; most older days sit slightly below. Overall a healthy profile — the run's result is carried by many mediocre-but-decent days plus the strong Day 0.

May 5: only Day 0 is clearly above average (57 %). Days 1 – 7 are almost uniformly net losers — especially Day 7 (23 %) and Days 1, 2, 3 (22 – 39 %). Those registrations pulled the run's conversion rate down rather than lifting it. Reading: the ad phase 2 – 7 days before the event delivered volume, but volume of poor quality. Day-of (Day 0) remains the only reliable contributor.

Replication thesis: run quality is not determined by total registration count, but by how many bars cross the line into green. More registrants per day whose lifestyle brings them here — not more registrants through ad-volume.

Attendance rate by registration lead time

Share of attendees per bucket — pooled across all three runs.

Lead-time bucket
Attendance rate (%)
≥ 14 days
31 %
177 regs
7 – 14 days
33 %
155 regs
3 – 7 days
42 %
221 regs
1 – 3 days
36 %
95 regs
1 – 24 h
55 %
87 regs
< 1 h
83 %
18 regs
after start
91 %
11 regs · walk-ins
Implication

Roughly 25 % of registrations per run land on Day 0 — and convert most reliably. The Apr 14 run had about 3× more Day-0 registrations than the others, at a 76 % attendance rate (vs. 47 / 57 %). That is the replicable lever.

Recommendation: dedicated same-day flow for Day-0 registrants ("welcome + calendar + join link in one click"); don't give up on the ≥7-day cohort — it provides the volume but needs the full reminder cadence; tag returning attendees in CRM (58 % attendance rate — high intent).

— 03

Punctuality is a property of the run, not of the person.

Drop-off · same-day sequence

Share of attendees present within 5 minutes of start

Punctuality correlates with the density of same-day communication — not with the individual's registration behavior.

Mar 3
53 %
28.8 % arrived 15+ min late · 18 % arrived 30+ min late
Apr 14
73 %
14 % arrived 15+ min late · 8 % arrived 30+ min late
May 5
71 %
18 % arrived 15+ min late · 10 % arrived 30+ min late
Punctual · stay duration
95 – 116 min
30+ min late · stay duration
50 – 68 min
Takeaway
Late arrival predicts early departure. "% of attendees present in first 5 min" is the cleanest KPI for same-day activation.
Likely cause · Mar 3

For Mar 3, at the host's request, a deliberately reduced nurture and countdown sequence was run. That exact difference shows up in punctuality — later arrival, identical stay behavior from min 60 onward. Recommendation: establish the full same-day sequence as a systematic default for future runs.

Chat engagement per run

Question density as a possible conversion proxy: share of chatters who wrote a "?".

Run
msgs / attendee
% chatters
% question-chatters
Mar 3
1.55
52 %
38 %
Apr 14
1.59
43 %
86 %
May 5
1.01
32 %
28 %

Chat density on Apr 14 shows a spike at min 0 – 10 (greetings), quiet at min 20 – 50, then steady growth from min 60 onward with a peak at min 100 (29 messages, 16 questions). This spike coincides with the transition from program (value content with embedded pitch) into open Q&A — it measures the format change, not the depth of content engagement. What remains robust is question density as an interest indicator: Apr 14 sits at 86 % question-chatters, 2 – 3× above Mar 3 and May 5.

— 04

Whoever registers Tue at 19:00 shows up Tue at 19:00.

Hour · day of week · domain

Attendance rate by registration hour (Berlin)

Pooled across all three runs · the sweet spot is the weekday-evening window.

0 – 6 night
19 %
68 regs
7 – 8
36 %
95 regs
9 – 11
44 %
128 regs
12 – 13
40 %
68 regs
14 – 16
47 %
93 regs
17 – 19 ★
52 %
157 regs · hour 19:00 → 57 %
20 – 22
39 %
132 regs
23
30 %
23 regs

Business vs. freemail registrants

The most stable cross-run pattern in the analysis · internal domains excluded.

Business
45.1 %
attendance rate · 215 regs
30.7 % reach min 90 · 14.4 % stay ≥120 min
Freemail
38.6 %
attendance rate · 542 regs
25.5 % reach min 90 · 11.6 % stay ≥120 min
Top freemail domainRegistrationsAttendance rateAvg attendanceSignal
web.de12043.3 %97.4 minstrong
gmx.de10137.6 %93.9 minsolid
gmail.com12436.3 %83.2 minsolid
googlemail.com2437.5 %71.5 minsolid
t-online.de4725.5 %101.6 minbelow avg.
icloud.com2828.6 %64.6 minbelow avg.
Hypothesis test

The assumption "pros come to learn, leave before the pitch" is not supported by the attendance data — median stay is essentially identical (104 vs. 106 min). If anything, business registrants stay slightly longer at p25 (80 vs. 68 min) and convert better at every funnel depth.

But: stay duration is not the real conversion criterion — bookings are. The engagement signal (question-asking) on Apr 14 correlates with a high freemail share (74 %). Definitively answerable only through a Typeform / CRM cross-reference per attendee.

— 05

One email decides almost everything.

T-1h reminder · click → attendance · funnel
Method note

Open rates in email have been systematically distorted since the introduction of Apple Mail Privacy Protection (2021) — some mail clients pre-fetch tracking pixels server-side, others block them entirely. Industry-wide: click rates are the reliable engagement measure, open rates at best provide a lower bound. This section works consistently with click data.

T-1h reminder: click → attendance

May 5 run · the one email most strongly correlated with attendance in the entire dataset.

Clickers
84 %
attended · 54 of 64
Non-clickers
13 %
attended · 16 of 120
Spread
+71 pp
strongest single signal in the dataset

The T-1h reminder contains the Zoom join link. A click on it is effectively the first step toward attendance — accordingly, the clicker group nearly coincides with the eventual attendee group. Causal reading: the email is half trigger (rescuing wobblers), half symptom (click = attendance intent). In either case, it is the highest-leverage touchpoint in the entire sequence.

Notable: only 35 % of recipients click. For an email that arrives exactly one hour before the event and delivers nothing but the join link, 65 % non-clickers is significant lost reach — and roughly identical to the 60 % of registrants who ultimately don't show up.

Click rates by funnel stage

What shows real funnel engagement — via clicks, since opens were unreliable.

Funnel stageTemplateSendsClick rateReading
DOI confirmationtpl 4414918 %funnel entry
Registration confirmationtpl 4319617 %funnel entry
T-3 day remindertpl 461235 %no signal
T-1 day remindertpl 451466 %no signal
T-1 h remindertpl 4718535 %attendance trigger
Post-event bookingtpl 1507643 %conversion signal
Last-chance bookingtpl 1061547 %conversion signal
No-show recoverytpl 1051028 %flat & constant

Three clear clusters: (a) T-3d and T-1d reminders deliver 5 – 6 % click and zero predictive power for attendance — the effort would be better invested elsewhere. (b) The T-1h reminder is the decisive lever. (c) Post-event booking emails achieve the highest click rates in the funnel at 43 – 47 % — the audience is warm and click-eager right after the webinar. A reliable conversion statement is only possible with a booking cross-reference.

Implication

The reminder sequence is not uniformly valuable. T-1h carries almost the entire effect; T-3d and T-1d, in their current form, are awareness touches with no measurable conversion signal. Recommendation: A/B-test the subject line and send time of the T-1h reminder — any improvement there acts directly on the attendance rate. T-3d / T-1d either redesign (e.g. with concrete preparation content) or drop entirely.

— 06

What to do next.

Action items · prioritized

Marketing · ready to ship

M1

Build a same-day flow for Day-0 registrants.

"Welcome + calendar + join link in one click." This cohort converts at 50 – 87 % and accounts for ~25 % of all registrations.

M2

Run ads Tue 9 – 19 Berlin time.

Registrants from this cell both show up AND stay. Avoid Sat/Sun pushes, deprioritize night slots (0 – 6).

M3

Take the T-1h reminder seriously as a lever.

A click on the T-1h reminder predicts attendance at 84 % vs. 13 % (n=184). The subject line, send time and join link of this one email are the highest-yield knobs in the funnel — an A/B test pays for itself immediately.

M4

Tag returning attendees in CRM.

Past attendees show up at a 58 % attendance rate. Dedicated messaging ("welcome back") is warranted.

M5

Deconstruct April 14.

What triggered ~3× more Day-0 registrations on Apr 14? Ad set, creative, promo push — identify and replicate.

M6

Cross-reference booking conversion.

Cross Typeform / CRM bookings with the domain bucket. Only then can "pros vs. privates" be answered definitively.

What ManyMeet enables next for the host

A1

Registration lead time → attendance rate.

Bucket chart per run · strongest single predictor.

A2

"% in the first 5 min" as a KPI.

Clean signal for reminder quality.

A3

T-1h reminder lift on the dashboard.

Click on T-1h → attendance. Currently 84 % vs. 13 % — measurable as a live KPI per run for every host.

A4

Hour / day of week / domain.

Three marketing-efficiency charts per host.

How this report came to be.

The host runs the same 90-minute live webinar every few weeks on Zoom Meetings, with the funnel orchestrated through Brevo (registration, double-opt-in, reminders, post-event nurture). ManyMeet observes the runs end-to-end and stores every signal in one place: who registered when, who joined when, how long they stayed, what they chatted, which emails fired, which clicks landed.

For this report, the host's marketing partner asked an AI assistant: "Do a deep-dive analysis of the last three runs of webinar 4 — find what's working, what's not, and what's repeatable."

The assistant queried ManyMeet via the MCP interface, iterated for several rounds with the partner correcting interpretations (the "Q&A spike at minute 100 is not engagement — it's the format change"), and produced this report. Same path is available to any ManyMeet host.

What it surfaced.

  • Registration lead time is the strongest predictor of attendance — last-24h registrants attend at 50–87 %, ≥7-day-out at 25–38 %. The chart on Day 0 vs. earlier days reframed the entire ad strategy.
  • The T-1h reminder click rate predicts attendance at 84 % vs. 13 % — the single strongest signal in the dataset. Subject line and send time of that one email is the highest-leverage knob in the funnel.
  • The Apr 14 outlier run had 3× more Day-0 registrations at 76 % same-day conversion — a replicable pattern, not luck.
  • Punctuality was a run-level property, not a person property — driven by the density of the same-day reminder sequence. One run with a reduced cadence had 53 % punctual joiners vs. 73 % in the others.
  • Pre-existing assumption falsified: business-email registrants do not leave before the pitch — their attendance and stay duration matched freemail at every depth.

Want reports like this for your own webinars?

Connect your next Zoom Meeting to ManyMeet. We observe it silently, capture every signal, and let you ask your AI assistant any question about it afterwards — without spreadsheets, without SQL, without changing your stack.

Analyze my next webinar — free

Free for your first webinar. No credit card. No commitment.

See how the Observer mode works →