数据可视化模式描述

0 浏览
0 试用
0 购买
Sep 28, 2025更新

清晰描述图表中的数据模式,提供专业可视化建议。

示例1

以下为在“周活(WAU)与留存”趋势解读中,常见图表所呈现的模式与对应的技术性解读要点。适用于折线图(WAU/新增/回流分解)、留存曲线图(按周龄)、以及周度 Cohort 热力图。

一、总体趋势模式
- 平稳上升:WAU持续上行,同时各周 Cohort 的W1/W4留存稳定或上行。解读为“质量驱动增长”,增长来自更高的使用黏性而非一次性拉新。
- 拉新驱动型上升:WAU上行但留存持平或下行,新增用户波动显著。解读为“规模扩张覆盖质量不足”,短期营销或渠道放量支撑了活跃,但长期留存与LTV压力增大。
- 下行或平台期:WAU走平/回落,伴随早期留存(W1/W2)下滑。解读为“前置体验问题”或“核心价值弱化”,需优先定位首周环节(注册、首单/首关键行为)转化与回访触点。
- 结构性跃迁:特定时间点出现WAU台阶式抬升/回落,与版本发布、定价调整或节假日一致。解读为“事件驱动的结构突变”,需在图上进行事件标注并评估后续回落幅度与留存面的变化。

二、季节性与周期性
- 周内周期:若产品有明显工作日/周末使用差异,WAU的周度锚定可掩盖日内波动,但在DAU曲线中可见;留存曲线若受周内习惯影响,可能在W1/W2形成轻微起伏。解读为“使用场景型季节性”,需用对齐周的 Cohort 以避免错配。
- 节假日效应:假期节点WAU与回流(Returning)上升,同时 Cohort 热力图在对应起始周更“浅色”(较高留存)。需在图表上显式打点注释,以便与常规周区分。

三、留存曲线形态(按周龄的Retention Curve)
- 快速下坠后稳定平台:前1-2周显著流失,随后趋于稳定形成“留存底”(如10%-20%)。解读为“核心用户群明确但新手引导不足”;优化应聚焦首周关键行为达成与召回。
- 凹形改善(早期更稳):近几期 Cohort 的前2周留存抬升,平台位相近。解读为“早期体验优化有效”,但长期价值未同步改善。
- 凸形拖累(长期留存走弱):W1改善而W4/W8下降。解读为“前期激励或营销提升了短期回访,但价值闭环/内容供给不足”。
- 平台位抬升:跨多个 Cohort 的中后期留存全面上移。解读为“产品价值或网络效应增强”,为最优质的改善信号。

四、Cohort 热力图模式(行=首周,列=周龄)
- 纵向整体变浅:新近 Cohort 比历史 Cohort 全列更浅(留存更高)。解读为产品/获客质量提升,通常与渠道结构优化或版本升级同步。
- 仅前几列变浅:W1-W2变浅但W4+无变化。解读为“早期体验修复”,应继续打通长期使用场景。
- 斜带条纹:隔周明暗交替,常见于营销节奏或周内季节性。需与投放/活动日历核对。
- 局部暗沉块:个别起始周开始的整行偏暗。解读为“异常周”(服务不稳定、版本缺陷或异常投放渠道),应在仪表盘中加异常标记与过滤。

五、WAU分解与关联解读
- WAU = 存量留存 + 回流(Reactivation) + 新增活跃。图表若显示回流带明显抬升,而留存率未改善,解读为“召回/活动驱动的短期回暖”,可持续性有限。
- 留存领先于可持续增长:当W1/W4留存先行上行,往往在2-6周后体现在WAU层面的自然增长(滞后效应)。建议在仪表盘增加“留存改善→WAU贡献”的预测带。
- 渠道/地域结构效应:若新增构成中低留存渠道占比提高,WAU短涨但后续面临更快回落。需叠加“按渠道的分布堆叠图+各渠道留存曲线”判断结构性风险。
- 产品版本/功能分群:功能采纳用户的 Cohort 留存显著高于未采纳人群。解读为“功能价值验证”,应在留存图中叠加分层曲线以强化因果指向。

六、异常与数据质量提示
- 基期错位:跨时区或周起始日不同会造成留存列错位,需统一周粒度定义(如ISO周)。
- 指标口径差异:滚动留存(Rolling/Return Rate)与经典留存(Fixed Cohort)不可混用解读。务必在图例中标明口径。
- 激活定义漂移:若“活跃”或“激活”口径调整(例如新增关键行为要求),曲线会出现伪跃迁,图中需版本线标注。

七、可视化与注释最佳实践
- 在WAU上叠加事件标注(版本发布、营销投放、节假日)与不确定区间(置信带),减少误判。
- Cohort 热力图采用感知均匀的连续色盘,并在色标上标出关键阈值(如W1=40%、W4=20%)。
- 提供维度筛选与分面(渠道/地域/版本/平台),支持对比最近N个Cohort的曲线叠加。
- 引入留存驱动分解视图:展示“Wk留存变化×Cohort规模”对未来WAU的贡献,帮助从“率”转化为“量”的影响评估。

通过上述模式与解读框架,可在不依赖具体数值的前提下,系统识别周活与留存图表中的信号,定位增长的来源与可持续性,并指导后续的产品与获客优化。

示例2

Below is a concise taxonomy of pattern types you should look for—and how to describe them—when reviewing an A/B conversion funnel one‑pager. Use these pattern descriptions directly or adapt them with your experiment’s numbers and confidence intervals.

1) Overall outcome and uncertainty
- Pattern: Net lift/decline in final conversion (e.g., Purchase) with confidence intervals.
- Description: Variant B increases final conversion by +X.xx pp (+Y.y% relative) vs A; 95% CI [L, U]. The credible interval excludes zero, indicating a statistically reliable effect.

2) Funnel shape and step-level effects
- Pattern: Consistent lift across steps vs localized bottlenecks.
- Description: The uplift is concentrated at [Step N], where step conversion improves by +X.pp; earlier/later steps are flat (CIs include zero), indicating a targeted effect rather than a funnel-wide shift.

- Pattern: Late-stage divergence.
- Description: Variants are similar through Add-to-Cart, but B underperforms at Payment (-X.pp step CVR), signaling increased friction in the checkout flow.

- Pattern: Early engagement boost with downstream loss.
- Description: B increases Product Views and Add-to-Cart, but Purchase does not improve due to degradation at [Shipping/Payment], producing leakage downstream (classic “front-load” effect).

3) Leakage redistribution and drop-off reasons
- Pattern: Drop-off migration between steps.
- Description: Total drop-off is not reduced but moves from [Step i] to [Step j]. Sankey/flow shows more exits at [specific reason], implying friction localized to that stage.

- Pattern: Micro-conversion vs macro-conversion mismatch.
- Description: Micro KPIs (e.g., clicks, add-to-cart) improve while macro conversion is flat/negative; indicates insufficient quality of progressed sessions or added cognitive load later.

4) Temporal stability
- Pattern: Day-over-day stability vs novelty/learning effects.
- Description: Uplift is front-loaded in the first N days and regresses toward zero; weekend behavior differs from weekdays. Time-series shows consistent direction but widening/narrowing CIs over time.

- Pattern: Delayed conversions (lag).
- Description: Same-session conversion improves immediately, but multi-session conversion catches up for A after day N. This lag sensitivity suggests waiting for the attribution window to close.

5) Segment heterogeneity
- Pattern: Device split divergence.
- Description: B outperforms on desktop (+X.pp) but is neutral/negative on mobile; pooled uplift is driven by desktop weight.

- Pattern: Traffic source dependence.
- Description: Organic traffic shows positive lift; paid social is flat/negative. Interaction between variant and acquisition channel is significant.

- Pattern: New vs returning users.
- Description: Returning users show uplift; new users do not. Indicates the change benefits familiarity rather than first-time comprehension.

- Pattern: Geography or locale variance.
- Description: Markets with [payment method/fulfillment] constraints show reduced step CVR at Payment/Shipping; consistent with localized friction.

6) Value vs rate trade-offs
- Pattern: Conversion rate vs AOV trade-off.
- Description: B increases conversion rate but decreases AOV by -X%; net revenue per visitor is [positive/negative] depending on margin assumptions.

- Pattern: Cart composition shifts.
- Description: Higher units per order but lower premium SKU share; revenue uplift is muted relative to conversion uplift.

7) Experiment health indicators (visible on one-pager)
- Pattern: Sample ratio mismatch (SRM).
- Description: Observed allocation deviates from expected (p < 0.01); results may be biased—investigate traffic bucketing or filters.

- Pattern: Imbalanced exposure by step.
- Description: Post-click filters (e.g., eligibility) reduce B exposure at deeper steps; interpret step CVRs with caution.

- Pattern: Underpowered segments.
- Description: Wide CIs in small-multiple panels; avoid over-interpreting segment-level directionality.

8) Latency and friction diagnostics
- Pattern: Increased time-to-next-step.
- Description: Median time from Add-to-Cart to Payment increases in B by +X seconds; aligns with lower Payment step CVR.

- Pattern: Error/validation spikes.
- Description: Error rate chart shows uptick at [form field/payment processor] for B, explaining step-level drop.

9) Consistency across visual modules
- Pattern: Alignment of topline, funnel bars, and uplift bars.
- Description: Final purchase uplift equals cumulative effect implied by step-level uplifts; no arithmetic inconsistencies.

- Pattern: Cohort consistency.
- Description: Weekly cohorts show similar effect sizes; absence of cohort drift supports generalizability.

How to phrase a complete, evidence-based summary (template)
- “Variant B increases Purchase by +X.xx pp (+Y.y% rel.; 95% CI [L, U]). The effect is localized to the [Step → Step] transition (+A.pp), while earlier steps are unchanged. Mobile underperforms (-B.pp), offset by a desktop gain (+C.pp), yielding a positive pooled effect due to desktop weight. Paid social shows no lift; organic search is positive. Time-series indicates stable uplift post day 3 without novelty decay. No SRM detected; power is sufficient at the final step. AOV is slightly lower (-D%), but net revenue/visitor remains +E%.”

If you provide the actual one-pager (or the per-step counts and conversion rates), I can replace placeholders with precise figures, compute absolute/relative lifts, and indicate which effects are statistically reliable.

示例3

I don’t have the chart or underlying data. To provide an accurate, chart-specific description, please share the visualization or a table of the key metrics by Channel × Creative (e.g., CTR, CVR, CPA/CPP, ROAS, Spend, Reach, Frequency, CI/error bars).

In the meantime, use the following structured observation template to describe patterns in a Channel × Creative effectiveness comparison. Replace placeholders with your chart’s values.

1) Channel-level performance gradient
- Ranking by primary KPI: <Channel A> leads on <KPI>, followed by <Channel B> and <Channel C>. The relative gap between the top and median channel is <X%>.
- Efficiency vs scale trade-off: <Channel A> achieves lower cost per result (CPA/CAC) but with lower reach, while <Channel B> provides higher reach at higher CPA. This indicates a scale–efficiency trade-off.

2) Within-channel creative dispersion
- Creative spread: In <Channel A>, creatives show a wide spread on <KPI> (range <min–max> or CV <X%>), indicating strong creative sensitivity. In <Channel B>, dispersion is narrow, suggesting creative choice has less impact than placement or audience.
- Top/bottom creatives: <Creative X> outperforms the channel median by <+Y%> on <KPI>, while <Creative Y> underperforms by <−Z%>.

3) Channel–creative interaction (ranking flips)
- Cross-over effects: <Creative X> ranks first in <Channel A> but drops below median in <Channel B>, indicating interaction effects rather than purely additive creative quality.
- Consistency: <Creative Z> performs consistently across channels (variance <low/medium/high>), making it a safer always-on asset.

4) Cost–outcome alignment
- Efficiency frontier: Plotting Spend vs <Outcome KPI>, <Channel A>/<Creative X> sits on the efficiency frontier (highest outcome per unit spend). <Channel C>/<Creative Y> is dominated (higher cost, lower outcome), a candidate for reallocation.
- Diminishing returns: Within <Channel B>, marginal performance declines at higher spend tiers (flattening slope), suggesting saturation or frequency fatigue.

5) Engagement vs conversion trade-offs
- Funnel divergence: <Channel/Creative> shows high engagement (CTR, VTR) but low conversion efficiency (CVR/CPA), implying message attracts clicks/views without qualification. Conversely, <Channel/Creative> with lower CTR but higher CVR indicates qualified traffic.
- Post-click quality: Time-on-site/bounce rate (if available) corroborates differences in traffic quality.

6) Uncertainty and stability
- Confidence intervals: Performance differences between <Creative M> and <Creative N> in <Channel A> overlap within CI, so the observed gap may not be statistically significant.
- Temporal stability: Week-over-week metrics for <Channel/Creative> show <stable/volatile> trends; volatility suggests sensitivity to auction dynamics or audience fatigue.

7) Audience/placement effects (if segmented)
- Segment heterogeneity: In <Segment S>, the lift for <Creative X> is <+Y%> vs baseline; in <Segment T>, the effect reverses, supporting targeted routing of creatives to segments where they resonate.

8) Actionable implications
- Scale winners: Increase budget for <Channel/Creative> combinations on the frontier; maintain frequency caps to avoid saturation.
- Test focus: In channels with high creative dispersion, prioritize iterative creative testing; in channels with low dispersion, optimize audience/placement/bid strategy.
- Remove dominated units: Pause <Channel/Creative> combinations below the 25th percentile on primary KPI with adequate spend/CI coverage.

Optional phrasing examples (replace placeholders):
- “Across channels, <Channel A> delivers the lowest CPA (−<X>% vs median) at moderate reach, while <Channel B> scales better but at +<Y>% higher CPA.”
- “Within <Channel A>, creative performance is highly dispersed (IQR = <value>), with <Creative X> outperforming the channel median by <+Y%> CVR.”
- “<Creative X> ranks top in <Channel A> but bottom in <Channel B>, indicating a channel–creative interaction that warrants channel-specific creative routing.”
- “At higher spend deciles in <Channel B>, ROAS plateaus, suggesting diminishing returns; shifting marginal budget to <Channel A>/<Creative X> is likely accretive.”

If you share the chart or a small table of Channel × Creative × Metrics, I will produce a precise, chart-specific pattern summary and recommendations.

适用用户

数据分析师/BI工程师

快速判断用什么图、怎么摆放;一键生成配色与标注;输出趋势与异常解读,用作周报与复盘的文字说明。

产品经理

将漏斗、留存、A/B结果可视化成一页图;自动突出关键结论与行动点;定义仪表盘结构,统一团队指标视角。

增长/市场运营

对比渠道与素材效果,生成可读竞品与投放报告;定位异常波动并给出验证路径,指导下一步实验。

电商/零售经营者

把销售、库存、复购率以看板呈现;按节日活动一键切换主题与语言;每日生成门店或类目趋势解读。

财务分析师/CFO

可视化营收、成本、现金流与周期性;生成董事会材料中的核心图块与配文;强调风险点与情景分析。

学术研究者/数据记者

将复杂数据讲得清楚不失严谨;选择适合期刊或媒体的图型与配色;输出图注与方法说明,避免误读。

教培从业者/讲师

把案例数据转为教学友好图示;按学生水平自动调整表述;快速生成课件中的图表解说稿。

解决的问题

用一次输入,得到“更会说话的图表”。本提示词面向数据分析、产品、运营、市场与BI团队,帮助你在分钟级产出三件事: - 清晰、可复用的图表模式解读(发生了什么、证据是什么、可能为什么) - 可直接落地的可视化设计建议(图表类型、编码方式、配色与布局、交互与注释) - 针对业务场景的行动建议与仪表盘结构梳理 核心价值:让复杂数据更易懂、报告更有说服力、决策更快更稳,并支持按主题与语言自定义输出,保持输出客观、精准、结构化。

特征总结

自动解读图表趋势、周期与异常,用通俗语言说明业务含义,助力快速同步团队认知
智能推荐最合适图表与布局,一键生成配色、标注与图例,减少反复试错
以业务目标为导向,自动提出可视化叙事结构,突出重点与结论行动点
针对不同受众自动调整专业度与细节,轻松生成高层简报或一线执行视图
模板化参数设置,按主题与语言一键调用,快速复用到月报、竞品分析与复盘
自动识别异常与离群点,提供可能成因与验证路径,支持后续实验或决策
输出具体的配色、字体与留白建议,自动优化阅读层次与对比度,提升可读性
生成可执行的仪表盘结构与图块清单,明确关键指标、交互与刷新节奏
提供数据清洗与汇总前的可视化思路,帮助团队先看问题再投入建模资源
保持专业、客观与可验证的表达,减少误读与夸大,增强报告的可信度

如何使用购买的提示词模板

1. 直接在外部 Chat 应用中使用

将模板生成的提示词复制粘贴到您常用的 Chat 应用(如 ChatGPT、Claude 等),即可直接对话使用,无需额外开发。适合个人快速体验和轻量使用场景。

2. 发布为 API 接口调用

把提示词模板转化为 API,您的程序可任意修改模板参数,通过接口直接调用,轻松实现自动化与批量处理。适合开发者集成与业务系统嵌入。

3. 在 MCP Client 中配置使用

在 MCP client 中配置对应的 server 地址,让您的 AI 应用自动调用提示词模板。适合高级用户和团队协作,让提示词在不同 AI 工具间无缝衔接。

¥20.00元
平台提供免费试用机制,
确保效果符合预期,再付费购买!

您购买后可以获得什么

获得完整提示词模板
- 共 235 tokens
- 2 个可调节参数
{ 图表主题 } { 输出语言 }
自动加入"我的提示词库"
- 获得提示词优化器支持
- 版本化管理支持
获得社区共享的应用案例
限时免费

不要错过!

免费获取高级提示词-优惠即将到期

17
:
23
小时
:
59
分钟
:
59