jpskill.com
💬 コミュニケーション コミュニティ 🟢 非エンジニアでもOK 👤 管理職・人事・カスタマー対応

💬 Competitive Teardown

competitive-teardown

競合他社の製品や企業を、価格情報、

⏱ メール返信10件 30分 → 3分

📺 まず動画で見る(YouTube)

▶ 【最新版】Claude(クロード)完全解説!20以上の便利機能をこの動画1本で全て解説 ↗

※ jpskill.com 編集部が参考用に選んだ動画です。動画の内容と Skill の挙動は厳密には一致しないことがあります。

📜 元の英語説明(参考)

Analyzes competitor products and companies by synthesizing data from pricing pages, app store reviews, job postings, SEO signals, and social media into structured competitive intelligence. Produces feature comparison matrices scored across 12 dimensions, SWOT analyses, positioning maps, UX audits, pricing model breakdowns, action item roadmaps, and stakeholder presentation templates. Use when conducting competitor analysis, comparing products against competitors, researching the competitive landscape, building battle cards for sales, preparing for a product strategy or roadmap session, responding to a competitor's new feature or pricing change, or performing a quarterly competitive review.

🇯🇵 日本人クリエイター向け解説

一言でいうと

競合他社の製品や企業を、価格情報、

※ jpskill.com 編集部が日本のビジネス現場向けに補足した解説です。Skill本体の挙動とは独立した参考情報です。

⚡ おすすめ: コマンド1行でインストール(60秒)

下記のコマンドをコピーしてターミナル(Mac/Linux)または PowerShell(Windows)に貼り付けてください。 ダウンロード → 解凍 → 配置まで全自動。

🍎 Mac / 🐧 Linux
mkdir -p ~/.claude/skills && cd ~/.claude/skills && curl -L -o competitive-teardown.zip https://jpskill.com/download/4618.zip && unzip -o competitive-teardown.zip && rm competitive-teardown.zip
🪟 Windows (PowerShell)
$d = "$env:USERPROFILE\.claude\skills"; ni -Force -ItemType Directory $d | Out-Null; iwr https://jpskill.com/download/4618.zip -OutFile "$d\competitive-teardown.zip"; Expand-Archive "$d\competitive-teardown.zip" -DestinationPath $d -Force; ri "$d\competitive-teardown.zip"

完了後、Claude Code を再起動 → 普通に「動画プロンプト作って」のように話しかけるだけで自動発動します。

💾 手動でダウンロードしたい(コマンドが難しい人向け)
  1. 1. 下の青いボタンを押して competitive-teardown.zip をダウンロード
  2. 2. ZIPファイルをダブルクリックで解凍 → competitive-teardown フォルダができる
  3. 3. そのフォルダを C:\Users\あなたの名前\.claude\skills\(Win)または ~/.claude/skills/(Mac)へ移動
  4. 4. Claude Code を再起動

⚠️ ダウンロード・利用は自己責任でお願いします。当サイトは内容・動作・安全性について責任を負いません。

🎯 このSkillでできること

下記の説明文を読むと、このSkillがあなたに何をしてくれるかが分かります。Claudeにこの分野の依頼をすると、自動で発動します。

📦 インストール方法 (3ステップ)

  1. 1. 上の「ダウンロード」ボタンを押して .skill ファイルを取得
  2. 2. ファイル名の拡張子を .skill から .zip に変えて展開(macは自動展開可)
  3. 3. 展開してできたフォルダを、ホームフォルダの .claude/skills/ に置く
    • · macOS / Linux: ~/.claude/skills/
    • · Windows: %USERPROFILE%\.claude\skills\

Claude Code を再起動すれば完了。「このSkillを使って…」と話しかけなくても、関連する依頼で自動的に呼び出されます。

詳しい使い方ガイドを見る →
最終更新
2026-05-17
取得日時
2026-05-17
同梱ファイル
6

💬 こう話しかけるだけ — サンプルプロンプト

  • Competitive Teardown で、お客様への返信文を作って
  • Competitive Teardown を使って、社内向けアナウンスを書いて
  • Competitive Teardown で、メールテンプレートを整備して

これをClaude Code に貼るだけで、このSkillが自動発動します。

📖 Claude が読む原文 SKILL.md(中身を展開)

この本文は AI(Claude)が読むための原文(英語または中国語)です。日本語訳は順次追加中。

Competitive Teardown

Tier: POWERFUL
Category: Product Team
Domain: Competitive Intelligence, Product Strategy, Market Analysis


When to Use

  • Before a product strategy or roadmap session
  • When a competitor launches a major feature or pricing change
  • Quarterly competitive review
  • Before a sales pitch where you need battle card data
  • When entering a new market segment

Teardown Workflow

Follow these steps in sequence to produce a complete teardown:

  1. Define competitors — List 2–4 competitors to analyze. Confirm which is the primary focus.
  2. Collect data — Use references/data-collection-guide.md to gather raw signals from at least 3 sources per competitor (website, reviews, job postings, SEO, social).
    Validation checkpoint: Before proceeding, confirm you have pricing data, at least 20 reviews, and job posting counts for each competitor.
  3. Score using rubric — Apply the 12-dimension rubric below to produce a numeric scorecard for each competitor and your own product.
    Validation checkpoint: Every dimension should have a score and at least one supporting evidence note.
  4. Generate outputs — Populate the templates in references/analysis-templates.md (Feature Matrix, Pricing Analysis, SWOT, Positioning Map, UX Audit).
  5. Build action plan — Translate findings into the Action Items template (quick wins / medium-term / strategic).
  6. Package for stakeholders — Assemble the Stakeholder Presentation using outputs from steps 3–5.

Data Collection Guide

Full executable scripts for each source are in references/data-collection-guide.md. Summaries of what to capture are below.

1. Website Analysis

Key things to capture:

  • Pricing tiers and price points
  • Feature lists per tier
  • Primary CTA and messaging
  • Case studies / customer logos (signals ICP)
  • Integration logos
  • Trust signals (certifications, compliance badges)

2. App Store Reviews

Review sentiment categories:

  • Praise → what users love (defend / strengthen these)
  • Feature requests → unmet needs (opportunity gaps)
  • Bugs → quality signals
  • UX complaints → friction points you can beat them on

Sample App Store query (iTunes Search API):

GET https://itunes.apple.com/search?term=<competitor_name>&entity=software&limit=1
# Extract trackId, then:
GET https://itunes.apple.com/rss/customerreviews/id=<trackId>/sortBy=mostRecent/json?l=en&limit=50

Parse entry[].content.label for review text and entry[].im:rating.label for star rating.

3. Job Postings (Team Size & Tech Stack Signals)

Signals from job postings:

  • Engineering volume → scaling vs. consolidating
  • Specific tech mentions → stack (React/Vue, Postgres/Mongo, AWS/GCP)
  • Sales/CS ratio → product-led vs. sales-led motion
  • Data/ML roles → upcoming AI features
  • Compliance roles → regulatory expansion

4. SEO Analysis

SEO signals to capture:

  • Top 20 organic keywords (intent: informational / navigational / commercial)
  • Domain Authority / backlink count
  • Blog publishing cadence and topics
  • Which pages rank (product pages vs. blog vs. docs)

5. Social Media Sentiment

Capture recent mentions via Twitter/X API v2, Reddit, or LinkedIn. Look for recurring praise, complaints, and feature requests. See references/data-collection-guide.md for API query examples.


Scoring Rubric (12 Dimensions, 1-5)

# Dimension 1 (Weak) 3 (Average) 5 (Best-in-class)
1 Features Core only, many gaps Solid coverage Comprehensive + unique
2 Pricing Confusing / overpriced Market-rate, clear Transparent, flexible, fair
3 UX Confusing, high friction Functional Delightful, minimal friction
4 Performance Slow, unreliable Acceptable Fast, high uptime
5 Docs Sparse, outdated Decent coverage Comprehensive, searchable
6 Support Email only, slow Chat + email 24/7, great response
7 Integrations 0-5 integrations 6-25 26+ or deep ecosystem
8 Security No mentions SOC2 claimed SOC2 Type II, ISO 27001
9 Scalability No enterprise tier Mid-market ready Enterprise-grade
10 Brand Generic, unmemorable Decent positioning Strong, differentiated
11 Community None Forum / Slack Active, vibrant community
12 Innovation No recent releases Quarterly Frequent, meaningful

Example completed row (Competitor: Acme Corp, Dimension 3 – UX):

Dimension Acme Corp Score Evidence
UX 2 App Store reviews cite "confusing navigation" (38 mentions); onboarding requires 7 steps before TTFV; no onboarding wizard; CC required at signup.

Apply this pattern to all 12 dimensions for each competitor.


Templates

Full template markdown is in references/analysis-templates.md. Abbreviated reference below.

Feature Comparison Matrix

Rows: core features, pricing tiers, platform capabilities (web, iOS, Android, API).
Columns: your product + up to 3 competitors.
Score each cell 1–5. Sum to get total out of 60.
Score legend: 5=Best-in-class, 4=Strong, 3=Average, 2=Below average, 1=Weak/Missing

Pricing Analysis

Capture per competitor: model type (per-seat / usage-based / flat rate / freemium), entry/mid/enterprise price points, free trial length.
Summarize: price leader, value leader, premium positioning, your position, and 2–3 pricing opportunity bullets.

SWOT Analysis

For each competitor: 3–5 bullets per quadrant (Strengths, Weaknesses, Opportunities for us, Threats to us). Anchor every bullet to a data signal (review quote, job posting count, pricing page, etc.).

Positioning Map

2x2 axes (e.g., Simple ↔ Complex / Low Value ↔ High Value). Place each competitor and your product. Bubble size = market share or funding. See references/analysis-templates.md for ASCII and editable versions.

UX Audit Checklist

Onboarding: TTFV (minutes), steps to activation, CC-required, onboarding wizard quality.
Key workflows: steps, friction points, comparative score (yours vs. theirs).
Mobile: iOS/Android ratings, feature parity, top complaint and praise.
Navigation: global search, keyboard shortcuts, in-app help.

Action Items

Horizon Effort Examples
Quick wins (0–4 wks) Low Add review badges, publish comparison landing page
Medium-term (1–3 mo) Moderate Launch free tier, improve onboarding TTFV, add top-requested integration
Strategic (3–12 mo) High Enter new market, build API v2, achieve SOC2 Type II

Stakeholder Presentation (7 slides)

  1. Executive Summary — Threat level (LOW/MEDIUM/HIGH/CRITICAL), top strength, top opportunity, recommended action
  2. Market Position — 2x2 positioning map
  3. Feature Scorecard — 12-dimension radar or table, total scores
  4. Pricing Analysis — Comparison table + key insight
  5. UX Highlights — What they do better (3 bullets) vs. where we win (3 bullets)
  6. Voice of Customer — Top 3 review complaints (quoted or paraphrased)
  7. Our Action Plan — Quick wins, medium-term, strategic priorities; Appendix with raw data

Related Skills

  • Product Strategist (product-team/product-strategist/) — Competitive insights feed OKR and strategy planning
  • Landing Page Generator (product-team/landing-page-generator/) — Competitive positioning informs landing page messaging

同梱ファイル

※ ZIPに含まれるファイル一覧。`SKILL.md` 本体に加え、参考資料・サンプル・スクリプトが入っている場合があります。