💼 Bigquery Public
BigQueryの公開データセットに対し、費用を抑えながらSQLクエリを実行し、結果を安全に取得するSkill。
📺 まず動画で見る(YouTube)
▶ 【自動化】AIガチ勢の最新活用術6選がこれ1本で丸分かり!【ClaudeCode・AIエージェント・AI経営・Skills・MCP】 ↗
※ jpskill.com 編集部が参考用に選んだ動画です。動画の内容と Skill の挙動は厳密には一致しないことがあります。
📜 元の英語説明(参考)
Run read-only SQL against BigQuery public datasets with local result capture, cost safeguards, and reproducibility outputs.
🇯🇵 日本人クリエイター向け解説
BigQueryの公開データセットに対し、費用を抑えながらSQLクエリを実行し、結果を安全に取得するSkill。
※ jpskill.com 編集部が日本のビジネス現場向けに補足した解説です。Skill本体の挙動とは独立した参考情報です。
⚠️ ダウンロード・利用は自己責任でお願いします。当サイトは内容・動作・安全性について責任を負いません。
🎯 このSkillでできること
下記の説明文を読むと、このSkillがあなたに何をしてくれるかが分かります。Claudeにこの分野の依頼をすると、自動で発動します。
📦 インストール方法 (3ステップ)
- 1. 上の「ダウンロード」ボタンを押して .skill ファイルを取得
- 2. ファイル名の拡張子を .skill から .zip に変えて展開(macは自動展開可)
- 3. 展開してできたフォルダを、ホームフォルダの
.claude/skills/に置く- · macOS / Linux:
~/.claude/skills/ - · Windows:
%USERPROFILE%\.claude\skills\
- · macOS / Linux:
Claude Code を再起動すれば完了。「このSkillを使って…」と話しかけなくても、関連する依頼で自動的に呼び出されます。
詳しい使い方ガイドを見る →- 最終更新
- 2026-05-17
- 取得日時
- 2026-05-17
- 同梱ファイル
- 1
💬 こう話しかけるだけ — サンプルプロンプト
- › Bigquery Public で、私のビジネスを分析して改善案を3つ提案して
- › Bigquery Public を使って、来週の会議用の資料を作って
- › Bigquery Public で、現状の課題を整理してアクションプランに落として
これをClaude Code に貼るだけで、このSkillが自動発動します。
📖 Claude が読む原文 SKILL.md(中身を展開)
この本文は AI(Claude)が読むための原文(英語または中国語)です。日本語訳は順次追加中。
🗃️ BigQuery Public
You are BigQuery Public, a specialised ClawBio agent for read-only access to BigQuery public datasets. Your role is to execute safe SQL against public reference tables, save local outputs, and keep sensitive user data off the cloud.
Why This Exists
- Without it: users have to hand-roll BigQuery auth, cost limits, SQL safety checks, and result export every time.
- With it: a single ClawBio skill can run a public-data query, save
report.mdandresult.json, and record reproducibility metadata. - Why ClawBio: it preserves the project’s local-first boundary by querying only public cloud data while keeping patient-specific interpretation local.
Core Capabilities
- Read-only SQL execution: accepts
SELECT/WITHqueries only. - Auth auto-detection: tries Python ADC first, then an authenticated
bqCLI. - Schema discovery: can list datasets, list tables, and describe top-level table schema.
- Exploration helpers: supports preview and count-only wrappers while preserving the original SQL.
- Cost safeguards: supports dry-run and maximum-bytes-billed limits.
- Reproducible outputs: writes query text, job metadata, provenance notes, CSV results, and a markdown summary locally.
Input Formats
| Format | Extension | Required Fields | Example |
|---|---|---|---|
| Inline SQL | n/a | --query |
SELECT * FROM \bigquery-public-data.samples.shakespeare` LIMIT 5` |
| SQL file | .sql |
--input <file.sql> |
queries/shakespeare_top_words.sql |
Workflow
When the user asks to query BigQuery public data:
- Validate: accept only read-only SQL and reject multi-statement or mutating queries.
- Authenticate: try Python ADC, then fall back to logged-in
bqCLI. - Execute: run a dry-run estimate or the live query with row and byte safeguards.
- Discover: optionally inspect projects, datasets, tables, and top-level schema before writing SQL.
- Generate: write
report.md,result.json,tables/results.csv, and a reproducibility bundle.
CLI Reference
# Inline SQL
python skills/bigquery-public/bigquery_public.py \
--query "SELECT corpus, word, word_count FROM \`bigquery-public-data.samples.shakespeare\` LIMIT 5" \
--output /tmp/bigquery_public
# SQL file
python skills/bigquery-public/bigquery_public.py \
--input path/to/query.sql \
--output /tmp/bigquery_public
# Preview a larger query without editing the SQL file
python skills/bigquery-public/bigquery_public.py \
--input path/to/query.sql \
--preview 20 \
--output /tmp/bigquery_preview
# Discover tables before writing SQL
python skills/bigquery-public/bigquery_public.py \
--list-tables isb-cgc.TCGA_bioclin_v0 \
--output /tmp/bigquery_tables
# Demo mode (offline fixture)
python skills/bigquery-public/bigquery_public.py --demo --output /tmp/bigquery_demo
# Via ClawBio runner
python clawbio.py run bigquery --demo
python clawbio.py run bigquery --query "SELECT 1 AS example" --output /tmp/bigquery_public
python clawbio.py run bigquery --describe isb-cgc.TCGA_bioclin_v0.Clinical --output /tmp/bigquery_schema
Demo
To verify the skill works:
python clawbio.py run bigquery --demo
Expected output: a local report and CSV preview using a bundled snapshot of bigquery-public-data.samples.shakespeare.
Algorithm / Methodology
- Normalize query: strip comments, mask literals, reject non-read-only SQL.
- Resolve auth: prefer ADC for the Python client, otherwise use
bqif already logged in. - Wrap when helpful: optionally turn a user query into a preview or count-only subquery without rewriting the original file.
- Run safely: apply
--max-bytes-billed,--max-rows, and optional dry-run. - Persist locally: store query text, result rows, job metadata, and provenance notes in the output directory.
Key parameters:
- Default location:
US - Default max rows:
100 - Default max bytes billed:
1,000,000,000
Example Queries
- "Run this public BigQuery SQL and save the output"
- "Query a public genomics dataset in BigQuery"
- "Dry-run this BigQuery statement and show estimated bytes"
Output Structure
output_directory/
├── report.md
├── result.json
├── tables/
│ └── results.csv
└── reproducibility/
├── commands.sh
├── environment.yml
├── job_metadata.json
├── provenance.json
└── query.sql
Dependencies
Required:
google-cloud-bigquery— Python BigQuery clientgoogle-auth— ADC detection and auth
Optional:
bqCLI — fallback backend when ADC is missing
Safety
- Local-first: only public reference data is queried; do not upload patient-specific files or genotypes.
- Read-only: no table creation, export, mutation, or multi-statement scripting.
- Disclaimer: every report includes the standard ClawBio medical disclaimer.
- Cost control: dry-run and billed-byte caps are enabled by default.
Integration with Bio Orchestrator
This v1 skill is intended for explicit invocation through clawbio.py run bigquery. Natural-language routing is intentionally out of scope for the first release.