🏢Databases and Warehouses Skills
Browse skills in the Databases and Warehouses category.
Streamlit Snowflake
A powerful skill for Claude agents.
Cloudflare Hyperdrive
A powerful skill for Claude agents.
Cloudflare Vectorize
A powerful skill for Claude agents.
wine-data-importer
Imports wine data from CSV files, spreadsheets, or structured text. Use when user mentions "import wines", "bulk import", "import from CSV", "import from spreadsheet", or "add wines from file".
data-engineer
Build ETL pipelines, data warehouses, and streaming architectures. Implements Spark jobs, Airflow DAGs, and Kafka streams. Use PROACTIVELY for data pipeline design or analytics infrastructure.
cloudbase-document-database-in-wechat-miniprogram
Use CloudBase document database WeChat MiniProgram SDK to query, create, update, and delete data. Supports complex queries, pagination, aggregation, and geolocation queries.
supabase-helper
Common Supabase operations and workflows
supabase
Supabase open-source Firebase alternative with PostgreSQL database, authentication, storage, real-time subscriptions, and edge functions
building-with-supabase
AI agent builds full-stack applications with Supabase PostgreSQL, authentication, Row Level Security, Edge Functions, and real-time subscriptions. Use when building apps with Supabase, implementing RLS policies, or setting up Supabase Auth.
developing-with-redis
The agent implements Redis caching, data structures, and real-time messaging patterns. Use when implementing caching layers, session storage, rate limiting, pub/sub messaging, or distributed data structures.
mongoose
Mongoose ODM for MongoDB and Node.js. Use for schemas, models, queries, validation, middleware, population, aggregation, and MongoDB database operations in JavaScript/TypeScript applications.
lamindb
This skill should be used when working with LaminDB, an open-source data framework for biology that makes data queryable, traceable, reproducible, and FAIR. Use when managing biological datasets (scRNA-seq, spatial, flow cytometry, etc.), tracking computational workflows, curating and validating data with biological ontologies, building data lakehouses, or ensuring data lineage and reproducibility in biological research. Covers data management, annotation, ontologies (genes, cell types, diseases, tissues), schema validation, integrations with workflow managers (Nextflow, Snakemake) and MLOps platforms (W&B, MLflow), and deployment strategies.
supabase-patterns
Use when working with Supabase database queries, RLS policies, Edge Functions, migrations, real-time subscriptions, or storage. Ensures best practices for backend operations and data security.
cloudbase-document-database-web-sdk
Use CloudBase document database Web SDK to query, create, update, and delete data. Supports complex queries, pagination, aggregation, and geolocation queries.
dst-data
Fetch actual data from Danmarks Statistik API and store in DuckDB. Use when user wants to download and store specific DST table data for analysis.
apache-nifi
Expert guidance for Apache NiFi data integration platform including flow design, processors, controller services, process groups, NiFi Registry integration, and cluster configuration. Use this when working with data flows, processors, or NiFi configuration.
redis
Redis caching, pub/sub, and data structures. Activate for Redis operations, caching strategies, session management, and message queuing.
supabase-service-usage
アプリ全域の Supabase 利用ルール。サービス層の統一、Service Role の安全な使い分け、ログ付与、Server Actions 連携を規定。
firebase-operations
Performs Firebase Firestore operations. Use when querying collections, creating/updating/deleting documents, using batch writes, or working with Timestamps. Includes pagination, transactions, and security rules patterns.
python-data-engineering
Comprehensive Python data engineering patterns for AWS Data Lake, including PySpark, Pandas, Apache Airflow, AWS Glue, ETL pipelines, data quality, schema management, performance optimization, FastAPI services, streaming with Kafka/Kinesis, data validation with Great Expectations, testing strategies, error handling, logging, and production deployment on AWS EMR and Glue.
databricks-notebook-manager
Create and manage Databricks notebooks programmatically. Use when generating ingestion code, creating ETL notebooks, executing Databricks workflows, or when user mentions notebook creation, job automation, or data pipeline implementation in Databricks. Handles notebook generation, execution, and results retrieval.
supabase-integration
Supabase 프로젝트 설정, 데이터베이스 설계, RLS 정책, Edge Functions, 인증/권한 통합 전문 스킬. Supabase CLI 및 클라이언트 라이브러리 활용.
local-clickhouse
Install, configure, and validate local ClickHouse for gapless-crypto-clickhouse development and backtesting. Use when setting up local development environment, enabling offline mode, improving query performance for backtesting, or running E2E validation. Includes mise/Homebrew/apt installation, mode detection, connection validation, and E2E workflow scripts.
relational-database-web-cloudbase
Use when building frontend Web apps that talk to CloudBase Relational Database via @cloudbase/js-sdk – provides the canonical init pattern so you can then use Supabase-style queries from the browser.
unity-catalog-tagger
Manage Unity Catalog metadata tags for data governance and classification. Use when applying tags to tables and columns, classifying data sensitivity (PII, PHI), marking data quality attributes, or when user mentions Unity Catalog tagging, metadata management, data governance, or compliance workflows.
databricks-query
Execute SQL queries against Databricks using the DBSQL MCP server. Use when querying Unity Catalog tables, running SQL analytics, exploring Databricks data, or when user mentions Databricks queries, SQL execution, Unity Catalog, or data warehouse operations. Handles query execution, result formatting, and error handling.
qdrant-doc-deleter
Deletes specific PDF documents from Qdrant vector database collection. Use when user wants to remove, delete, or clean up PDF documents from the vector database, Qdrant collection, or needs to manage document versions.
data-lake-management
Data Lake architecture and management including medallion architecture (bronze/silver/gold zones), data catalog with AWS Glue, partitioning strategies, schema evolution, data quality, governance, cost optimization, S3 lifecycle policies, data retention, compliance, query optimization with Athena, data formats (Parquet, ORC, Avro), incremental processing, CDC patterns, and production best practices for scalable data lakes.
notion-mcp-developer
Develop Notion templates and databases using Notion MCP tools in Claude Code. Orchestrates escape room design skills (narrative-architect, puzzle-designer, formula-master, localizer, playtester) and implements their output via Notion MCP API. Use when creating Notion templates, building databases, implementing game mechanics, or managing complex multi-step Notion development workflows. Handles rate limits, session persistence via Serena MCP, and iterative development cycles.
julien-infra-hostinger-database
Manage shared database instances on Hostinger VPS srv759970 - PostgreSQL, Redis, MongoDB operations. Use for database connections, backups, user management, performance checks, or troubleshooting database issues.
genie-integration
Integrate Databricks Genie rooms as tools in agent workflows. Use when integrating Genie spaces with AI agents, querying Genie rooms programmatically via SDK or MCP, managing Genie conversations and polling, handling Genie API responses and errors, or building tool-calling agents that use Genie as a data source. Covers SDK patterns, MCP tool integration, conversation management, error handling, and performance optimization for Genie-based agent tools.
supabase-auth
Manage authentication and user operations in Supabase. Use for sign up, sign in, sign out, password resets, and user management.
supabase-realtime
Subscribe to realtime changes in Supabase using WebSocket connections. Use for listening to database changes, presence tracking, and broadcast messaging.
redis
Redis in-memory data structure store - use for caching, session management, pub/sub, real-time analytics, and key-value storage
nexla-data-flows-operator
Build, deploy, monitor, and troubleshoot production Nexla data pipelines via Python SDK or REST API. Use for flow setup, data transformation pipelines, schema management, access control, credential rotation, batch operations, error recovery, monitoring, CI/CD integration, and operational troubleshooting.
moai-platform-firestore
Firebase Firestore specialist covering NoSQL patterns, real-time sync, offline caching, and Security Rules. Use when building mobile-first apps with offline support, implementing real-time listeners, or configuring Firestore security.
say-ducklake-xor
Parallel thread/DuckLake discovery with XOR uniqueness from gay_seed. Finds "say" or MCP usage, cross-refs with all DuckDB sources, launches bounded parallel ops.
geohash-coloring
Geohash Coloring Skill
supabase-db
Supabase database, auth, and storage operations
redis-cache
Redis caching operations and data structures
pinecone-vector
Pinecone vector database for similarity search and RAG
neon-database-manager
Tool for managing Neon Postgres databases. Use for listing projects, inspecting schema, and running queries.
data-warehouse
Data warehouse (大数据/数仓) ops skill for designing and operating analytical data platforms. Use for tasks like defining source-of-truth, building ETL/ELT pipelines, dimensional modeling (star schema), data quality checks, partitioning strategies, cost/performance tuning, governance, lineage, and SLA monitoring.
algolia-search
Algolia search indexing and querying
mongodb-atlas
MongoDB Atlas cloud database operations
feature-stores
Master feature stores - Feast, data validation, versioning, online/offline serving
notion-llm-config
Manages the Notion LLM Config Table database tracking model configurations and parameters across multiple model providers. Use when working with the LLM config table, adding models, updating model parameters, or querying model architecture details.
supabase-expert
Advanced Supabase integration specialist for Auth, Database (PostgreSQL/RLS), Storage, Realtime, Edge Functions, and AI/Vector features. Use when implementing Supabase features, debugging Supabase issues, setting up RLS policies, creating database schemas, building auth flows, optimizing Supabase queries, migrating to Supabase, or architecting Supabase-based applications. Invoke for Supabase client setup, type generation, migration creation, performance tuning, security audits, or Supabase best practices. Handles Next.js, React, Vue, Svelte, and server-side integrations.
data-transformation
データ変換パイプラインの設計・実装・検証を整理するスキル。スキーママッピング、ETL設計、品質確認までの実務フローを提供する。Anchors:• Designing Data-Intensive Applications / 適用: データモデリング / 目的: 変換の整合性確保• Designing Data-Intensive Applications / 適用: スキーマ設計 / 目的: マッピングの明確化• Designing Data-Intensive Applications / 適用: パイプライン設計 / 目的: 伸縮性と監視性の確保Trigger:Use when designing data transformation pipelines, defining schema mappings, implementing ETL processes, or optimizing data flows.data transformation, schema mapping, etl design, pipeline optimization, data modeling
agentuity-cli-cloud-vector-delete
Delete one or more vectors by key. Requires authentication. Use for Agentuity cloud platform operations