senior-data-scientist

Senior Data Scientist

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "senior-data-scientist" with this command: npx skills add rickydwilson-dcs/claude-skills/rickydwilson-dcs-claude-skills-senior-data-scientist

Senior Data Scientist

World-class senior data scientist skill for production-grade AI/ML/Data systems.

Overview

This skill provides world-class data science capabilities through three core Python automation tools and comprehensive reference documentation. Whether designing experiments, building predictive models, performing causal inference, or driving data-driven decisions, this skill delivers expert-level statistical modeling and analytics solutions.

Senior data scientists use this skill for A/B testing, experiment design, statistical modeling, causal inference, time series analysis, feature engineering, model evaluation, and business intelligence. Expertise covers Python (NumPy, Pandas, Scikit-learn), R, SQL, statistical methods, hypothesis testing, and advanced analytics techniques.

Core Value: Accelerate analytics and experimentation by 65%+ while improving model accuracy, statistical rigor, and business impact through proven methodologies and automated pipelines.

Quick Start

Main Capabilities

Core Tool 1

python scripts/experiment_designer.py --input data/ --output results/

Core Tool 2

python scripts/feature_engineering_pipeline.py --target project/ --analyze

Core Tool 3

python scripts/model_evaluation_suite.py --config config.yaml --deploy

Core Capabilities

  • Experiment Design & A/B Testing - Statistical power analysis, sample size calculation, multi-armed bandits, sequential testing

  • Statistical Modeling - Regression, classification, time series, causal inference, Bayesian methods

  • Feature Engineering - Automated feature generation, selection, transformation, interaction terms, dimensionality reduction

  • Model Evaluation - Cross-validation, hyperparameter tuning, bias-variance tradeoff, model interpretation (SHAP, LIME)

  • Business Analytics - Customer segmentation, churn prediction, lifetime value, attribution modeling, forecasting

  • Causal Inference - Propensity score matching, difference-in-differences, instrumental variables, regression discontinuity

Python Tools

  1. Experiment Designer

Design statistically rigorous experiments with power analysis.

Key Features:

  • A/B test design with sample size calculation

  • Statistical power analysis

  • Multi-variant testing setup

  • Sequential testing frameworks

  • Bayesian experiment design

Common Usage:

Design A/B test

python scripts/experiment_designer.py --effect-size 0.05 --power 0.8 --alpha 0.05

Multi-variant test

python scripts/experiment_designer.py --variants 4 --mde 0.03 --output experiment_plan.json

Sequential testing

python scripts/experiment_designer.py --sequential --stopping-rule obf

Help

python scripts/experiment_designer.py --help

Use Cases:

  • Designing product experiments before launch

  • Calculating required sample sizes

  • Planning sequential testing strategies

  1. Feature Engineering Pipeline

Automate feature generation, selection, and transformation.

Key Features:

  • Automated feature generation (polynomial, interaction terms)

  • Feature selection (mutual information, recursive elimination)

  • Encoding (one-hot, target, frequency)

  • Scaling and normalization

  • Dimensionality reduction (PCA, t-SNE, UMAP)

Common Usage:

Generate features

python scripts/feature_engineering_pipeline.py --input data.csv --generate --interactions

Feature selection

python scripts/feature_engineering_pipeline.py --input data.csv --select --top-k 20

Full pipeline

python scripts/feature_engineering_pipeline.py --input data.csv --pipeline full --output features.csv

Help

python scripts/feature_engineering_pipeline.py --help

Use Cases:

  • Preparing features for model training

  • Reducing feature dimensionality

  • Discovering important feature interactions

  1. Model Evaluation Suite

Comprehensive model evaluation with interpretability.

Key Features:

  • Cross-validation strategies (k-fold, stratified, time-series)

  • Hyperparameter optimization (grid search, random search, Bayesian)

  • Model interpretation (SHAP values, feature importance, partial dependence)

  • Performance metrics (accuracy, precision, recall, F1, AUC, MAE, RMSE)

  • Model comparison and statistical testing

Common Usage:

Evaluate model

python scripts/model_evaluation_suite.py --model model.pkl --data test.csv --metrics all

Hyperparameter tuning

python scripts/model_evaluation_suite.py --model sklearn.ensemble.RandomForestClassifier --tune --data train.csv

Model interpretation

python scripts/model_evaluation_suite.py --model model.pkl --interpret --shap

Help

python scripts/model_evaluation_suite.py --help

Use Cases:

  • Comparing multiple model architectures

  • Finding optimal hyperparameters

  • Explaining model predictions to stakeholders

See statistical_methods_advanced.md for comprehensive tool documentation and advanced examples.

Core Expertise

This skill covers world-class capabilities in:

  • Advanced production patterns and architectures

  • Scalable system design and implementation

  • Performance optimization at scale

  • MLOps and DataOps best practices

  • Real-time processing and inference

  • Distributed computing frameworks

  • Model deployment and monitoring

  • Security and compliance

  • Cost optimization

  • Team leadership and mentoring

Tech Stack

Languages: Python, SQL, R, Scala, Go ML Frameworks: PyTorch, TensorFlow, Scikit-learn, XGBoost Data Tools: Spark, Airflow, dbt, Kafka, Databricks LLM Frameworks: LangChain, LlamaIndex, DSPy Deployment: Docker, Kubernetes, AWS/GCP/Azure Monitoring: MLflow, Weights & Biases, Prometheus Databases: PostgreSQL, BigQuery, Snowflake, Pinecone

Key Workflows

  1. A/B Test Design and Analysis

Time: 2-3 hours for design, ongoing for analysis

  • Define Hypothesis - State null and alternative hypotheses, success metrics

  • Design Experiment - Calculate sample size, randomization strategy

Design A/B test with power analysis

python scripts/experiment_designer.py --effect-size 0.05 --power 0.8 --alpha 0.05 --output test_plan.json

  • Run Experiment - Implement randomization, collect data

  • Analyze Results - Statistical significance testing, confidence intervals

  • Report Findings - Effect size, business impact, recommendations

See experiment_design_frameworks.md for detailed methodology.

  1. Predictive Model Development

Time: 1-2 days for initial model, ongoing refinement

  • Exploratory Data Analysis - Understand distributions, correlations, missing data

  • Feature Engineering - Generate and select features

Automated feature engineering

python scripts/feature_engineering_pipeline.py --input data.csv --pipeline full --output features.csv

  • Model Training - Train multiple model types (linear, tree-based, neural nets)

  • Model Evaluation - Cross-validation, hyperparameter tuning

Evaluate and tune model

python scripts/model_evaluation_suite.py --model sklearn.ensemble.RandomForestClassifier --tune --data train.csv

  • Model Interpretation - SHAP values, feature importance, business insights
  1. Causal Inference Analysis

Time: 3-5 hours for setup and analysis

  • Define Causal Question - Treatment, outcome, confounders

  • Select Method - Propensity score matching, diff-in-diff, instrumental variables

  • Implement Analysis - Control for confounders, estimate treatment effect

  • Validate Assumptions - Check overlap, parallel trends, instrument validity

  • Report Causal Estimates - Average treatment effect, confidence intervals, sensitivity analysis

See statistical_methods_advanced.md for causal inference techniques.

  1. Time Series Forecasting

Time: 4-6 hours for model development

  • Data Preparation - Handle missing values, detect seasonality, stationarity tests

  • Feature Engineering - Lag features, rolling statistics, external variables

Generate time series features

python scripts/feature_engineering_pipeline.py --input timeseries.csv --temporal --lags 7,14,30

  • Model Selection - ARIMA, Prophet, LSTM, XGBoost for time series

  • Cross-Validation - Time-series split, walk-forward validation

  • Forecast & Monitor - Generate forecasts, track accuracy over time

Reference Documentation

  1. Statistical Methods Advanced

Comprehensive guide available in references/statistical_methods_advanced.md covering:

  • Advanced patterns and best practices

  • Production implementation strategies

  • Performance optimization techniques

  • Scalability considerations

  • Security and compliance

  • Real-world case studies

  1. Experiment Design Frameworks

Complete workflow documentation in references/experiment_design_frameworks.md including:

  • Step-by-step processes

  • Architecture design patterns

  • Tool integration guides

  • Performance tuning strategies

  • Troubleshooting procedures

  1. Feature Engineering Patterns

Technical reference guide in references/feature_engineering_patterns.md with:

  • System design principles

  • Implementation examples

  • Configuration best practices

  • Deployment strategies

  • Monitoring and observability

Production Patterns

Pattern 1: Scalable Data Processing

Enterprise-scale data processing with distributed computing:

  • Horizontal scaling architecture

  • Fault-tolerant design

  • Real-time and batch processing

  • Data quality validation

  • Performance monitoring

Pattern 2: ML Model Deployment

Production ML system with high availability:

  • Model serving with low latency

  • A/B testing infrastructure

  • Feature store integration

  • Model monitoring and drift detection

  • Automated retraining pipelines

Pattern 3: Real-Time Inference

High-throughput inference system:

  • Batching and caching strategies

  • Load balancing

  • Auto-scaling

  • Latency optimization

  • Cost optimization

Best Practices

Development

  • Test-driven development

  • Code reviews and pair programming

  • Documentation as code

  • Version control everything

  • Continuous integration

Production

  • Monitor everything critical

  • Automate deployments

  • Feature flags for releases

  • Canary deployments

  • Comprehensive logging

Team Leadership

  • Mentor junior engineers

  • Drive technical decisions

  • Establish coding standards

  • Foster learning culture

  • Cross-functional collaboration

Performance Targets

Latency:

  • P50: < 50ms

  • P95: < 100ms

  • P99: < 200ms

Throughput:

  • Requests/second: > 1000

  • Concurrent users: > 10,000

Availability:

  • Uptime: 99.9%

  • Error rate: < 0.1%

Security & Compliance

  • Authentication & authorization

  • Data encryption (at rest & in transit)

  • PII handling and anonymization

  • GDPR/CCPA compliance

  • Regular security audits

  • Vulnerability management

Common Commands

Development

python -m pytest tests/ -v --cov python -m black src/ python -m pylint src/

Training

python scripts/train.py --config prod.yaml python scripts/evaluate.py --model best.pth

Deployment

docker build -t service:v1 . kubectl apply -f k8s/ helm upgrade service ./charts/

Monitoring

kubectl logs -f deployment/service python scripts/health_check.py

Resources

  • Advanced Patterns: references/statistical_methods_advanced.md

  • Implementation Guide: references/experiment_design_frameworks.md

  • Technical Reference: references/feature_engineering_patterns.md

  • Automation Scripts: scripts/ directory

Senior-Level Responsibilities

As a world-class senior professional:

Technical Leadership

  • Drive architectural decisions

  • Mentor team members

  • Establish best practices

  • Ensure code quality

Strategic Thinking

  • Align with business goals

  • Evaluate trade-offs

  • Plan for scale

  • Manage technical debt

Collaboration

  • Work across teams

  • Communicate effectively

  • Build consensus

  • Share knowledge

Innovation

  • Stay current with research

  • Experiment with new approaches

  • Contribute to community

  • Drive continuous improvement

Production Excellence

  • Ensure high availability

  • Monitor proactively

  • Optimize performance

  • Respond to incidents

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

senior-data-scientist

No summary provided by upstream source.

Repository SourceNeeds Review
General

senior-data-scientist

No summary provided by upstream source.

Repository SourceNeeds Review
General

senior-flutter

No summary provided by upstream source.

Repository SourceNeeds Review
General

senior-java

No summary provided by upstream source.

Repository SourceNeeds Review