Senior Data Scientist
World-class senior data scientist skill for production-grade AI/ML/Data systems.
Overview
This skill provides world-class data science capabilities through three core Python automation tools and comprehensive reference documentation. Whether designing experiments, building predictive models, performing causal inference, or driving data-driven decisions, this skill delivers expert-level statistical modeling and analytics solutions.
Senior data scientists use this skill for A/B testing, experiment design, statistical modeling, causal inference, time series analysis, feature engineering, model evaluation, and business intelligence. Expertise covers Python (NumPy, Pandas, Scikit-learn), R, SQL, statistical methods, hypothesis testing, and advanced analytics techniques.
Core Value: Accelerate analytics and experimentation by 65%+ while improving model accuracy, statistical rigor, and business impact through proven methodologies and automated pipelines.
Quick Start
Main Capabilities
Core Tool 1
python scripts/experiment_designer.py --input data/ --output results/
Core Tool 2
python scripts/feature_engineering_pipeline.py --target project/ --analyze
Core Tool 3
python scripts/model_evaluation_suite.py --config config.yaml --deploy
Core Capabilities
-
Experiment Design & A/B Testing - Statistical power analysis, sample size calculation, multi-armed bandits, sequential testing
-
Statistical Modeling - Regression, classification, time series, causal inference, Bayesian methods
-
Feature Engineering - Automated feature generation, selection, transformation, interaction terms, dimensionality reduction
-
Model Evaluation - Cross-validation, hyperparameter tuning, bias-variance tradeoff, model interpretation (SHAP, LIME)
-
Business Analytics - Customer segmentation, churn prediction, lifetime value, attribution modeling, forecasting
-
Causal Inference - Propensity score matching, difference-in-differences, instrumental variables, regression discontinuity
Python Tools
- Experiment Designer
Design statistically rigorous experiments with power analysis.
Key Features:
-
A/B test design with sample size calculation
-
Statistical power analysis
-
Multi-variant testing setup
-
Sequential testing frameworks
-
Bayesian experiment design
Common Usage:
Design A/B test
python scripts/experiment_designer.py --effect-size 0.05 --power 0.8 --alpha 0.05
Multi-variant test
python scripts/experiment_designer.py --variants 4 --mde 0.03 --output experiment_plan.json
Sequential testing
python scripts/experiment_designer.py --sequential --stopping-rule obf
Help
python scripts/experiment_designer.py --help
Use Cases:
-
Designing product experiments before launch
-
Calculating required sample sizes
-
Planning sequential testing strategies
- Feature Engineering Pipeline
Automate feature generation, selection, and transformation.
Key Features:
-
Automated feature generation (polynomial, interaction terms)
-
Feature selection (mutual information, recursive elimination)
-
Encoding (one-hot, target, frequency)
-
Scaling and normalization
-
Dimensionality reduction (PCA, t-SNE, UMAP)
Common Usage:
Generate features
python scripts/feature_engineering_pipeline.py --input data.csv --generate --interactions
Feature selection
python scripts/feature_engineering_pipeline.py --input data.csv --select --top-k 20
Full pipeline
python scripts/feature_engineering_pipeline.py --input data.csv --pipeline full --output features.csv
Help
python scripts/feature_engineering_pipeline.py --help
Use Cases:
-
Preparing features for model training
-
Reducing feature dimensionality
-
Discovering important feature interactions
- Model Evaluation Suite
Comprehensive model evaluation with interpretability.
Key Features:
-
Cross-validation strategies (k-fold, stratified, time-series)
-
Hyperparameter optimization (grid search, random search, Bayesian)
-
Model interpretation (SHAP values, feature importance, partial dependence)
-
Performance metrics (accuracy, precision, recall, F1, AUC, MAE, RMSE)
-
Model comparison and statistical testing
Common Usage:
Evaluate model
python scripts/model_evaluation_suite.py --model model.pkl --data test.csv --metrics all
Hyperparameter tuning
python scripts/model_evaluation_suite.py --model sklearn.ensemble.RandomForestClassifier --tune --data train.csv
Model interpretation
python scripts/model_evaluation_suite.py --model model.pkl --interpret --shap
Help
python scripts/model_evaluation_suite.py --help
Use Cases:
-
Comparing multiple model architectures
-
Finding optimal hyperparameters
-
Explaining model predictions to stakeholders
See statistical_methods_advanced.md for comprehensive tool documentation and advanced examples.
Core Expertise
This skill covers world-class capabilities in:
-
Advanced production patterns and architectures
-
Scalable system design and implementation
-
Performance optimization at scale
-
MLOps and DataOps best practices
-
Real-time processing and inference
-
Distributed computing frameworks
-
Model deployment and monitoring
-
Security and compliance
-
Cost optimization
-
Team leadership and mentoring
Tech Stack
Languages: Python, SQL, R, Scala, Go ML Frameworks: PyTorch, TensorFlow, Scikit-learn, XGBoost Data Tools: Spark, Airflow, dbt, Kafka, Databricks LLM Frameworks: LangChain, LlamaIndex, DSPy Deployment: Docker, Kubernetes, AWS/GCP/Azure Monitoring: MLflow, Weights & Biases, Prometheus Databases: PostgreSQL, BigQuery, Snowflake, Pinecone
Key Workflows
- A/B Test Design and Analysis
Time: 2-3 hours for design, ongoing for analysis
-
Define Hypothesis - State null and alternative hypotheses, success metrics
-
Design Experiment - Calculate sample size, randomization strategy
Design A/B test with power analysis
python scripts/experiment_designer.py --effect-size 0.05 --power 0.8 --alpha 0.05 --output test_plan.json
-
Run Experiment - Implement randomization, collect data
-
Analyze Results - Statistical significance testing, confidence intervals
-
Report Findings - Effect size, business impact, recommendations
See experiment_design_frameworks.md for detailed methodology.
- Predictive Model Development
Time: 1-2 days for initial model, ongoing refinement
-
Exploratory Data Analysis - Understand distributions, correlations, missing data
-
Feature Engineering - Generate and select features
Automated feature engineering
python scripts/feature_engineering_pipeline.py --input data.csv --pipeline full --output features.csv
-
Model Training - Train multiple model types (linear, tree-based, neural nets)
-
Model Evaluation - Cross-validation, hyperparameter tuning
Evaluate and tune model
python scripts/model_evaluation_suite.py --model sklearn.ensemble.RandomForestClassifier --tune --data train.csv
- Model Interpretation - SHAP values, feature importance, business insights
- Causal Inference Analysis
Time: 3-5 hours for setup and analysis
-
Define Causal Question - Treatment, outcome, confounders
-
Select Method - Propensity score matching, diff-in-diff, instrumental variables
-
Implement Analysis - Control for confounders, estimate treatment effect
-
Validate Assumptions - Check overlap, parallel trends, instrument validity
-
Report Causal Estimates - Average treatment effect, confidence intervals, sensitivity analysis
See statistical_methods_advanced.md for causal inference techniques.
- Time Series Forecasting
Time: 4-6 hours for model development
-
Data Preparation - Handle missing values, detect seasonality, stationarity tests
-
Feature Engineering - Lag features, rolling statistics, external variables
Generate time series features
python scripts/feature_engineering_pipeline.py --input timeseries.csv --temporal --lags 7,14,30
-
Model Selection - ARIMA, Prophet, LSTM, XGBoost for time series
-
Cross-Validation - Time-series split, walk-forward validation
-
Forecast & Monitor - Generate forecasts, track accuracy over time
Reference Documentation
- Statistical Methods Advanced
Comprehensive guide available in references/statistical_methods_advanced.md covering:
-
Advanced patterns and best practices
-
Production implementation strategies
-
Performance optimization techniques
-
Scalability considerations
-
Security and compliance
-
Real-world case studies
- Experiment Design Frameworks
Complete workflow documentation in references/experiment_design_frameworks.md including:
-
Step-by-step processes
-
Architecture design patterns
-
Tool integration guides
-
Performance tuning strategies
-
Troubleshooting procedures
- Feature Engineering Patterns
Technical reference guide in references/feature_engineering_patterns.md with:
-
System design principles
-
Implementation examples
-
Configuration best practices
-
Deployment strategies
-
Monitoring and observability
Production Patterns
Pattern 1: Scalable Data Processing
Enterprise-scale data processing with distributed computing:
-
Horizontal scaling architecture
-
Fault-tolerant design
-
Real-time and batch processing
-
Data quality validation
-
Performance monitoring
Pattern 2: ML Model Deployment
Production ML system with high availability:
-
Model serving with low latency
-
A/B testing infrastructure
-
Feature store integration
-
Model monitoring and drift detection
-
Automated retraining pipelines
Pattern 3: Real-Time Inference
High-throughput inference system:
-
Batching and caching strategies
-
Load balancing
-
Auto-scaling
-
Latency optimization
-
Cost optimization
Best Practices
Development
-
Test-driven development
-
Code reviews and pair programming
-
Documentation as code
-
Version control everything
-
Continuous integration
Production
-
Monitor everything critical
-
Automate deployments
-
Feature flags for releases
-
Canary deployments
-
Comprehensive logging
Team Leadership
-
Mentor junior engineers
-
Drive technical decisions
-
Establish coding standards
-
Foster learning culture
-
Cross-functional collaboration
Performance Targets
Latency:
-
P50: < 50ms
-
P95: < 100ms
-
P99: < 200ms
Throughput:
-
Requests/second: > 1000
-
Concurrent users: > 10,000
Availability:
-
Uptime: 99.9%
-
Error rate: < 0.1%
Security & Compliance
-
Authentication & authorization
-
Data encryption (at rest & in transit)
-
PII handling and anonymization
-
GDPR/CCPA compliance
-
Regular security audits
-
Vulnerability management
Common Commands
Development
python -m pytest tests/ -v --cov python -m black src/ python -m pylint src/
Training
python scripts/train.py --config prod.yaml python scripts/evaluate.py --model best.pth
Deployment
docker build -t service:v1 . kubectl apply -f k8s/ helm upgrade service ./charts/
Monitoring
kubectl logs -f deployment/service python scripts/health_check.py
Resources
-
Advanced Patterns: references/statistical_methods_advanced.md
-
Implementation Guide: references/experiment_design_frameworks.md
-
Technical Reference: references/feature_engineering_patterns.md
-
Automation Scripts: scripts/ directory
Senior-Level Responsibilities
As a world-class senior professional:
Technical Leadership
-
Drive architectural decisions
-
Mentor team members
-
Establish best practices
-
Ensure code quality
Strategic Thinking
-
Align with business goals
-
Evaluate trade-offs
-
Plan for scale
-
Manage technical debt
Collaboration
-
Work across teams
-
Communicate effectively
-
Build consensus
-
Share knowledge
Innovation
-
Stay current with research
-
Experiment with new approaches
-
Contribute to community
-
Drive continuous improvement
Production Excellence
-
Ensure high availability
-
Monitor proactively
-
Optimize performance
-
Respond to incidents