Nagaraju Anthati
Senior Data Scientist
Experience
Senior Data Scientist
JPMC
- Designed and deployed Bayesian Marketing Mix Models (MMM) using PyMC-3+ and PySpark, quantifying ROI and channel-level elasticity across retail and asset management portfolios.
- Engineered ETL and feature pipelines in Airflow and AWS Databricks, automating ingestion of terabyte-scale marketing, transaction, and behavioral data from S3, Hive, Postgres, and Kafka.
- Built Delta Lake and Apache Iceberg architecture supporting adstock, carry-over, and seasonal ETL transformations for model input.
- Implemented hierarchical Bayesian structures and regression-based MMMs using NumPy, PyMC, and TensorFlow Probability to model multi-region effects.
- Optimized PySpark jobs with liquid clustering and adaptive partitioning, cutting MMM data-prep runtime by ~40%.
- Automated model training, versioning, and deployment via MLflow and Databricks Asset Bundles, ensuring reproducibility and compliance.
- Streamed near-real-time ad-exposure and conversion data from multi-tenant Kafka clusters into model pipelines.
- Deployed probabilistic inference workflows on AWS EMR using distributed MCMC sampling, reducing convergence time significantly.
- Delivered model explainability dashboards in Plotly Dash visualizing posterior distributions, channel effects, and uncertainty intervals.
- Applied Bayesian regularization and feature selection techniques to optimize MMM performance.
- Integrated MMM outputs into Snowflake and AWS RDS for BI and marketing analytics consumption.
- Implemented data quality monitoring using Great Expectations and integrated validation across ETL workflows.
- Collaborated with quant research teams to embed MMM-driven elasticities into financial forecasting models.
- Automated CI/CD pipelines using Jules and ServiceNow for model retraining and deployment.
- Delivered cross-functional MMM insights to marketing, finance, and analytics teams to support budget optimization.
- Defined and implemented advanced eCommerce tracking for online transactions, allowing granular reporting on product performance and customer journey analysis.
- Integrated Google Analytics with third-party platforms, such as Google Ads and CRM systems, enabling cross-platform attribution and seamless data flow.
- Utilized regression analysis, decision trees, and clustering to predict customer behavior and segment audiences for targeted marketing.
- Developed a multi-touch attribution model to accurately assign conversion credit across digital touchpoints.
- Implemented resiliency, reliability, and availability of various asset and wealth management tools both on premises and in the Cloud.
- Worked on reconciliation and reporting integration of fund positions, instruments, cash, or money markets.
- Managed change management, release process, and release management using Jules pipelines and ServiceNow.
- Produced AFX merchant reports, P&L validation, and reporting related to various funds, assets, and instruments.
- Implemented real-time daily load status solution using Geneos dashboards.
- Provided customer-facing support for various asset and wealth management MMM and ML activities.
Data Scientist - MMM/ML
Glaxo Smith Kline
- Designed and implemented Bayesian MMM frameworks in PyMC to evaluate ROI across multichannel marketing campaigns in consumer health and pharma domains.
- Built end-to-end ETL pipelines using Airflow, Kafka, Azure Data Factory, and Databricks Spark integrating CRM, sales, scheduling, and process data (>100 TB).
- Developed probabilistic regression models with hierarchical priors to capture campaign, region, and HCP-level heterogeneity.
- Built schema-evolving data models using open table formats and ADLS Gen2 integration.
- Implemented Bayesian inference workflows with MCMC sampling on Azure Databricks for channel elasticity estimation.
- Developed custom priors to reflect domain knowledge such as decay rates, carry-over, and saturation effects.
- Automated training and evaluation pipelines using Azure ML and MLflow with version-controlled experiments.
- Implemented streaming analytics using Kafka and Flink to continuously refresh MMM datasets from digital and field systems.
- Built PySpark feature stores and validation layers to ensure data quality and consistency.
- Conducted model diagnostics using WAIC, LOO-CV, and posterior predictive checks.
- Created Power BI and Plotly Dash dashboards for marketing teams to visualize MMM insights and posterior ROI curves.
- Ensured data governance, lineage tracking, and GDPR/GxP compliance across all Azure data pipelines.
- Migrated legacy MMM workloads from on-prem HDP to Azure Databricks, improving scalability and reducing processing time by 60%.
- Built budget optimization simulators in Python using Bayesian Decision Theory principles.
- Partnered with commercial analytics teams to operationalize MMM insights into forecasting and promotional planning models.
- Worked on cloud-hosted Kafka data sources and streamed using Kafka connectors and Flink.
- Created standardized SQL engine clusters using Presto DB.
- Developed virtual cloud data warehouses using Snowflake and querying data using SnowSQL, Spark jobs, and Tez.
- Maintained documentation on Confluence, conducted code reviews, and managed builds with Groovy on Jenkins.
Data Engineer
Visa Europe
- Performed data analysis on CDH5 and CDH6 clusters using Apache Hue.
- Managed autoscaling and maintenance of AWS EMR clusters.
- Implemented massive data warehouse solutions to offload 800 TB of data from DB2 storage to Hadoop.
- Set up streaming processes for transactional and clearance data using Kinesis.
- Implemented workflow schedules using Airflow and Oozie.
- Deployed streaming ingestions using Kafka Confluent platform consisting of 10 broker nodes from various data sources.
Hadoop/Big Data Engineer
Solera Holdings
- Worked with Hadoop, Sqoop, Hive, HBase, Spark, Akka, Lucene, Solr, Pig, Pentaho, Hue, and Scala to build big data solutions.
Big Data Hadoop Developer
Silicon Integra Limited
- Developed Hadoop and big data pipelines using Hadoop, Sqoop, R, Kite SDK, Kudu, Hive (CDH5.4, CDH5.6), HBase, Impala, Hue, Spark, Oozie, AWS EMR, Azure, Solr, Pig, Paxata, Scala, and Presto DB.
- Applied valuation and estimation algorithms.
Hadoop Developer / Analyst Consultant
Nortech Solutions
- Built big data solutions using Hadoop, Sqoop, Hive, HBase, Spark, Akka, Lucene, Solr, Pig, Pentaho, Hue, and Scala.
Bigdata Developer/Engineer
Nextgen Solutions
- Developed big data applications using Hadoop, Hive, Scala, JSF, MongoDB, HBase, ActiveMQ, and multithreading.
Bigdata/Hadoop Engineer
Tata Telecom
- Implemented big data analytics and Hadoop ecosystem solutions using Hadoop Analytics, Pentaho, Java, Python, and J2EE.
Summary
Having overall 13+ years of experience in planning, building, implementation, and integration of full-scale commercial projects in different verticals like financial, retail, insurance, banking, high-tech, social media, oil and gas and networking/telecom. Having started career as Graduate Systems Engineer with TCS, I have been involved in large scale Java and Hadoop projects which are highly scalable, distributed, and available. Worked with various cloud environments like AWS, Azure and open-source cloud deployment and configuration tools like OpenStack and OpenNebula. Gained hands on experience on NoSQL databases like MongoDB, HBase and Cassandra. Worked on various Agile practices like TDD, BDD, pair programming, continuous integration and Scrum. Worked on programming languages like Java, Scala, Python, Golang, C, PySpark, shell scripting, J2EE, JSF, Apache Hadoop ecosystem, Hortonworks, Cloudera, Accel Data ODP, ETL practices and analytics platforms.
Skills
- Pymc
- Pymc-marketing
- Bayesian Modelling
- Regression
- Classification/clustering
- Timeseries Forecasting
- Google Analytics
- Genai
- Lang Chain/lang Graph
- Milvus
- Neilson Marketing Cloud
- Python
- Sql
- Java
- Git
- Docker
- Mongodb
- R
- Presto Db
- Linux/unix
- Github
- Spring Boot
- Artificial Intelligence
- Etl
- Cloud Services
- Bash
- Ansible
- Graphql
- Nosql
- Eks
- Jupyter Hub
- Scala
- Kubernetes
- Apache Hadoop
- Conflient/kafka
- Oracle Database
- Azure Adf
- Azure Datalake
- Databricks
- Azure Synapse
- Dataiku
- Sagemaker
- Azure Dsvm
- Slurm/lsf
- Data Analysis
- Statistical Modelling
- Model Deployment
- Cloud Data Engineering (Aws, Azure)
- Advanced Analytics
- Machine Learning
- Generative Ai
- Solution Development
- Streaming Data Pipelines
- Staging Tables
- Low Latency Solutions
- Multitasking Abilities
- Decision-making
- Self-motivated
Languages
Education
Northumbria University
Master of Science · Computer Science · Newcastle upon Tyne, United Kingdom
JNTU
Bachelor of Technology · Electrical, Electronics and Communications Engineering · India
Similar Freelancers
Discover other experts with similar qualifications and experience