0 viewsjobseeker
Sanket K. — Senior Data Engineer from United States

Sanket K.

Senior Data Engineer

United States 6+ years
Open to offersNew to Platform
Languages
EnglishHindi
Video Introduction
No video introduction yet
The candidate has not added a video.
Contact information and social networks are private. Connect to unlock.
Hidden

About

Sanket K. is a seasoned Senior Data Engineer with over seven years of experience specializing in building scalable data platforms and machine learning pipelines across Azure and AWS environments. He has a robust skill set in Python, PySpark, Databricks, and Spark SQL, with significant experience in developing enterprise data pipelines and large-scale analytics systems. His expertise extends to implementing LLM-powered applications, orchestrating AI workflows using LangChain, and ensuring deployment of secure, production-ready solutions in regulated environments. At Alpha Net Corporation, Sanket built AI-driven automation workflows and designed Databricks Lakehouse pipelines to transform extensive datasets. His previous engagements with companies like The Dotcom Team and ICT Zoom Inc. highlight his role in developing ELT and analytics pipelines, predictive analytics, and data quality frameworks. With a Master of Science in Computer Engineering from Southern Methodist University, Sanket is dedicated to delivering reliable, high-performance AI and analytics solutions.

Experience

  • Data Engineer – Data Engineering & AI Enablement

    ALPHA NET CORPORATION LLC · 2025 — Present
    Developed automation workflows on Azure Databricks utilizing Python, LangChain, and CrewAI for multi-step enterprise data processes, document analysis, and retrieval-based AI applications. Created Databricks Lakehouse pipelines with PySpark, Spark SQL, and Delta Lake, following Medallion architecture principles for data processing and transformation. Designed workflows for feature engineering and ML data preparation in conjunction with Azure ML and MLflow, facilitating experiment tracking and model retraining support. Integrated LLM-powered RAG pipelines in production environments, ensuring governance, access controls, and compliance for regulated enterprise data.
  • Data Engineer – Operations and Analytics

    THE DOTCOM TEAM LLC · 2024 — 2025
    Engineered scalable ELT and analytics pipelines using Azure Data Factory, Databricks, dbt, and PySpark to assist in enterprise reporting and subsequent machine learning tasks. Developed both streaming and batch processing workflows leveraging Kafka, Spark Structured Streaming, and Delta Lake for analytics and operational reporting. Established transformation pipelines and datasets suitable for predictive analytics and NLP applications while using MLflow to track model lifecycles. Streamlined cloud deployment and data integration processes through Azure DevOps, Docker, Terraform, and Snowflake, enhancing scalability and reliability.
  • Analytics Engineer II – Product Analytics and Development

    ICT ZOOM INC · 2023 — 2024
    Created distributed data pipelines using PySpark on Databricks for processing extensive customer and marketing datasets for analytics and machine learning. Built workflows for data preparation that incorporated APIs, vector embeddings, and structured datasets, which supported internal AI experimentation and retrieval-based applications. Enhanced Delta Lake transformation pipelines and Spark SQL workloads, optimizing performance for high-volume data processing. Implemented frameworks for automated data quality and monitoring with Python and Azure services, thus improving reporting reliability and consistency.
  • Data Engineer II – Marketing Science

    OMNICOM GROUP via GENISYS GROUP · 2022 — 2022
    Constructed scalable PySpark and Spark pipelines to support audience analytics and campaign optimization within enterprise marketing frameworks. Developed dimensional models and feature engineering workflows using SparkSQL and HiveQL, enhancing analytical and predictive modeling abilities. Automated reporting processes and integrated dashboards with Tableau, Python, and API-based data pipelines to increase visibility and efficiency in reporting. Worked alongside analytics and data science teams to refine data consistency and improve machine learning dataset preparation.
  • Senior Data Analyst – Digital Transformation

    JP MORGAN CHASE & CO. via PRIDE GLOBAL · 2021 — 2021
    Engineered production data reconciliation workflows using Python, Alteryx, and AWS Redshift in a regulated financial services setting. Managed job scheduling pipelines with Apache Airflow, configuring DAGs and task dependencies while performing root cause analysis on failures to alleviate manual reconciliation efforts. Spearheaded data automation initiatives for EMEA in a production support context, facilitating cross-functional incident resolution and ensuring compliance with governance standards across production data workflows utilizing AWS infrastructure. Created comprehensive Tableau dashboards for executive reporting and operational oversight.
  • Data Analyst II – Information Technology

    DAVIES – CONSULTING DIVISION · 2018 — 2020
    Automated reporting pipelines for regulatory purposes using Python and Bash scripts that handled data extraction, file processing, and validation, significantly reducing reporting time. Developed data pipelines employing AWS S3 and Redshift for financial compliance and taxonomy reporting. Designed frameworks for risk modeling aimed at financial compliance and fraud detection.

Skills & Expertise

Education

  • Master of Science in Computer Engineering
    SOUTHERN METHODIST UNIVERSITY (SMU)