Jeevan Gaire

|

I build data pipelines that scale, transform raw chaos into clean insights, and help businesses make decisions that actually matter.

1+
Year Experience
5+
Projects & POCs Built
6+
Certifications
50TB+
Data Processed
scroll
Who I Am

Turning raw data into business gold

Jeevan Gaire

Hey! I'm a Data Engineer passionate about building robust, scalable data infrastructure. Over the past year I've worked on ETL pipelines, data warehousing, and real-time streaming systems that power analytics at scale.

My stack leans heavy on the modern data stack — dbt, Airflow, Spark — and I love the challenge of squeezing milliseconds out of a slow query or redesigning a broken pipeline from the ground up.

When I'm not writing PySpark jobs, I'm writing about data engineering on my blog or contributing to open-source projects.

🏢

Current Company

Data Engineer · 2025 – present

🎓

B.Tech Computer Science

MMDU · 2021-2025

📍

Based in Jaipur, India

Available for remote roles

What I Use

My Tech Stack

Big Data & Processing

Apache SparkPySparkBatch ProcessingReal-time ProcessingStructured Streaming
🏗️

Data Lakehouse & Architecture

DatabricksDelta LakeMedallion ArchitectureData Lakehouse ArchitectureData Modeling
🔄

Orchestration & ETL

Apache AirflowAirflow DAG DevelopmentdbtAzure Data FactoryETL / ELTData Pipelines
☁️

Azure Cloud Services

AzureAzure DatabricksADLS Gen2Azure Key VaultAzure SQL Database
📡

Streaming Technologies

Spark StreamingStructured StreamingAutoLoader
📊

Data Engineering Practices

Data GovernanceData QualityData ValidationPipeline Optimization
My Journey

Work Experience

Data Engineer

Celebal Technologies · Full-time · Jaipur

Dec 2025 – Present
  • Providing end-to-end data engineering and cloud solutions for one of the world’s leading banking clients using Azure-based technologies.
  • Developing and maintaining scalable ETL/ELT pipelines using Azure Data Factory, Azure Databricks, and PySpark for enterprise-level data processing.
  • Automating existing ADF workflows and optimizing data pipelines to improve execution efficiency, reduce failures, and enhance overall performance.
  • Designing and implementing robust data transformation and ingestion solutions using Databricks Autoloader, PySpark, and SQL-based processing.
  • Collaborating with cross-functional analytics and business teams to deliver reliable, scalable, and production-ready data solutions.
  • Working on enhancement, monitoring, debugging, and optimization of existing cloud data infrastructure and reporting systems.
PySparkAzure DatabricksAzureADFSQL ServerSSMSAzure DevOpsAutoloader

Junior Associate Data Engineer

Celebal Technologies · Apprenticeship · On-Site

May 2025 – Nov 2025
  • Built hands-on expertise in modern data engineering technologies including SQL, PySpark, Databricks, Python, Azure Data Factory (ADF), Unity Catalog, Delta Live Tables (DLT), Auto Loader, Spark Streaming, Structured Streaming, and Data Modeling.
  • Contributed to an enterprise migration project focused on converting existing Jupyter-based workflows into scalable PySpark notebooks on Databricks.
  • Modified and optimized notebook logic for distributed data processing using PySpark and Spark transformations.
  • Validated migrated notebooks by testing them against sample datasets to ensure data accuracy, transformation correctness, and workflow reliability.
  • Worked with cloud-based data pipelines and learned best practices for scalable ETL architecture and data lake processing.
  • Collaborated with senior engineers to understand real-world data engineering workflows, debugging strategies, and production-level data processing practices.
PythonSQLPySparkDatabricksAzureAzure Data Factory (ADF)
Let's Talk

Get In Touch

Let's build something great together

I'm currently open to full-time Data Engineer roles and interesting freelance projects. If you have a pipeline to build or a data problem to solve, let's talk!