Summary
Overview
Work History
Education
Skills
Accomplishments
Work Availability
Timeline
Professional Accreditations
Professional Accreditations
Core Competencies
Hi, I’m

Ved Prakash

Staff Data Engineer
Almere Stad
The real test is not whether you avoid this failure, because you won’t. It’s whether you let it harden or shame you into inaction, or whether you learn from it; whether you choose to persevere.
Barack Obama

Summary

Technical Data Leader with 15+ years of experience architecting enterprise-scale data platforms and leading high-performing engineering teams. Currently operating as a Staff Data Engineer and recently recognized as a Snowflake Data Superhero. Demonstrated success in:

- Leading cross-functional teams to implement complex data architectures, including successful synchronization between Postgres, Clickhouse, and Snowflake systems
- Optimizing cloud infrastructure and data processing pipelines, achieving 70% improvement in processing time and 30% reduction in infrastructure costs
- Establishing data governance frameworks and architectural guidelines for analytical feature development
- Mentoring and developing subject matter experts across critical domains including Snowplow infrastructure and data pipeline development

Core expertise spans modern data stack implementation (GCP, Snowflake, Apache Iceberg), real-time data processing, and infrastructure automation. Proven track record of translating complex technical requirements into scalable solutions while fostering a culture of engineering excellence.

Technical certifications include Google Cloud Data Engineer, Snowflake Certified Architect, and Snowflake Certified Developer. Regular speaker at technical conferences on data engineering best practices and infrastructure scaling.

Overview

2025
years of professional experience
4
years of post-secondary education

Work History

Gitlab
Almere Stad, Provincie Flevoland

Staff Data Engineer
6 2022 - Current

Job overview

  • Designed and implemented scalable and reliable data pipelines using NATS, PySpark, Go and Apache Iceberg, resulting in a 70% improvement in data processing time.
  • Managed cloud computing platforms GCP and snowflake resulting in 30% reduction in infrastructure and snowflake cost.
  • Demonstrated commitment to mentoring and developing team members, fostering a culture of growth and collaboration.
  • Working closely with cross-functional teams, including data analysts, data scientists, and business stakeholders to ensure data solutions meet business requirements
  • Led end-to-end implementation of multiple high-impact projects from requirements gathering through deployment and post-launch support stages

Gitlab

Senior Data Engineer
02.2021 - 06.2022

Job overview

  • Principle developer on Meltano Taps to build Connector to different data source like Adaptive, Xactly, Edcast , Zendesk.
  • Responsible for Deployment of Meltano In GKE setup to dynamically scale.
  • Enable monitoring of Airflow by integrating it into GitLab main infrastructure monitoring system.
  • Host and manage Airflow in GKE using Terraform and helm packages , which improved testing data pipeline timeline by 80%.
  • Built scalable data pipeline which is responsible to perform heavy data pull from SaaS Data Platform.
  • Develop and maintain DBT workflows to automate ETL processes, data transformations, and data quality checks.

Infosys Limited

Senior Data Engineer
08.2016 - 02.2021

Job overview

  • Led technology transformation initiatives with Product Owner and Engineering Head, successfully migrating data lake to Snowflake while reducing infrastructure costs by 30% and improving performance
  • Architected and implemented end-to-end ELT pipelines using Talend and Snowflake across AWS/Azure, enabling seamless processing of multi-cloud data sources
  • Reduced deployment timeline by 70% through CI/CD automation for AWS Lambda functions and containerized batch processes using Docker
  • Built and managed migration framework moving 100TB+ data from on-premise to AWS cloud using Python, Apache Airflow, and Pandas
  • Established robust data governance through Snowflake RBAC implementation and standardized DevOps practices, ensuring security compliance and code quality

Infosys Limited

Data Engineer
07.2014 - 07.2016

Job overview

  • Designed STAR Schema Data Lake for assortment reporting by defining dimensions and fact tables and key attributes.
  • Designed and Developed ETL (Pentaho) to integrate Planogram with Point of Sale Data.
  • Build API in Python (Flask) to call ETL Job in back end.
  • Work in Scrum Team of 8 and worked pro actively with product Owner for story grooming.
  • Designed and Developed MySQL data model to store data in coming in from DataMart.
  • Implemented CI/CD of MySQL changes using Liquibase and Bamboo.
  • Built a framework in Python (Pandas) to scramble customer PII information for Dev/Test environment.
  • Performed clustered installation, design, backup, recovery, security for MySQL database.
  • Docker Container for MySQL database reducing over all server commissioning time by 90%.

Infosys Limited

Senior Systems Engineer
02.2010 - 07.2013

Job overview

  • Designed and Migrated legacy scheduler Control_M to SKYBOT.
  • Developed batch auto recovery job leveraging shell and SQL to reduce recovery time by 30%.
  • SME in doing RCA for production incident , perform routine tests on databases (Oracle , Kognitio MPP) and provide extended support to all ETL applications(Pentaho).

Education

VTU
Bijapur, Karnataka

Bachelor of Engineering from Computer Science
01.2005 - 01.2009

Skills

Cloud Platforms : GCP, AWS

Accomplishments

  • Reduced data processing latency by 70% through architecting new data pipelines using Google PubSub, PySpark, and Apache Iceberg, enabling near real-time analytics for 100K+ daily users.
  • Reduced data quality incidents by 75% through implementation of data observability framework.
  • Saved Licensed cost of 200k £ by building entire ETL pipeline in Pentaho Community Edition.
  • Accelerated deployment cycles from 2 hours to 15 minutes (87% improvement) through CI/CD automation
  • Decreased infrastructure costs by $80K annually through strategic architecture decisions and resource optimization.
  • Drove 40% increase in team velocity by leading and mentoring a cross-functional team of 20 engineers
  • Automated 90% of routine infrastructure tasks through Terraform and helm packages implementation
  • Developed DB refresh tool saving 80 days DBA effort in a year.
Availability
See my work availability
Not Available
Available
monday
tuesday
wednesday
thursday
friday
saturday
sunday
morning
afternoon
evening
swipe to browse

Timeline

Senior Data Engineer

Gitlab
02.2021 - 06.2022

Senior Data Engineer

Infosys Limited
08.2016 - 02.2021

Data Engineer

Infosys Limited
07.2014 - 07.2016

Senior Systems Engineer

Infosys Limited
02.2010 - 07.2013

VTU

Bachelor of Engineering from Computer Science
01.2005 - 01.2009

Staff Data Engineer

Gitlab
6 2022 - Current

Professional Accreditations

  • Google Certified Data Engineer
  • Snowflake Certified Developer
  • Infosys Outstanding Performer for year 2016 ,2017, 2018 & 2019.
  • AIMIA Engineering Star Player Of the year 2018.
  • Infosys MOST-VALUABLE award for excellent performance 2015.

Professional Accreditations

  • Snowflake Data Super Hero
  • Google Certified Data Engineer
  • Snowflake Certified Architect
  • Snowflake Certified Developer
  • Infosys Outstanding Performer for year 2016 ,2017, 2018 & 2019.

Core Competencies

Technical Leadership

  • Strategic Planning & Execution
  • Cross-functional Team Leadership
  • Technical Architecture & Design
  • Data Engineering & Analytics

Professional Skills

  • Problem-solving & Critical Thinking
  • Stakeholder Management
  • Technical Mentorship
  • Effective Communication
  • Project Management
  • Change Management
Ved PrakashStaff Data Engineer