About

Data Engineer

Hello, I'm Prayag Verma.

Currently pursuing my Master’s in Information Technology and Management at UT Dallas, set to graduate in December 2025.

  • www.aimtocode.com
  • +1 945-274-7200

I'm an independent data engineer and data architect with over four years of experience, working with global companies like Infosys, Amdocs, and Briston Infotech.

Currently, I'm committed for continuous learning, growth, and aspire to assist organizations in achieving data-driven objectives.

Skills Rating

% Eager to Learn

Core Skills

Oracle & MySQL98%
Teradata & SQL Server75%
ETL & Data Warehousing 95%
Informatica power center85%
Data Marts & SCD 100%
Facts, Dimensions & Star Schema100%
DataOps Suit 95%
Python 75%
NumPy & Pandas 85%
R Lang Basics
UNIX 75%
Power BI 50%
Microsoft Excel 75%
Jira Project Management 90%


Cloud Skills

Azure Databricks 90%
Snowflake 75%
ADLS & ADF 90%

Interests

Data Engineering

ETL/ Data Warehousing

Solution Architect

Data Architect

Resume

Summary

Prayag Verma

I'm an innovative, independent, and deadline-driven Data Engineer and Architect with over four years of hands-on experience designing, developing, and testing user-centered enterprise data warehouse solutions.

I've worked across diverse domains, including telecommunications, health insurance, and retail, bringing expertise in data engineering, solution architecture, etl processing, data pipeline streamlining, and data warehousing to every project.

Education

Master's In Information Technology and Management

2023 - 2025

The University of Texas at Dallas, TX

Completed courses include AWS Cloud Solution Architect, OOP with Python, Advanced Statistics for Data Science, Business Analytics with R, Database Foundations for Business Analytics, Business Data Warehousing, Big Data, Technology and New Product Development, System Analysis and Project Management (SAPM), Predictive Analytics for Data Science, and Prescriptive Analytics.

Bachelor of Engineering in Computer Science and Engineering

2015 - 2019

Anna University, Chennai, India

Completed courses including rogramming langues, databases like C, C++, PHP, JavaScript, PHP, Java, Database Management, HTML, CSS, and Python which led me comple my final year project in "General TCP State Inference Model From Passive Measurements Using Machine Learning Techniques".

Certifications

2015 - 2023

  • Azure Data Engineer Associate (DP-203)

  • Azure Fundamentals (AZ-900)

  • AWS Certified Solutions Architect - Associate (SAA-C03)

  • Python (by Kaggle)

  • Python Essentials 1 (by Cisco)

  • Python 101 for Data Science (by IBM)

  • Data analysis with Python (by IBM)

  • Data Science Methodology (by IBM)

  • SQL and Relational Databases 101 (by IBM)

  • NoSQL and DBaaS 101 (by IBM)

  • ETL Tester 1 (by Datagaps)

Domain Knowledge

Data engineering, Data and Solution Architect

  • I’m currently seeking for a full time opportunities where I can apply my skills in designing, implementing, and managing robust data systems that drive informed business decisions. I’m ready to take on new challenges in data engineering and architecture roles.

Relevant Coursework

  • AWS Cloud Solution Architecture, Big Data, Business Data Warehousing, Advance Statistics for Data Science, Business Analytics with R, Database Foundation for Business Analytics, Predictive Analytics for Data Science, and Prescriptive Analytics.

Projects

  • Developed an RBAC based digital license management web app with PHP Laravel and Node.js for API, capable of handling 5M+ daily APIs calls with responses in under 100 milliseconds. Click here
  • Created an RBAC-enabled ETL Validator tool using Python Flask, validating flat files (CSV, Excel, Parquet, JSON) up to 15GBs data with 99% accuracy and supporting Azure Data Lake connections. Click here

Professional Experience

Data Engineer

Jan 2023 – Jul 2023

Infosys, Bengalore, India

  • Designed and maintained scalable data pipelines to process and integrate large volumes of structured and unstructured healthcare data using Azure Data Factory and Databricks, reducing data latency by 30% and ensuring real-time analytics readiness.
  • Orchestrated and automated ETL/ELT workflows to integrate healthcare coverage data from flat files, Oracle, and SQL Server into Snowflake, reducing manual data preparation tasks by 95% and ensuring real-time availability for analytics.
  • Optimized database architectures by migrating on-premises flat files and databases data, Informatica PowerCenter workflows to SQL Server, Snowflake, and Azure Data Lake (ADLS), reducing query execution times by 35%.
  • Collaborated with data analysts and business stakeholders to integrate data from APIs and third-party systems, delivering tailored datasets that enhanced operational reporting accuracy by 25%.
  • Established data governance frameworks, including data validation, error handling, and compliance mechanisms, ensuring adherence to privacy standards such as HIPAA and GDPR.
  • Built monitoring solutions using Azure Monitor, deploying auto-scaling mechanisms that reduced compute costs by 35% while maintaining performance SLAs.
  • Documented complex data architectures and processes, created interactive dashboards using Power BI, and provided actionable insights that improved claims processing efficiency by 30%.

Data Engineer

Dec 2021 – Jan 2023

Amdocs Development Center India, Pune, India

  • Designed scalable ADF data pipelines for ingesting and processing terabytes of organized and non-relational customer usage, billing, and purchase data using Teradata, ADLS, Databricks, Redshift, and Snowflake, increasing efficiency by 35%.
  • Migrated KSH script-based legacy systems for telecom payment and transaction data systems to modern cloud infrastructure, integrating Azure Data Lake and Snowflake, achieving a 60% improvement in data integration accuracy.
  • Developed real-time streaming solutions with Azure Event Hubs and Kafka, for analytics and collaborated with cross-functional teams to implement data quality checks, ensuring 99% accuracy.
  • Orchestrated IPC workflows to capture slowly changing dimensions (SCDs) and flat files data; processed data cleansing, scrubbing, and handling missing value, ensuring 100% of business requirement and data migration covered.
  • Enhanced analytics performance by optimizing SQL queries with indexing and partitioning, leading to 50% faster execution times for key reporting workflows.
  • Automated pipeline monitoring and alert systems using Python and Azure Monitor, reducing downtime incidents by 20% and ensuring SLA compliance.

Data Engineer

Jul 2019 – Nov 2021

Briston Infotech Pvt Ltd, New Delhi, India

  • Developed cross-domain ETL solutions for telecom, retail, and healthcare industries using Informatica PowerCenter, ensuring seamless integration of tabular and non-tabular data into Snowflake with 99% reliability.
  • Implemented advanced data modeling techniques, including hybrid star-snowflake schemas, tailored to telecom billing and healthcare claims, reducing report generation times by 40%.
  • Implemented high-performance fact and dimension tables using star schema, improving query speed for 3UK’s customer insights by 40%.
  • Streamlined flat-file data ingestion workflows by automating preprocessing and validating data using Python, reducing manual intervention by 60% and ensuring seamless integration with data warehouse.
  • Executed comprehensive data migration projects, transforming healthcare claims data from OLTP systems (Oracle, MySQL) to Snowflake, increasing query speeds by 60%.
  • Built scalable QA frameworks across domains and ensured data accuracy for analytics dashboards and compliance reports.

Contact

Contact Me

Social Profiles

Loading
Your message has been sent. Thank you!
Designed by Prayag Verma