You are on page 1of 1

Vamshi Krishna Repala

vamshi.r14@iiits.in · +91 9010894585


github.com/ishmav16
linkedin.com/in/vamshir14
Kodad, Suryapet (Dt.), Telangana - 508206.

SUMMARY
Skilled data engineer having 3+ years of experience in designing, developing and maintaining
data pipelines in an agile manner. Seeking for a role to deliver excellent products having scope of
much learning technical and managerial skills.

WORK EXPERIENCE TECHNICAL SKILLS


• DBS Bank, Hyderabad • Languages
Data Engineer July 2019 - June 2022 Python, C/C++, Java, JS & Bash
Developed Spark based data pipelines for
• Frameworks
 Migrating the data from Teradata to HDFS and cut Spark, Spring & Angular
down the 70% of team effort by automation.
 Reconciling the data between Teradata and HDFS • Databases
that helps users for 100% data trusty worthiness. MySQL, Hive & Teradata
 Ingesting the data from 6 different sources (V+,
Mainframes & other systems) into Hive and built the
Data Lakes and Data Warehouses. AWARDS & RECOGNITION
 Building the data marts for a business use case DBS
Card Rewards. • Super Techie, WeTQ Awards
DBS Asia Hub 2, Hyderabad
Tools & Technologies: Python, Spark, SQL, Hive, 2020 & 2021
Alluxio, S3, Shell, Jenkins, Airflow, Tivoli and Collibra.
• Dean’s Research Award
• Ernst & Young Associates (EY), Hyderabad
IIIT Sri City, A.P
Data Analyst July - Oct 2018
Monsoon 2016
Performed data analysis and developed different
dashboard reports that will help users to gain more • Best Paper Award
insights with respect to audit related financial fraud LAMSYS Conference organized by ISRO
analysis. 2016
Tools & Technologies: SQL, SSAS & Power BI.
PROJECTS

EDUCATION • Data Reconciler


Pipeline for data reconciliation between Teradata & Hive.
• Indian Institute of Information Technology Keywords: Teradata, Hive, PySpark, AWS S3, Data
(IIIT), Sri City, Andhra Pradesh Governance, JDBC, on-premise, Shell Scripting.

B.Tech in Computer Science


• Project Merlion
Pipelines for data ingestions and migrations into Hive.
AREAS OF INTEREST Keywords: PySpark, HDFS, Parquet, Mainframes, S3,
Data Engineering & Data Analysis Airflow, SCD, Metadata driven, ETL, Data Lake.
Programming
Data Structures & Algorithms • Data Feeder for Rewards System
Machine Learning Pipelines for building Data Marts and Data Warehouses.
Micro Services & Backend Development Keywords: Spark, SQL, Compute, CDC, AWS S3, Tivoli
Scheduler, Validation-Metrics, Bash, DBS Card Rewards.

You might also like