Professional Documents
Culture Documents
PROFESSIONAL SUMMARY
Experienced Data Architect with over 16 years of expertise in leading development teams, technical
leadership, and architecting solutions.
Proficient in requirements analysis, design, development, testing, and implementation within the
Health and Financial domains.
Extensive knowledge of SDLC methodologies, Kubernetes, AWS, Data Warehousing, and Data
Modeling.
Senior Software Application Developer with proficiency in Requirements Analysis, Design,
Development, Testing, and Implementation.
Extensive exposure to the Health and Financial domains & Solid understanding of SDLC
methodologies.
Tools and Version Control: Skilled in using tools like JIRA, GIT, and CI/CD for streamlined
development and collaborative work.
Data Handling:
Vast experience in data processing, including reconciliation processes and loading data into
External Tables from Flat Files using Unix shell scripts (5000+ scripts developed).
Proficient in working with Data Warehousing, Data Marts, Star Schema/Snowflake Schema
Modeling, Dimensional Modeling, Entity-Relationship Modeling, and OLAP concepts.
Key Competencies
Data Architecture
Data Warehousing & Data Marts
Dimensional & Entity-Relationship Modeling
SDLC Methodologies
Database Design (Logical & Physical)
Performance Tuning (Oracle)
SQL Tuning & Optimization
ETL Processes
Agile & Waterfall Methodologies
DevOps (JIRA, GIT, CI/CD)
Kubernetes & AWS
Data Analysis (Pandas, NumPy)
Technical Proficiency
Database Expertise: Oracle (SQL, PL/SQL)
Performance tuning, indexing, explain plan monitoring
Object-oriented PL/SQL design and development
Other Tools:
Autosys (Job Configuration)
JIRA (Project Management)
Git (Version Control)
CI/CD (Continuous Integration/Continuous Deployment)
EDUCATION
Masters in Computer Science, Sambalpur University, India
Certifications
Certified Scrum Product Owner (CSPO)
PROFESSIONAL EXPERIENCE
Client: Harman Connected Services
Location: Arizona, USA
Data Architect/Lead | April 2018- Jun2023
Project 1: Creating Excel Reports using Python and Autosys jobs
Responsibilities:
Developed an automated solution using Python and Autosys jobs to generate QC reports, delivering a
high-level summary of key metrics critical for evaluating data quality.
Employed Autosys jobs to trigger Python scripts, streamlining the report generation process and
significantly reducing processing time.
Achieved remarkable efficiency by automating the creation of 20-30 Excel sheets within less than 2
minutes, a substantial improvement over manual efforts that would have taken days.
The automation initiative led to a two-fold increase in productivity, reducing errors, and cutting
manual efforts.
Resulted in cost savings, enhanced visibility, and improved customer service, enhancing
organizational competitiveness in today’s fast-evolving business landscape.
Designed and implemented over 500 Python scripts utilizing Pandas, NumPy, Matplotlib, and other
libraries for efficient data analysis and report generation.
Established seamless connections between Python scripts, Oracle Database, and Unix systems to
extract data from Oracle tables and files, enhancing data retrieval capabilities.
Integrated the Python scripts with Autosys jobs using a .bat file, streamlining the automation process
and ensuring timely report generation.
Utilized Excel to generate graphs, facilitating trend and behavioral analysis of drugs, markets, and
accounts, providing valuable insights to stakeholders.
Proficient in analyzing information using various graphical methods, identifying patterns in data, and
comparing trends of drugs within the same generic to derive the impact on drug sales, especially
during natural calamities.
Designed and developed data visualizations to present a weekly graphical representation of drug
script activity trends using Python, enhancing data comprehension and decision-making.
Provided essential production support, swiftly resolving production issues encountered by end-users,
ensuring smooth operations of critical systems.
Significantly optimized report generation time by fine-tuning SQL statements, reducing processing
time from over 2 hours to just 10 minutes, improving overall system efficiency.
Collaborated closely with QA, UAT, and Production teams to resolve JIRA UAT defects and
production issues, ensuring the quality and reliability of applications.
Played a key role in the application deployment process across various environments (DEV, QA,
UAT, Performance Testing, Production) using GitHub, effectively managing and processing
transactions.
Environment: Python, Oracle 19c, Oracle SQL, PL/SQL, TOAD, Unix, GitHub, Autosys.
Responsibilities:
Led a critical project involving the migration of historical data from an upstream system, known as
Holding Hub, to its Local Database.
Initiated the migration due to the planned decommissioning of the upstream system, necessitating the
transfer of historical data to its corresponding database, including loading data into FACT tables.
Played a pivotal role in capturing business requirements, creating comprehensive data models, and
formulating DDL scripts for deployment.
Designed and developed Business Rules, defined project scope, and crafted technical specifications
to guide the project.
Developed PL/SQL procedures, functions, and objects and orchestrated the loading of data files
utilizing SQL*Loader, Control files, and export/import utilities for seamless data import from the
Legacy System.
Implemented efficient data organization within database partitions and actively participated in
database development.
Created essential Oracle tables, Procedures, and Functions to facilitate the data migration process.
Utilized advanced features of Oracle, including Bulk Collections, Autonomous Transactions,
Dynamic SQL, Object types, and Records, optimizing data handling and processing.
Constructed Oracle views encapsulating critical business logic for ease of use by Business End-
Users.
Generated the Application Design Document (ADD), undergoing design review and obtaining sign-
off from key stakeholders.
Collaborated with QA, UAT teams to resolve defects identified during testing, ensuring a robust and
error-free application.
Managed the deployment of the application across various environments, including DEV, QA, UAT,
and Performance Testing (PT), handling voluminous data of over 55 million daily transactions.
Proficiently identified and optimized long-running SQLs using AWR report, contributing to
enhanced performance.
Environment: Oracle 11c, Oracle SQL, PL/SQL, TOAD, ERWIN, Shell Scripts
Responsibilities:
Managed the GCIX system, focusing on data consolidation from diverse feeds, mapping business
area data to a standardized CIB-wide data definition, and storing it in optimized structures for
efficient reporting, analysis, and utilization by CIB users.
Designed and developed a comprehensive PL/SQL API (Application Programming Interfaces),
contributing to over 50,000 lines of PL/SQL code, ensuring seamless data handling and processing.
Held responsibilities encompassing the review of Functional Specification Documents (FSD),
preparation and review of Technical Specification Documents (TSD), and active development and
modification of PL/SQL code to meet project requirements.
Key involvement in capturing business requirements, enhancing Data Models, and orchestrating the
creation and deployment of DDL scripts, business rules, and Data Quality (DQ) checks.
Created, tested, and fine-tuned over 250 complex database procedures using PL/SQL Cursors, REF
Cursors, Pipelined Functions, and Object Types to optimize data processing.
Actively participated in the application deployment process across various environments such as
DEV, QA, UAT, and Production, ensuring smooth transitions and optimal functionality.
Provided validation scripts critical for the successful production deployment of the application,
validating its reliability and accuracy.
Employed optimization techniques to fine-tune Oracle PL/SQL APIs, improving response times
through the implementation of Bulk Collection methods.
Contributed to a structured SDLC process by adding JIRA Stories, effectively defining requirements,
facilitating design, development, and smooth deployment of changes.
Environment: Oracle 11g, Oracle SQL, PL/SQL, TOAD, Erwin, Shell Scripts, SVN.
Project: 2
MIS4GTB (Management Information System for Global Transaction Bank)
Responsibilities:
Led the MIS4GTB project, focusing on sourcing and processing 'finance revenue' from GTB front
office systems after posting, encompassing revenue, balances, and their conversion into reporting
currency.
Reviewed Functional Specification Documents (FSD), prepared and reviewed Technical
Specification Documents (TSD), and actively participated in the development and modification of
PL/SQL code to align with project requirements.
Developed PL/SQL packages, procedures, and functions utilizing a variety of advanced
programming constructs such as Records, Bulk Collections, Pipelined functions, Object Types,
Arrays, Cursors, and REF Cursor.
Designed and implemented an API to dynamically create SPLIT Partitions on Tables and manage
Truncate Partition operations, optimizing data management.
Employed optimization techniques, fine-tuning Oracle PL/SQL APIs for enhanced response times
using Bulk Collection methods.
Successfully promoted application functionality from Lower Environments (DEV, SIT, UAT, Pre-
PROD) to Production for Go-Live, ensuring a smooth transition and minimal disruption.
Contributed to a structured SDLC process by adding JIRA Stories, effectively defining requirements,
facilitating design, development, and smooth deployment of changes.
Environment: Oracle 11g, Oracle SQL, PL/SQL, TOAD, Erwin, Shell Scripts, SVN.