You are on page 1of 6

SHARMISTHA BOHIDAR

Data Architect/Lead | Python | AWS


CARY, NC-25719
data.engineer981@gmail.com
848.283.3031
https://www.linkedin.com/in/sharmistha-bohidar-639b7474/

PROFESSIONAL SUMMARY
 Experienced Data Architect with over 16 years of expertise in leading development teams, technical
leadership, and architecting solutions.
 Proficient in requirements analysis, design, development, testing, and implementation within the
Health and Financial domains.
 Extensive knowledge of SDLC methodologies, Kubernetes, AWS, Data Warehousing, and Data
Modeling.
 Senior Software Application Developer with proficiency in Requirements Analysis, Design,
Development, Testing, and Implementation.
 Extensive exposure to the Health and Financial domains & Solid understanding of SDLC
methodologies.

Skills and Expertise:


 SDLC and Methodologies: Proficient in Agile and Waterfall methodologies, ensuring effective
project management and development lifecycles.

 Tools and Version Control: Skilled in using tools like JIRA, GIT, and CI/CD for streamlined
development and collaborative work.

 Database Expertise (Oracle):


 Demonstrated skills in performance tuning at both application and database levels, including
optimization, index management, and explain plan monitoring.
 Proficient in SQL Tuning and Object-Oriented PL/SQL design and development.

 Data Handling:
 Vast experience in data processing, including reconciliation processes and loading data into
External Tables from Flat Files using Unix shell scripts (5000+ scripts developed).
 Proficient in working with Data Warehousing, Data Marts, Star Schema/Snowflake Schema
Modeling, Dimensional Modeling, Entity-Relationship Modeling, and OLAP concepts.

 PL/SQL and Scripting:


 Strong backend development skills utilizing PL/SQL Cursors, PL/SQL Collection, Packages,
Stored Procedures, Pipelined functions, and Triggers.
 Proficient in designing and implementing PL/SQL procedures, functions, and objects, as well
as loading data files using SQL*Loader, Control files, and export/import utilities from
Legacy Systems.
 Extensive use of Oracle Standard packages like DBMS_SQL, UTL_FILE, DBMS_LOB,
UTL_RAW, and others, along with records and bulk collections, external tables, merge
statements, multi-table inserts, etc.
 Data Analysis and Reporting:
 Skilled in data analysis, leveraging Pandas and NumPy, and proficient in generating QC
reports.

Key Competencies
 Data Architecture
 Data Warehousing & Data Marts
 Dimensional & Entity-Relationship Modeling
 SDLC Methodologies
 Database Design (Logical & Physical)
 Performance Tuning (Oracle)
 SQL Tuning & Optimization
 ETL Processes
 Agile & Waterfall Methodologies
 DevOps (JIRA, GIT, CI/CD)
 Kubernetes & AWS
 Data Analysis (Pandas, NumPy)

Technical Proficiency
Database Expertise: Oracle (SQL, PL/SQL)
 Performance tuning, indexing, explain plan monitoring
 Object-oriented PL/SQL design and development

Development Tools & Languages:


 Shell Scripting (Unix)
 Erwin (Data Modeling)
 Python (Pandas, NumPy)

Other Tools:
 Autosys (Job Configuration)
 JIRA (Project Management)
 Git (Version Control)
 CI/CD (Continuous Integration/Continuous Deployment)

EDUCATION
 Masters in Computer Science, Sambalpur University, India

Certifications
 Certified Scrum Product Owner (CSPO)

PROFESSIONAL EXPERIENCE
Client: Harman Connected Services
Location: Arizona, USA
Data Architect/Lead | April 2018- Jun2023
Project 1: Creating Excel Reports using Python and Autosys jobs

Responsibilities:
 Developed an automated solution using Python and Autosys jobs to generate QC reports, delivering a
high-level summary of key metrics critical for evaluating data quality.
 Employed Autosys jobs to trigger Python scripts, streamlining the report generation process and
significantly reducing processing time.
 Achieved remarkable efficiency by automating the creation of 20-30 Excel sheets within less than 2
minutes, a substantial improvement over manual efforts that would have taken days.
 The automation initiative led to a two-fold increase in productivity, reducing errors, and cutting
manual efforts.
 Resulted in cost savings, enhanced visibility, and improved customer service, enhancing
organizational competitiveness in today’s fast-evolving business landscape.
 Designed and implemented over 500 Python scripts utilizing Pandas, NumPy, Matplotlib, and other
libraries for efficient data analysis and report generation.
 Established seamless connections between Python scripts, Oracle Database, and Unix systems to
extract data from Oracle tables and files, enhancing data retrieval capabilities.
 Integrated the Python scripts with Autosys jobs using a .bat file, streamlining the automation process
and ensuring timely report generation.
 Utilized Excel to generate graphs, facilitating trend and behavioral analysis of drugs, markets, and
accounts, providing valuable insights to stakeholders.
 Proficient in analyzing information using various graphical methods, identifying patterns in data, and
comparing trends of drugs within the same generic to derive the impact on drug sales, especially
during natural calamities.
 Designed and developed data visualizations to present a weekly graphical representation of drug
script activity trends using Python, enhancing data comprehension and decision-making.
 Provided essential production support, swiftly resolving production issues encountered by end-users,
ensuring smooth operations of critical systems.
 Significantly optimized report generation time by fine-tuning SQL statements, reducing processing
time from over 2 hours to just 10 minutes, improving overall system efficiency.
 Collaborated closely with QA, UAT, and Production teams to resolve JIRA UAT defects and
production issues, ensuring the quality and reliability of applications.
 Played a key role in the application deployment process across various environments (DEV, QA,
UAT, Performance Testing, Production) using GitHub, effectively managing and processing
transactions.

Environment: Python, Oracle 19c, Oracle SQL, PL/SQL, TOAD, Unix, GitHub, Autosys.

Project 2: Adhoc project


Responsibilities:
 Successfully executed ad-hoc projects to generate customized data analysis reports based on client
requirements, focusing on drug market studies and prescribing trends.
 Developed and maintained software applications, optimizing performance and usability to meet
specific project requirements.
 Implemented complex PL/SQL packages, procedures, functions, and analytical functions to enhance
data handling and report generation.
 Designed and implemented APIs to dynamically manage table partitions, significantly improving
data organization and retrieval processes.
 Utilized Oracle Analytical Functions and Shell scripts to efficiently process voluminous data and
identify and fine-tune long-running SQL queries.

Environment: Oracle 19c, Oracle SQL, PL/SQL, TOAD, Unix

Fidelity Investment Banking, Bangalore, India


Lead Software Engineer, Oracle Developer | Apr 2017 – Nov 2017

Project: Migration of Historic Data


The aim of this project was to export data from an upstream named Holding Hub to its Local Database. The
reason for the export was that the upstream was getting decommissioned. The historic data needs to be
moved to its database and data should be loaded to its FACT tables.

Responsibilities:
 Led a critical project involving the migration of historical data from an upstream system, known as
Holding Hub, to its Local Database.
 Initiated the migration due to the planned decommissioning of the upstream system, necessitating the
transfer of historical data to its corresponding database, including loading data into FACT tables.
 Played a pivotal role in capturing business requirements, creating comprehensive data models, and
formulating DDL scripts for deployment.
 Designed and developed Business Rules, defined project scope, and crafted technical specifications
to guide the project.
 Developed PL/SQL procedures, functions, and objects and orchestrated the loading of data files
utilizing SQL*Loader, Control files, and export/import utilities for seamless data import from the
Legacy System.
 Implemented efficient data organization within database partitions and actively participated in
database development.
 Created essential Oracle tables, Procedures, and Functions to facilitate the data migration process.
 Utilized advanced features of Oracle, including Bulk Collections, Autonomous Transactions,
Dynamic SQL, Object types, and Records, optimizing data handling and processing.
 Constructed Oracle views encapsulating critical business logic for ease of use by Business End-
Users.
 Generated the Application Design Document (ADD), undergoing design review and obtaining sign-
off from key stakeholders.
 Collaborated with QA, UAT teams to resolve defects identified during testing, ensuring a robust and
error-free application.
 Managed the deployment of the application across various environments, including DEV, QA, UAT,
and Performance Testing (PT), handling voluminous data of over 55 million daily transactions.
 Proficiently identified and optimized long-running SQLs using AWR report, contributing to
enhanced performance.
Environment: Oracle 11c, Oracle SQL, PL/SQL, TOAD, ERWIN, Shell Scripts

Deutsche Bank(HCL Bangalore, India)


Technical Lead, Oracle Developer | Nov 2009 – April 2017
Project: 1
GCIX (Global Customer Information Exchange)

Responsibilities:
 Managed the GCIX system, focusing on data consolidation from diverse feeds, mapping business
area data to a standardized CIB-wide data definition, and storing it in optimized structures for
efficient reporting, analysis, and utilization by CIB users.
 Designed and developed a comprehensive PL/SQL API (Application Programming Interfaces),
contributing to over 50,000 lines of PL/SQL code, ensuring seamless data handling and processing.
 Held responsibilities encompassing the review of Functional Specification Documents (FSD),
preparation and review of Technical Specification Documents (TSD), and active development and
modification of PL/SQL code to meet project requirements.
 Key involvement in capturing business requirements, enhancing Data Models, and orchestrating the
creation and deployment of DDL scripts, business rules, and Data Quality (DQ) checks.
 Created, tested, and fine-tuned over 250 complex database procedures using PL/SQL Cursors, REF
Cursors, Pipelined Functions, and Object Types to optimize data processing.
 Actively participated in the application deployment process across various environments such as
DEV, QA, UAT, and Production, ensuring smooth transitions and optimal functionality.
 Provided validation scripts critical for the successful production deployment of the application,
validating its reliability and accuracy.
 Employed optimization techniques to fine-tune Oracle PL/SQL APIs, improving response times
through the implementation of Bulk Collection methods.
 Contributed to a structured SDLC process by adding JIRA Stories, effectively defining requirements,
facilitating design, development, and smooth deployment of changes.

Environment: Oracle 11g, Oracle SQL, PL/SQL, TOAD, Erwin, Shell Scripts, SVN.

Project: 2
MIS4GTB (Management Information System for Global Transaction Bank)
Responsibilities:
 Led the MIS4GTB project, focusing on sourcing and processing 'finance revenue' from GTB front
office systems after posting, encompassing revenue, balances, and their conversion into reporting
currency.
 Reviewed Functional Specification Documents (FSD), prepared and reviewed Technical
Specification Documents (TSD), and actively participated in the development and modification of
PL/SQL code to align with project requirements.
 Developed PL/SQL packages, procedures, and functions utilizing a variety of advanced
programming constructs such as Records, Bulk Collections, Pipelined functions, Object Types,
Arrays, Cursors, and REF Cursor.
 Designed and implemented an API to dynamically create SPLIT Partitions on Tables and manage
Truncate Partition operations, optimizing data management.
 Employed optimization techniques, fine-tuning Oracle PL/SQL APIs for enhanced response times
using Bulk Collection methods.
 Successfully promoted application functionality from Lower Environments (DEV, SIT, UAT, Pre-
PROD) to Production for Go-Live, ensuring a smooth transition and minimal disruption.
 Contributed to a structured SDLC process by adding JIRA Stories, effectively defining requirements,
facilitating design, development, and smooth deployment of changes.

Environment: Oracle 11g, Oracle SQL, PL/SQL, TOAD, Erwin, Shell Scripts, SVN.

British Telecom(Tata Consultancy Services (TCS), Bhubaneswar, India)


Assistant System Engineer | Sep 2006 – Apr 2009
Project: British Telecom
 Designed comprehensive test scenarios for the Post Pilot Project based on Functional Requirement
Document (FRD), ensuring efficient testing processes.
 Identified gaps in test cases prepared by the ETL Team, providing valuable insights for enhanced
testing coverage.
 Developed over 50 Oracle PL/SQL Stored Procedures for in-depth analysis, optimizing data handling
and processing.
 Created and optimized 150+ Oracle SQL statements for Scenario Testing, contributing to a thorough
and rigorous testing phase.
 Developed 15+ Shell scripts, streamlining various processes and enhancing automation within the
project.
 Actively executed test cases, meticulously identifying defects between requirements and
implementation to ensure high-quality deliverables.
Environment:
 Technologies: Oracle 12c, Oracle SQL, PL/SQL, TOAD
 Operating System: Unix
 Scripting: Shell Scripts

You might also like