0% found this document useful (0 votes)
32 views5 pages

Rubina Data Engineer

Uploaded by

itstaffing198
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
32 views5 pages

Rubina Data Engineer

Uploaded by

itstaffing198
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd

Senior Data engineer

RUBINA SHAIK
Senior Data Engineer/BI Engineer
Email: rubina.zun0417@gmail.com
Contact No: (571) 402-2169
---------------------------------------------------------------------------------------------------------------------------------------------------------------
PROFESSIONAL SUMMARY:
 Innovative and results-driven Senior Data Engineer and BI Engineer with over 13 years of experience in
architecting, developing, and optimizing large-scale data infrastructure and analytics ecosystems across leading
organizations, including Door Dash, Moody’s Analytics, Capital One, IBM, and Oracle.
 Demonstrates deep expertise in building and maintaining high-performance data pipelines, data lakes, and data
warehouses that enable enterprise-wide analytics, predictive modeling, and data-driven decision-making.
 Highly skilled in cloud technologies, distributed data processing, and business intelligence development with
proficiency in AWS (S3, Redshift, EMR, Glue, Cloud Formation), Snowflake, Data bricks, Apache Spark, Airflow,
and DBT.
 Adept in modern data orchestration and streaming platforms such as Kafka and Flink. Strong command of
programming languages including Python, SQL, and Shell scripting for developing scalable ETL/ELT frameworks
and automation workflows.
 Proven success in delivering robust BI solutions using Tableau, Looker, Power BI, and IBM Cognos to transform
complex data into actionable business insights.
 Experienced in leading data quality initiatives, implementing monitoring and governance frameworks, and
ensuring compliance with security and regulatory standards.
 Recognized for mentoring and guiding engineering teams, driving continuous improvement, and fostering a
culture of data excellence and innovation.
 Provided leadership and mentorship to junior engineers, cultivating best practices in coding standards, data
engineering methodologies, and cloud technologies.
 Championed data governance initiatives, ensuring regulatory compliance (GDPR, HIPAA, PCI-DSS) and securing
sensitive data within enterprise systems.
 Optimized complex SQL queries, data models, and storage architectures in Snowflake, Redshift, and PostgreSQL,
improving query performance by up to 30%.
 Adopted agile methodologies and CI/CD pipelines with Jenkins and Circle CI to streamline development cycles
and enhance deployment consistency.

TECHNICAL SUMMARY:

Programming Scripting &


Languages Python, Java, SQL, Shell Scripting
Apache Airflow, DBT, Informatics, Alteryx, Apache Flink, Apache NiFi, Apache
Data Engineering & ETL Tools Kafka, Dragster
Big Data & Distributed
Computing Apache Spark, Spark SQL, Data bricks, Hive, Presto, Hadoop Ecosystem
Snowflake, Amazon Redshift, PostgreSQL, Dremio, Oracle Data Warehouse,
Data Warehousing Teradata

Databases Dynamo DB, RDS, Oracle Database, PostgreSQL, SQL Server

Cloud Platforms Amazon Web Services (AWS), ORACLE Cloud

AWS Services S3, Redshift, Dynamo DB, Cloud Formation, EC2, EMR, Glue, Cloud Watch
Containerization & CI/CD
Tools Docker, Jenkins, Circle CI

Workflow & Orchestration Apache Airflow, Dagster, Flink


Tools
Business Intelligence (BI) &
Visualization Tools Tableau, Looker, Power BI, IBM Cognos, OBIEE, Sigma, Mode, Sisense
Data Modeling, Data Quality Checks, Data Governance, Metadata Management,
Data Modeling & Governance Data Validation
Version Control & Project
Management Git, JIRA, Confluence, Service Now

Monitoring & Logging Tools AWS Cloud Watch, Data dog


Automation & Infrastructure
as Code (IaC) AWS Cloud Formation, Shell Scripts

PROFESSIONAL EXPERIENCE

Client: Door Dash, INC | Jul 2024 to Present


Role: Senior Data Engineer
Responsibilities:
 Designed and rolled out ICEX, a migration framework adopted company-wide for high volume data
modernization enabling cross-platform query-ability between Datalake, Snowflake, and Databricks. leveraging
Apache Iceberg, Delta Lake. 170 datasets migrated using ICEX eliminating redundant data lake copies, reducing
compute and storage costs by ~40%, enhancing dataset discoverability across DE pipelines.
 Redesigned vendor data ingestion workflows for daily mail and call-center partners by replacing legacy SFTP and
PGP-based processes with modern data lake architecture. Reduced run times from 11 hours to 1 hour,
eliminated vendor infra dependency and improved data reliability, scalability across growth marketing systems.
 Developed advanced attribution models for delivery event tracking, improving measurement of intra-funnel
conversions and campaign impact. Enabled onboarding of new marketing channels into the DAC optimization
engine, reducing DAC per delivery by ~30% and driving smarter allocation of media spend.
 Designed GenAI-powered observability dashboards using LangChain + OpenAI API, automating incident
summaries, root cause explanations, and alert prioritization for on-call engineers.
 Strengthened data observability and reliability by implementing DQ checks, assigning clear ownership for
incident response, and setting up focused alert routing for critical issues. Audited and deprecated legacy ETLs to
cut alert noise by ~35%, developed runbooks to streamline on-call operations, and reduced incident resolution
time by ~40% across DE Marketing pipelines.
 Partnered with Marketing and DSML teams to design robust data integration pipelines integrated with Feature
Store and curated high-quality datasets by implementing anomaly detection and validation checks significantly
improving campaign performance visibility and data-driven decision-making.
 Established data governance standards using Hubble (inhouse framework) + AWS Glue Data Catalog, tagging
datasets for lineage, PII classification, and compliance visibility across 500+ tables.
 Automated CI/CD data pipeline workflows using GitHub Actions + Copilot Review for syntax +Terraform +
CircleCI, resulting in 60% faster release cycles and consistent environment parity across Dev/UAT/Prod.

Environment: Amazon Web Services (S3, Redshift, Dynamo DB, Cloud Formation, EC2, EMR, Cloud Watch), Apache
Airflow, Snowflake, DBT, Data bricks, Apache Spark, Spark SQL, Python, Shell Scripting, Tableau, Sigma, Mode, Docker,
Jenkins, JIRA.

Client: Moody’s Analytics | Nov 2021 to Jun 2024


Role: Senior Data Engineer
Responsibilities:
 Design, develop, and maintain scalable and robust data pipelines that support large-scale data processing and
analytics across multiple data sources and platforms.
 Lead continuous improvement initiatives for data infrastructure, including tuning applications, upgrading data
platforms, and enhancing process controls to ensure optimal performance and reliability.
 Provide hands-on development and operational support for deploying database scripts and changes across
development, testing, and production environments while ensuring seamless, error-free report delivery.
 Collaborate closely with cross-functional teams—including technical engineers, data scientists, and business
stakeholders—to gather requirements, design data models, and implement effective data solutions aligned with
business needs.
 Perform data analysis, cleansing, wrangling, and validation to guarantee data quality and integrity, while
troubleshooting and resolving production issues rapidly to minimize downtime.
 Optimize existing reports and data queries to improve efficiency and timeliness using best practices in SQL, ETL,
and big data technologies.
 Mentor and guide junior engineers by conducting code reviews, leading workshops, and maintaining
comprehensive documentation to uphold high-quality engineering standards.
 Ensure adherence to data governance, security protocols, and compliance standards, including implementing
data access controls and contributing to policy development.

Environment: Amazon Web Services (S3, Redshift, Dynamo DB, Cloud Formation, EC2, EMR, Cloud Watch), Apache
Airflow, Snowflake, DBT, Data bricks, Apache Spark, Spark SQL, Python, Sisense, Dremio, Circle CI, JIRA.

Client: Capital One | Jun 2021 to Nov 2021


Role: Senior Data Engineer
Responsibilities:
 Designing, developing, and maintaining highly scalable and efficient data pipelines to support complex analytics
and reporting needs.
 Supporting continuous improvement efforts in data applications, infrastructure, and process controls, including
tuning and upgrades.
 Providing hands-on development and operational support for deploying database scripts and managing changes
across multiple environments.
 Collaborating with technical teams, business owners, and cross-functional stakeholders during design and
implementation phases.
 Debugging and resolving production report failures to ensure seamless data delivery and availability.
 Performing data analysis, cleansing, and wrangling from various sources to ensure high data quality and
reliability.
 Designing and developing reports with rigorous testing, data validation, and quality checks.
 Optimizing existing reports and database queries to improve performance and timeliness.
 Implementing solutions and processes to enhance deliverables' quality and team efficiency.
 Developed or mentoring junior data engineers, including conducting code reviews, workshops, and
documentation reviews.
 Following best data, engineering practices such as requirement analysis, specification, design, development,
testing, and deployment.

 Applying data governance principles and ensuring compliance with security and regulatory standards.
 Utilizing programming and scripting languages like Python, SQL, Shell scripting, and working with ETL and BI
tools such as Informatics, Alteryx, and Tableau.
 Engaging in root cause analysis and ongoing troubleshooting to strengthen operational controls and improve
business processes.
 Staying current with emerging technologies and trends to continuously innovate data solutions at Moody’s.

Environment: Amazon Web Services (S3, Redshift, Dynamo DB, Cloud Formation, EC2, EMR, Cloud Watch), Apache
Airflow, Snowflake, DBT, Data bricks, Apache Spark, Spark SQL, Python, Sisense, Dremio, Circle CI, JIRA.

Client: IBM | Dec 2017 to Jan 2020


Role: BI Engineer
Responsibilities:
 Developed, design, and implement scalable, robust BI solutions, including data warehouses, data lakes, and
reporting platforms, ensuring alignment with IBM's business objectives and technology standards.
 Oversee and optimize complex ETL processes for extracting, transforming, and loading data from multiple
heterogeneous sources, ensuring data quality, integrity, and timely availability.
 Develop and manage sophisticated data models, schemas, and metadata for enterprise data warehouses and
marts to facilitate complex analytics and reporting needs.
 Ensure high performance of BI systems by tuning SQL queries, optimizing database architectures, and
implementing best practices for data retrieval and processing efficiency.
 Guide the creation of insightful dashboards and reports using BI tools (e.g., IBM Cognos, Power BI, Tableau) to
support data-driven decision-making at all organizational levels.
 Collaborate with business analysts, data scientists, and development teams to translate business requirements
into technical specifications and BI deliverables.
 Mentor junior BI engineers, lead technical discussions, and foster best practices in BI development and data
governance.
 Implement and enforce data governance policies, ensuring compliance with organizational standards and
industry regulations regarding data privacy and security.
 Keep abreast of emerging BI technologies and trends to proactively enhance IBM’s BI capabilities and drive
innovation in data processing and analytics.
 Provide expert-level support for BI systems, troubleshooting complex data issues and coordinating with cross-
functional teams for resolution.

Environment: ORACLE Cloud, SQL, Shell Scripting, OBIEE, Informatics, DAC, IPM, SOA, Sisense, Info web, Service Now,
Hyperion, Essbase, EBS R12.

Client: ORACLE | Jun 2012 to Dec 2017


Role: IT Consultant
Responsibilities:
 Designing, developing, and maintaining robust data architectures, including data warehouses and data marts, to
support Oracle's business intelligence initiatives.
 Developing and managing complex ETL (Extract, Transform, Load) processes to ensure data from multiple
sources is integrated, cleansed, and transformed effectively for analysis.
 Creating and optimizing data models and schemas to ensure data consistency, accuracy, and availability for
analytical and reporting purposes.
 Building and maintaining high-performance reporting systems, dashboards, and visualizations using Oracle BI
tools and other industry-standard BI platforms.
 Collaborating with business stakeholders to gather and translate business requirements into scalable technical
solutions and actionable BI reports.
 Ensuring data quality, integrity, and security by implementing data governance, access controls, and compliance
with Oracle's policies.
 Troubleshooting, performance tuning, and optimizing databases and BI systems for efficient data retrieval and
analysis.
 Developed or mentoring junior BI engineers and developers in best practices aligned with Oracle's technology
stack.
 Staying updated with the latest BI technologies and trends, particularly within Oracle's suite of BI and cloud
analytics products, to continuously improve BI capabilities.
 Documenting BI architectures, data flows, ETL processes, and technical specifications for maintenance and
knowledge sharing.

Environment: ORACLE Cloud, SQL, Shell Scripting, OBIEE, Informatics, DAC, IPM, SOA, Sisense, Info web, Service Now,
Hyperion, Essbase, EBS R12.

You might also like