You are on page 1of 7

Ram Manohar Bheemana

Senior Big Data/Cloud Architect

Contact About Me
A Computer Science under-graduate with extensive
experience as Senior Architect in Big-data and cloud
technologies. Has 13 years of professional experience in field
of Data-Engineering. Has working knowledge on data-science
and machine-learning. Frequent blogger and contributor to
open-source Community. Follower of technical scholarly
articles related to Artificial Intelligence, Machine Learning and
Computer Vision.

Work Experience

Skills Hilton
10/2017 - Now
As a Sr. Lead Architect, designed and
implemented Ingestion, Data-Lake,
Data-warehousing and Data-Science
capabilities
Big Data Cloud – AWS
Spark Redshift, S3 Global Atlantic Architect and build cloud infrastructure
03/2017 – 10/2017 on AWS with disaster recovery model
Kafka Machine Learning
HBase Data Science Putnam Investments Built and Supported Data-Science
08/2016 – 02/2017 Capabilities
Certifications PNC Bank Built Big-data capabilities from ground-
02/2015 – 08/2016 up
AWS Certified Solutions Architect –
Associate Royal Bank of Developed and supported Investment
Canada Products
Cloudera Certified Developer for 09/2010 – 02/2015
Apache Hadoop (CCDH)
Machine Learning Computer Sciences Developed and Supported Farmers
Computer Vision Corporation Insurance products
RBS Certified AGILE Practitioner 06/2007 – 08/2010

Education
2007 Bachelor of Technology, Computer Science
Ram Manohar Bheemana
Senior Big Data/Cloud Architect

Summary

Around 13 years of IT experience with over 8 years in Big data technologies and 5 years in Amazon cloud.

✓ Experience in architecting, designing and implementing data-lakes with Hadoop Infrastructure and AWS
cloud infrastructure
✓ Educated and helped various clients in
¨ building Hadoop centric projects from ground-up
¨ setting Big-data environment and laying down the code standards and best practices
✓ Working experience on Hadoop Distributions
¨ Cloudera distribution
¨ Hortonworks distribution
¨ Amazon Webservices.
✓ Experience in building
¨ Scripts/applications with Shell, Java, Scala and Python languages.
¨ Data-science propensity and forecast models
¨ Hadoop Batch applications with SPARK, MapReduce, PIG, HIVE, SQOOP, AVRO, PARQUET
¨ Near real-time streaming applications with NIFI, Flume, Kafka, HBASE and SPARK streaming
¨ Workflow and coordinator jobs using OOZIE and AirFlow
¨ Platforms for Business-Intelligence using Impala, Hive, Zeppelin and Jupiter Notebooks
¨ Realtime dashboards and centralized logging framework using SOLR
¨ Automation scripts of infrastructure using Cloud Formation and Boto3 APIs
¨ tailor-made low-level applications based on client requirements using Custom NIFI Processors and
Custom Hive SERDE for data ingestion
✓ Working knowledge on
¨ Machine learning
¨ Computer vision
✓ Adept at gathering and documenting requirements, assessing business objectives, and managing all stages
of software development and implementation.
✓ Excellent interpersonal skills and communication skills.
✓ Strong domain experience in
¨ Banking Industry
¨ Investments Industry
¨ Hospitality Industry
✓ Strong leader with people management skills and be able to influence, facilitate, gain consensus and resolve
conflict.
✓ Proven track record of completing and managing large, complex, multiple and aggressive projects on
schedule with limited resources and environment with tight schedules.

Social/Community Contribution

✓ Built open-source project to copy data and transform the complex mainframe (VB and FB) format files to
Hadoop-Hive using FTP and Hive custom SERDE. See more at http://rbheemana.github.io/Cobol-to-Hive/
✓ Active blogger at github.com, recent being http://rbheemana.github.io/reality-of-sqoop-mainframe/
✓ Active contributor to stackoverflow.com
✓ Developed an automatic ingestion framework for import of data from SQL database to Hadoop through
Python, OOZIE, Hive and SQOOP which can be easily integrated with scheduling tools like ZEKE,
AUTOSYS, CRONTAB etc.
https://github.com/rbheemana/Sqoop-Automated
✓ Proposed a better solution to sample the data than the InputSampler.java in an open source map reduce
project.
https://issues.apache.org/jira/browse/MAPREDUCE-6423
✓ Built an iOS application called ‘Row Master’
Ram Manohar Bheemana
Senior Big Data/Cloud Architect

Skills

Hadoop AWS Databases: Languages: Frame works:


Core Java PMF and AGILE
SPARK / Streaming S3 Teradata Python
Kafka Aurora Red Shift Scala Operating Systems
Map Reduce Dynamo DB My SQL Shell Linux
SQOOP Redshift SQL Server C# Windows XP
HIVE CloudFormation Oracle Java Script OS/390
OOZIE Lambda IBM DB2 COBOL
SOLR Kinesis IMS DB VSAM
PIG
HBASE
NIFI

Education
ANDHRA University
Bachelor of Technology (B.TECH.), Computer Science, 8.18 GPA – 2007

Certifications
AWS Certified Solutions Architect – Associate
September 2019

Cloudera Certified Developer for Apache Hadoop (CCDH)


Cloudera, License 100-011-634 December 2014

Machine Learning
License 3H3ZE96C7VDV January 2017

Computer Vision
courses.learnopencv.com

IBM Certified Database Associate DB2 Universal Database V8.1 Family


IBM August 2009

RBS Certified AGILE Practitioner


Royal Bank of Scotland Jan 2013
Ram Manohar Bheemana
Senior Big Data/Cloud Architect

Professional Experience
Sr. Big Data/Cloud Architect
Concept Software & Services Inc.
Client: Hilton Hotels
Oct 2017 – Present (2-year 5 months) | Mclean VA, US

Enterprise Information Management (EIM) is an initiative to gather all planet wide Hilton data into one single
application. EIM acts a central application to derive business intelligence which will enable Hilton to run campaigns,
derive pace and position of revenue, etc. Various data-science principles are applied on the EIM data to get insights.
EIM is a complex application which stores Peta Bytes of Data which includes wide-variety of data formats (csv,
JSON, xml, streaming data, etc.)

Responsibilities:
→ Designing and implement Big data Cloud environment using AWS to handle Peta Bytes of data.
→ Implement and govern code standards and best practices
→ Architect and develop applications with Hadoop Infrastructure like SPARK, SPARK-streaming, MapReduce,
HIVE, SQOOP, OOZIE, Kafka, SOLR, HBASE, MYSQL, Red Shift, NIFI, GIT and Maven tools
→ Built Data-Science Propensity and revenue forecast models and deployed on Big data Cloud
→ Performance tuning of SPARK applications jobs
→ Lead the team towards achieving the project goals
→ Implement solutions to support data retention policies
→ Implement solutions to save cost by data segregation (hot/cold data)
→ Mentor the team members in achieving knowledge on Big Data technologies
→ Support AWS QuickSight dashboards and reporting solutions over S3, Presto, Redshift.

Achievements:
→ Small File Problem: Spark applications are creating too many small problems with skewed data and
repartitioning. Implemented a solution which accepts the number of output files as parameter and normalizes the
output files size with optimal performance.
→ Unzip Large Files: Hilton receives huge volume of data in zip files (5 GB) from 5400 properties. Unzipping and
processing the files became a problem as unzipping has to be done on a local file system and consuming too
much time. Designed and Implemented a solution to separate unzipping and processing of files. Handled
unzipping using map-reduce application and processing with SPARK. This design increased the performance by
300%.
→ Unit Testing Framework: Implemented testing framework into code base so that testing of Big Data applications
can be done on local machine without the need of cluster. This is achieved by creating SPARK, HIVE,
ZOOKEEPER, S3 servers in local.
→ Logging Framework: Reading logs in Hadoop application is tedious and time-consuming task. Implemented
SOLR-logging for all big-data applications and created Dashboards on Banana to retrieve information seamlessly
→ NIFI Large File transformations: Faced memory issues while transforming large files to JSON using NIFI
processors which reads full file in memory. Implemented NIFI custom processors for transformations using output
streams and efficient row-wise operations.
→ With increase in volume of the data, NIFI was quickly becoming the bottleneck. Designed and implemented the
ingestion solution using Spark. As solution NIFI cluster is a costly affair comparatively.

Environment: Amazon web services, NIFI, Hive, SQOOP, FLUME, AWS Redshift, AWS S3, Apache SPARK, Scala,
JAVA, HBASE, JDK 1.8, SOLR, Red hat Linux, Python, Shell scripting, Oracle, Hortonworks distribution
Ram Manohar Bheemana
Senior Big Data/Cloud Architect

Hadoop Architect
Concept Software & Services Inc.
Client: Global Atlantic Financial Group
Mar 2017 – Oct 2017 (8 months) | Boston, MA, US

Global Atlantic Financial Group, through its subsidiaries, offers a broad range of retirement, life and reinsurance
products designed to help our customers address financial challenges with confidence. A variety of options help
Americans customize a strategy to fulfill their protection, accumulation, income, wealth transfer and end-of-life needs

Worked as a Hadoop Architect on various Big Data initiatives to meet various business requirements in the area of
Big Data analytics and reporting

Responsibilities:

→ Provide Road map for Big-Data applications


→ Architect and build cloud infrastructure on AWS with disaster recovery model
→ Design and developed various Big Data workflows using custom Spark, Hive and SQOOP.
→ Implement and govern industrial guidelines
→ Benchmarking various tools and applications like Hive, Spark, Hive LLAP, INFORMATICA BDM
→ Performance tuning of SPARK applications jobs
→ Lead the team towards achieving the project goals
→ Implement solutions to support data retention policies
→ Implement solutions to save cost by data segregation (hot/cold data)
→ Mentor the team members in achieving knowledge on big data technologies

Achievements:
→ Designed and implemented AWS cloud formation techniques for auto-scaling needs
→ Built data-ingestion streams using SQOOP from Oracle/MySQL/SYBASE databases.
→ Built big data workflows to ingest the data from various sources to Hadoop using OOZIE and these workflows
comprises of heterogeneous jobs like Hive, SPARK, SQOOP and Python Script.
→ Designed end-to-end architecture for Ingesting various sources to Business layer.
→ Leveraged Hadoop SFTP file system to implement custom map-reduce jobs to transfer the files from SFTP
server into HDFS.
→ Enhanced the open source Spark-ftp connector and implemented to read files directly from the SFTP server.
→ Educated the clients on the usage of Hadoop and AWS cloud services to meet their business needs
→ Educated the client on the usage of RDS service for the reliable Database servers.

Environment: Amazon web services, Hive, SQOOP, FLUME, Apache SPARK, HBASE, JDK 1.7, SOLR, Red hat
Linux, Python, Shell scripting, Oracle, Hortonworks distribution
Ram Manohar Bheemana
Senior Big Data/Cloud Architect

Hadoop Solution Architect


Concept Software & Services Inc.
Client: Putnam investments
October 2016 – Feb 2017 (5 months) | Boston, MA, US

Putnam offers actively-managed mutual funds and Institutional strategies across equities, fixed income, absolute
return, and alternatives.

Currently working as a Hadoop solution architect on various Big Data initiatives to meet various business
requirements in the area of Big Data analytics and reporting.

Responsibilities:

→ Design technical architecture and develop various Big Data workflows using custom Spark, Hive and SQOOP.
→ Support and build the Data Science projects on to Hadoop
→ Advising the industrial guidelines on the Hadoop applications setup
→ Benchmarking various options that are available in Hive, Spark and Impala.
→ Performance tuning big-data applications.
→ Evaluating Hadoop ingest tools like TALEND, RCG framework
→ Converted various SAS scripts into their equivalent Hive QLs
Achievements:
→ Created Custom FTP job to import the Freddie Mac and Fannie Mae data directly into HDFS and to do
preprocessing steps in hive.
→ Created various scripts to import the data various internal sources using Curl commands.
→ Built SOLR Dashboards for audit data by build indexes on the Oracle audit data.
→ Created custom python/shell scripts to import data via SQOOP from Oracle databases.
→ Created big data workflows to ingest the data from various sources to Hadoop using OOZIE and these workflows
comprises of heterogeneous jobs like Hive, SQOOP and Python Script.

Environment: Hive, SQOOP, FLUME, Apache SPARK, HBASE, JDK 1.7, SOLR, Red hat Linux, Python, Shell
scripting, Oracle, Cloudera distribution

Hadoop Lead/Architect
Concept Software & Services Inc.
Client: PNC Bank
Feb 2015 – October 2016 (1 year 8 months) | Cleveland, OH, US

Hadoop is being used as data central repository for all PNC data. BDH is the application which pulls the data from various
data sources and stores into Hive database. Other Applications such as APD, IAE, and FRAUD etc. use the centralized
data for their particular needs.

Worked as a Hadoop lead/architect on various Big Data initiatives to meet various business requirements in the area
of Big Data analytics and reporting.

Environment: Hive, SQOOP, FLUME, Apache SPARK, HBASE, JDK 1.6, Maven, Red hat Linux, Python, Shell
scripting, Teradata, SQL Server, Oracle, MySQL, Cloudera distribution
Ram Manohar Bheemana
Senior Big Data/Cloud Architect

Hadoop Developer/Lead Technical System Analyst


IGATE.
Client: Royal Bank of Canada
November 2012 – February 2015 (2 year 4 months) | Hyderabad, India|
Toronto, Canada
Environment: MapReduce, Pig, Hive, SQOOP, FLUME, HBASE, JDK 1.6, Maven, OS/390, Linux.

Technical Lead
IGATE.
Client: Royal Bank of Canada
September 2010 – November 2012 (2 years 3 months) | Hyderabad
Area, India
Environment: Mainframe, COBOL, JCL, REXX, FOCUS, WEB FOCUS OS/390, Linux.

Software Engineer
CSC - Computer Sciences Corporation
Client: Zurich Farmers
June 2007 – August 2010 (3 years 3 months) | Hyderabad Area, India
Environment: Mainframe, COBOL, JCL, CICS, IMS-DB OS/390, Linux.

You might also like