You are on page 1of 4

Sai Thikkavarapu 🕾 +1 774-519-0973

Westborough, MA tsaik917@gmail.com

Career Objective

Seeking assignments on the latest technology with a challenging environment and a growth-oriented
organization.

Summary of Expertise

● 10+ years of Experience in Cloud/SRE/DevOps/DB, Application Deployment, and development lifecycle.

● An effective communicator with excellent relationship-building & interpersonal skills.

● Strong analytical, problem-solving & organizational abilities. Possess a flexible & detail-oriented attitude.

Academic Background

● B.E in Computer Science from ANNA UNIVERSITY, India

● Masters in Embedded Software Engineering (Program of Computer Software) from Gannon University.

PROJECTS:

Bose Corporation
Duration: Sep 16 to Current
Technologies: AWS, APIGEE, GCP, Kubernetes, SonarQube, Qualys, Jenkins, Python, Prometheus, Grafana, Splunk,
JMeter, Linux (RHEL), Docker, Terraform, Ansible and EFK stack, Jaeger/Dynatrace (Tracing), BASH, KAFKA,
Various Databases (Oracle, Aurora - MYSQL/POSTGRES)

Roles: Linux Engineer

Responsibilities:
✔ Gathers and documents requirements from a diverse user community and Translates requirements to
functional design documents for complex systems/processes.
✔ Work with the development teams to ensure alignment of development needs and timelines with
infrastructure capacity and capabilities.
✔ Worked in Kubernetes operations managing various activities like deployment, logs viewing using stern,
creating docker images for apps, Scaling of applications, archival, Secrets management, and role-binding
features.
✔ POC for migration of APIGEE Edge to APIGEE X.

✔ Worked on GCP components like Compute Engine, Cloud Storage (for performance testing of applications), and
Cloud SQL.
✔ Worked on AWS features including security extensively as admin i.e. EC2, ELB, SES, VPC, RDS, S3, SNS, Billing
dashboards, CloudWatch, IAM, DNS, Route 53, WAF, Secrets Manager, Certificate Manager, and worked with
AWS support for various maintenance activities.

Page 1 of 4
✔ Used Terraform for automated provisioning of resources on native AWS stack and Ansible to manage various
nodes and software installations.
✔ Worked on security components of infrastructure like vulnerability remediation using Qualys and deployments
of newly created AMI to PROD.
✔ Experienced in developing Splunk queries and dashboards for various applications.

✔ Supported Legacy infrastructure (Java, Tomcat, Ngnix) as tier 1 ops, installation, and troubleshooting of issues
like restarting ad updating Tomcat configurations as required.
✔ Worked on Terraform and Ansible to provision deployments to QA, NonPROD, and PROD envs.

✔ Worked on Qualys to monitor and patch vulnerabilities and AMI creations for higher envs.

✔ Installed and configured Qualys agent to various worker nodes for automated node scans and reported them to
various teams using EXCEL and Power BI Dashboards
✔ Configure and Test High Availability (HA) and Disaster Recovery (DR) scenarios for Databases like aurora
MSQL, DB, etc. for customer-facing applications.
✔ Created various custom dashboards for service monitoring using metrics from APIGEE Stats and monitoring
API
✔ Various APIGEE Admin tasks like deploying proxies, and API keys, adding policies like JWT validations, Spike
Arrests, etc.
✔ Custom Report Jobs creation to analyze various metrics like errors/tps/latencies on APIGEE proxies in
production.
✔ Develop and run Containerized Microservices through the Orchestration tool Kubernetes.

✔ Worked Heavily on JMETER scripts for Performance.

✔ Configured many JMETER scripts on various environments and was responsible for managing the performance
aspect of services in production
✔ Did Performance testing for many applications (Java, Python) and scaled the resources for holiday traffic.

✔ Managed entire soak test environment including deployments and execution of soak test and providing
recommendations/reported various bugs.
✔ Installed and Configured blazemeter plugin for JMeter tests to have better reporting of test results for
stakeholders.
✔ Develop and improve existing Cloud Platform by writing Continuous integration and continuous delivery
scripts in Python and shell scripting.
✔ Used the Blue/Green deployment method to upgrade various RDS instances of Aurora and Postgres databases
with almost zero downtime.
✔ Configured various cloudwatch alerts for S3, WAF, and RDS resources in AWS.

✔ Support the Production platform and be involved in 24*7 On-Call Support on a Rotation basis to handle any
issues that arise with production applications and doing RCA.
✔ Used BOTO3 to write various automation for features like S3 encryption, resource management like reserved
instances, Default VPC and their components deletion, etc.

Page 2 of 4
✔ Create and maintain Runbooks and Troubleshooting guides for internal applications for debugging issues in
various Environments.
✔ Created various Grafana dashboards for metrics like CPU/MEM, TPS, Latency, and error codes for various
applications and guided developers on best practices.
✔ Configured various data sources like Prometheus, Aurora MySQL, etc. to pull various metrics and create
dashboards in Grafana for visibility of key KPIs
✔ I wrote complex Prometheus queries and constructed Grafana dashboards for anomaly detection across
various applications.
✔ Implement Production monitoring using Tools like Prometheus and Grafana.

✔ Wrote various alert rules to manage the anomalies in CPU/MEM, TPS, Latency, and error codes.

✔ Develop Python Scripts for Performance/Stress Testing applications newly being Onboarded to benchmark
and provide resource recommendations.
✔ Support Jenkins to create build/deploy Jobs and create schedules to run them in an automated way and
manage access for various development teams.
✔ Support SonarQube Tool to run Security Analysis and develop code-coverage scripts for applications.

DC HBX (District of Columbia – Health Benefit Exchange)


Duration: March 2015 to Sep 2016
Technologies: LINUX, Oracle RAC, PL/SQL, IBM-CURAM, Kubernetes, Web logic, Dynatrace, SVN, SonarQube,
Jenkins (BUILD-AUTOMATION), Groovy, Grafana, BASH

Roles: SRE/DevOps

Responsibilities:
✔ Gathers and documents requirements from a diverse user community and Translates requirements to
functional design documents for complex systems/processes.
✔ Planned and Implemented CURAM DATABASE upgrade from version 6.0.5.4 to 6.2 supporting DC-HBX and
DCAS related Applications.
✔ Did POC on Dockerizing applications and building custom Docker images for developers.

✔ Creating, configuring, and maintaining Oracle Automated Database Maintenance using Shell scripting
techniques.
✔ Supported Applications running in Kubernetes on AWS cloud involved entire Cluster Setup, deployments, Role-
based Access-controls, Auto-Scaling, and Namespace Management using Python scripts.
✔ Controls migrations of programs, database changes, reference data changes, and menu changes through the
development life cycle.
✔ Used Dynatrace to debug high-latency calls coming from the app to the database.

✔ Performance tuning and Optimization (PTO) of Oracle Databases by Generating AWR (Automatic Workload
Repository) and ASH (Active Session History) reports using Oracle SQL DEVELOPER (Tool) and analyzing
them.

Page 3 of 4
✔ Creation of AWS Development (CENTOS, RHEL) and Production stack and deploy the developed Code onto
AWS stack to ensure operational readiness of Code.
✔ AWS RDS Provisioning and upgrades using terraform.

✔ Managing Backup/Recovery and Cloning Operations of the Databases using RMAN and export/import Utilities
of Oracle and integrated them with Jenkins Jobs to run and notify their success on Grafana dashboards.

AIRTEL
Duration: Dec 2012 to Nov 2013
Technologies: LINUX, RHEL, UBUNTU, Shell scripting(bash)
Roles: Software Analyst

Responsibilities:

✔ Involved in patching of Linux systems

✔ Establish monitoring, backup, and logging procedures for all environments

✔ Administer and maintain multiple Centos/RHEL and other Linux environments, establish and enforce
configuration management controls, and perform security patches.
✔ Database Optimization and Tuning using Execution Plans, SQL Trace, and AWR Reports.

✔ Shell scripting to automate various monitoring of file systems and other system resources

✔ Software installation and configu\rations.

Page 4 of 4

You might also like