Nothing Special   »   [go: up one dir, main page]

Naveen - Srigakollu - Devops 12+

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 8

Naveen Srigakollu

Email: naveensrigakollu3007@gmail.com AWS Certified Senior SRE


+1 469-740-5125 Engineer
CICD, Dynatrace, Splunk
Summary:

 Senior SRE Engineer with over 12 years of extensive experience in Cloud (AWS), Java applications,
Configuration management, Infrastructure automation, Continuous Integration and Delivery (CI/CD).
As SRE Engineer focused on Build and Deployment Automation, Writing and Managing CI/CD pipelines
using AWS DevOps, Git, GitHub Actions, Jenkins, Containerization and Orchestration, and monitoring
to help teams deliver a better, reliable Production environment.
 Implemented Terraform Scripts to build AWS Infrastructure.
 Worked as a production support engineer and supported Java based P1/P2 applications which are
hosted in AWS
 Develop and optimize SQL queries.
 Implement and manage IT Service Management (ITSM) processes to ensure efficient incident,
problem, and change management.
 Experience working with Apache Hadoop, Kafka, Spark and Log stash.
 Worked with Apache Kafka for High throughput for both publishing and subscribing, with disk
structures that provide constant
 My core competences are Infrastructure Management, Operations Management, Change
Management, SLA Management, Application Availability Management, Application Cost
Optimization, Customer Engagement.
 Experienced in working on DevOps/Agile operations process and tools area (Code review, unit test
automation, Build & Release automation, Environment, Service, Incident and Change Management).
 Provisioning of AWS resources like EC2, VPC, EBS, AMI, S3 buckets, creation of subnets and all other
operational tasks
 Experience in Monitoring server performance with tools like Splunk, Datadog, Grafana, Prometheus,
new relic and Cloud Watch, BigPanda, Dynatrace.
 Configured custom dashboards in Dynatrace to provide real-time insights into application health and
performance metrics.
 Expertise in Querying RDBMS such as Oracle, MYSQL and SQL Server by using SQL for data integrity.
 Utilized Dynatrace alerts and notifications to proactively identify and resolve performance issues
before they impact end-users.
 Collaborated with development and operations teams to optimize application performance based on
Dynatrace insights.
 Experienced in setting up and configuring Splunk environments for various use cases.
 Deployed Splunk for log aggregation and analysis across all my applications in Dev, UAT and
Production environments.
 Created and optimized Splunk queries and search strings to extract actionable insights from large
datasets.
 Developed custom Splunk dashboards and reports to monitor system performance, security events,
and operational metrics.
 Experience in implementation and setting up the tools in high availability. (SVN, GIT, ARTIFACTORY,
NEXUS,JENKINS, JIRA).
 Managed environments DEV, QA, UAT and PROD for various releases and managed using Blue-
Green and Canary deployment strategies.
 Highly organized, detailed oriented, able to plan, prioritize work and meet deadlines. work well under
tight deadlines.
 Troubleshooting the application issues using the Dynatrace Tool for finding the Root Cause Analysis of
P1/P2 tickets.
 Ability to work directly with all levels of Management to gather user requirements.
 Experience of Jenkins, Apache Tomcat, JBoss, Subversion, Git, Maven.
 Experience on implementation of SonarQube for Continuous static code analysis with CI/CD systems
such as Jenkins & build tool Maven.
 Monitoring profiles for AWS Services - EC2 Parameters (CPU, Memory, Disk, Response time, etc.)
 Technical expertise in facilitating Cloud Infrastructure Management for entire org, Experienced in
Amazon Web Services like Amazon EC2, EBS, ELK, ECS, S3, Glacier, RDS, ELB, VPC, Route 53,Cloud
trail, Lambda, Code Deploy, Elastic Cache, SNS, SQS, SES, Cloud Formation, Cloud Front, Cloud
watch, IAM, Import, Directory Service, Cognito.
 In-depth knowledge of IAM security features such as password policies, identity federation, and
role-based access control (RBAC), effectively configured and managed via the AWS CLI.
 Experience in scripting languages Python.
 Experience in Windows/ Linux Administration (Installation, Configuration and Upgrades of
Linux (Redhat, Centos, Ubuntu, Suse).
 Experience in creating, reimaging and cloning datacenters (VMs) on ESXI platform.
 Experience on configuration management tool Ansible, Terraform.
 Expertise in using Terraform for deploying Cloud Infrastructure in AWS/Azure.
 Experience in writing Groovy & bash scripts for automation of build and infrastructure automation.
 Experience is using Tomcat, JBOSS and Nginx servers for deployments.
 In-depth understanding of the principles and best practices of Software Configuration Management
(SCM) processes, which include compiling, packaging, deploying and Application configurations.
 Worked on Container management using Docker by writing Docker files and setting up the
automated build on Docker, installing and configuring Kubernetes.
 Experience in configuring and managing the Kubernetes Clusters to Deploy, Scale, Load Balance and
manage Containers which includes creation of Pods, Replica Set, Labels, Deployments, Services,
Ingress, Config Map, Secret and Health Checks using Liveness Probe, Readiness Probe by writing YAML
Scripts.
 Excellent communication, interpersonal and managerial skills.

Technical Skills:

Operating System Windows Variants, Linux, UNIX.


AWS Services EC2, VPC, IAM, EBS, S3, ELB, Auto Scaling, EKS, ECR, AWS CLI,
Route53,Lambda,CloudWatch AWS Database Migration Service (DMS),
AWS Application
Migration Service (AMS), Cloud Formation Templates
Server Apache Tomcat, JBoss, Nginx
Issue Tracking JIRA, Service Now
Database My SQL, Oracle, DynamoDB
Version Control GitHub, Bit Bucket
CI Tools Jenkins, Bamboo, GITLAB
Build Tools Maven
Repository Tools Nexus, Artifactory, Archiva
Quality & Test Automation SonarQube
Cloud AWS
CM Tools Ansible, Terraform
Containers Docker
Orchestration Docker swarm, Kubernetes
Monitoring Tool Splunk, Dynatrace, Grafana, Kafka, Prometheus, Nagios, AWS Cloud
Watch, New relic, Datadog
Programming & Scripting Java,Groovy, Bash scripting, Python

Educational Qualification:
 Master of Computer applications (MCA), JNTU Kakinada, 2011
 Bachelor in science Andra Univesity, India 2008

Projects Execution

Project # 5 HKS – Hitachi Kubernetes Service


Duration Sept-2023 to Till Date
Client Hitachi Vantara
Role SRE Lead
AWS (Cloud Technology) EC2, Elastic Load Balancer, Auto Scaling Group, VPC, Route 53 and
RDS
Monitoring Tools Cloud Watch, Grafana & Kibana

Description:
Hitachi Kubernetes Service, a cloud hosted SaaS control plane for managing Kubernetes clusters and their
workloads. Hitachi Kubernetes Engine, a Kubernetes distribution maintained by Hitachi Vantara.

Environment: AWS (EC2, IAM, S3, VPC, Route53, EKS, Lambda, Cloud Watch, Cloud Front, ), Jenkins, GIT,
Kubernetes, Elastic Load Balancers, Amazon RDS (SQL Server, PostgreSQL), Nexus Repository, Terraform,
Shell/Python Scripting, Groovy, WebLogic, Python Programming.

Responsibilities:
 Managing and supporting production releases with at most precision and delivered value.
Maintained GIT workflows for production control. Provisioned servers and deployed playbooks for
Linux servers patching. Kubernetes nodes, pods, config-maps, routes, and secrets.
 Launching EC2 instances and involved in AWS RDS, S3, Load Balancing, IAM, VPC, Lambda, and
Cloud Watch.
Monitor and report on ITSM metrics to identify trends, areas for improvement, and compliance with
SLAs.
Worked on SQL Server, DynamoDB, RDS, and Aurora DB by deploying cloud databases using CI/CD
pipelines in AWS.
 Worked on AWS Lambda for Cost Optimization of resources using Python.
 Used AWS Route53, to route the traffic between different availability zones. Deployed and
supported Memory-cache/AWS Elastic-Cache and then Configured Elastic Load Balancer (ELB) for
routing traffic between zones.
 Working closely with Cassandra loading activity on history load and incremental loads from
Teradata and Oracle Databases and resolving loading issues and tuning the loader for optimal
performance. Responsible for Installing, setup and Configuring Apache Kafka and Apache
Zookeeper.
Provide training and support to IT staff on ITSM tools and processes to ensure effective adoption
and usage.
 Configured CI/CD pipeline in Jenkins to implement Continuous Integration and Continuous
Delivery process, accommodate software teams with compilation and artifact deployment
requests in an AWS cloud environment.
 Experience in creating and managing IAM users, groups, and roles, and configuring fine-grained
permissions for AWS resources using Python scripts.
 Launching WebLogic servers for developing, deploying, and running enterprise applications,
such as Java, in the cloud
 Integrated Kafka with Flume in sand box Environment using Kafka source and Kafka sink
 Worked with application teams to install operating system, Hadoop updates, patches, version
upgrades as required. Integrated Kafka with Spark in sand box Environment.
 Responsible for Installing, setup and Configuring Apache Kafka and Apache Zookeeper
 Used Kafka to collect Website activity and Stream processing.
 Used WebLogic for troubleshooting, enhancing performance, and constantly improving the
server and applications.
Configured and fine-tuned MySQL and MS SQL Server databases to meet specific application
requirements and performance benchmarks.
 Setting up new Micro services in all environments from Dev to production. Monitored
Server’s health and micro services health using tools like Grafana, Prometheus.
 Worked with Development and Testing teams to design / develop robust solutions to meet the
client requirements for functionality and performance.
 Implementation of the DevOps roadmap through the definition of a common set of design
standards, driving the adoption of these standards across the enterprise.
 Designing, implementing, and maintaining CI/CD pipelines for cloud-based applications using
tools like Jenkins, Gitlabs and AWS Code Pipeline.
 Automated day-to-day activities like Kubernetes troubleshooting and health checks using
Python/Shell.

Project # 4 VStaaS - Hitachi Virtual Storage as a Service


Duration Jan-2022 to Aug - 2023
Client Hitachi Vantara
Role Senior SRE
AWS (Cloud Technology) EC2, Elastic Load Balancer, Auto Scaling Group, VPC, Route 53 and
RDS

Description:
Hitachi Virtual Storage as a Service is designed for midsize organizations looking for a private cloud that
offers a "public-like cloud experience" for storage consumption. By providing a cloud-based management
interface, Hitachi Virtual Storage as a Service delivers control of on-premises or cloud hosted enterprise-
class storage infrastructure.

Environment: AWS (EC2, IAM, S3, VPC, Route53, EKS, Lambda, Cloud Watch, Cloud Front), Jenkins, GIT,
GitHub, SonarQube, Maven, Tomcat, Docker, Kubernetes, Terraform, Grafana, Prometheus

Responsibilities:
 Led efforts to achieve HIPAA compliance for cloud-based healthcare systems, including
implementing encryption mechanisms and access controls to protect patient data.
 Managing and supporting production releases for Enterprise applications. Implemented CI/CD
pipeline in Development lifecycle.
 Migrated the production MySQL schema to the new AWS RDS Aurora instance.
 Continuously evaluate and improve ITSM processes to enhance service quality and efficiency.
 Worked on the Analytics Infrastructure team to develop a stream filtering system on top of
Apache Kafka.
 Implemented Kafka Storm topologies, which are capable of handling and channelizing high stream
of data and integrating the storm topologies with Esper to filter and process that data across
multiple clusters for complex event processing.
 Have built the entire infra from scratch through Terraform.
 Written groovy scripts to use multi branch pipeline modules in Jenkins to configure it as per
product requirements
 Implemented a distributed messaging queue to integrate with Cassandra using Apache Kafka and
ZooKeeper
 Worked with application teams to install operating system, Hadoop updates, patches, Kafka
version upgrades as required.
 Collaborate with vendors and external partners to align ITSM practices and ensure integrated
service delivery.
 Developed Maven Scripts to automate the build process. Managed the Maven Repository using
Nexus tool and used the same to share the snapshots and releases of internal projects.
 Developed, maintained and enhanced pre and post build scripts (Shell)
 Container management using Docker by writing Dockerfiles and set up the automated build on
Docker HUB and installed and configured Kubernetes.
 Developed a stream filtering system using Spark streaming on top of Apache Kafka.
 Designed a system using Kafka to auto - scale the backend servers based on the events
throughput.
 Administered and maintained MS SQL Server databases on-premises, handling routine tasks
such as backups, restores, and performance tuning.
 Participate in ITSM tool selection, implementation, and upgrades to enhance service management
capabilities.
 Virtualized the servers using Docker for the test environments and dev-environments needs, also
configuration automation using Docker containers.
 Responsible for installation & configuration of Jenkins to support various Java builds and Jenkins
plugins to automate continuous builds and publishing Docker images to the Nexus repository.
 Used CI/CD tools Jenkins, Git/GitHub, Jira and Docker registry/daemon for configuration
management and automation using Ansible.
 Implemented SoanrQube setup from scratch.
 Written groovy scripts for continuous delivery framework using Jenkins, Maven on multiple
environments.
 Assisted in the migration of legacy databases to newer versions of MS SQL Server, ensuring
compatibility and data integrity.
 Developed scripts for executing validation/test cases automatically after builds.

Project # 3 Optum Provisioning


Duration Feb-2018 to Jan-2022
Client Optum Global Solutions
Role SRE
AWS (Cloud Technology) S3, EC2, Elastic Load Balancer, Auto Scaling Group, VPC, Route 53
and RDS
DevOps Tools Git, Docker, Ansible and Jenkins.
Technology Java Application
Monitoring Tools Cloud Watch, New Relic, Splunk, Dynatrace etc..
Description:
Optum Provisioning provides centralized management of organizations, products, users, and roles,
creating a one-stop shop for an administrator of an Optum Cloud application to add users and configure
their access. Provisioning refers to granting access to functionality within an Optum service (this refers to
both services and products licensed from Optum). Once a user is provisioned, the user has access to the
appropriate functionality within one or more Optum services.

Environment: AWS (EC2, IAM, S3, VPC, Route53, EKS, Lambda, Cloud Watch, Cloud Front), Jenkins, GIT,
GitHub, SonarQube, Maven, Tomcat, Docker, Kubernetes, Terraform, Grafana, Prometheus

Responsibilities:

 Interacting with partners to capture requirements.


 Gathered user requirement and performing functional and detailed design analysis.
 Analyzing the cloud requirement document and Design the AWS Services like EC2, Load Balancer,
Auto Scaling Group etc in track with change tickets.
 Installing, Configure the Apache, Tomcat Server.
 Generate regular reports on ITSM performance, incidents, changes, and problem resolution for management
review.
 Implemented Cloud watch alarms where ever possible for taking proactive actions.
 Analyzing the cloud watch alarms trigged and taking actions immediately to avoid impact to
business.
 Maintained the user accounts (IAM), RDS (SQL), Route 53, VPC, Dynamo DB (NOSQL), SES, SQS and
SNS services in AWS cloud.
 Manage security groups in specific to VPC for providing high security.
 Create & Manage the AMI/Snapshots/Volumes.
 Troubleshooting network problems.
 Created and configured S3 Bucket for the users to store project related documents.
 Resolving Cloud Incidents within the provided SLA.
 On Call support for 24/7 Production environments.
 Handling Critical Production environment issue effectively with in the time.

Project # 2 One Healthcare ID


Duration Oct-2014 to Jan- 2018
Client United Healthcare
Role Software Engineer
AWS (Cloud Technology) S3, EC2, Elastic Load Balancer, Auto Scaling Group, VPC, Route 53
and RDS
DevOps Tools Git, Docker, Ansible and Jenkins.
Technology Java Application
Monitoring Tools Cloud Watch, New Relic, Splunk, Dynatrace etc..

Description:
OneHealthcareID is a centralized login application utilized to authenticate user logins across various
applications within UnitedHealthcare. Serving as a single authentication point, it streamlines the login
process for users across multiple applications. This platform is responsible for verifying user credentials
and seamlessly redirecting them to their respective applications upon authentication.

Environment: AWS (EC2, IAM, S3, VPC, Route53, EKS, Lambda, Cloud Watch, Cloud Front, ), Jenkins, GIT,
Kubernetes, Elastic Load Balancers, Amazon RDS (SQL Server, PostgreSQL), Nexus Repository, Terraform.

Responsibilities:

 Managing and supporting production releases with at most precision and delivered value.
Maintained GIT workflows for production control. Provisioned servers and deployed playbooks for
Linux servers patching. Kubernetes nodes, pods, config-maps, routes, and secrets.
 Launching EC2 instances and involved in AWS RDS, S3, Load Balancing, IAM, VPC, Lambda, and
Cloud Watch.
Migrated MSSQL Server database from Rackspace to AWS and provided day to day support.
Maintain detailed documentation of ITSM processes, incident reports, change requests, and
problem resolutions.
 Worked on AWS Lambda for Cost Optimization of resources using Python.
 Used AWS Route53, to route the traffic between different availability zones. Deployed and
supported Memory-cache/AWS Elastic-Cache and then Configured Elastic Load Balancer (ELB) for
routing traffic between zones.
 Working closely with Cassandra loading activity on history load and incremental loads from
Teradata and Oracle Databases and resolving loading issues and tuning the loader for optimal
performance.
 Configured CI/CD pipeline in Jenkins to implement Continuous Integration and Continuous
Delivery process, accommodate software teams with compilation and artifact deployment
requests in an AWS cloud environment.
 Experience in creating and managing IAM users, groups, and roles, and configuring fine-grained
permissions for AWS resources using Python scripts.
 Launching WebLogic servers for developing, deploying, and running enterprise applications,
such as Java, in the cloud
 Used WebLogic for troubleshooting, enhancing performance, and constantly improving the
server and applications.
 Setting up new Micro services in all environments from Dev to production. Monitored
Server’s health and micro services health using tools like Grafana, Prometheus.
 Worked with Development and Testing teams to design / develop robust solutions to meet the
client requirements for functionality and performance.

Project # 1 Poetic - Software License Management System


Duration July-2011 to Sept-2014
Client Moduslink, Dedham, USA
Role Software Engineer
Technology Java
Tools SVN, OTRS etc.
Description:
Poetic™ the flagship product offering from U & I informatics pvt ltd. It is a comprehensive, sophisticated,
and highly configurable web-based license management system that publishers, customers and channel
partners use to manage the entire lifecycle of their volume license agreements. It simplifies the
complexity, reduces administrative and call center costs, and improves the results gained from managing,
acquiring, and deploying license programs.

Environment: AWS (EC2, IAM, S3, VPC, Route53, EKS, Lambda, Cloud Watch, Cloud Front), Jenkins, GIT,
GitHub, SonarQube, Maven, Tomcat, Docker, Kubernetes, Terraform, Grafana, Prometheus

Responsibilities:

 Supporting users in solving their regular day to day problems by using the OTRS tool.
 Prioritizing and managing many open tickets at one time.
 Resolving the tickets raised by client on the production system.
 Taking ownership of technical issues, and working with our Development group to resolve more
advanced issues when necessary.
 Troubleshoot and reproduce problems and define workarounds.
 Work with the Global IT Operation team when Production system is down.
 Participation in providing training to customers as required
 Implementing DB related changes to improve the performance of the Product.

Naveen Srigakollu

You might also like