4894 Senior Devops Engineers jobs in Bengaluru
Principal Devops Engineer
Posted today
Job Viewed
Job Description
Job Ref:
VL BanPDE
Location:
Bangalore, Karnataka, India
Postcode:
Type of contract:
Permanent
Salary:
Posted Date:
Friday, September 19, 2025
Closing Date:
Friday, October 17, 2025
Job Description – Principal Devops Engineer
Experience Level: 7 to 10+ years
Location: (Bangalore)
Key Responsibilities
- Design and manage scalable, secure, and cost-optimized cloud infrastructure (AWS/GCP/Azure).
- Lead Kubernetes (multi-cluster) deployments and implement Infrastructure as Code (Terraform/Helm).
- Build and optimize CI/CD pipelines (Jenkins, GitLab, GitHub Actions) with GitOps practices.
- Ensure observability with Prometheus, Grafana, ELK, and lead incident management.
- Drive DevSecOps—security automation, IAM, and vulnerability management.
- Mentor DevOps/SRE team, foster automation culture, and align with AI product goals.
- Enable AI/ML workflows (MLOps) for model deployment, monitoring, and lifecycle management.
Key Qualifications
- 7–12 years in DevOps/SRE/Platform Engineering, with startup or AI/ML experience preferred.
- Strong expertise in Kubernetes, IaC (Terraform, Helm), and cloud platforms (AWS/GCP/Azure).
- Proficiency in automation/scripting (Python, Go, Bash) and modern CI/CD practices.
- Experience with observability tools (Prometheus, Grafana, ELK, NewRelic) and reliability engineering.
- Knowledge of DevSecOps principles, IAM, RBAC, and cloud security.
- Proven leadership in mentoring teams, cross-functional collaboration, and scaling systems.
- Exposure to MLOps workflows (model CI/CD, monitoring, reproducibility) and GPU workloads.
Principal DevOps Engineer
Posted 13 days ago
Job Viewed
Job Description
Join our dynamic team at the forefront of cutting-edge technology as we seek a seasoned Principal DevOps Engineer (Remote, Bangalore, Coimbatore). Embark on a journey where your deep-rooted expertise in computer science fundamentals, alongside an intricate understanding of data structures, algorithms, and system design, becomes the cornerstone of innovative solutions. This pivotal role not only demands your proficiency in developing and elevating compute and I/O-intensive applications but also ensures their peak performance and unwavering reliability.
Responsibilities:
Own design and implementation of infrastructure services including GCP cloud architecture.
Build and maintain highly automated infrastructure using tools like Terraform, Ansible, and scripting (Python/Go).
Drive improvements in system reliability, performance monitoring, logging, and alerting systems.
Ensure systems are secure by design—manage access controls, secrets, and network segmentation with a strong security-first mindset.
Build internal tooling and CI/CD pipelines to accelerate developer productivity.
Collaborate with engineering leaders and mentor a growing team of infrastructure engineers.
Work closely with application teams, product managers, and leadership to align infrastructure efforts with Arcana’s technical roadmap.
Requirements:
10+ years of hands-on experience in DevOps/SRE/Infrastructure Engineering roles, preferably at high-growth tech startups or product companies.
Deep expertise in Linux systems, distributed systems, containers, Kubernetes, and cloud (preferably GCP/AWS).
Strong programming skills in Python, Go, or similar languages; scripting expertise is a must.
Proven experience with infrastructure-as-code, automation, CI/CD (e.g., GitHub Actions, ArgoCD, Jenkins), and monitoring tools like Prometheus, ELK, or Grafana.
Strong focus on security, reliability, and performance in high-availability environments.
Leadership experience or strong mentorship history, comfortable driving architectural decisions and influencing engineering culture.
Excellent communication skills; ability to work cross-functionally with both technical and non-technical teams.
Experience scaling infrastructure for fast-growing startups
Exposure to cloud-native security practices and zero-trust architectures
Background in developer platforms or internal tooling
Familiarity with data-heavy or ML-based system
Passion for elegant tooling and developer enablement
Principal DevOps Engineer
Posted today
Job Viewed
Job Description
About the job
Join our dynamic team at the forefront of cutting-edge technology as we seek a seasoned Principal DevOps Engineer (Remote, Bangalore, Coimbatore). Embark on a journey where your deep-rooted expertise in computer science fundamentals, alongside an intricate understanding of data structures, algorithms, and system design, becomes the cornerstone of innovative solutions. This pivotal role not only demands your proficiency in developing and elevating compute and I/O-intensive applications but also ensures their peak performance and unwavering reliability.
Responsibilities:
- Own design and implementation of infrastructure services including GCP cloud architecture.
- Build and maintain highly automated infrastructure using tools like Terraform, Ansible, and scripting (Python/Go).
- Drive improvements in system reliability, performance monitoring, logging, and alerting systems.
- Ensure systems are secure by design—manage access controls, secrets, and network segmentation with a strong security-first mindset.
- Build internal tooling and CI/CD pipelines to accelerate developer productivity.
- Collaborate with engineering leaders and mentor a growing team of infrastructure engineers.
- Work closely with application teams, product managers, and leadership to align infrastructure efforts with Arcana’s technical roadmap.
Requirements:
- 10+ years of hands-on experience in DevOps/SRE/Infrastructure Engineering roles, preferably at high-growth tech startups or product companies.
- Deep expertise in Linux systems, distributed systems, containers, Kubernetes, and cloud (preferably GCP/AWS).
- Strong programming skills in Python, Go, or similar languages; scripting expertise is a must.
- Proven experience with infrastructure-as-code, automation, CI/CD (e.g., GitHub Actions, ArgoCD, Jenkins), and monitoring tools like Prometheus, ELK, or Grafana.
- Strong focus on security, reliability, and performance in high-availability environments.
- Leadership experience or strong mentorship history, comfortable driving architectural decisions and influencing engineering culture.
- Excellent communication skills; ability to work cross-functionally with both technical and non-technical teams.
- Experience scaling infrastructure for fast-growing startups
- Exposure to cloud-native security practices and zero-trust architectures
- Background in developer platforms or internal tooling
- Familiarity with data-heavy or ML-based system
- Passion for elegant tooling and developer enablement
Principal DevOps Engineer
Posted 2 days ago
Job Viewed
Job Description
About the job
Join our dynamic team at the forefront of cutting-edge technology as we seek a seasoned Principal DevOps Engineer (Remote, Bangalore, Coimbatore). Embark on a journey where your deep-rooted expertise in computer science fundamentals, alongside an intricate understanding of data structures, algorithms, and system design, becomes the cornerstone of innovative solutions. This pivotal role not only demands your proficiency in developing and elevating compute and I/O-intensive applications but also ensures their peak performance and unwavering reliability.
Responsibilities:
- Own design and implementation of infrastructure services including GCP cloud architecture.
- Build and maintain highly automated infrastructure using tools like Terraform, Ansible, and scripting (Python/Go).
- Drive improvements in system reliability, performance monitoring, logging, and alerting systems.
- Ensure systems are secure by design—manage access controls, secrets, and network segmentation with a strong security-first mindset.
- Build internal tooling and CI/CD pipelines to accelerate developer productivity.
- Collaborate with engineering leaders and mentor a growing team of infrastructure engineers.
- Work closely with application teams, product managers, and leadership to align infrastructure efforts with Arcana’s technical roadmap.
Requirements:
- 10+ years of hands-on experience in DevOps/SRE/Infrastructure Engineering roles, preferably at high-growth tech startups or product companies.
- Deep expertise in Linux systems, distributed systems, containers, Kubernetes, and cloud (preferably GCP/AWS).
- Strong programming skills in Python, Go, or similar languages; scripting expertise is a must.
- Proven experience with infrastructure-as-code, automation, CI/CD (e.g., GitHub Actions, ArgoCD, Jenkins), and monitoring tools like Prometheus, ELK, or Grafana.
- Strong focus on security, reliability, and performance in high-availability environments.
- Leadership experience or strong mentorship history, comfortable driving architectural decisions and influencing engineering culture.
- Excellent communication skills; ability to work cross-functionally with both technical and non-technical teams.
- Experience scaling infrastructure for fast-growing startups
- Exposure to cloud-native security practices and zero-trust architectures
- Background in developer platforms or internal tooling
- Familiarity with data-heavy or ML-based system
- Passion for elegant tooling and developer enablement
Principal DevOps Engineer
Posted today
Job Viewed
Job Description
About the job
Join our dynamic team at the forefront of cutting-edge technology as we seek a seasoned Principal DevOps Engineer (Remote, Bangalore, Coimbatore). Embark on a journey where your deep-rooted expertise in computer science fundamentals, alongside an intricate understanding of data structures, algorithms, and system design, becomes the cornerstone of innovative solutions. This pivotal role not only demands your proficiency in developing and elevating compute and I/O-intensive applications but also ensures their peak performance and unwavering reliability.
Responsibilities:
- Own design and implementation of infrastructure services including GCP cloud architecture.
- Build and maintain highly automated infrastructure using tools like Terraform, Ansible, and scripting (Python/Go).
- Drive improvements in system reliability, performance monitoring, logging, and alerting systems.
- Ensure systems are secure by design—manage access controls, secrets, and network segmentation with a strong security-first mindset.
- Build internal tooling and CI/CD pipelines to accelerate developer productivity.
- Collaborate with engineering leaders and mentor a growing team of infrastructure engineers.
- Work closely with application teams, product managers, and leadership to align infrastructure efforts with Arcana’s technical roadmap.
Requirements:
- 10+ years of hands-on experience in DevOps/SRE/Infrastructure Engineering roles, preferably at high-growth tech startups or product companies.
- Deep expertise in Linux systems, distributed systems, containers, Kubernetes, and cloud (preferably GCP/AWS).
- Strong programming skills in Python, Go, or similar languages; scripting expertise is a must.
- Proven experience with infrastructure-as-code, automation, CI/CD (e.g., GitHub Actions, ArgoCD, Jenkins), and monitoring tools like Prometheus, ELK, or Grafana.
- Strong focus on security, reliability, and performance in high-availability environments.
- Leadership experience or strong mentorship history, comfortable driving architectural decisions and influencing engineering culture.
- Excellent communication skills; ability to work cross-functionally with both technical and non-technical teams.
- Experience scaling infrastructure for fast-growing startups
- Exposure to cloud-native security practices and zero-trust architectures
- Background in developer platforms or internal tooling
- Familiarity with data-heavy or ML-based system
- Passion for elegant tooling and developer enablement
Principal DevOps Engineer
Posted today
Job Viewed
Job Description
Job Title: Senior engineer (SDE-2)– Platform Engineering
Location: Bengaluru
Employment Type: Full-time
Team: Platform Engineering
About the Role:
We are looking for a passionate and hands-on DevOps Engineer to join our Platform
Engineering team and accelerate our platform modernization journey. This role is ideal for
engineers who thrive in automation-heavy environments and want to contribute to the evolution
of scalable, secure and intelligent infrastructure platforms. You will work on next-gen
cloud-native platforms, implement AI-powered automation and help build self-healing, resilient
systems that support our applications at scale.
Key Responsibilities
- Manage platform modernization initiatives including containerization, service mesh
adoption and migration to microservices and serverless infrastructure.
- Design and implement robust CI/CD pipelines and self-service DevOps platforms to
streamline software delivery across environments.
- Develop and manage Infrastructure as a Code (IaaC) using tools like Terraform or
CloudFormation for scalable and repeatable deployments.
- Automate infrastructure provisioning, configuration management and operations using
tools like Ansible, Chef or Puppet.
- Leverage AI/ML-driven automation for predictive alerting, anomaly detection,
auto-scaling and intelligent incident response.
Build intelligent, self-healing systems focused on resilience, cost optimization and
performance forecasting in public cloud environments.
- Drive observability by establishing standards for monitoring, logging and alerting using
tools such as Prometheus, Grafana, OpenTelemetry and AIOps platforms.
- Embed security and compliance into DevOps workflows by adopting DevSecOps
practices throughout the software development lifecycle.
- Evaluate emerging technologies and methodologies to improve system reliability,
developer experience and platform scalability.
- Participate in SRE on-call rotations, production support and post-incident reviews to
continuously improve system resilience.
- Build internal tools and automation solutions to enhance platform observability and
operational efficiency.
- Identify and resolve performance bottlenecks and lead root cause analysis efforts for
critical incidents.
- Collaborate cross-functionally with engineering, architecture and security teams to drive
best practices and architectural alignment.
- Support disaster recovery planning, backup strategy implementation and compliance
initiatives (e.G., SOC2, ISO).
- Mentor junior engineers, promote knowledge sharing and foster a culture of engineering
excellence.
Qualifications
- 4–6 years of experience in DevOps, SRE or platform engineering roles with a software
engineering mindset.
- Hands-on expertise in Kubernetes, Docker and service mesh architectures (Istio,
Linkerd).
- Expertise in CI/CD tools such as Jenkins, ArgoCD, Spinnaker or similar tools for
automating and managing deployment workflows.
- Experience with observability stacks (Prometheus, Grafana, ELK, Loki or Datadog) for
monitoring, logging and alerting.
- Good understanding of AIOps and ML-driven automation, including anomaly detection,
intelligent alerting and predictive incident response
- Strong problem-solving and debugging skills, particularly in complex, production-grade
distributed systems
- Expertise in Infrastructure as a Code (IaaC) using tools like Terraform or Pulumi and
proficiency in configuration management with Ansible, Puppet or similar tools.
- Familiarity with event-driven architectures using tools like Kafka or cloud-native pub/sub
messaging systems.
- Good understanding of cloud cost optimization and efficiency practices through
automation and resource management.
- Experience integrating security scanning and compliance checks into CI/CD pipelines
using tools like Trivy, Snyk or arnica.
Why Join Us?
- Work on impactful infrastructure and DevOps challenges at scale.
- Build infrastructure that enables fast, reliable and responsible deployment of AI
solutions.
- Be part of a culture that champions engineering excellence, ownership and continuous
learning.
- Help shape the future of DevOps and AI integration in a fast-moving, innovation-focused
environment.
- Collaborate with architects and DevOps leaders on strategic initiatives.
- Be part of a team building intelligent, resilient platforms using cutting-edge DevOps and
AI technologies.
Principal DevOps Engineer (Prisma AIRS) - NetSec

Posted 4 days ago
Job Viewed
Job Description
At Palo Alto Networks® everything starts and ends with our mission:
Being the cybersecurity partner of choice, protecting our digital way of life.
Our vision is a world where each day is safer and more secure than the one before. We are a company built on the foundation of challenging and disrupting the way things are done, and we're looking for innovators who are as committed to shaping the future of cybersecurity as we are.
**Who We Are**
We take our mission of protecting the digital way of life seriously. We are relentless in protecting our customers and we believe that the unique ideas of every member of our team contributes to our collective success. Our values were crowdsourced by employees and are brought to life through each of us everyday - from disruptive innovation and collaboration, to execution. From showing up for each other with integrity to creating an environment where we all feel included.
As a member of our team, you will be shaping the future of cybersecurity. We work fast, value ongoing learning, and we respect each employee as a unique individual. Knowing we all have different needs, our development and personal wellbeing programs are designed to give you choice in how you are supported. This includes our FLEXBenefits wellbeing spending account with over 1,000 eligible items selected by employees, our mental and financial health resources, and our personalized learning opportunities - just to name a few!
At Palo Alto Networks, we believe in the power of collaboration and value in-person interactions. This is why our employees generally work full time from our office with flexibility offered where needed. This setup fosters casual conversations, problem-solving, and trusted relationships. Our goal is to create an environment where we all win with precision.
**Your Career**
As a Senior Staff DevOps Engineer on the Prisma AIRS Red Teaming team, you'll be crucial in designing, developing and maintaining the advanced infrastructure for our product. This role demands strong DevOps knowledge, focusing on building scalable, secure systems and bridging development and operations through CI/CD and automation.
Key responsibilities include leading infrastructure development, optimizing CI/CD pipelines, implementing Infrastructure as Code, setting up monitoring, collaborating with teams, ensuring security, and optimizing performance and scalability.
This is a unique opportunity to significantly impact cybersecurity within a leading-edge product and a skilled, collaborative team.
**Your Impact**
+ Passionate in building highly robus and scalable systems
+ Work on complex problems that require in-depth evaluation of the requirements and apply engineering techniques to obtain quality results
+ Identify problems and opportunities, propose and develop simple and attainable solutions to improve the team's development process and product quality.
+ Be the role model in establishing and implementing engineering best practices.
+ Drive collaborative discussion and create synergy among product feature stakeholders to create an amazing product
**Your Experience**
+ 9+ Years of total experience. 3+ years of experience in a role focused on CI/CD, automation or DevOps engineering with a strong drive for problem-solving and ownership
+ Test automation infrastructure development and maintenance.
+ GitLab based CI
+ Terraform based cloud infrastructure management in GCP
+ Proven expertise in designing, developing, and managing CI/CD pipelines for complex software products within continuous delivery and agile environments
+ Solid understanding of cloud infrastructure concepts, particularly with GCP and/or AWS, and practical experience with Infrastructure as Code using Terraform
+ Familiarity with containerization technologies like Docker and orchestration with Kubernetes
**The Team**
The Prisma AIRS Red Teaming team is dedicated to building an industry-defining product in AI security. This product helps organizations understand and mitigate risks within their AI systems. By providing insights and actionable solutions, the team empowers customers to secure their AI deployments effectively.
**Our Commitment**
We're problem solvers that take risks and challenge cybersecurity's status quo. It's simple: we can't accomplish our mission without diverse teams innovating, together.
We are committed to providing reasonable accommodations for all qualified individuals with a disability. If you require assistance or accommodation due to a disability or special need, please contact us at .
Palo Alto Networks is an equal opportunity employer. We celebrate diversity in our workplace, and all qualified applicants will receive consideration for employment without regard to age, ancestry, color, family or medical care leave, gender identity or expression, genetic information, marital status, medical condition, national origin, physical or mental disability, political affiliation, protected veteran status, race, religion, sex (including pregnancy), sexual orientation, or other legally protected characteristics.
All your information will be kept confidential according to EEO guidelines.
Be The First To Know
About the latest Senior devops engineers Jobs in Bengaluru !
Principal Associate, DevOps Engineer
Posted today
Job Viewed
Job Description
Voyager , India, Bangalore, Karnataka
Principal Associate, DevOps Engineer
Job Description
Do you love building and pioneering in the technology space? Do you enjoy solving complex business problems in a fast-paced, collaborative, inclusive, and iterative delivery environment? At Capital One, you'll be part of a big group of makers, breakers, doers and disruptors, who love to solve real problems and meet real customer needs. We are seeking
DevOps Engineers
who are passionate about marrying data with emerging technologies to join our team. As a DevOps Engineer, you'll have the opportunity to be on the forefront of driving a major transformation within Capital One.
Team Info:
The EMP (Enterprise Model Platform) SRE team is responsible for building, maintaining, and governing the model development Kubernetes fleet. This includes, but is not limited to, the Kubernetes control plane, workers, storage, addons, custom operators, and applications within each cluster as our team is the very foundation of the ML platform at Capital One. Additionally, this team owns microservices that enable self-service onboarding and interactions with the ML fleet of clusters. We work closely with various open source projects and internal teams to ensure a seamless experience for our internal customers building machine learning models.
What You'll Do:
- Collaborate with and across Agile teams to design, develop, test, implement, and support technical solutions in full-stack development tools and technologies
- Share your passion for staying on top of tech trends, experimenting with and learning new technologies, participating in internal & external technology communities, and mentoring other members of the engineering community
- Collaborate with digital product managers, and deliver robust cloud-based solutions that drive powerful experiences to help millions of Americans achieve financial empowerment
- Utilize programming languages like Java, Python, SQL, Ruby and Go, Container Orchestration services including Docker and Kubernetes, CM tools including Ansible and Terraform, and a variety of AWS tools and services
Basic Qualifications:
- Bachelor's degree
- At least 4 years of experience in DevOps Engineering (Internship experience does not apply)
- At least 2 years of experience with Cloud Native technologies (Amazon Web Services, Microsoft Azure, Google Cloud Platform)
- At least 2 years of Unix or Linux system administration experience
Preferred Qualifications:
- 2+ years of experience with Terraform or Ansible
- 2+ years of experience with container orchestration services including Docker or Kubernetes
- 2+ years of experience with coding and scripting (Python, SQL, Java, JavaScript, Golang, Bash, Perl or Ruby)
- 2+ years of experience working with Agile Development Practices
At this time, Capital One will not sponsor a new applicant for employment authorization for this position.
No agencies please. Capital One is an equal opportunity employer (EOE, including disability/vet) committed to non-discrimination in compliance with applicable federal, state, and local laws. Capital One promotes a drug-free workplace. Capital One will consider for employment qualified applicants with a criminal history in a manner consistent with the requirements of applicable laws regarding criminal background inquiries, including, to the extent applicable, Article 23-A of the New York Correction Law; San Francisco, California Police Code Article 49, Sections ; New York City's Fair Chance Act; Philadelphia's Fair Criminal Records Screening Act; and other applicable federal, state, and local laws and regulations regarding criminal background inquiries.
If you have visited our website in search of information on employment opportunities or to apply for a position, and you require an accommodation, please contact Capital One Recruiting at or via email at All information you provide will be kept confidential and will be used only to the extent required to provide needed reasonable accommodations.
For technical support or questions about Capital One's recruiting process, please send an email to
Capital One does not provide, endorse nor guarantee and is not liable for third-party products, services, educational tools or other information available through this site.
Capital One Financial is made up of several different entities. Please note that any position posted in Canada is for Capital One Canada, any position posted in the United Kingdom is for Capital One Europe and any position posted in the Philippines is for Capital One Philippines Service Corp. (COPSSC).
DevOps Engineer/Lead DevOps Engineer
Posted today
Job Viewed
Job Description
Designation : DevOps Engineer/Lead DevOps Engineer
Experience : 8-15 years
Location: Remote
Summary
Work with talented DevOps and Cloud operations engineers and architects to deliver Sycamore SaaS product offerings to our Bio-Pharma customers using exciting, cutting-edge technologies. Develop, execute, maintain, and improve procedures, automation scripts, and infrastructure implementations to support Sycamore SaaS Operations.
Roles and Responsibility
Specific roles and responsibilities include:
- Provide technical expertise and leadership when needed to SaaS Operations Production Operations teams.
- Help Implement the Cloud Operations team's goals and deliverables as determined by Sycamore Leadership
- Ensure smooth operations of Sycamore SaaS products
- Take Complete ownership of Customer Implementations, including SLA and SLO.
- Automate, enhance and maintain critical processes in Cloud Operations, such as Change Control, Monitoring & Alerting
- Drive critical processes in SaaS Operations such as Change Control, Problem & Incident Management, and Reporting, as well as key tools for Monitoring & Alerting
- Drive Disaster Recovery and failover procedures, training, testing, and team readiness
- Coordinate focus groups across all teams on process improvements and technical improvements that lead to better stability and reliability
- Contribute to process improvements and technical improvements that lead to increased stability and reliability
- Support continuous improvements in SaaS Operations by
- developing platform services and tooling for modern cloud operations, including metrics monitoring, CI/CD pipelines, etc.
- improving automation of provisioning, deployment, monitoring, alerting, and escalation
- Support Secure operations by
- implementing best-in-class recommendations for secure operations
- Carry out ongoing Production Ops activities with precision and quality
- Define, build, and deliver a high-quality SaaS Platform for Work with third-party vendors and partners to help develop a complete solution set on the SaaS platform
- Representing Cloud Operations in InfoSec meetings and developing and driving secure procedures
- Help obtain and maintain various certifications
- Being a good team player & a leader when needed for a high-performance Cloud/SaaS delivery team by
- Reviewing personal/team performance, quality reviews,
- Manage operations and operational issues.
- Establish a culture of high performance, ownership, delivery focus, and continuous improvement.
Excellence in Operations
- Implement and carry out procedures and policies to ensure high-quality SaaS operations with appropriate levels of management controls.
- Act as an internal contact for platform services issues for a customer
- Work with cross-functional departments: Sales, Professional Service, Customer Support, Engineering, and QA
Desired Experience
- Has experience in implementing, managing, maintaining, and decommissioning complex cloud-based Information system components in a secure and controlled manner.
- Must be experienced in coordinating cross-functional teams such as support, escalation, and engineering software teams to address product issues successfully.
- Strong understanding of how to build, scale, and manage complex multi-product/service environments
- Record of building lean, automated, scalable support structures versus labor-intensive environments.
- Strong innovation mindset, analytical skills, excellent oral and written communication skills, and experience effectively communicating project/program mission and objectives.
- Must exhibit a practical customer service attitude and lead a team in resolving difficult customer situations.
Skills Required
- Very Strong Linux Knowledge & Troubleshooting Skills
- Scripting using – Bash, Python, PowerShell, etc
- Kubernetes, helm Charts
- Terraform, Ansible
- Windows Terminal Services, AD, LDAP
- Hands-on experience in cloud technology – AWS, Azure – AWS preferred
- Change, Problem & Incident Management
- Implementation awareness of Vulnerability/Penetration Testing, Security
- Strong Networking Skills
- Tools and frameworks used for monitoring, performance management, logging
- CI/CD pipeline
- SRE – Including Datadog.
- Datadog
Certification
- RHEL
- AWS
- Kubernetes
DevOps Engineer/Lead DevOps Engineer
Posted today
Job Viewed
Job Description
Designation : DevOps Engineer/Lead DevOps Engineer
Experience : 8-15 years
Location: Remote
Summary
Work with talented DevOps and Cloud operations engineers and architects to deliver Sycamore SaaS product offerings to our Bio-Pharma customers using exciting, cutting-edge technologies. Develop, execute, maintain, and improve procedures, automation scripts, and infrastructure implementations to support Sycamore SaaS Operations.
Roles and Responsibility
Specific roles and responsibilities include:
- Provide technical expertise and leadership when needed to SaaS Operations Production Operations teams.
- Help Implement the Cloud Operations team's goals and deliverables as determined by Sycamore Leadership
- Ensure smooth operations of Sycamore SaaS products
- Take Complete ownership of Customer Implementations, including SLA and SLO.
- Automate, enhance and maintain critical processes in Cloud Operations, such as Change Control, Monitoring & Alerting
- Drive critical processes in SaaS Operations such as Change Control, Problem & Incident Management, and Reporting, as well as key tools for Monitoring & Alerting
- Drive Disaster Recovery and failover procedures, training, testing, and team readiness
- Coordinate focus groups across all teams on process improvements and technical improvements that lead to better stability and reliability
- Contribute to process improvements and technical improvements that lead to increased stability and reliability
- Support continuous improvements in SaaS Operations by
- developing platform services and tooling for modern cloud operations, including metrics monitoring, CI/CD pipelines, etc.
- improving automation of provisioning, deployment, monitoring, alerting, and escalation
- Support Secure operations by
- implementing best-in-class recommendations for secure operations
- Carry out ongoing Production Ops activities with precision and quality
- Define, build, and deliver a high-quality SaaS Platform for Work with third-party vendors and partners to help develop a complete solution set on the SaaS platform
- Representing Cloud Operations in InfoSec meetings and developing and driving secure procedures
- Help obtain and maintain various certifications
- Being a good team player & a leader when needed for a high-performance Cloud/SaaS delivery team by
- Reviewing personal/team performance, quality reviews,
- Manage operations and operational issues.
- Establish a culture of high performance, ownership, delivery focus, and continuous improvement.
Excellence in Operations
- Implement and carry out procedures and policies to ensure high-quality SaaS operations with appropriate levels of management controls.
- Act as an internal contact for platform services issues for a customer
- Work with cross-functional departments: Sales, Professional Service, Customer Support, Engineering, and QA
Desired Experience
- Has experience in implementing, managing, maintaining, and decommissioning complex cloud-based Information system components in a secure and controlled manner.
- Must be experienced in coordinating cross-functional teams such as support, escalation, and engineering software teams to address product issues successfully.
- Strong understanding of how to build, scale, and manage complex multi-product/service environments
- Record of building lean, automated, scalable support structures versus labor-intensive environments.
- Strong innovation mindset, analytical skills, excellent oral and written communication skills, and experience effectively communicating project/program mission and objectives.
- Must exhibit a practical customer service attitude and lead a team in resolving difficult customer situations.
Skills Required
- Very Strong Linux Knowledge & Troubleshooting Skills
- Scripting using – Bash, Python, PowerShell, etc
- Kubernetes, helm Charts
- Terraform, Ansible
- Windows Terminal Services, AD, LDAP
- Hands-on experience in cloud technology – AWS, Azure – AWS preferred
- Change, Problem & Incident Management
- Implementation awareness of Vulnerability/Penetration Testing, Security
- Strong Networking Skills
- Tools and frameworks used for monitoring, performance management, logging
- CI/CD pipeline
- SRE – Including Datadog.
- Datadog
Certification
- RHEL
- AWS
- Kubernetes