Senior DevOps Engineer
Location
Chennai/Pune
Work format
Full-Time
Experience level
7+ years
About Finzly:
Headquartered in Charlotte, NC, and founded in 2012 by visionary minds shaping the future of banking and payments in the US, Finzly makes bank transformation radically simple. Our core-independent platform is trusted by some of the market’s leading financial institutions to transform operations and launch new products and services at speed.
We’ve been recognized with 20+ industry awards in the last two years - including three years in a row as one of the Best and Brightest Places to Work as well as accolades for Best Parallel Core Technology, Best Corporate Payments Solution, Best Payments-as-a-Service Provider, and Best Trading System. Quite simply, Finzly is known for being the best in everything we do, giving you the perfect opportunity to grow your career with impact.
About the Role:
Finzly is looking for an experienced Senior DevOps Engineer who thrives on ownership, accountability, and delivering reliable infrastructure at scale.
You’ll take charge of building, automating, and optimizing our systems — ensuring stability, scalability, and security across all environments.
This role requires a hands-on technologist with deep DevOps expertise, strong problem-solving skills, and a delivery-focused mindset.
Key Responsibilities:
- Lead the design, automation, and management of large-scale AWS infrastructure across multi-account environments using services such as EKS, EC2, VPC, Route53, CloudFront, RDS, S3, CloudWatch, and others.
- Architect, implement, and manage CI/CD pipelines using Jenkins, including Groovy-based shared libraries and custom pipeline templates to support build, deploy, and security scanning stages.
- Enhance pipeline efficiency by integrating DevSecOps tools (SonarQube, Trivy, Veracode, or AWS Inspector) to automate code quality and security checks.
- Manage and maintain Kubernetes clusters — including Helm-based application deployments, version upgrades, patch management, autoscaling, and performance tuning.
- Develop and enforce Infrastructure-as-Code (IaC) standards using Terraform, ensuring consistency and reusability across teams.
- Implement and govern branching strategies (Git Flow, feature/hotfix branches) to enable parallel development and controlled releases.
- Apply strong Linux system administration fundamentals to troubleshoot, secure, and optimize workloads.
- Design and support event-driven architectures using the Serverless Framework, AWS Lambda, SNS, SQS, and API Gateway for scalable asynchronous workflows.
- Manage and optimize relational databases (MySQL, Oracle) for connectivity, tuning, and availability.
- Build and maintain observability stacks (Datadog, Prometheus, Grafana, CloudWatch) for proactive monitoring, logging, and alerting.
- Explore and integrate AI and search technologies in AWS, including Bedrock Knowledge bases and Agents, into existing DevOps automation and monitoring pipelines.
- Champion best practices around security, cost optimization, and operational reliability across environments.
Required Skills & Experience:
- 7+ years of hands-on DevOps or Site Reliability Engineering experience.
- 3+ years working with Kubernetes in production (upgrades, patches, scaling, monitoring).
- Strong hands-on experience with Jenkins and Groovy, creating and maintaining shared libraries and custom pipelines.
- Deep expertise with AWS core services (EKS, EC2, VPC, Lambda, API Gateway, RDS, S3, SNS, SQS, CloudWatch).
- Experience implementing and maintaining event-driven architectures and serverless systems.
- Solid understanding of branching strategies (Git Flow, trunk-based, hotfix flows).
- Strong hands-on experience with Helm for packaging, versioning, and managing Kubernetes deployments.
- Proficiency with Terraform or equivalent IaC tools.
- Strong scripting experience in Python and Bash.
- Experience with MySQL and Oracle databases for backup, performance tuning, and failover management.
- Excellent understanding of Linux administration, networking, and system troubleshooting.
- Experience building and maintaining robust monitoring and alerting stacks using Datadog, Prometheus, Grafana, or CloudWatch.
- Familiarity with AI and data processing tools in AWS (e.g., Bedrock, OpenSearch).
- Strong understanding of security, cost management, and scalability principles.
Soft Skills and Team Culture:
- A positive mindset and a strong can-do attitude.
- Proven ability to work collaboratively in Agile / Scrum teams.
- Demonstrated success working on large, complex distributed systems.
- A sense of extreme ownership — taking full responsibility for the quality, reliability, and delivery of your work.
- A genuine passion for meeting deadlines and delivering outcomes as promised.
- Strong alignment with client goals and dedication to delivering value and reliability in every sprint and release.
- Excellent communication and teamwork skills — both verbal and written.
Education Qualification:
- BE in Information Technology, Computer Science, related field or equivalent work experience
Apply:
Interested candidates can send their resume to Finzly at careers@finzly.com