Deployed and managed ClickHouse clusters to support scalable, low-latency logs data analytics
Provisioned and maintained Apache Kafka clusters, ensuring high availability and fault tolerance for event streaming
Engineered end-to-end logging pipelines using Vector.dev (VRL) to ingest application logs into ClickHouse for real-time querying
Designed and implemented a logs archival system leveraging Kafka Connect with an S3 sink to offload historical data efficiently
Automated key DevOps workflows using Apache Airflow and n8n, reducing manual intervention and improving pipeline reliability
Installed, configured, and maintained HashiCorp Vault clusters; integrated Vault secrets management into CI/CD pipelines and guided application teams on Vault adoption
Set up and configured monitoring tools to track metrics and performance of high-throughput Kafka clusters
Developed an enhanced monitoring system using Prometheus and Grafana in a hybrid (on-premises and cloud) environment
Automated Terraform module deployments by templating configurations with Jinja and Python scripts to enforce consistency and accelerate infrastructure provisioning
DevOps Engineer
Karkinos Healthcare
12.2020 - 01.2024
Proficient in deploying and managing applications on cloud platforms such as Amazon Web Services (AWS) and Google Cloud Platform (GCP).
Implemented end-to-end CI/CD pipelines using Jenkins, ArgoCD, and Tekton, ensuring smooth and automated delivery of software applications from development to production environments.
Utilized containerization technologies such as Docker and container orchestration platforms like Kubernetes for building and deploying applications in the cloud environment.
Implemented and managed ArgoCD for GitOps-based continuous deployment, enabling automated and auditable deployments of containerized applications to Kubernetes clusters.
Implemented Change Data Capture (CDC) solutions using Debezium for both MongoDB and PostgreSQL databases, enabling real-time data streaming and synchronization with downstream systems.
Conducted regular vulnerability assessments and penetration testing in coordination with the security team, identifying and prioritizing critical vulnerabilities and developing effective remediation plans.
Leveraged GCP's commitment discounts, reserved instances, and sustained usage discounts to optimize pricing for long-term resource usage, resulting in substantial cost savings for the organization.
Successfully designed and implemented Kafka cluster architectures, including multi-broker setups, ensuring scalability, resilience, and optimal performance for data streaming workloads.
Implemented end-to-end monitoring solutions using Prometheus, Grafana, and GCP monitoring tools to ensure comprehensive visibility and proactive management of application and infrastructure health.
Designed and implemented a robust log management system using the EFK stack, enabling centralized log collection, storage, analysis, and visualization for efficient troubleshooting and monitoring.
Improved code deployment efficiency by automating processes with CI/CD pipelines.
Maintained version control systems like Git or SVN for seamless collaboration among developers and engineers during project lifecycles.
DevOps Engineer
SalesboxAI
10.2018 - 12.2020
Proficient in managing Docker orchestration and containerization using Kubernetes and Docker Swarm, enabling efficient deployment, scaling, and management of containerized applications.
Proficient in AWS Services like ECR, ECS, EC2, S3, Route53, ELB, VPC, RDS.
Successfully led the migration of production MySQL servers to Amazon RDS (Relational Database Service), ensuring a seamless transition and improved manageability of database infrastructure.
Successfully installed and set up Nagios Core, configuring the necessary dependencies and components for effective monitoring of network devices, servers, applications, and services.
Successfully designed, implemented, and maintained MySQL replication for high availability and data redundancy, ensuring continuous availability and disaster recovery capabilities.
Actively participated in architectural discussions and provided valuable insights on migrating a monolithic application to a microservices architecture, contributing to the overall strategy and decision-making process.
Installed, configured, and managed Jetty and Tomcat servers, including the setup of server instances, virtual hosts, connectors, and SSL/TLS configurations, ensuring proper server operation and secure communication.
Implemented SonarQube and image scanning as part of a comprehensive DevSecOps strategy, integrating security checks and code quality assessments seamlessly into the CI/CD pipeline workflow.
Big Data Engineer
Knowledge Lens: A Rockwell Automation Company
04.2018 - 08.2018
Optimized Docker images and container configurations for improved performance, security, and resource efficiency, leveraging techniques like multi-stage builds, caching, and minimal base images.
Installed and configured various Linux distributions, such as CentOS, Ubuntu, or Red Hat, ensuring proper system setup, partitioning, and filesystem configurations.
Proficient in maintaining and configuring web servers, such as Apache HTTP Server or Nginx, ensuring secure and reliable delivery of web applications and websites.
Created custom Bash scripts to automate system administration tasks, such as log rotation, backup and restore procedures, file management, or software installations, saving significant time and effort.
DevOps Engineer / Site Reliability Engineer (SRE) at Nicoll Curtin / Julius BaerDevOps Engineer / Site Reliability Engineer (SRE) at Nicoll Curtin / Julius Baer