Cloudflare Logo

Cloudflare

Distributed Systems Engineer - Data Platform - Logs and Audit Logs

Posted 2 Days Ago
Hybrid
6 Locations
Mid level
Hybrid
6 Locations
Mid level
As a Distributed Systems Engineer, you will design and build logging platforms, develop data connectors, ensure compliance of audit logs, optimize data delivery, and maintain operational health through monitoring and collaboration with other teams.
The summary above was generated by AI
Locations Available: London (UK), Lisbon (Portugal), Austin (US), Denver (US), Atlanta (US)
About Role
We are actively seeking experienced and highly motivated Distributed Systems Engineers to join Cloudflare's dynamic DATA Organisation. This is a pivotal opportunity to contribute to the future of data at Cloudflare, working on systems that are fundamental to our global operations and customer insights. Our organisation is responsible for the entire data lifecycle, encompassing everything from initial ingestion and sophisticated processing to robust storage and efficient retrieval. These systems are the backbone that power critical logs and analytics, providing our customers with real-time, actionable visibility into the health, performance, and security of their online properties.
Our overarching mission is to empower customers to leverage their data effectively, enabling them to drive superior outcomes for their businesses. To achieve this, we design, build, and maintain a suite of high-performance, massively scalable distributed systems that are engineered to handle an unprecedented scale - processing well over a billion events per second. As an engineer within our organisation, you will be presented with unique challenges across various critical parts of our intricate data stack. This role offers the chance to work on cutting-edge technologies and contribute to solutions that operate at the very edge of internet infrastructure.
Our Data Organisation is strategically composed of several key teams, each focusing on a distinct aspect of our comprehensive data platform:
  • Data Delivery / Data Pipeline: This team is responsible for the design, development, and operation of our distributed data delivery pipeline. This system is a high-throughput, low-latency powerhouse, primarily written in Go, and is tasked with ingesting, processing, and intelligently routing massive volumes of data originating from across Cloudflare's vast global network to multiple core destinations. This involves handling diverse data types and ensuring reliable, timely delivery to various downstream systems.

  • Analytical Database Platform: Engineers on this team contribute to and evolve our core analytical platform, which is powered by ClickHouse. This team is dedicated to building and maintaining a high-performance, scalable database platform meticulously optimised for the immense analytical workloads generated by all of Cloudflare's products and services. This includes ensuring data integrity, query optimisation, and continuous platform scalability to meet ever-growing demands.

  • Data Retrieval (Customer-Facing Products): This department is focused on building and continuously improving our customer-facing products, making data not only accessible but also genuinely actionable for our users. This department comprises two main groups:
    • Analytics and Alerts: Members of this group are at the forefront of developing our public APIs such as the GraphQL Analytics API, providing customers and internal Cloudflare teams with flexible access to their data. They will also work on our alerting platform, empowering users to configure and receive near real-time alerts based on the critical logs and metrics observed by our robust data platform. This includes designing intuitive alerting mechanisms and ensuring the reliability of notification systems.
    • Logs and Audit Logs: This specialised team is dedicated to building a robust and easy-to-use logging platform that powers reliable data delivery and seamless integrations with customer destinations. The team's mission is to make it simple for customers to access, manage, and use their log data - ensuring that critical datasets, including comprehensive audit logs, are delivered securely and efficiently to their preferred storage and analysis platforms. The work spans developing intuitive connectors, ensuring data integrity, optimising delivery pipelines, and upholding strict standards for compliance, performance, and usability.

Responsibilities
This particular role is focused on the Logs and Audit Logs group. As a Software Engineer you will focus on the following areas:
  • Design, build, and operate a robust logging platform, ensuring reliable logging, and secure data transfer to a wide array of customer destinations and third-party integrations.
  • Develop and maintain high-performance data connectors and integrations for our log-shipping products, focusing on usability, scalability and data integrity.
  • Create and manage systems for handling comprehensive audit logs, ensuring they are delivered securely and adhere to strict compliance and performance standards.
  • Scale and optimise the data delivery pipeline to handle massive data volumes with low latency, identifying and removing bottlenecks in data processing and routing.
  • Work closely with Product and other engineering teams to define requirements for a new logging platform and integrations.
  • Maintain the operational health of our log delivery platform through comprehensive monitoring and participation in an on-call rotation (with flexibility for out-of-hours technical issue resolution).
  • Collaborate on the architectural evolution of our data egress platform, researching and implementing new technologies to improve efficiency and reliability.
Key Qualifications
  • 3+ years of experience working in software development covering distributed systems and data pipelines.
  • Strong programming skills (Go is preferable), with a deep understanding of software development best practices for building resilient, high-throughput systems.
  • Hands-on experience with modern observability stacks, including Prometheus, Grafana, and a strong understanding of handling high-cardinality metrics at scale.
  • Strong knowledge of SQL, including experience with query optimisation.
  • A solid foundation in computer science, including algorithms, data structures, distributed systems, and concurrency.
  • Strong analytical and problem-solving skills, with a willingness to debug, troubleshoot, and learn about complex problems at high scale.
  • Ability to work collaboratively in a team environment and communicate effectively with other teams across Cloudflare.
  • Experience with data streaming technologies (e.g., Kafka, Flink) is a strong plus.
  • Experience with various logging platforms or SIEMs (e.g., Splunk, Datadog, Sumo Logic) and storage destinations (e.g., S3, R2, GCS) is a plus.
  • Experience with Infrastructure as Code tools like SALT or Terraform is a plus.
  • Experience with Linux container technologies, such as Docker and Kubernetes, is a plus.

If you're passionate about building scalable and performant data platforms using cutting-edge technologies and want to work with a world-class team of engineers, then we want to hear from you! Join us in our mission to help build a better internet for everyone!

Top Skills

Datadog
Docker
Flink
Gcs
Go
Grafana
Kafka
Kubernetes
Prometheus
R2
S3
Salt
Splunk
SQL
Sumo Logic
Terraform

Similar Jobs at Cloudflare

16 Hours Ago
Hybrid
Austin, TX, USA
Mid level
Mid level
Cloud • Information Technology • Security • Software • Cybersecurity
The People Operations Specialist will manage employee lifecycle processes, payroll activities, and ensure compliance with local regulations while supporting HR operations across the Americas.
Top Skills: ExcelWorkdayZendesk
16 Hours Ago
Hybrid
Austin, TX, USA
Senior level
Senior level
Cloud • Information Technology • Security • Software • Cybersecurity
The Legal Operations Analyst will optimize contract lifecycle management, implement technology solutions, and enhance operational efficiencies within the Legal, Policy, and Trust & Safety department.
Top Skills: AIAutomation ToolsEnterprise PlatformsIronclad ClmLegal Systems
Yesterday
Hybrid
Austin, TX, USA
Senior level
Senior level
Cloud • Information Technology • Security • Software • Cybersecurity
Join the Enterprise Integrations Team to build and maintain integration workflows between Cloudflare and various SaaS applications, managing projects from conception to deployment.
Top Skills: Cloudflare Developer'S PlatformGitlabGoGrafanaKibanaKubernetesPostgresPrometheusRestful ApisSentry

What you need to know about the Sydney Tech Scene

From opera to comedy shows, the Sydney Opera House hosts more than 1,600 performances a year, yet its entertainment sector isn't the only one taking center stage. The city's tech sector has earned a reputation as one of the fastest-growing in the region. More specifically, its IT sector stands out as the country's third-largest, growing at twice the rate of overall employment in the past decade as businesses continue to digitize their operations to stay competitive.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account