Senior Data Engineer
XM Cyber
Description
XM Cyber is the leader in hybrid-cloud security posture management, using the attacker’s perspective to find and remediate critical attack paths across on-premises and multi-cloud networks. XM Cyber is looking for a talented Senior Data Engineer
Join a core team of experts responsible for developing innovative cyber-attack techniques for Cloud-based environments (AWS, Azure, GCP, Kubernetes) that integrate into XM Cyber’s fully automated attack simulation.
About the Role:We are seeking an experienced Senior Data Engineer to join our dynamic data team. The ideal candidate will be responsible for designing, building, and maintaining scalable data pipelines and infrastructure, ensuring the availability, reliability, and quality of our data. This role requires strong technical expertise, problem-solving skills, and the ability to collaborate across teams to deliver data-driven solutions.Key Responsibilities:
- Design, implement, and maintain robust, scalable, and high-performance data pipelines and ETL processes.
- Develop and optimize data models, schemas, and storage solutions to support analytics and machine learning initiatives.
- Collaborate with software engineers and product managers to understand data requirements and deliver high-quality solutions.
- Ensure data quality, integrity, and governance across multiple sources and systems.
- Monitor and troubleshoot data workflows, resolving performance and reliability issues.
- Evaluate and implement new data technologies and frameworks to improve the data platform.
- Document processes, best practices, and data architecture.
- Mentor junior data engineers and contribute to team knowledge sharing.
Requirements
Required Qualifications:
- Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field.
- 5+ years of experience in data engineering, ETL development, or a similar role.
- Strong proficiency in SQL and experience with relational and NoSQL databases.
- Experience with data pipeline frameworks and tools such as: Apache Spark, Airflow & Kafka. - MUST
- Familiarity with cloud platforms (AWS, GCP, or Azure) and their data services.
- Solid programming skills in Python, Java, or Scala.
- Strong problem-solving, analytical, and communication skills.
- Knowledge of data governance, security, and compliance standards.
- Experience with data warehousing, big data technologies, and data modeling best practices such as ClickHouse, SingleStore, StarRocks.
Preferred Qualifications (Advantage):
- Familiarity with machine learning workflows and MLOps practices.
- Work with Data Lakehouse architectures and technologies such as Apache Iceberg.
- Experience working with data ecosystems in Open Source/On-Premise environments.
Why Join Us:
- Work with cutting-edge technologies and large-scale data systems.
- Collaborate with a talented and innovative team.
- Opportunities for professional growth and skill development.
- Make a direct impact on data-driven decision-making across the organization.