This role involves developing and automating infrastructure features for large-scale, fault-tolerant data platforms using Python and distributed systems technologies like Kafka and Spark, with a focus on managing data operations rather than data processing.
Key Responsibilities
Automate data platform operations, including fault-tolerant replication, TLS, installation, and backups
Develop and maintain infrastructure features for data platforms using Python
Collaborate with a distributed team to design and implement automation solutions
Debug issues and interact with open-source communities to improve automation tools
Provide domain-specific expertise on data systems to other teams within Canonical
Requirements
Proven hands-on experience in software development using Python
Proven hands-on experience in distributed systems, such as Kafka and Spark
Have a Bachelor’s or equivalent degree in Computer Science, STEM, or a similar field
Willingness to travel up to 4 times a year for internal and external events
Benefits & Perks
Competitive base pay depending on location, experience, knowledge, and skills
Annual compensation review
Recognition rewards
Annual holiday leave
Parental Leave
Employee Assistance Programme
Fully remote working environment
Personal learning and development budget of 2,000 USD per annum
Opportunity to travel to new locations to meet colleagues twice a year
Travel upgrades for long haul company events
Ready to Apply?
Join Canonical and make an impact in renewable energy