Canonical Logo

Canonical

Software Engineer - Data Infrastructure - Kafka

Posted 14 Hours Ago
Remote
8 Locations
Entry level
Remote
8 Locations
Entry level
The Software Engineer in Data Infrastructure will work on automating data platform operations, ensuring fault-tolerance, and managing Big Data platforms at scale. Responsibilities include collaborating with a distributed team, writing Python code to create features, debugging issues, and interaction with upstream communities, along with domain expertise for various data systems.
The summary above was generated by AI

Canonical is building a comprehensive automation suite to provide multi-cloud and on-premise data solutions for the enterprise. The data platform team is a collaborative team that develops a managed solutions for a full range of data stores and data technologies, spanning from big data, through NoSQL,  cache-layer capabilities, and analytics; all the way to structured SQL engines (similar to Amazon RDS approach). 

We are facing the interesting problem of fault-tolerant mission-critical distributed systems and intend to deliver the world's best automation solution for delivering managed data platforms. 

We are looking for candidates from junior to senior level with interests, experience and willingness to learn around Big Data technologies, such as distributed event-stores (Kafka) and parallel computing frameworks (Spark). Engineers who thrive at Canonical are mindful of open-source community dynamics and equally aware of the needs of large, innovative organisations.

Location: This is a Globally remote role

What your day will look like

The data platform team is responsible for the automation of data platform operations, with the mission of managing and integrating Big Data platforms at scale. This includes ensuring fault-tolerant replication, TLS, installation, backups and much more; but also provides domain-specific expertise on the actual data system to other teams within Canonical. This role is focused on the creation and automation of infrastructure features of data platforms, not analysing and/or processing the data in them.

  • Collaborate proactively with a distributed team
  • Write high-quality, idiomatic Python code to create new features
  • Debug issues and interact with upstream communities publicly
  • Work with helpful and talented engineers including experts in many fields
  • Discuss ideas and collaborate on finding good solutions
  • Work from home with global travel for 2 to 4 weeks per year for internal and external events

What we are looking for in you

  • Proven hands-on experience in software development using Python
  • Proven hands-on experience in distributed systems, such as Kafka and Spark
  • Have a Bachelor’s or equivalent in Computer Science, STEM, or a similar degree
  • Willingness to travel up to 4 times a year for internal events

Additional skills that you might also bring

You might also bring a subset of experience from the followings that can help Data Platform to achieve its challenging goals and determine the level we will consider you for:

  • Experience operating and managing other data platform technologies, SQL (MySQL, PostgreSQL, Oracle, etc) and/or NoSQL (MongoDB, Redis, ElasticSearch, etc), similar to DBA level expertise
  • Experience with Linux systems administration, package management, and infrastructure operations
  • Experience with the public cloud or a private cloud solution like OpenStack
  • Experience with operating Kubernetes clusters and a belief that it can be used for serious persistent data services

What we offer you

Your base pay will depend on various factors including your geographical location, level of experience, knowledge and skills. In addition to the benefits above, certain roles are also eligible for additional benefits and rewards including annual bonuses and sales incentives based on revenue or utilisation. Our compensation philosophy is to ensure equity right across our global workforce.  

In addition to a competitive base pay, we provide all team members with additional benefits, which reflect our values and ideals. Please note that additional benefits may apply depending on the work location and, for more information on these, please ask your Talent Partner.

  • Fully remote working environment - we’ve been working remotely since 2004!
  • Personal learning and development budget of 2,000USD per annum
  • Annual compensation review
  • Recognition rewards
  • Annual holiday leave
  • Parental Leave
  • Employee Assistance Programme
  • Opportunity to travel to new locations to meet colleagues twice a year
  • Priority Pass for travel and travel upgrades for long haul company events

About Canonical

Canonical is a pioneering tech firm that is at the forefront of the global move to open source. As the company that publishes Ubuntu, one of the most important open source projects and the platform for AI, IoT and the cloud, we are changing the world on a daily basis. We recruit on a global basis and set a very high standard for people joining the company. We expect excellence - in order to succeed, we need to be the best at what we do.

Canonical has been a remote-first company since its inception in 2004.​ Work at Canonical is a step into the future, and will challenge you to think differently, work smarter, learn new skills, and raise your game. Canonical provides a unique window into the world of 21st-century digital business.

Canonical is an equal-opportunity employer

We are proud to foster a workplace free from discrimination. Diversity of experience, perspectives, and background create a better work environment and better products. Whatever your identity, we will give your application fair consideration.

#li-remote

Top Skills

Python

Similar Jobs

5 Hours Ago
Remote
Bengaluru, Karnataka, IND
Senior level
Senior level
Cloud • Information Technology • Productivity • Security • Software • App development • Automation
As a Principal Data Platform Engineer, you will build a world-class data platform solution, collaborating with analysts, engineers, and data scientists to identify gaps and implement innovative data tool capabilities while ensuring data quality and seamless experiences for users.
Top Skills: JavaKotlinPython
8 Hours Ago
Remote
India
Senior level
Senior level
Fintech • Information Technology • Insurance • Financial Services • Big Data Analytics
As a Unit Manager Java Developer, you will lead and develop cloud-based hybrid mobile applications and scalable microservices. You will oversee the configuration of CI/CD pipelines, collaborate with stakeholders, and write high-quality code adhering to best practices. The role demands strong leadership, problem-solving skills, and high levels of ownership and professionalism.
Top Skills: JavaJavaScript
8 Hours Ago
Remote
India
Junior
Junior
Fintech • Information Technology • Insurance • Financial Services • Big Data Analytics
As an Associate Big Data Engineer at MetLife, you will contribute to building and implementing data ingestion processes using various big data tools, support model operationalization, collaborate with teams on data enablement solutions, and optimize development efforts with reusable frameworks.
Top Skills: PythonScala

What you need to know about the Mumbai Tech Scene

From haggling for the best price at Chor Bazaar to the bustle of Crawford Market, the energy of Mumbai's traditional markets is a key part of the city's charm. And while these markets will always have their place, the city also boasts a thriving e-commerce scene, ranking among the largest in the region. Driven by online sales in everything from snacks to licensed sports merchandise to children's apparel, the local industry is worth billions, with companies actively recruiting to meet the demands of continued growth.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account