Proximity Works Logo

Proximity Works

Data Engineer

Posted 7 Days Ago
Be an Early Applicant
In-Office
Navi Mumbai, Thane, Maharashtra
Mid level
In-Office
Navi Mumbai, Thane, Maharashtra
Mid level
The Data Engineer role involves building scalable data pipelines, managing datasets, and collaborating with teams to enable data-driven insights and decision-making.
The summary above was generated by AI

We are looking for insight-driven Data Engineers, not pure ETL builders, to help build and scale the data pipelines and core analytical datasets that power analytics, AI model evaluation, safety systems, and business decision-making across Bharat AI’s agentic AI platform.

This role sits at the heart of how data flows through the organization — from raw events to insights, feedback loops, and decisions. You will work closely with Product, Data Science, Infrastructure, Marketing, Finance, and AI/Research teams to ensure data is reliable, interpretable, and outcome-oriented as the platform scales rapidly.

At Proximity, you won’t just move data — your work will directly influence how AI systems learn, adapt, and improve, and how leaders make product and business decisions.

Responsibilities
  • Design, build, and manage scalable data pipelines, ensuring high-fidelity user event and system data is reliably ingested into the data warehouse and streaming systems.
  • Develop and maintain canonical, analytics-ready datasets to track key product and business metrics such as:
    • user growth
    • engagement & retention
    • cohorts & behavioral patterns
    • revenue & conversion funnels
  • Think beyond pipelines — structure data to enable insights, experimentation, and feedback loops that improve both models and product decisions.
  • Collaborate with Infrastructure, Data Science, Product, Marketing, Finance, and Research teams to translate business questions into well-modeled, trustworthy datasets.
  • Implement robust, fault-tolerant systems for batch and streaming ingestion, transformation, and processing.
  • Participate actively in data architecture decisions, balancing scalability, cost, latency, and analytical flexibility.
  • Ensure data security, integrity, and compliance in line with company policies and industry standards.
  • Monitor pipeline and streaming job health, troubleshoot failures, and continuously improve reliability, performance, and data quality.
What Matters (Non-Negotiable)
  • Thinks in architecture, feedback loops, and outcomes, not pipelines alone.
  • Can structure data models that improve:
    • AI model evaluation
    • decision-making
    • experimentation and learning velocity

Communicates in terms of impact and insights, not only jobs and tables.


Requirements
  • 3–5 years of professional experience working as a Data Engineer or in a similar role.
  • Strong proficiency in Python for data processing and orchestration.
  • Solid hands-on experience with Apache Spark, including writing, debugging, and optimizing Spark jobs.
  • Experience with distributed data processing and storage systems.
  • Strong understanding of data pipeline design for analytics, reporting, and ML workflows.
  • Experience working closely with cross-functional teams with diverse data needs.
  • Experience designing and maintaining data pipelines for analytics, reporting, or ML use cases.
  • Strong problem-solving skills and the ability to work across teams with varied data requirements.
Desired Skills
  • Hands-on experience working with Databricks in production environments.
  • Familiarity with the GCP data stack, including Pub/Sub, Dataflow, BigQuery, and Google Cloud Storage (GCS).
  • Exposure to data quality frameworks, data validation, or schema management tools.
  • Understanding of analytics use cases, experimentation, or ML data workflows.
  • Strong Plus - GCP Dataflow, BigQuery, Google Analytics / Data Analytics tooling
  • Experience designing datasets for experimentation or ML feedback loops
  • Exposure to data quality, validation, or schema management frameworks

Benefits
  • Best in class compensation: We hire only the best, and we pay accordingly.
  • Proximity Talks: Learn from experienced engineers, data scientists, and product leaders.
  • High-impact work: Build data systems that directly power AI models and business decisions.

Continuous learning: Work with a strong, collaborative team and grow your data engineering skills every day.

About us

We are Proximity — a global team of coders, designers, product managers, geeks, and experts. We solve complex problems and build cutting-edge technology at scale.

Our team of Proxonauts is growing quickly, which means your impact on the company’s success will be significant. You’ll work with experienced leaders who have built and led high-performing tech, data, and product teams.

Here’s a quick guide to getting to know us better:

  • Watch our CEO, Hardik Jagda, tell you all about Proximity.
  • Read about Proximity’s values and meet some of our Proxonauts.
  • Explore our website, blog, and design wing — Studio Proximity.
  • Get behind the scenes with us on Instagram — follow @ProxWrks and @H.Jagda.

Top Skills

Spark
BigQuery
Databricks
Dataflow
GCP
Google Cloud Storage
Pub/Sub
Python

Similar Jobs

3 Days Ago
Remote or Hybrid
India
Senior level
Senior level
Fintech • Information Technology • Insurance • Financial Services • Big Data Analytics
Design, develop, implement and maintain Big Data solutions for Disability & Absence products. Build batch, speed, and serving layers using Spark, Hive, NoSQL, SOLR and Kafka; optimize Hadoop jobs; automate with Shell/Python; support deployments, CI/CD, runbooks, and cross-team collaboration.
Top Skills: Spark,Hive,Nosql,Solr,Scala,Pig,Kafka,Hbase,Nifi,Change-Data-Capture,Hadoop,Ci/Cd,Shell,Python,Azure,Google Cloud
5 Days Ago
Remote or Hybrid
India
Senior level
Senior level
Fintech • Professional Services • Consulting • Energy • Financial Services • Cybersecurity • Generative AI
As a Data Engineer, you will work with GCP technologies, managing big data, utilizing ETL/ELT processes, and collaborating across teams while advancing in your career.
Top Skills: AirflowBigQueryBigtableCloud RunDatastoreETLGCPGcsJavaPysparkPythonSpannerSparkSQL
6 Days Ago
Remote or Hybrid
India
Senior level
Senior level
Fintech • Professional Services • Consulting • Energy • Financial Services • Cybersecurity • Generative AI
The Data Engineer will design and implement data solutions using GCP services, focusing on big data processing, ETL/ELT processes, and analytical problem-solving.
Top Skills: AirflowBigQueryBigtableCloud FunctionsDatastoreGCPJavaPythonSpannerSparkSQL

What you need to know about the Mumbai Tech Scene

From haggling for the best price at Chor Bazaar to the bustle of Crawford Market, the energy of Mumbai's traditional markets is a key part of the city's charm. And while these markets will always have their place, the city also boasts a thriving e-commerce scene, ranking among the largest in the region. Driven by online sales in everything from snacks to licensed sports merchandise to children's apparel, the local industry is worth billions, with companies actively recruiting to meet the demands of continued growth.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account