Honeywell Logo

Honeywell

Advanced Data Scientist

Posted Yesterday
Be an Early Applicant
In-Office or Remote
Hiring Remotely in India
Senior level
In-Office or Remote
Hiring Remotely in India
Senior level
The role involves developing end-to-end data science and ML solutions, collaborating with stakeholders, designing ML systems, and mentoring junior scientists. Strong expertise in ML, data engineering, MLOps, and cloud platforms is essential.
The summary above was generated by AI

Job Description

Advanced Data Scientist

Location

Bangalore, India (Hybrid / Remote as applicable)

Role Overview

We are looking for a Senior / Lead Data Scientist who can own end‑to‑end data science and machine learning solutions, from problem formulation to production deployment.
This role requires a strong blend of machine learning expertise, data engineering, MLOps, cloud platforms, and technical leadership.

You will work closely with product, engineering, and business stakeholders to design scalable data and ML systems that drive measurable business impact.

Responsibilities

Key Responsibilities

Data Science & Machine Learning

  • Translate business problems into data science and ML solutions
  • Perform advanced EDA, feature engineering, and model development
  • Build and optimize: 
    • Classical ML models (regression, classification, tree‑based models)
    • Time‑series, anomaly detection, and recommendation systems
  • Develop and fine‑tune deep learning models using PyTorch / TensorFlow
  • Design and evaluate experiments (A/B testing, statistical validation)
 

GenAI, NLP & LLM Solutions

  • Build NLP and GenAI applications using modern LLMs
  • Implement RAG pipelines, prompt engineering, and vector search
  • Integrate LLMs using OpenAI / Azure OpenAI APIs
  • Evaluate model quality, latency, and cost for production LLM systems
 

Data Engineering & Pipelines (Good to Have)

  • Design and build scalable data pipelines for batch and streaming use cases
  • Work with distributed processing frameworks like Apache Spark
  • Orchestrate workflows using Airflow / Dagster / Prefect/ Azure Data Factory / Databricks 
  • Handle real‑time data using Kafka or cloud‑native streaming services
  • Ensure data reliability, quality, and performance at scale
 

MLOps, Deployment & Production

  • Own the full ML lifecycle: experimentation → training → deployment → monitoring
  • Implement model versioning, reproducibility, and CI/CD pipelines
  • Deploy models using REST APIs or batch inference pipelines
  • Monitor model performance, drift, and data quality in production
  • Work with Docker and Kubernetes for scalable deployments
 

Cloud & Platform Engineering

  • Build solutions on AWS / Azure / GCP (at least one in depth)
  • Work with cloud data platforms like Databricks, Snowflake, BigQuery
  • Optimize system performance and cloud costs
  • Ensure security, access control, and compliance best practices
 

Architecture, Collaboration & Leadership

  • Design end‑to‑end data and ML architectures
  • Make tradeoffs between batch vs streaming, cost vs performance
  • Mentor junior data scientists and review code and models
  • Set data science and ML best practices across teams
  • Communicate insights clearly to technical and non‑technical stakeholders
 

Required Skills & Qualifications

Core Technical Skills

  • Strong proficiency in Python and advanced SQL
  • Solid foundation in statistics, probability, and linear algebra
  • Hands‑on experience with XGBoost, LightGBM
  • Experience with PyTorch or TensorFlow

Data Engineering (Good to have)

  • Strong experience with Spark / PySpark
  • Pipeline orchestration using Airflow or similar tools
  • Experience with relational, NoSQL, and analytical databases
  • Understanding of data lakes and warehouse architectures

MLOps & DevOps (Optional)

  • Experience with MLflow, DVC, or W&B
  • Model deployment using FastAPI
  • Containers and orchestration: Docker, Kubernetes
  • CI/CD and monitoring tools

Cloud Platforms

  • Deep expertise in at least one cloud provider: 
    • AWS, Azure, or GCP
  • Experience with managed ML and data services
 

Preferred / Nice‑to‑Have

  • Experience with LLM frameworks (LangChain, LlamaIndex)
  • Vector databases (FAISS, Pinecone, Weaviate)
  • Streaming frameworks (Flink)
  • Knowledge of data governance, privacy, and compliance
  • Experience leading cross‑functional technical initiatives

Machine Learning Algorithms & Techniques (Hands‑On)

Supervised Learning

  • Linear Models 
    • Linear Regression
    • Logistic Regression
    • Regularization (L1, L2, Elastic Net)
  • Tree‑Based Models 
    • Decision Trees
    • Random Forest
    • Gradient Boosting (XGBoost, LightGBM, CatBoost)
  • Clustering Techniques 
    • K‑Means
    • Hierarchical Clustering
    • DBSCAN
    • PCA (feature reduction)
    • t‑SNE / UMAP (visualization & analysis)

Dimensionality Reduction 

 

Time Series & Forecasting (Basic–Intermediate)

  • Statistical forecasting: 
    • Moving averages
    • ARIMA / SARIMA (conceptual + basic use)
  • ML‑based forecasting using regression and tree‑based models

 

Model Evaluation & Optimization

  • Cross‑validation techniques
  • Hyperparameter tuning (Grid Search, Random Search)
  • Bias–variance tradeoff
  • Handling class imbalance
  • Selection of appropriate evaluation metrics
Qualifications

Experience

8–12 years 

About UsHoneywell helps organizations solve the world's most complex challenges in automation, the future of aviation and energy transition. As a trusted partner, we provide actionable solutions and innovation through our Aerospace Technologies, Building Automation, Energy and Sustainability Solutions, and Industrial Automation business segments – powered by our Honeywell Forge software – that help make the world smarter, safer and more sustainable.

Top Skills

Airflow
AWS
Azure
Docker
GCP
Kubernetes
Lightgbm
Python
PyTorch
Spark
SQL
TensorFlow
Xgboost

Similar Jobs

9 Days Ago
Remote or Hybrid
India
Senior level
Senior level
Fintech • Information Technology • Insurance • Financial Services • Big Data Analytics
The Senior Data Scientist role involves developing data analytics solutions, mentoring other data scientists, and utilizing machine learning to solve business problems in the banking sector.
Top Skills: Azure MlMl OpsNlpPythonPyTorchScikit-LearnTensorFlow
8 Days Ago
In-Office or Remote
Mumbai, Maharashtra, IND
Senior level
Senior level
Blockchain • Fintech • Payments • Financial Services • Cryptocurrency • Web3
As a Senior Data Scientist, you'll analyze data, support strategic decision-making in payments, build visualizations, and enhance data products by collaborating with business partners.
Top Skills: PythonRSQL
Yesterday
In-Office or Remote
India
Mid level
Mid level
Aerospace • Security • Energy • Industrial
The Data Scientist II will analyze data, build models, and support deployment, focusing on machine learning and NLP, collaborating with cross-functional teams.
Top Skills: AWSAzureDockerGCPJupyter NotebooksMlflowPysparkPythonSparkSQL

What you need to know about the Mumbai Tech Scene

From haggling for the best price at Chor Bazaar to the bustle of Crawford Market, the energy of Mumbai's traditional markets is a key part of the city's charm. And while these markets will always have their place, the city also boasts a thriving e-commerce scene, ranking among the largest in the region. Driven by online sales in everything from snacks to licensed sports merchandise to children's apparel, the local industry is worth billions, with companies actively recruiting to meet the demands of continued growth.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account