946 Data Engineer jobs in New Delhi

Big Data Engineer

Noida, Uttar Pradesh ₹1992000 - ₹2280000 Y DCS IT Solution Pvt. Ltd

Posted today

Job Viewed

Tap Again To Close

Job Description

Job Title: Python Databricks Engineer

Experience: 7–10 Years

Location: Noida

Job Summary

We are seeking a highly skilled Python Databricks Engineer with strong expertise in data engineering and cloud technologies. The ideal candidate will have a solid background in building scalable data pipelines, working with Databricks/Spark, and leveraging AWS services for data processing. Experience in healthcare data systems is a plus.

Key Responsibilities-

Design, develop, and optimize scalable data pipelines and workflows using Databricks, Spark, and Python.

Work with AWS services (S3, Lambda) to integrate and process large-scale datasets.

Develop, schedule, and monitor ETL/ELT processes using Airflow and other orchestration tools.

Collaborate with cross-functional teams to ensure data quality, security, and governance.

Requirements-

7–10 years of overall IT experience with at least 5+ years in Data Engineering.

2+ years of hands-on Python development.

Strong experience with Databricks and Spark.

Proficiency in SQL for querying and transforming data.

Hands-on experience with AWS technologies (S3, Lambda).

Experience with DataStage or similar ETL tools.

Strong knowledge of Airflow for workflow orchestration.

Exposure to healthcare data processing systems is highly desirable.

Ability to work in hybrid teams, mentor juniors, and provide technical guidance.

Technical Stack-

Databricks, Python, Spark, SQL

AWS (S3, Lambda)

Healthcare Data Processing Systems

Batch Processing Frameworks

Airflow, DataStage (or similar ETL tools)

Job Type: Full-time

Pay: ₹166, ₹190,000.00 per month

Work Location: In person

This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

Delhi, Delhi Alef Education

Posted 1 day ago

Job Viewed

Tap Again To Close

Job Description

Who we are

Alef Education began with a bold idea: that every learner deserves a personalised and meaningful education experience. What started in 2016 as a small pilot programme in Abu Dhabi has evolved into one of the world’s most dynamic EdTech companies—reshaping how millions of students engage with learning across the globe.

Today, Alef is proudly headquartered in the UAE, working hand-in-hand with ministries of education, schools, and teachers to bring smart, data-powered platforms into classrooms in over 14,000 schools.

Supporting over 1.1 million students and 50,000 teachers across the UAE, Indonesia & Morocco our AI-driven platforms generate 16+ million data points every day, helping drive smarter learning decisions. Whether it’s improving national exam results, boosting classroom engagement, or supporting educators with world-class tools, Alef is committed to impact at scale.

In 2024, Alef made history as the first EdTech company to list on the Abu Dhabi Securities Exchange (ADX), cementing our role as a regional innovator with global reach.

About The Role

As an ALEF Big Data Engineer you will have a strong understanding of big data technologies with an exceptional ability to code. You will provide technical leadership, working closely with the wider team to ensure high quality code is delivered in line with the project goals and delivery cycles. You will work closely with other teams to deliver rapid prototypes as well as production code for which you will ensure high accessibility standards are upheld. We expect familiarity with modern frameworks and languages, as well as working practices such as Clean Code, TDD, BDD, continuous integration, continuous delivery, and DevOps.

Key Responsibilities

Defining and developing services and solutions

- Define, design, and develop services and solutions around large data ingestion, storage, and management such as withRDBMS, No SQL DBs, Log Files, Events.
- Define, design, and run robust data pipelines/batch jobs in a production environment.
- Architecting highly scalable, highly concurrent, and low latency systems

Maintain, support, and enhance current systems.

- Contribute to paying down technical debt and use development approaches that minimize the growth of new technical debt.
- Contribute feedback to improve the quality, readability, and testability of the code base within your team.
- Mentor and train other developers in a non-line management capacity.
- Build tools (One of SBT, Gradle, Maven).
- Ensure all software built is robust and scalable.

Collaborating with Internal and external stakeholders

- Participating in sprint planning to work with developers and project teams to ensure projects are deployable and monitorable from the outside.
- Work with third-party and other internal providers to support a variety of integrations.
- As part of the team, you may be expected to participate in some of the 2nd line in-house support and Out-of-Hours support rotas.
- Proactively advise on best practices.

To Be The Right Fit, You'll Need

- Degree in Computer Science, Software Engineering or related preferred
- Minimum of 5 years experience in a Big Data
- Follow Clean Code/Solid principles
- Adhere and use TDD/BDD.
- Outstanding ability to develop efficient, readable, highly optimized/maintainable and clear code.
- Highly Proficient in either Functional Java or Scala, Python
- Knowledge of Azure Big Data/Analytics services – ADLS (Azure Data Lake Storage), HDInsight, Azure Data Factory, Azure Synapse Analytics, Azure Fabric, Azure Event Hubs, Azure Stream Analytics, Azure Databricks
- Experience of Storing Data in systems such as Hadoop HDFS, ADLS, Event Hubs
- Experience of designing, setting up and running big data tech stacks such as Hadoop, Azure Databricks, Spark and distributed datastores such as Cassandra, DocumentDBs, MongoDB, Event Hubs
- In-depth knowledge of Hadoop technology ecosystem – HDFS, Spark, Hive, HBase, Event Hubs, Flume, Sqoop, Oozie, SPARK, Avro, Parquet
- Experience debugging a complex multi-server service.
- In depth knowledge and experience in IaaS/PaaS solutions (eg AWS Infrastructure hosting and managed services)
- Familiarity with network protocols - TCP/IP, HTTP, SSL, etc.
- Knowledge of relational and non-relational database systems
- Understanding continuous integration and delivery.
- Mocking (any of the following Mockito, ScalaTest Spock, Jasmine, Mocha).
- IDE Intellij or Eclipse.
- Build tools (One of SBT, Gradle, Maven).
- Ensure all software built is robust and scalable.
- An ability to communicate technical concepts to a non-technical audience.
- Working knowledge of unix-like operating systems such as Linux and/or Mac OS X.
- Knowledge of the git version control system.
- Ability to quickly research and learn new programming tools and techniques.
This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

Ghaziabad, Uttar Pradesh Alef Education

Posted today

Job Viewed

Tap Again To Close

Job Description

Who we are

Alef Education began with a bold idea: that every learner deserves a personalised and meaningful education experience. What started in 2016 as a small pilot programme in Abu Dhabi has evolved into one of the world’s most dynamic EdTech companies—reshaping how millions of students engage with learning across the globe.


Today, Alef is proudly headquartered in the UAE, working hand-in-hand with ministries of education, schools, and teachers to bring smart, data-powered platforms into classrooms in over 14,000 schools.


Supporting over 1.1 million students and 50,000 teachers across the UAE, Indonesia & Morocco our AI-driven platforms generate 16+ million data points every day, helping drive smarter learning decisions. Whether it’s improving national exam results, boosting classroom engagement, or supporting educators with world-class tools, Alef is committed to impact at scale.


In 2024, Alef made history as the first EdTech company to list on the Abu Dhabi Securities Exchange (ADX), cementing our role as a regional innovator with global reach.


About The Role

As an ALEF Big Data Engineer you will have a strong understanding of big data technologies with an exceptional ability to code. You will provide technical leadership, working closely with the wider team to ensure high quality code is delivered in line with the project goals and delivery cycles. You will work closely with other teams to deliver rapid prototypes as well as production code for which you will ensure high accessibility standards are upheld. We expect familiarity with modern frameworks and languages, as well as working practices such as Clean Code, TDD, BDD, continuous integration, continuous delivery, and DevOps.


Key Responsibilities

Defining and developing services and solutions

  • Define, design, and develop services and solutions around large data ingestion, storage, and management such as withRDBMS, No SQL DBs, Log Files, Events.
  • Define, design, and run robust data pipelines/batch jobs in a production environment.
  • Architecting highly scalable, highly concurrent, and low latency systems


Maintain, support, and enhance current systems.

  • Contribute to paying down technical debt and use development approaches that minimize the growth of new technical debt.
  • Contribute feedback to improve the quality, readability, and testability of the code base within your team.
  • Mentor and train other developers in a non-line management capacity.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.


Collaborating with Internal and external stakeholders


  • Participating in sprint planning to work with developers and project teams to ensure projects are deployable and monitorable from the outside.
  • Work with third-party and other internal providers to support a variety of integrations.
  • As part of the team, you may be expected to participate in some of the 2nd line in-house support and Out-of-Hours support rotas.
  • Proactively advise on best practices.


To Be The Right Fit, You'll Need

  • Degree in Computer Science, Software Engineering or related preferred
  • Minimum of 5 years experience in a Big Data
  • Follow Clean Code/Solid principles
  • Adhere and use TDD/BDD.
  • Outstanding ability to develop efficient, readable, highly optimized/maintainable and clear code.
  • Highly Proficient in either Functional Java or Scala, Python
  • Knowledge of Azure Big Data/Analytics services – ADLS (Azure Data Lake Storage), HDInsight, Azure Data Factory, Azure Synapse Analytics, Azure Fabric, Azure Event Hubs, Azure Stream Analytics, Azure Databricks
  • Experience of Storing Data in systems such as Hadoop HDFS, ADLS, Event Hubs
  • Experience of designing, setting up and running big data tech stacks such as Hadoop, Azure Databricks, Spark and distributed datastores such as Cassandra, DocumentDBs, MongoDB, Event Hubs
  • In-depth knowledge of Hadoop technology ecosystem – HDFS, Spark, Hive, HBase, Event Hubs, Flume, Sqoop, Oozie, SPARK, Avro, Parquet
  • Experience debugging a complex multi-server service.
  • In depth knowledge and experience in IaaS/PaaS solutions (eg AWS Infrastructure hosting and managed services)
  • Familiarity with network protocols - TCP/IP, HTTP, SSL, etc.
  • Knowledge of relational and non-relational database systems
  • Understanding continuous integration and delivery.
  • Mocking (any of the following Mockito, ScalaTest Spock, Jasmine, Mocha).
  • IDE Intellij or Eclipse.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.
  • An ability to communicate technical concepts to a non-technical audience.
  • Working knowledge of unix-like operating systems such as Linux and/or Mac OS X.
  • Knowledge of the git version control system.
  • Ability to quickly research and learn new programming tools and techniques.
This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

Gurgaon, Haryana Alef Education

Posted 1 day ago

Job Viewed

Tap Again To Close

Job Description

Who we are

Alef Education began with a bold idea: that every learner deserves a personalised and meaningful education experience. What started in 2016 as a small pilot programme in Abu Dhabi has evolved into one of the world’s most dynamic EdTech companies—reshaping how millions of students engage with learning across the globe.


Today, Alef is proudly headquartered in the UAE, working hand-in-hand with ministries of education, schools, and teachers to bring smart, data-powered platforms into classrooms in over 14,000 schools.


Supporting over 1.1 million students and 50,000 teachers across the UAE, Indonesia & Morocco our AI-driven platforms generate 16+ million data points every day, helping drive smarter learning decisions. Whether it’s improving national exam results, boosting classroom engagement, or supporting educators with world-class tools, Alef is committed to impact at scale.


In 2024, Alef made history as the first EdTech company to list on the Abu Dhabi Securities Exchange (ADX), cementing our role as a regional innovator with global reach.


About The Role

As an ALEF Big Data Engineer you will have a strong understanding of big data technologies with an exceptional ability to code. You will provide technical leadership, working closely with the wider team to ensure high quality code is delivered in line with the project goals and delivery cycles. You will work closely with other teams to deliver rapid prototypes as well as production code for which you will ensure high accessibility standards are upheld. We expect familiarity with modern frameworks and languages, as well as working practices such as Clean Code, TDD, BDD, continuous integration, continuous delivery, and DevOps.


Key Responsibilities

Defining and developing services and solutions

  • Define, design, and develop services and solutions around large data ingestion, storage, and management such as withRDBMS, No SQL DBs, Log Files, Events.
  • Define, design, and run robust data pipelines/batch jobs in a production environment.
  • Architecting highly scalable, highly concurrent, and low latency systems


Maintain, support, and enhance current systems.

  • Contribute to paying down technical debt and use development approaches that minimize the growth of new technical debt.
  • Contribute feedback to improve the quality, readability, and testability of the code base within your team.
  • Mentor and train other developers in a non-line management capacity.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.


Collaborating with Internal and external stakeholders


  • Participating in sprint planning to work with developers and project teams to ensure projects are deployable and monitorable from the outside.
  • Work with third-party and other internal providers to support a variety of integrations.
  • As part of the team, you may be expected to participate in some of the 2nd line in-house support and Out-of-Hours support rotas.
  • Proactively advise on best practices.


To Be The Right Fit, You'll Need

  • Degree in Computer Science, Software Engineering or related preferred
  • Minimum of 5 years experience in a Big Data
  • Follow Clean Code/Solid principles
  • Adhere and use TDD/BDD.
  • Outstanding ability to develop efficient, readable, highly optimized/maintainable and clear code.
  • Highly Proficient in either Functional Java or Scala, Python
  • Knowledge of Azure Big Data/Analytics services – ADLS (Azure Data Lake Storage), HDInsight, Azure Data Factory, Azure Synapse Analytics, Azure Fabric, Azure Event Hubs, Azure Stream Analytics, Azure Databricks
  • Experience of Storing Data in systems such as Hadoop HDFS, ADLS, Event Hubs
  • Experience of designing, setting up and running big data tech stacks such as Hadoop, Azure Databricks, Spark and distributed datastores such as Cassandra, DocumentDBs, MongoDB, Event Hubs
  • In-depth knowledge of Hadoop technology ecosystem – HDFS, Spark, Hive, HBase, Event Hubs, Flume, Sqoop, Oozie, SPARK, Avro, Parquet
  • Experience debugging a complex multi-server service.
  • In depth knowledge and experience in IaaS/PaaS solutions (eg AWS Infrastructure hosting and managed services)
  • Familiarity with network protocols - TCP/IP, HTTP, SSL, etc.
  • Knowledge of relational and non-relational database systems
  • Understanding continuous integration and delivery.
  • Mocking (any of the following Mockito, ScalaTest Spock, Jasmine, Mocha).
  • IDE Intellij or Eclipse.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.
  • An ability to communicate technical concepts to a non-technical audience.
  • Working knowledge of unix-like operating systems such as Linux and/or Mac OS X.
  • Knowledge of the git version control system.
  • Ability to quickly research and learn new programming tools and techniques.
This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

New Delhi, Delhi Alef Education

Posted 1 day ago

Job Viewed

Tap Again To Close

Job Description

Who we are

Alef Education began with a bold idea: that every learner deserves a personalised and meaningful education experience. What started in 2016 as a small pilot programme in Abu Dhabi has evolved into one of the world’s most dynamic EdTech companies—reshaping how millions of students engage with learning across the globe.


Today, Alef is proudly headquartered in the UAE, working hand-in-hand with ministries of education, schools, and teachers to bring smart, data-powered platforms into classrooms in over 14,000 schools.


Supporting over 1.1 million students and 50,000 teachers across the UAE, Indonesia & Morocco our AI-driven platforms generate 16+ million data points every day, helping drive smarter learning decisions. Whether it’s improving national exam results, boosting classroom engagement, or supporting educators with world-class tools, Alef is committed to impact at scale.


In 2024, Alef made history as the first EdTech company to list on the Abu Dhabi Securities Exchange (ADX), cementing our role as a regional innovator with global reach.


About The Role

As an ALEF Big Data Engineer you will have a strong understanding of big data technologies with an exceptional ability to code. You will provide technical leadership, working closely with the wider team to ensure high quality code is delivered in line with the project goals and delivery cycles. You will work closely with other teams to deliver rapid prototypes as well as production code for which you will ensure high accessibility standards are upheld. We expect familiarity with modern frameworks and languages, as well as working practices such as Clean Code, TDD, BDD, continuous integration, continuous delivery, and DevOps.


Key Responsibilities

Defining and developing services and solutions

  • Define, design, and develop services and solutions around large data ingestion, storage, and management such as withRDBMS, No SQL DBs, Log Files, Events.
  • Define, design, and run robust data pipelines/batch jobs in a production environment.
  • Architecting highly scalable, highly concurrent, and low latency systems


Maintain, support, and enhance current systems.

  • Contribute to paying down technical debt and use development approaches that minimize the growth of new technical debt.
  • Contribute feedback to improve the quality, readability, and testability of the code base within your team.
  • Mentor and train other developers in a non-line management capacity.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.


Collaborating with Internal and external stakeholders


  • Participating in sprint planning to work with developers and project teams to ensure projects are deployable and monitorable from the outside.
  • Work with third-party and other internal providers to support a variety of integrations.
  • As part of the team, you may be expected to participate in some of the 2nd line in-house support and Out-of-Hours support rotas.
  • Proactively advise on best practices.


To Be The Right Fit, You'll Need

  • Degree in Computer Science, Software Engineering or related preferred
  • Minimum of 5 years experience in a Big Data
  • Follow Clean Code/Solid principles
  • Adhere and use TDD/BDD.
  • Outstanding ability to develop efficient, readable, highly optimized/maintainable and clear code.
  • Highly Proficient in either Functional Java or Scala, Python
  • Knowledge of Azure Big Data/Analytics services – ADLS (Azure Data Lake Storage), HDInsight, Azure Data Factory, Azure Synapse Analytics, Azure Fabric, Azure Event Hubs, Azure Stream Analytics, Azure Databricks
  • Experience of Storing Data in systems such as Hadoop HDFS, ADLS, Event Hubs
  • Experience of designing, setting up and running big data tech stacks such as Hadoop, Azure Databricks, Spark and distributed datastores such as Cassandra, DocumentDBs, MongoDB, Event Hubs
  • In-depth knowledge of Hadoop technology ecosystem – HDFS, Spark, Hive, HBase, Event Hubs, Flume, Sqoop, Oozie, SPARK, Avro, Parquet
  • Experience debugging a complex multi-server service.
  • In depth knowledge and experience in IaaS/PaaS solutions (eg AWS Infrastructure hosting and managed services)
  • Familiarity with network protocols - TCP/IP, HTTP, SSL, etc.
  • Knowledge of relational and non-relational database systems
  • Understanding continuous integration and delivery.
  • Mocking (any of the following Mockito, ScalaTest Spock, Jasmine, Mocha).
  • IDE Intellij or Eclipse.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.
  • An ability to communicate technical concepts to a non-technical audience.
  • Working knowledge of unix-like operating systems such as Linux and/or Mac OS X.
  • Knowledge of the git version control system.
  • Ability to quickly research and learn new programming tools and techniques.
This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

Faridabad, Haryana Alef Education

Posted 1 day ago

Job Viewed

Tap Again To Close

Job Description

Who we are

Alef Education began with a bold idea: that every learner deserves a personalised and meaningful education experience. What started in 2016 as a small pilot programme in Abu Dhabi has evolved into one of the world’s most dynamic EdTech companies—reshaping how millions of students engage with learning across the globe.


Today, Alef is proudly headquartered in the UAE, working hand-in-hand with ministries of education, schools, and teachers to bring smart, data-powered platforms into classrooms in over 14,000 schools.


Supporting over 1.1 million students and 50,000 teachers across the UAE, Indonesia & Morocco our AI-driven platforms generate 16+ million data points every day, helping drive smarter learning decisions. Whether it’s improving national exam results, boosting classroom engagement, or supporting educators with world-class tools, Alef is committed to impact at scale.


In 2024, Alef made history as the first EdTech company to list on the Abu Dhabi Securities Exchange (ADX), cementing our role as a regional innovator with global reach.


About The Role

As an ALEF Big Data Engineer you will have a strong understanding of big data technologies with an exceptional ability to code. You will provide technical leadership, working closely with the wider team to ensure high quality code is delivered in line with the project goals and delivery cycles. You will work closely with other teams to deliver rapid prototypes as well as production code for which you will ensure high accessibility standards are upheld. We expect familiarity with modern frameworks and languages, as well as working practices such as Clean Code, TDD, BDD, continuous integration, continuous delivery, and DevOps.


Key Responsibilities

Defining and developing services and solutions

  • Define, design, and develop services and solutions around large data ingestion, storage, and management such as withRDBMS, No SQL DBs, Log Files, Events.
  • Define, design, and run robust data pipelines/batch jobs in a production environment.
  • Architecting highly scalable, highly concurrent, and low latency systems


Maintain, support, and enhance current systems.

  • Contribute to paying down technical debt and use development approaches that minimize the growth of new technical debt.
  • Contribute feedback to improve the quality, readability, and testability of the code base within your team.
  • Mentor and train other developers in a non-line management capacity.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.


Collaborating with Internal and external stakeholders


  • Participating in sprint planning to work with developers and project teams to ensure projects are deployable and monitorable from the outside.
  • Work with third-party and other internal providers to support a variety of integrations.
  • As part of the team, you may be expected to participate in some of the 2nd line in-house support and Out-of-Hours support rotas.
  • Proactively advise on best practices.


To Be The Right Fit, You'll Need

  • Degree in Computer Science, Software Engineering or related preferred
  • Minimum of 5 years experience in a Big Data
  • Follow Clean Code/Solid principles
  • Adhere and use TDD/BDD.
  • Outstanding ability to develop efficient, readable, highly optimized/maintainable and clear code.
  • Highly Proficient in either Functional Java or Scala, Python
  • Knowledge of Azure Big Data/Analytics services – ADLS (Azure Data Lake Storage), HDInsight, Azure Data Factory, Azure Synapse Analytics, Azure Fabric, Azure Event Hubs, Azure Stream Analytics, Azure Databricks
  • Experience of Storing Data in systems such as Hadoop HDFS, ADLS, Event Hubs
  • Experience of designing, setting up and running big data tech stacks such as Hadoop, Azure Databricks, Spark and distributed datastores such as Cassandra, DocumentDBs, MongoDB, Event Hubs
  • In-depth knowledge of Hadoop technology ecosystem – HDFS, Spark, Hive, HBase, Event Hubs, Flume, Sqoop, Oozie, SPARK, Avro, Parquet
  • Experience debugging a complex multi-server service.
  • In depth knowledge and experience in IaaS/PaaS solutions (eg AWS Infrastructure hosting and managed services)
  • Familiarity with network protocols - TCP/IP, HTTP, SSL, etc.
  • Knowledge of relational and non-relational database systems
  • Understanding continuous integration and delivery.
  • Mocking (any of the following Mockito, ScalaTest Spock, Jasmine, Mocha).
  • IDE Intellij or Eclipse.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.
  • An ability to communicate technical concepts to a non-technical audience.
  • Working knowledge of unix-like operating systems such as Linux and/or Mac OS X.
  • Knowledge of the git version control system.
  • Ability to quickly research and learn new programming tools and techniques.
This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

Delhi, Delhi Alef Education

Posted 1 day ago

Job Viewed

Tap Again To Close

Job Description

Who we are

Alef Education began with a bold idea: that every learner deserves a personalised and meaningful education experience. What started in 2016 as a small pilot programme in Abu Dhabi has evolved into one of the world’s most dynamic EdTech companies—reshaping how millions of students engage with learning across the globe.


Today, Alef is proudly headquartered in the UAE, working hand-in-hand with ministries of education, schools, and teachers to bring smart, data-powered platforms into classrooms in over 14,000 schools.


Supporting over 1.1 million students and 50,000 teachers across the UAE, Indonesia & Morocco our AI-driven platforms generate 16+ million data points every day, helping drive smarter learning decisions. Whether it’s improving national exam results, boosting classroom engagement, or supporting educators with world-class tools, Alef is committed to impact at scale.


In 2024, Alef made history as the first EdTech company to list on the Abu Dhabi Securities Exchange (ADX), cementing our role as a regional innovator with global reach.


About The Role

As an ALEF Big Data Engineer you will have a strong understanding of big data technologies with an exceptional ability to code. You will provide technical leadership, working closely with the wider team to ensure high quality code is delivered in line with the project goals and delivery cycles. You will work closely with other teams to deliver rapid prototypes as well as production code for which you will ensure high accessibility standards are upheld. We expect familiarity with modern frameworks and languages, as well as working practices such as Clean Code, TDD, BDD, continuous integration, continuous delivery, and DevOps.


Key Responsibilities

Defining and developing services and solutions

  • Define, design, and develop services and solutions around large data ingestion, storage, and management such as withRDBMS, No SQL DBs, Log Files, Events.
  • Define, design, and run robust data pipelines/batch jobs in a production environment.
  • Architecting highly scalable, highly concurrent, and low latency systems


Maintain, support, and enhance current systems.

  • Contribute to paying down technical debt and use development approaches that minimize the growth of new technical debt.
  • Contribute feedback to improve the quality, readability, and testability of the code base within your team.
  • Mentor and train other developers in a non-line management capacity.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.


Collaborating with Internal and external stakeholders


  • Participating in sprint planning to work with developers and project teams to ensure projects are deployable and monitorable from the outside.
  • Work with third-party and other internal providers to support a variety of integrations.
  • As part of the team, you may be expected to participate in some of the 2nd line in-house support and Out-of-Hours support rotas.
  • Proactively advise on best practices.


To Be The Right Fit, You'll Need

  • Degree in Computer Science, Software Engineering or related preferred
  • Minimum of 5 years experience in a Big Data
  • Follow Clean Code/Solid principles
  • Adhere and use TDD/BDD.
  • Outstanding ability to develop efficient, readable, highly optimized/maintainable and clear code.
  • Highly Proficient in either Functional Java or Scala, Python
  • Knowledge of Azure Big Data/Analytics services – ADLS (Azure Data Lake Storage), HDInsight, Azure Data Factory, Azure Synapse Analytics, Azure Fabric, Azure Event Hubs, Azure Stream Analytics, Azure Databricks
  • Experience of Storing Data in systems such as Hadoop HDFS, ADLS, Event Hubs
  • Experience of designing, setting up and running big data tech stacks such as Hadoop, Azure Databricks, Spark and distributed datastores such as Cassandra, DocumentDBs, MongoDB, Event Hubs
  • In-depth knowledge of Hadoop technology ecosystem – HDFS, Spark, Hive, HBase, Event Hubs, Flume, Sqoop, Oozie, SPARK, Avro, Parquet
  • Experience debugging a complex multi-server service.
  • In depth knowledge and experience in IaaS/PaaS solutions (eg AWS Infrastructure hosting and managed services)
  • Familiarity with network protocols - TCP/IP, HTTP, SSL, etc.
  • Knowledge of relational and non-relational database systems
  • Understanding continuous integration and delivery.
  • Mocking (any of the following Mockito, ScalaTest Spock, Jasmine, Mocha).
  • IDE Intellij or Eclipse.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.
  • An ability to communicate technical concepts to a non-technical audience.
  • Working knowledge of unix-like operating systems such as Linux and/or Mac OS X.
  • Knowledge of the git version control system.
  • Ability to quickly research and learn new programming tools and techniques.
This advertiser has chosen not to accept applicants from your region.
Be The First To Know

About the latest Data engineer Jobs in New Delhi !

Big Data Engineer

Noida, Uttar Pradesh Alef Education

Posted 1 day ago

Job Viewed

Tap Again To Close

Job Description

Who we are

Alef Education began with a bold idea: that every learner deserves a personalised and meaningful education experience. What started in 2016 as a small pilot programme in Abu Dhabi has evolved into one of the world’s most dynamic EdTech companies—reshaping how millions of students engage with learning across the globe.


Today, Alef is proudly headquartered in the UAE, working hand-in-hand with ministries of education, schools, and teachers to bring smart, data-powered platforms into classrooms in over 14,000 schools.


Supporting over 1.1 million students and 50,000 teachers across the UAE, Indonesia & Morocco our AI-driven platforms generate 16+ million data points every day, helping drive smarter learning decisions. Whether it’s improving national exam results, boosting classroom engagement, or supporting educators with world-class tools, Alef is committed to impact at scale.


In 2024, Alef made history as the first EdTech company to list on the Abu Dhabi Securities Exchange (ADX), cementing our role as a regional innovator with global reach.


About The Role

As an ALEF Big Data Engineer you will have a strong understanding of big data technologies with an exceptional ability to code. You will provide technical leadership, working closely with the wider team to ensure high quality code is delivered in line with the project goals and delivery cycles. You will work closely with other teams to deliver rapid prototypes as well as production code for which you will ensure high accessibility standards are upheld. We expect familiarity with modern frameworks and languages, as well as working practices such as Clean Code, TDD, BDD, continuous integration, continuous delivery, and DevOps.


Key Responsibilities

Defining and developing services and solutions

  • Define, design, and develop services and solutions around large data ingestion, storage, and management such as withRDBMS, No SQL DBs, Log Files, Events.
  • Define, design, and run robust data pipelines/batch jobs in a production environment.
  • Architecting highly scalable, highly concurrent, and low latency systems


Maintain, support, and enhance current systems.

  • Contribute to paying down technical debt and use development approaches that minimize the growth of new technical debt.
  • Contribute feedback to improve the quality, readability, and testability of the code base within your team.
  • Mentor and train other developers in a non-line management capacity.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.


Collaborating with Internal and external stakeholders


  • Participating in sprint planning to work with developers and project teams to ensure projects are deployable and monitorable from the outside.
  • Work with third-party and other internal providers to support a variety of integrations.
  • As part of the team, you may be expected to participate in some of the 2nd line in-house support and Out-of-Hours support rotas.
  • Proactively advise on best practices.


To Be The Right Fit, You'll Need

  • Degree in Computer Science, Software Engineering or related preferred
  • Minimum of 5 years experience in a Big Data
  • Follow Clean Code/Solid principles
  • Adhere and use TDD/BDD.
  • Outstanding ability to develop efficient, readable, highly optimized/maintainable and clear code.
  • Highly Proficient in either Functional Java or Scala, Python
  • Knowledge of Azure Big Data/Analytics services – ADLS (Azure Data Lake Storage), HDInsight, Azure Data Factory, Azure Synapse Analytics, Azure Fabric, Azure Event Hubs, Azure Stream Analytics, Azure Databricks
  • Experience of Storing Data in systems such as Hadoop HDFS, ADLS, Event Hubs
  • Experience of designing, setting up and running big data tech stacks such as Hadoop, Azure Databricks, Spark and distributed datastores such as Cassandra, DocumentDBs, MongoDB, Event Hubs
  • In-depth knowledge of Hadoop technology ecosystem – HDFS, Spark, Hive, HBase, Event Hubs, Flume, Sqoop, Oozie, SPARK, Avro, Parquet
  • Experience debugging a complex multi-server service.
  • In depth knowledge and experience in IaaS/PaaS solutions (eg AWS Infrastructure hosting and managed services)
  • Familiarity with network protocols - TCP/IP, HTTP, SSL, etc.
  • Knowledge of relational and non-relational database systems
  • Understanding continuous integration and delivery.
  • Mocking (any of the following Mockito, ScalaTest Spock, Jasmine, Mocha).
  • IDE Intellij or Eclipse.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.
  • An ability to communicate technical concepts to a non-technical audience.
  • Working knowledge of unix-like operating systems such as Linux and/or Mac OS X.
  • Knowledge of the git version control system.
  • Ability to quickly research and learn new programming tools and techniques.
This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

Ghaziabad, Uttar Pradesh Alef Education

Posted 1 day ago

Job Viewed

Tap Again To Close

Job Description

Who we are

Alef Education began with a bold idea: that every learner deserves a personalised and meaningful education experience. What started in 2016 as a small pilot programme in Abu Dhabi has evolved into one of the world’s most dynamic EdTech companies—reshaping how millions of students engage with learning across the globe.


Today, Alef is proudly headquartered in the UAE, working hand-in-hand with ministries of education, schools, and teachers to bring smart, data-powered platforms into classrooms in over 14,000 schools.


Supporting over 1.1 million students and 50,000 teachers across the UAE, Indonesia & Morocco our AI-driven platforms generate 16+ million data points every day, helping drive smarter learning decisions. Whether it’s improving national exam results, boosting classroom engagement, or supporting educators with world-class tools, Alef is committed to impact at scale.


In 2024, Alef made history as the first EdTech company to list on the Abu Dhabi Securities Exchange (ADX), cementing our role as a regional innovator with global reach.


About The Role

As an ALEF Big Data Engineer you will have a strong understanding of big data technologies with an exceptional ability to code. You will provide technical leadership, working closely with the wider team to ensure high quality code is delivered in line with the project goals and delivery cycles. You will work closely with other teams to deliver rapid prototypes as well as production code for which you will ensure high accessibility standards are upheld. We expect familiarity with modern frameworks and languages, as well as working practices such as Clean Code, TDD, BDD, continuous integration, continuous delivery, and DevOps.


Key Responsibilities

Defining and developing services and solutions

  • Define, design, and develop services and solutions around large data ingestion, storage, and management such as withRDBMS, No SQL DBs, Log Files, Events.
  • Define, design, and run robust data pipelines/batch jobs in a production environment.
  • Architecting highly scalable, highly concurrent, and low latency systems


Maintain, support, and enhance current systems.

  • Contribute to paying down technical debt and use development approaches that minimize the growth of new technical debt.
  • Contribute feedback to improve the quality, readability, and testability of the code base within your team.
  • Mentor and train other developers in a non-line management capacity.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.


Collaborating with Internal and external stakeholders


  • Participating in sprint planning to work with developers and project teams to ensure projects are deployable and monitorable from the outside.
  • Work with third-party and other internal providers to support a variety of integrations.
  • As part of the team, you may be expected to participate in some of the 2nd line in-house support and Out-of-Hours support rotas.
  • Proactively advise on best practices.


To Be The Right Fit, You'll Need

  • Degree in Computer Science, Software Engineering or related preferred
  • Minimum of 5 years experience in a Big Data
  • Follow Clean Code/Solid principles
  • Adhere and use TDD/BDD.
  • Outstanding ability to develop efficient, readable, highly optimized/maintainable and clear code.
  • Highly Proficient in either Functional Java or Scala, Python
  • Knowledge of Azure Big Data/Analytics services – ADLS (Azure Data Lake Storage), HDInsight, Azure Data Factory, Azure Synapse Analytics, Azure Fabric, Azure Event Hubs, Azure Stream Analytics, Azure Databricks
  • Experience of Storing Data in systems such as Hadoop HDFS, ADLS, Event Hubs
  • Experience of designing, setting up and running big data tech stacks such as Hadoop, Azure Databricks, Spark and distributed datastores such as Cassandra, DocumentDBs, MongoDB, Event Hubs
  • In-depth knowledge of Hadoop technology ecosystem – HDFS, Spark, Hive, HBase, Event Hubs, Flume, Sqoop, Oozie, SPARK, Avro, Parquet
  • Experience debugging a complex multi-server service.
  • In depth knowledge and experience in IaaS/PaaS solutions (eg AWS Infrastructure hosting and managed services)
  • Familiarity with network protocols - TCP/IP, HTTP, SSL, etc.
  • Knowledge of relational and non-relational database systems
  • Understanding continuous integration and delivery.
  • Mocking (any of the following Mockito, ScalaTest Spock, Jasmine, Mocha).
  • IDE Intellij or Eclipse.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.
  • An ability to communicate technical concepts to a non-technical audience.
  • Working knowledge of unix-like operating systems such as Linux and/or Mac OS X.
  • Knowledge of the git version control system.
  • Ability to quickly research and learn new programming tools and techniques.
This advertiser has chosen not to accept applicants from your region.

Senior Data Engineer - Big Data

201301 Noida, Uttar Pradesh ₹115000 Annually WhatJobs

Posted 15 days ago

Job Viewed

Tap Again To Close

Job Description

full-time
Our client is seeking a highly skilled and experienced Senior Data Engineer to join their fully remote team, specializing in Big Data solutions. This role is crucial for designing, building, and maintaining robust data pipelines and infrastructure that support advanced analytics and machine learning initiatives. The ideal candidate will have a deep understanding of distributed systems, data warehousing, ETL processes, and cloud-based data platforms. You will be responsible for architecting scalable data solutions, ensuring data quality and integrity, and collaborating with data scientists and analysts to enable data-driven decision-making across the organization.

Key Responsibilities:
  • Design, develop, and optimize scalable data pipelines using technologies such as Spark, Kafka, and Hadoop ecosystem.
  • Build and manage data warehouses and data lakes on cloud platforms like AWS, Azure, or GCP.
  • Implement robust ETL/ELT processes to ingest, transform, and load data from various sources.
  • Ensure data quality, integrity, and reliability across all data systems.
  • Collaborate closely with data scientists, analysts, and business stakeholders to understand data requirements and deliver solutions.
  • Develop and maintain data models that support analytical needs and reporting.
  • Implement monitoring and alerting systems for data pipelines and infrastructure.
  • Stay current with emerging big data technologies and best practices, and evaluate their potential adoption.
  • Write clean, maintainable, and well-documented code.
  • Contribute to the overall data strategy and architecture of the organization.

Qualifications:
  • Bachelor's or Master's degree in Computer Science, Engineering, or a related quantitative field.
  • Minimum of 7 years of experience in data engineering, with a strong focus on Big Data technologies.
  • Proficiency in programming languages like Python, Scala, or Java.
  • Extensive experience with distributed data processing frameworks (e.g., Apache Spark, Hadoop).
  • Hands-on experience with cloud data platforms (AWS, Azure, GCP) and their services (e.g., S3, Redshift, BigQuery, Data Factory).
  • Solid understanding of SQL and NoSQL databases.
  • Experience with data warehousing concepts and ETL/ELT tools.
  • Familiarity with containerization technologies (e.g., Docker, Kubernetes) is a plus.
  • Excellent problem-solving, analytical, and communication skills.

This is a fantastic opportunity to work remotely on challenging Big Data projects and contribute to a data-centric culture. We offer a competitive salary, comprehensive benefits, and a flexible work environment that fosters growth and innovation.
This advertiser has chosen not to accept applicants from your region.
 

Nearby Locations

Other Jobs Near Me

Industry

  1. request_quote Accounting
  2. work Administrative
  3. eco Agriculture Forestry
  4. smart_toy AI & Emerging Technologies
  5. school Apprenticeships & Trainee
  6. apartment Architecture
  7. palette Arts & Entertainment
  8. directions_car Automotive
  9. flight_takeoff Aviation
  10. account_balance Banking & Finance
  11. local_florist Beauty & Wellness
  12. restaurant Catering
  13. volunteer_activism Charity & Voluntary
  14. science Chemical Engineering
  15. child_friendly Childcare
  16. foundation Civil Engineering
  17. clean_hands Cleaning & Sanitation
  18. diversity_3 Community & Social Care
  19. construction Construction
  20. brush Creative & Digital
  21. currency_bitcoin Crypto & Blockchain
  22. support_agent Customer Service & Helpdesk
  23. medical_services Dental
  24. medical_services Driving & Transport
  25. medical_services E Commerce & Social Media
  26. school Education & Teaching
  27. electrical_services Electrical Engineering
  28. bolt Energy
  29. local_mall Fmcg
  30. gavel Government & Non Profit
  31. emoji_events Graduate
  32. health_and_safety Healthcare
  33. beach_access Hospitality & Tourism
  34. groups Human Resources
  35. precision_manufacturing Industrial Engineering
  36. security Information Security
  37. handyman Installation & Maintenance
  38. policy Insurance
  39. code IT & Software
  40. gavel Legal
  41. sports_soccer Leisure & Sports
  42. inventory_2 Logistics & Warehousing
  43. supervisor_account Management
  44. supervisor_account Management Consultancy
  45. supervisor_account Manufacturing & Production
  46. campaign Marketing
  47. build Mechanical Engineering
  48. perm_media Media & PR
  49. local_hospital Medical
  50. local_hospital Military & Public Safety
  51. local_hospital Mining
  52. medical_services Nursing
  53. local_gas_station Oil & Gas
  54. biotech Pharmaceutical
  55. checklist_rtl Project Management
  56. shopping_bag Purchasing
  57. home_work Real Estate
  58. person_search Recruitment Consultancy
  59. store Retail
  60. point_of_sale Sales
  61. science Scientific Research & Development
  62. wifi Telecoms
  63. psychology Therapy
  64. pets Veterinary
View All Data Engineer Jobs View All Jobs in New Delhi