33,429 Big Data jobs in India

Big Data Engineer

Chennai, Tamil Nadu Citigroup

Posted 2 days ago

Job Viewed

Tap Again To Close

Job Description

The Applications Development Intermediate Programmer Analyst is an intermediate level position responsible for participation in the establishment and implementation of new or revised application systems and programs in coordination with the Technology team. The overall objective of this role is to contribute to applications systems analysis and programming activities.
**Responsibilities:**
+ Design, development of BigData applications/ pipelines using Spark, Scala, SQL, Pyspark, Python, Java
+ Consult with users, clients, and other technology groups on issues, and recommend programming solutions, install, and support customer exposure systems
+ Appropriately assess risk when business decisions are made, demonstrating particular consideration for the firm's reputation and safeguarding Citigroup, its clients and assets, by driving compliance with applicable laws, rules and regulations, adhering to Policy, applying sound ethical judgment regarding personal behavior, conduct and business practices, and escalating, managing and reporting control issues with transparency.
**Qualifications:**
+ 4-8 years of experience in software development, building large scale distributed data processing systems or large-scale applications
+ Designing & developing Big Data solutions with at least one end to end implementation.
+ Strong Hands-on experience in following technical skills: Apache Spark, Scala/ Java, XML/ JSON/ Parquet/ Avro, SQL, Linux, Hadoop Ecosystem (HDFS, Spark, Impala, HIVE, HBASE etc.), Kafka.
+ Performance analysis, troubleshooting and issue resolution and Exposure to latest Cloudera offerings like Ozone, Iceberg.
+ Intermediate level experience in Applications Development role
+ Consistently demonstrates clear and concise written and verbal communication
+ Demonstrated problem-solving and decision-making skills
+ Ability to work under pressure and manage deadlines or unexpected changes in expectations or requirements
**Education:**
+ Bachelor's degree/University degree or equivalent experience
This job description provides a high-level review of the types of work performed. Other job-related duties may be assigned as required.
---
**Job Family Group:**
Technology
---
**Job Family:**
Applications Development
---
**Time Type:**
Full time
---
**Most Relevant Skills**
Please see the requirements listed above.
---
**Other Relevant Skills**
For complementary skills, please see above and/or contact the recruiter.
---
_Citi is an equal opportunity employer, and qualified candidates will receive consideration without regard to their race, color, religion, sex, sexual orientation, gender identity, national origin, disability, status as a protected veteran, or any other characteristic protected by law._
_If you are a person with a disability and need a reasonable accommodation to use our search tools and/or apply for a career opportunity review_ _Accessibility at Citi ( _._
_View Citi's_ _EEO Policy Statement ( _and the_ _Know Your Rights ( _poster._
Citi is an equal opportunity and affirmative action employer.
Minority/Female/Veteran/Individuals with Disabilities/Sexual Orientation/Gender Identity.
This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

Pune, Maharashtra Citigroup

Posted 2 days ago

Job Viewed

Tap Again To Close

Job Description

The Applications Development Programmer Analyst is an intermediate level position responsible for participation in the establishment and implementation of new or revised application systems and programs in coordination with the Technology team. The overall objective of this role is to contribute to applications systems analysis and programming activities.
**Responsibilities:**
+ Utilize knowledge of applications development procedures and concepts, and basic knowledge of other technical areas to identify and define necessary system enhancements
+ Identify and analyze issues, make recommendations, and implement solutions
+ Utilize knowledge of business processes, system processes, and industry standards to solve complex issues
+ Analyze information and make evaluative judgements to recommend solutions and improvements
+ Conduct testing and debugging, utilize script tools, and write basic code for design specifications
+ Assess applicability of similar experiences and evaluate options under circumstances not covered by procedures
+ Develop working knowledge of Citi's information systems, procedures, standards, client server application development, network operations, database administration, systems administration, data center operations, and PC-based applications
+ Appropriately assess risk when business decisions are made, demonstrating particular consideration for the firm's reputation and safeguarding Citigroup, its clients and assets, by driving compliance with applicable laws, rules and regulations, adhering to Policy, applying sound ethical judgment regarding personal behavior, conduct and business practices, and escalating, managing and reporting control issues with transparency.
**Qualifications:**
+ 3 to 5 years of relevant experience
+ Experience in programming/debugging used in business applications
+ Working knowledge of industry practice and standards
+ Comprehensive knowledge of specific business area for application development
+ Working knowledge of program languages
+ Consistently demonstrates clear and concise written and verbal communication
**Education:**
+ Bachelor's degree/University degree or equivalent experience
This job description provides a high-level review of the types of work performed. Other job-related duties may be assigned as required.
Additional Job Description
We are looking for a Big Data Engineer that will work on the collecting, storing, processing, and analyzing of huge sets of data. The primary focus will be on choosing optimal solutions to use for these purposes, then maintaining, implementing, and monitoring them. You will also be responsible for integrating them with the architecture used across the company.
Responsibilities
- Selecting and integrating any Big Data tools and frameworks required to provide requested capabilities
- Implementing data wrangling, scarping, cleaning using both Java or Python
Strong experience on data structure.
Skills and Qualifications
- Proficient understanding of distributed computing principles
- Proficient in Java or Python and some part of machine learning
- Proficiency with Hadoop v2, MapReduce, HDFS, Pyspark, Spark
- Experience with building stream-processing systems, using solutions such as Storm or Spark-Streaming
- Good knowledge of Big Data querying tools, such as Pig, Hive, and Impala
- Experience with Spark
- Experience with integration of data from multiple data sources
- Experience with NoSQL databases, such as HBase, Cassandra, MongoDB
- Knowledge of various ETL techniques and frameworks, such as Flume
- Experience with various messaging systems, such as Kafka or RabbitMQ
- Experience with Big Data ML toolkits, such as Mahout, SparkML, or H2O
- Good understanding of Lambda Architecture, along with its advantages and drawbacks
- Experience with Cloudera/MapR/Hortonworks
This job description provides a high-level review of the types of work performed. Other job-related duties may be assigned as required.
---
**Job Family Group:**
Technology
---
**Job Family:**
Applications Development
---
**Time Type:**
Full time
---
**Most Relevant Skills**
Please see the requirements listed above.
---
**Other Relevant Skills**
For complementary skills, please see above and/or contact the recruiter.
---
_Citi is an equal opportunity employer, and qualified candidates will receive consideration without regard to their race, color, religion, sex, sexual orientation, gender identity, national origin, disability, status as a protected veteran, or any other characteristic protected by law._
_If you are a person with a disability and need a reasonable accommodation to use our search tools and/or apply for a career opportunity review_ _Accessibility at Citi ( _._
_View Citi's_ _EEO Policy Statement ( _and the_ _Know Your Rights ( _poster._
Citi is an equal opportunity and affirmative action employer.
Minority/Female/Veteran/Individuals with Disabilities/Sexual Orientation/Gender Identity.
This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

Bengaluru, Karnataka People Prime Worldwide

Posted today

Job Viewed

Tap Again To Close

Job Description

About the Company

Our client is a trusted global innovator of IT and business services, present in 50+ countries. They specialize in digital & IT modernization, consulting, managed services, and industry-specific solutions. With a commitment to long-term success, they empower clients and society to move confidently into the digital future.


Title: Senior Data Engineer

Location: Hyderabad,Chennai

Experience: 6 - 12 years

Employment Type: Permanent

Notice Period: Immediate Joiners


Key Responsibilities:

  • Big Data Processing:
  • Design, develop, and maintain scalable ETL/ELT pipelines using PySpark, Scala, or Java.
  • Implement data transformation workflows to ensure clean and reliable datasets.
  • Cloud Integration:
  • Work extensively with AWS services such as EMR, S3, Glue, Lambda, Redshift, DynamoDB, and RDS to build and manage data solutions.
  • Data Modeling & Optimization:
  • Develop and optimize data models in Hadoop and Spark environments.
  • Ensure performance tuning and cost optimization of cloud-based data platforms.
  • Data Ingestion & Integration:
  • Ingest data from various sources including APIs, RDBMS, and NoSQL databases.
  • Implement data pipelines for batch and real-time processing.
  • Collaboration & Documentation:
  • Collaborate with cross-functional teams to understand data requirements and deliver solutions.
  • Document data models, transformation logic, and technical specifications.
  • Automation & Monitoring:
  • Automate data workflows and ensure data pipelines are reliable, scalable, and error-free.
  • Implement data monitoring, logging, and alerting to ensure pipeline reliability.

Required Skills & Qualifications:

  • Technical Expertise:
  • 6–12 years of experience in Big Data engineering with strong hands-on experience in AWS, PySpark, Scala, or Java.
  • Proficiency in SQL and experience with Hadoop ecosystem components like Hive, HDFS, and Kafka.
  • Cloud Services:
  • Extensive experience with AWS services such as EMR, S3, Glue, Lambda, Redshift, DynamoDB, and RDS.
  • Data Engineering Tools:
  • Familiarity with data orchestration tools like Apache Airflow or Oozie.
  • Experience with containerization technologies like Docker and Kubernetes is a plus.
  • Soft Skills:
  • Strong analytical and problem-solving skills.
  • Excellent communication and collaboration abilities.
  • Ability to work independently and in a team-oriented environment.

Preferred Skills (Good to Have):

  • Certifications: AWS Certified Solutions Architect, AWS Certified Big Data – Specialty, or Cloudera Certified Associate (CCA).
  • Experience with CI/CD pipelines using tools like Jenkins or GitLab.
  • Knowledge of data warehousing concepts and experience with platforms like Redshift or Snowflake.
  • Familiarity with streaming technologies such as Apache Kafka or Kinesis.
This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

Bengaluru, Karnataka Tata Consultancy Services

Posted today

Job Viewed

Tap Again To Close

Job Description

Job Title :- Data Engineer - Pyspark

Experience: 5 to 8 Years

Location: Pune/Hyderabad


Job Description


Required Skills:

5+ years of experience in Big data and pyspark

Must-Have

Good work experience on Big Data Platforms like Hadoop, Spark, Scala, Hive, Impala, SQL

Good-to-Have

Good Spark, Pyspark,Big Data experience

Spark UI/Optimization/debugging techniques

Good python scripting skills

Intermediate SQL exposure – Subquery, Joins, CTE’s

Database technologies

AWS EMR, S3, IAM, Lambda, SNS, SQS


Good work experience on Big Data Platforms like Hadoop, Spark, Scala, Hive, Impala, SQL

Experience working on Data Engineering projects

Good Understanding of SQL

Good understanding of Unix and HDFS commands

Communication

Experience working on Data Analytics and Pyspark



Thanks & Regards ,


Manasi Prava Natha

This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

Alef Education

Posted today

Job Viewed

Tap Again To Close

Job Description

Who we are

Alef Education began with a bold idea: that every learner deserves a personalised and meaningful education experience. What started in 2016 as a small pilot programme in Abu Dhabi has evolved into one of the world’s most dynamic EdTech companies—reshaping how millions of students engage with learning across the globe.


Today, Alef is proudly headquartered in the UAE, working hand-in-hand with ministries of education, schools, and teachers to bring smart, data-powered platforms into classrooms in over 14,000 schools.


Supporting over 1.1 million students and 50,000 teachers across the UAE, Indonesia & Morocco our AI-driven platforms generate 16+ million data points every day, helping drive smarter learning decisions. Whether it’s improving national exam results, boosting classroom engagement, or supporting educators with world-class tools, Alef is committed to impact at scale.


In 2024, Alef made history as the first EdTech company to list on the Abu Dhabi Securities Exchange (ADX), cementing our role as a regional innovator with global reach.


About The Role

As an ALEF Big Data Engineer you will have a strong understanding of big data technologies with an exceptional ability to code. You will provide technical leadership, working closely with the wider team to ensure high quality code is delivered in line with the project goals and delivery cycles. You will work closely with other teams to deliver rapid prototypes as well as production code for which you will ensure high accessibility standards are upheld. We expect familiarity with modern frameworks and languages, as well as working practices such as Clean Code, TDD, BDD, continuous integration, continuous delivery, and DevOps.


Key Responsibilities

Defining and developing services and solutions

  • Define, design, and develop services and solutions around large data ingestion, storage, and management such as withRDBMS, No SQL DBs, Log Files, Events.
  • Define, design, and run robust data pipelines/batch jobs in a production environment.
  • Architecting highly scalable, highly concurrent, and low latency systems


Maintain, support, and enhance current systems.

  • Contribute to paying down technical debt and use development approaches that minimize the growth of new technical debt.
  • Contribute feedback to improve the quality, readability, and testability of the code base within your team.
  • Mentor and train other developers in a non-line management capacity.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.


Collaborating with Internal and external stakeholders


  • Participating in sprint planning to work with developers and project teams to ensure projects are deployable and monitorable from the outside.
  • Work with third-party and other internal providers to support a variety of integrations.
  • As part of the team, you may be expected to participate in some of the 2nd line in-house support and Out-of-Hours support rotas.
  • Proactively advise on best practices.


To Be The Right Fit, You'll Need

  • Degree in Computer Science, Software Engineering or related preferred
  • Minimum of 5 years experience in a Big Data
  • Follow Clean Code/Solid principles
  • Adhere and use TDD/BDD.
  • Outstanding ability to develop efficient, readable, highly optimized/maintainable and clear code.
  • Highly Proficient in either Functional Java or Scala, Python
  • Knowledge of Azure Big Data/Analytics services – ADLS (Azure Data Lake Storage), HDInsight, Azure Data Factory, Azure Synapse Analytics, Azure Fabric, Azure Event Hubs, Azure Stream Analytics, Azure Databricks
  • Experience of Storing Data in systems such as Hadoop HDFS, ADLS, Event Hubs
  • Experience of designing, setting up and running big data tech stacks such as Hadoop, Azure Databricks, Spark and distributed datastores such as Cassandra, DocumentDBs, MongoDB, Event Hubs
  • In-depth knowledge of Hadoop technology ecosystem – HDFS, Spark, Hive, HBase, Event Hubs, Flume, Sqoop, Oozie, SPARK, Avro, Parquet
  • Experience debugging a complex multi-server service.
  • In depth knowledge and experience in IaaS/PaaS solutions (eg AWS Infrastructure hosting and managed services)
  • Familiarity with network protocols - TCP/IP, HTTP, SSL, etc.
  • Knowledge of relational and non-relational database systems
  • Understanding continuous integration and delivery.
  • Mocking (any of the following Mockito, ScalaTest Spock, Jasmine, Mocha).
  • IDE Intellij or Eclipse.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.
  • An ability to communicate technical concepts to a non-technical audience.
  • Working knowledge of unix-like operating systems such as Linux and/or Mac OS X.
  • Knowledge of the git version control system.
  • Ability to quickly research and learn new programming tools and techniques.
This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

Mumbai, Maharashtra HirePower Staffing Solution

Posted today

Job Viewed

Tap Again To Close

Job Description

Position Overview:

We are seeking a skilled Big Data Developer to join our growing delivery team, with a dual focus on hands-on project support and mentoring junior engineers. This role is ideal for a developer who not only thrives in a technical, fast-paced environment but is also passionate about coaching and developing the next generation of talent.

You will work on live client projects, provide technical support, contribute to solution delivery, and serve as a go-to technical mentor for less experienced team members.


Key Responsibilities:

  • Perform hands-on Big Data development work, including coding, testing, troubleshooting, and deploying solutions.
  • Support ongoing client projects, addressing technical challenges and ensuring smooth delivery.
  • Collaborate with junior engineers to guide them on coding standards, best practices, debugging, and project execution.
  • Review code and provide feedback to junior engineers to maintain high quality and scalable solutions.
  • Assist in designing and implementing solutions using Hadoop, Spark, Hive, HDFS, and Kafka.
  • Lead by example in object-oriented development, particularly using Scala and Java.
  • Translate complex requirements into clear, actionable technical tasks for the team.
  • Contribute to the development of ETL processes for integrating data from various sources.
  • Document technical approaches, best practices, and workflows for knowledge sharing within the team.

Required Skills and Qualifications:

  • 8+ years of professional experience in Big Data development and engineering.
  • Strong hands-on expertise with Hadoop, Hive, HDFS, Apache Spark, and Kafka.
  • Solid object-oriented development experience with Scala and Java.
  • Strong SQL skills with experience working with large data sets.
  • Practical experience designing, installing, configuring, and supporting Big Data clusters.
  • Deep understanding of ETL processes and data integration strategies.
  • Proven experience mentoring or supporting junior engineers in a team setting.
  • Strong problem-solving, troubleshooting, and analytical skills.
  • Excellent communication and interpersonal skills.


Preferred Qualifications:

  • Professional certifications in Big Data technologies (Cloudera, Databricks, AWS Big Data Specialty, etc.).
  • Experience with cloud Big Data platforms (AWS EMR, Azure HDInsight, or GCP Dataproc).
  • Exposure to Agile or DevOps practices in Big Data project environments.


What We Offer:

Opportunity to work on challenging, high-impact Big Data projects.

Leadership role in shaping and mentoring the next generation of engineers.

Supportive and collaborative team culture.

Flexible working environment

Competitive compensation and professional growth opportunities.

This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

Pune, Maharashtra Coforge

Posted today

Job Viewed

Tap Again To Close

Job Description

Coforge Ltd is Hiring for Big Data Engineer – AWS, Spark & Scala.


Must Have Skills:- AWS, Spark & Scala.

Experience Required: 3 to 6 Years

Job Locations: Pune, Hyderabad, Greater Noida Only.

Send your CV to:

For queries, contact via WhatsApp:



Key Responsibilities:-


• Design, develop, and optimize Big Data architectures leveraging AWS services for large-scale, complex data processing.


• Build and maintain data pipelines using Spark (Scala) for both structured and unstructured datasets.


• Architect and operationalize data engineering and analytics platforms (AWS preferred; Hortonworks, Cloudera, or MapR experience a plus).


• Implement and manage AWS services including EMR, Glue, Kinesis, DynamoDB, Athena, CloudFormation, API Gateway, and S3.


• Work on real-time streaming solutions using Kafka and AWS Kinesis.


• Support ML model operationalization on AWS (deployment, scheduling, and monitoring).


• Analyze source system data and data flows to ensure high-quality, reliable data delivery for business needs.


• Write highly efficient SQL queries and support data warehouse initiatives using Apache NiFi, Airflow, and Kylo.


• Collaborate with cross-functional teams to provide technical leadership, mentor team members, and strengthen the data engineering capability.


• Troubleshoot and resolve complex technical issues, ensuring scalability, performance, and security of data solutions.



Mandatory Skills & Qualifications:-


• Solid hands-on experience in Big Data Technologies (AWS, Scala, Hadoop, and Spark Mandatory)

• Proven expertise in Spark with Scala


• Hands-on experience with: AWS services (EMR, Glue, Lambda, S3, CloudFormation, API Gateway, Athena, Lake Formation)

This advertiser has chosen not to accept applicants from your region.
Be The First To Know

About the latest Big data Jobs in India !

Big Data Engineer

Alappuzha, Kerala Alef Education

Posted today

Job Viewed

Tap Again To Close

Job Description

Who we are

Alef Education began with a bold idea: that every learner deserves a personalised and meaningful education experience. What started in 2016 as a small pilot programme in Abu Dhabi has evolved into one of the world's most dynamic EdTech companies—reshaping how millions of students engage with learning across the globe.

Today, Alef is proudly headquartered in the UAE, working hand-in-hand with ministries of education, schools, and teachers to bring smart, data-powered platforms into classrooms in over 14,000 schools.

Supporting over 1.1 million students and 50,000 teachers across the UAE, Indonesia & Morocco our AI-driven platforms generate 16+ million data points every day, helping drive smarter learning decisions. Whether it's improving national exam results, boosting classroom engagement, or supporting educators with world-class tools, Alef is committed to impact at scale.

In 2024, Alef made history as the first EdTech company to list on the Abu Dhabi Securities Exchange (ADX), cementing our role as a regional innovator with global reach.

About The Role

As an ALEF Big Data Engineer you will have a strong understanding of big data technologies with an exceptional ability to code. You will provide technical leadership, working closely with the wider team to ensure high quality code is delivered in line with the project goals and delivery cycles. You will work closely with other teams to deliver rapid prototypes as well as production code for which you will ensure high accessibility standards are upheld. We expect familiarity with modern frameworks and languages, as well as working practices such as Clean Code, TDD, BDD, continuous integration, continuous delivery, and DevOps.

Key Responsibilities

Defining and developing services and solutions

  • Define, design, and develop services and solutions around large data ingestion, storage, and management such as withRDBMS, No SQL DBs, Log Files, Events.
  • Define, design, and run robust data pipelines/batch jobs in a production environment.
  • Architecting highly scalable, highly concurrent, and low latency systems

Maintain, support, and enhance current systems.

  • Contribute to paying down technical debt and use development approaches that minimize the growth of new technical debt.
  • Contribute feedback to improve the quality, readability, and testability of the code base within your team.
  • Mentor and train other developers in a non-line management capacity.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.

Collaborating with Internal and external stakeholders

  • Participating in sprint planning to work with developers and project teams to ensure projects are deployable and monitorable from the outside.
  • Work with third-party and other internal providers to support a variety of integrations.
  • As part of the team, you may be expected to participate in some of the 2nd line in-house support and Out-of-Hours support rotas.
  • Proactively advise on best practices.

To Be The Right Fit, You'll Need

  • Degree in Computer Science, Software Engineering or related preferred
  • Minimum of 5 years experience in a Big Data
  • Follow Clean Code/Solid principles
  • Adhere and use TDD/BDD.
  • Outstanding ability to develop efficient, readable, highly optimized/maintainable and clear code.
  • Highly Proficient in either Functional Java or Scala, Python
  • Knowledge of Azure Big Data/Analytics services – ADLS (Azure Data Lake Storage), HDInsight, Azure Data Factory, Azure Synapse Analytics, Azure Fabric, Azure Event Hubs, Azure Stream Analytics, Azure Databricks
  • Experience of Storing Data in systems such as Hadoop HDFS, ADLS, Event Hubs
  • Experience of designing, setting up and running big data tech stacks such as Hadoop, Azure Databricks, Spark and distributed datastores such as Cassandra, DocumentDBs, MongoDB, Event Hubs
  • In-depth knowledge of Hadoop technology ecosystem – HDFS, Spark, Hive, HBase, Event Hubs, Flume, Sqoop, Oozie, SPARK, Avro, Parquet
  • Experience debugging a complex multi-server service.
  • In depth knowledge and experience in IaaS/PaaS solutions (eg AWS Infrastructure hosting and managed services)
  • Familiarity with network protocols - TCP/IP, HTTP, SSL, etc.
  • Knowledge of relational and non-relational database systems
  • Understanding continuous integration and delivery.
  • Mocking (any of the following Mockito, ScalaTest Spock, Jasmine, Mocha).
  • IDE Intellij or Eclipse.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.
  • An ability to communicate technical concepts to a non-technical audience.
  • Working knowledge of unix-like operating systems such as Linux and/or Mac OS X.
  • Knowledge of the git version control system.
  • Ability to quickly research and learn new programming tools and techniques.
This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

Bikaner, Rajasthan Alef Education

Posted today

Job Viewed

Tap Again To Close

Job Description

Who we are

Alef Education began with a bold idea: that every learner deserves a personalised and meaningful education experience. What started in 2016 as a small pilot programme in Abu Dhabi has evolved into one of the world's most dynamic EdTech companies—reshaping how millions of students engage with learning across the globe.

Today, Alef is proudly headquartered in the UAE, working hand-in-hand with ministries of education, schools, and teachers to bring smart, data-powered platforms into classrooms in over 14,000 schools.

Supporting over 1.1 million students and 50,000 teachers across the UAE, Indonesia & Morocco our AI-driven platforms generate 16+ million data points every day, helping drive smarter learning decisions. Whether it's improving national exam results, boosting classroom engagement, or supporting educators with world-class tools, Alef is committed to impact at scale.

In 2024, Alef made history as the first EdTech company to list on the Abu Dhabi Securities Exchange (ADX), cementing our role as a regional innovator with global reach.

About The Role

As an ALEF Big Data Engineer you will have a strong understanding of big data technologies with an exceptional ability to code. You will provide technical leadership, working closely with the wider team to ensure high quality code is delivered in line with the project goals and delivery cycles. You will work closely with other teams to deliver rapid prototypes as well as production code for which you will ensure high accessibility standards are upheld. We expect familiarity with modern frameworks and languages, as well as working practices such as Clean Code, TDD, BDD, continuous integration, continuous delivery, and DevOps.

Key Responsibilities

Defining and developing services and solutions

  • Define, design, and develop services and solutions around large data ingestion, storage, and management such as withRDBMS, No SQL DBs, Log Files, Events.
  • Define, design, and run robust data pipelines/batch jobs in a production environment.
  • Architecting highly scalable, highly concurrent, and low latency systems

Maintain, support, and enhance current systems.

  • Contribute to paying down technical debt and use development approaches that minimize the growth of new technical debt.
  • Contribute feedback to improve the quality, readability, and testability of the code base within your team.
  • Mentor and train other developers in a non-line management capacity.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.

Collaborating with Internal and external stakeholders

  • Participating in sprint planning to work with developers and project teams to ensure projects are deployable and monitorable from the outside.
  • Work with third-party and other internal providers to support a variety of integrations.
  • As part of the team, you may be expected to participate in some of the 2nd line in-house support and Out-of-Hours support rotas.
  • Proactively advise on best practices.

To Be The Right Fit, You'll Need

  • Degree in Computer Science, Software Engineering or related preferred
  • Minimum of 5 years experience in a Big Data
  • Follow Clean Code/Solid principles
  • Adhere and use TDD/BDD.
  • Outstanding ability to develop efficient, readable, highly optimized/maintainable and clear code.
  • Highly Proficient in either Functional Java or Scala, Python
  • Knowledge of Azure Big Data/Analytics services – ADLS (Azure Data Lake Storage), HDInsight, Azure Data Factory, Azure Synapse Analytics, Azure Fabric, Azure Event Hubs, Azure Stream Analytics, Azure Databricks
  • Experience of Storing Data in systems such as Hadoop HDFS, ADLS, Event Hubs
  • Experience of designing, setting up and running big data tech stacks such as Hadoop, Azure Databricks, Spark and distributed datastores such as Cassandra, DocumentDBs, MongoDB, Event Hubs
  • In-depth knowledge of Hadoop technology ecosystem – HDFS, Spark, Hive, HBase, Event Hubs, Flume, Sqoop, Oozie, SPARK, Avro, Parquet
  • Experience debugging a complex multi-server service.
  • In depth knowledge and experience in IaaS/PaaS solutions (eg AWS Infrastructure hosting and managed services)
  • Familiarity with network protocols - TCP/IP, HTTP, SSL, etc.
  • Knowledge of relational and non-relational database systems
  • Understanding continuous integration and delivery.
  • Mocking (any of the following Mockito, ScalaTest Spock, Jasmine, Mocha).
  • IDE Intellij or Eclipse.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.
  • An ability to communicate technical concepts to a non-technical audience.
  • Working knowledge of unix-like operating systems such as Linux and/or Mac OS X.
  • Knowledge of the git version control system.
  • Ability to quickly research and learn new programming tools and techniques.
This advertiser has chosen not to accept applicants from your region.

Big Data Engineer

Kollam, Kerala Alef Education

Posted today

Job Viewed

Tap Again To Close

Job Description

Who we are

Alef Education began with a bold idea: that every learner deserves a personalised and meaningful education experience. What started in 2016 as a small pilot programme in Abu Dhabi has evolved into one of the world's most dynamic EdTech companies—reshaping how millions of students engage with learning across the globe.

Today, Alef is proudly headquartered in the UAE, working hand-in-hand with ministries of education, schools, and teachers to bring smart, data-powered platforms into classrooms in over 14,000 schools.

Supporting over 1.1 million students and 50,000 teachers across the UAE, Indonesia & Morocco our AI-driven platforms generate 16+ million data points every day, helping drive smarter learning decisions. Whether it's improving national exam results, boosting classroom engagement, or supporting educators with world-class tools, Alef is committed to impact at scale.

In 2024, Alef made history as the first EdTech company to list on the Abu Dhabi Securities Exchange (ADX), cementing our role as a regional innovator with global reach.

About The Role

As an ALEF Big Data Engineer you will have a strong understanding of big data technologies with an exceptional ability to code. You will provide technical leadership, working closely with the wider team to ensure high quality code is delivered in line with the project goals and delivery cycles. You will work closely with other teams to deliver rapid prototypes as well as production code for which you will ensure high accessibility standards are upheld. We expect familiarity with modern frameworks and languages, as well as working practices such as Clean Code, TDD, BDD, continuous integration, continuous delivery, and DevOps.

Key Responsibilities

Defining and developing services and solutions

  • Define, design, and develop services and solutions around large data ingestion, storage, and management such as withRDBMS, No SQL DBs, Log Files, Events.
  • Define, design, and run robust data pipelines/batch jobs in a production environment.
  • Architecting highly scalable, highly concurrent, and low latency systems

Maintain, support, and enhance current systems.

  • Contribute to paying down technical debt and use development approaches that minimize the growth of new technical debt.
  • Contribute feedback to improve the quality, readability, and testability of the code base within your team.
  • Mentor and train other developers in a non-line management capacity.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.

Collaborating with Internal and external stakeholders

  • Participating in sprint planning to work with developers and project teams to ensure projects are deployable and monitorable from the outside.
  • Work with third-party and other internal providers to support a variety of integrations.
  • As part of the team, you may be expected to participate in some of the 2nd line in-house support and Out-of-Hours support rotas.
  • Proactively advise on best practices.

To Be The Right Fit, You'll Need

  • Degree in Computer Science, Software Engineering or related preferred
  • Minimum of 5 years experience in a Big Data
  • Follow Clean Code/Solid principles
  • Adhere and use TDD/BDD.
  • Outstanding ability to develop efficient, readable, highly optimized/maintainable and clear code.
  • Highly Proficient in either Functional Java or Scala, Python
  • Knowledge of Azure Big Data/Analytics services – ADLS (Azure Data Lake Storage), HDInsight, Azure Data Factory, Azure Synapse Analytics, Azure Fabric, Azure Event Hubs, Azure Stream Analytics, Azure Databricks
  • Experience of Storing Data in systems such as Hadoop HDFS, ADLS, Event Hubs
  • Experience of designing, setting up and running big data tech stacks such as Hadoop, Azure Databricks, Spark and distributed datastores such as Cassandra, DocumentDBs, MongoDB, Event Hubs
  • In-depth knowledge of Hadoop technology ecosystem – HDFS, Spark, Hive, HBase, Event Hubs, Flume, Sqoop, Oozie, SPARK, Avro, Parquet
  • Experience debugging a complex multi-server service.
  • In depth knowledge and experience in IaaS/PaaS solutions (eg AWS Infrastructure hosting and managed services)
  • Familiarity with network protocols - TCP/IP, HTTP, SSL, etc.
  • Knowledge of relational and non-relational database systems
  • Understanding continuous integration and delivery.
  • Mocking (any of the following Mockito, ScalaTest Spock, Jasmine, Mocha).
  • IDE Intellij or Eclipse.
  • Build tools (One of SBT, Gradle, Maven).
  • Ensure all software built is robust and scalable.
  • An ability to communicate technical concepts to a non-technical audience.
  • Working knowledge of unix-like operating systems such as Linux and/or Mac OS X.
  • Knowledge of the git version control system.
  • Ability to quickly research and learn new programming tools and techniques.
This advertiser has chosen not to accept applicants from your region.
 

Nearby Locations

Other Jobs Near Me

Industry

  1. request_quote Accounting
  2. work Administrative
  3. eco Agriculture Forestry
  4. smart_toy AI & Emerging Technologies
  5. school Apprenticeships & Trainee
  6. apartment Architecture
  7. palette Arts & Entertainment
  8. directions_car Automotive
  9. flight_takeoff Aviation
  10. account_balance Banking & Finance
  11. local_florist Beauty & Wellness
  12. restaurant Catering
  13. volunteer_activism Charity & Voluntary
  14. science Chemical Engineering
  15. child_friendly Childcare
  16. foundation Civil Engineering
  17. clean_hands Cleaning & Sanitation
  18. diversity_3 Community & Social Care
  19. construction Construction
  20. brush Creative & Digital
  21. currency_bitcoin Crypto & Blockchain
  22. support_agent Customer Service & Helpdesk
  23. medical_services Dental
  24. medical_services Driving & Transport
  25. medical_services E Commerce & Social Media
  26. school Education & Teaching
  27. electrical_services Electrical Engineering
  28. bolt Energy
  29. local_mall Fmcg
  30. gavel Government & Non Profit
  31. emoji_events Graduate
  32. health_and_safety Healthcare
  33. beach_access Hospitality & Tourism
  34. groups Human Resources
  35. precision_manufacturing Industrial Engineering
  36. security Information Security
  37. handyman Installation & Maintenance
  38. policy Insurance
  39. code IT & Software
  40. gavel Legal
  41. sports_soccer Leisure & Sports
  42. inventory_2 Logistics & Warehousing
  43. supervisor_account Management
  44. supervisor_account Management Consultancy
  45. supervisor_account Manufacturing & Production
  46. campaign Marketing
  47. build Mechanical Engineering
  48. perm_media Media & PR
  49. local_hospital Medical
  50. local_hospital Military & Public Safety
  51. local_hospital Mining
  52. medical_services Nursing
  53. local_gas_station Oil & Gas
  54. biotech Pharmaceutical
  55. checklist_rtl Project Management
  56. shopping_bag Purchasing
  57. home_work Real Estate
  58. person_search Recruitment Consultancy
  59. store Retail
  60. point_of_sale Sales
  61. science Scientific Research & Development
  62. wifi Telecoms
  63. psychology Therapy
  64. pets Veterinary
View All Big Data Jobs