181 Cloud Data Engineer jobs in India
Cloud Data Engineer
Posted 5 days ago
Job Viewed
Job Description
About Position:
We are seeking a highly analytical and detail-oriented Cloud Data Engineer
- Role: Cloud Data Engineer
- Location: All Persistent Locations
- Experience: 5+ Years
- Job Type: Full Time Employment
What You'll Do:
- Minimum of 5+ years of experience as a Cloud Data Engineer.
- Hands-on experience with Azure Cloud data tools (ADF, SHIR, Logic Apps, ADLS Gen2, Blob Storage) and Snowflake.
- Strong experience in ETL development using on-premises databases
- ETL technologiesExperience with Python or other scripting languages for data processing.Proficiency in DevOps and CI/CD practices using GitActions.
- Experience with Agile methodologies.
- Excellent problem-solving skills and ability to work independently.Strong communication and collaboration skills.
- Strong analytical skills and attention to detail.Ability to adapt to new technologies and learn quickly.
Expertise You'll Bring:
- Analyze and organize raw data from diverse sources, including healthcare datasets.
- Design and implement ETL pipelines using Azure Data Factory.Develop and optimize SQL queries, stored procedures, and data models in Snowflake.
- Parse and flatten semi-structured data (XML, JSON) using Snowflake and SQL.Build scalable data systems and pipelines to support analytics and reporting.Migrate on-premise data workloads to cloud platforms, preferably Snowflake.
- Work with Databricks using PySpark, SparkSQL, and Python for data transformation.Ensure data quality through profiling, validation, and adherence to governance standards.
- Collaborate in Agile development environments and contribute to continuous improvement.
- Identify opportunities to enhance data architecture and storage efficiency.
Benefits:
- Competitive salary and benefits package
- Culture focused on talent development with quarterly growth opportunities and company-sponsored higher education and certifications
- Opportunity to work with cutting-edge technologies
- Employee engagement initiatives such as project parties, flexible work hours, and Long Service awards
- Annual health check-ups
- Insurance coverage: group term life, personal accident, and Mediclaim hospitalization for self, spouse, two children, and parents
Values-Driven, People-Centric & Inclusive Work Environment:
Persistent Ltd. is dedicated to fostering diversity and inclusion in the workplace. We invite applications from all qualified individuals, including those with disabilities, and regardless of gender or gender preference. We welcome diverse candidates from all backgrounds.
- We support hybrid work and flexible hours to fit diverse lifestyles.
- Our office is accessibility-friendly, with ergonomic setups and assistive technologies to support employees with physical disabilities.
- If you are a person with disabilities and have specific requirements, please inform us during the application process or at any time during your employment
Let’s unleash your full potential at Persistent - persistent.com/careers
“Persistent is an Equal Opportunity Employer and prohibits discrimination and harassment of any kind.”
Cloud Data Engineer
Posted 5 days ago
Job Viewed
Job Description
About Position:
We are seeking cloud data engineers with hands on experience in azure data bricks, data factory, SQL, etc.,
- Role: Cloud Data Engineer
- Job Location: All Persistent Locations
- Experience: 5+ Years
- Job Type: Full Time Employment
What You'll Do:
- Collaborate with product owners, systems analysts and software engineers to deliver against an agile roadmap
- Work with architects and other software engineers to evaluate functional and non-functional requirements and deliver creative, high-quality solutions
- Design, develop, test and maintain data intensive applications including Java, Azure or AWS, Databricks, Spark, relational and non-relational data storage, APIs
- Experience in Data as a Service and Container as a Service models.
- Experience in incident ticket tracking tools and processes like ServiceNow
- Collaborate on quality strategies that ensure our data platform is correct, resilient, scalable and secure
- Support applications throughout the SDLC from design to Production Deployment
- Participate in and provide input for system analysis, design walkthroughs and code reviews
- Participating in defect review and triage
- Adhere to design/coding standards and constantly improve the way we build and deliver software
Expertise You'll Bring:
- Experience in Azure technologies
- Experience in Databricks
- Experience in Spark
- Experience in Snowflake
- Working knowledge of RESTful APIs
- 2+ years of relational and non-relational database delivery
- 2+ years of experience in Agile Delivery
- 1+ years in Dev Ops Automation tools (Oozie, Python, Jenkins etc.)
- Experience in working with DevOps teams in an Agile Delivery/onshore-offshore model (5-10 headcount)
- Bachelor’s degree in engineering or equivalent experience
- Expertise on cloud migration activities.
Benefits:
- Competitive salary and benefits package
- Culture focused on talent development with quarterly growth opportunities and company-sponsored higher education and certifications
- Opportunity to work with cutting-edge technologies
- Employee engagement initiatives such as project parties, flexible work hours, and Long Service awards
- Annual health check-ups
- Insurance coverage: group term life, personal accident, and Mediclaim hospitalization for self, spouse, two children, and parents
Values-Driven, People-Centric & Inclusive Work Environment:
Persistent Ltd. is dedicated to fostering diversity and inclusion in the workplace. We invite applications from all qualified individuals, including those with disabilities, and regardless of gender or gender preference. We welcome diverse candidates from all backgrounds.
- We support hybrid work and flexible hours to fit diverse lifestyles.
- Our office is accessibility-friendly, with ergonomic setups and assistive technologies to support employees with physical disabilities.
- If you are a person with disabilities and have specific requirements, please inform us during the application process or at any time during your employment
Let’s unleash your full potential at Persistent - persistent.com/careers
“Persistent is an Equal Opportunity Employer and prohibits discrimination and harassment of any kind.”
Cloud Data Engineer
Posted 5 days ago
Job Viewed
Job Description
About Lemongrass
Lemongrass is a software-enabled services provider, synonymous with SAP on Cloud, focused on delivering superior, highly automated Managed Services to Enterprise customers. Our customers span multiple verticals and geographies across the Americas, EMEA and APAC. We partner with AWS, SAP, Microsoft and other global technology leaders.
We are seeking an experienced Cloud Data Engineer with a strong background in AWS, Azure, and GCP. The ideal candidate will have extensive experience with cloud-native ETL tools such as AWS DMS, AWS Glue, Kafka, Azure Data Factory, GCP Dataflow, and other ETL tools like Informatica, SAP Data Intelligence, etc. You will be responsible for designing, implementing, and maintaining robust data pipelines and building scalable data lakes. Experience with various data platforms like Redshift, Snowflake, Databricks, Synapse, Snowflake and others is essential. Familiarity with data extraction from SAP or ERP systems is a plus.
Key Responsibilities:
Design and Development:
- Design, develop, and maintain scalable ETL pipelines using cloud-native tools (AWS DMS, AWS Glue, Kafka, Azure Data Factory, GCP Dataflow, etc.).
- Architect and implement data lakes and data warehouses on cloud platforms (AWS, Azure, GCP).
- Develop and optimize data ingestion, transformation, and loading processes using Databricks, Snowflake, Redshift, BigQuery and Azure Synapse.
- Implement ETL processes using tools like Informatica, SAP Data Intelligence, and others.
- Develop and optimize data processing jobs using Spark Scala.
Data Integration and Management:
- Integrate various data sources, including relational databases, APIs, unstructured data, and ERP systems into the data lake.
- Ensure data quality and integrity through rigorous testing and validation.
- Perform data extraction from SAP or ERP systems when necessary.
Performance Optimization:
- Monitor and optimize the performance of data pipelines and ETL processes.
- Implement best practices for data management, including data governance, security, and compliance.
Collaboration and Communication:
- Work closely with data scientists, analysts, and other stakeholders to understand data requirements and deliver solutions.
- Collaborate with cross-functional teams to design and implement data solutions that meet business needs.
Documentation and Maintenance:
- Document technical solutions, processes, and workflows.
- Maintain and troubleshoot existing ETL pipelines and data integrations.
Qualifications
Education:
- Bachelor’s degree in Computer Science, Information Technology, or a related field. Advanced degrees are a plus.
Experience:
- 7+ years of experience as a Data Engineer or in a similar role.
- Proven experience with cloud platforms: AWS, Azure, and GCP.
- Hands-on experience with cloud-native ETL tools such as AWS DMS, AWS Glue, Kafka, Azure Data Factory, GCP Dataflow, etc.
- Experience with other ETL tools like Informatica, SAP Data Intelligence, etc.
- Experience in building and managing data lakes and data warehouses.
- Proficiency with data platforms like Redshift, Snowflake, BigQuery, Databricks, and Azure Synapse.
- Experience with data extraction from SAP or ERP systems is a plus.
- Strong experience with Spark and Scala for data processing.
Skills:
- Strong programming skills in Python, Java, or Scala.
- Proficient in SQL and query optimization techniques.
- Familiarity with data modeling, ETL/ELT processes, and data warehousing concepts.
- Knowledge of data governance, security, and compliance best practices.
- Excellent problem-solving and analytical skills.
- Strong communication and collaboration skills.
Preferred Qualifications:
- Experience with other data tools and technologies such as Apache Spark, or Hadoop.
- Certifications in cloud platforms (AWS Certified Data Analytics – Specialty, Google Professional Data Engineer, Microsoft Certified: Azure Data Engineer Associate).
- Experience with CI/CD pipelines and DevOps practices for data engineering
- Selected applicant will be subject to a background investigation, which will be conducted and the results of which will be used in compliance with applicable law.
What we offer in return:
- Remote Working: Lemongrass always has been and always will offer 100% remote work
- Flexibility: Work where and when you like most of the time
- Training: A subscription to A Cloud Guru and generous budget for taking certifications and other resources you’ll find helpful
- State of the art tech: An opportunity to learn and run the latest industry standard tools
- Team: Colleagues who will challenge you giving the chance to learn from them and them from you
Lemongrass Consulting is proud to be an Equal Opportunity and Affirmative Action employer. We do not discriminate on the basis of race, religion, color, national origin, religious creed, gender, sexual orientation, gender identity, gender expression, age, genetic information, status as a protected veteran, status as an individual with a disability, or other applicable legally protected characteristics
Full Stack Cloud Data Engineer- Large Asset Management Firm
Posted 569 days ago
Job Viewed
Job Description
Job Description:
Responsibilities -
• Work with development teams and product managers to ideate software solutions.
• Design client-side and server-side architecture • Build the front-end and back-end of applications through appealing visual design and design patterns.
• Develop and manage well-functioning databases and applications.
• Design, develop, conduct unit testing, and maintain complex Tableau reports for scalability, manageability, extensibility, performance, and re-use.
• Work with team members to create useful reports and dashboards that provide insight, improve/automate processes, or otherwise add value to the team
• Write effective APIs for consumption and getting consumed.
• Test software to ensure responsiveness and efficiency.
• Troubleshoot, debug and upgrade software.
• Create security and data protection settings.
• Build features and applications with a mobile responsive design.
• Write technical documentation.
• Work with data scientists and analysts to improve software.
Requirements and skills –
• Proven experience as a Full Stack Developer or similar role
. • Experience developing desktop and mobile applications.
• Familiarity with common stacks.
• Knowledge of multiple front-end languages and libraries (e.g. HTML/ CSS, JavaScript, XML, jQuery)
• Knowledge of multiple back-end languages (e.g. C#, Java, Python) and JavaScript frameworks (e.g. Angular, React, Node.js)
• Working knowledge of Tableau administrator/architecture.
• A solid understanding of SQL, relational database management systems, data modeling, and normalisation.
• Advanced Tableau development skills required.
• Familiarity with databases (e.g. SQL, NoSQL), web servers (e.g. Apache) and UI/UX design.
• Robust grasp of cloud platforms, architecture patterns, and infrastructure as code (IaC) principles.
Expertise in cloud providers (AWS, Azure, etc.) , networking, security protocols, and familiarity with serverless computing.
• Excellent communication and teamwork skills
• Great attention to detail
• Organizational skills
• An analytical mind
• Degree in Computer Science, Statistics or relevant field
• Strong problem-solving, project management, and communication skills on design scalability and reliable cloud environments.
Domain -
• Financial Services / Banking / Mutual Funds / AMC
Cloud Data Platform Engineer
Posted 5 days ago
Job Viewed
Job Description
About Position:
We are hiring for skills experience with setting up, configuring, and optimizing Databricks for data engineering and analytics workloads. Experience with other cloud platforms and data solutions. Knowledge of data governance and compliance standards.Familiarity with containerization technologies such as Docker and Kubernetes.
- Role: Cloud Data Platform Engineer
- Location: All Persistent Location
- Experience: 5+ Years
- Job Type: Full Time Employment
What You'll Do:
- 4+ years of hands-on experience in setting up and managing Cloud Data Platform infrastructure for Azure Cloud Data Stacks (ADF, SHIR, Logic Apps, ADLS Gen2, Blob Storage).
- Strong expertise in Snowflake configuration and optimization.Hands-on experience with DevOps tools and practices, including GitActions CI/CD and Terraform.
- Solid understanding of cloud architecture and infrastructure as code (IaC) principles.
- Excellent problem-solving skills and ability to work in a fast-paced environment.Strong communication and collaboration skills.
Expertise You'll Bring:
- Design, implement, and manage Cloud Data Platform infrastructure for Azure Data Factory (ADF), Self-hosted Integration Runtime (SHIR), Logic Apps, Azure Data Lake Storage Gen2 (ADLS Gen2), and Blob Storage.Configure and optimize Snowflake data warehouse solutions to meet business requirements.
- Develop and maintain CI/CD pipelines using GitActions to ensure seamless integration and deployment processes.Utilize Terraform for infrastructure as code (IaC) to automate and manage cloud resources.
- Collaborate with cross-functional teams to understand data requirements and deliver scalable, reliable cloud solutions.Monitor and troubleshoot Cloud Data Platform infrastructure to ensure high availability and performance.Implement security best practices to safeguard data and cloud resources.
Benefits:
- Competitive salary and benefits package
- Culture focused on talent development with quarterly growth opportunities and company-sponsored higher education and certifications
- Opportunity to work with cutting-edge technologies
- Employee engagement initiatives such as project parties, flexible work hours, and Long Service awards
- Annual health check-ups
- Insurance coverage: group term life, personal accident, and Mediclaim hospitalization for self, spouse, two children, and parents
Values-Driven, People-Centric & Inclusive Work Environment:
Persistent Ltd. is dedicated to fostering diversity and inclusion in the workplace. We invite applications from all qualified individuals, including those with disabilities, and regardless of gender or gender preference. We welcome diverse candidates from all backgrounds.
- We support hybrid work and flexible hours to fit diverse lifestyles.
- Our office is accessibility-friendly, with ergonomic setups and assistive technologies to support employees with physical disabilities.
- If you are a person with disabilities and have specific requirements, please inform us during the application process or at any time during your employment
Let’s unleash your full potential at Persistent - persistent.com/careers
“Persistent is an Equal Opportunity Employer and prohibits discrimination and harassment of any kind.”
Cloud Data Platform Engineer
Posted 5 days ago
Job Viewed
Job Description
About Position:
We are seeking a highly skilled and experienced Data Engineer – SQL / SSIS to join the Optum Advisory Team. The ideal candidate will bring deep expertise in cloud-based data engineering, healthcare data modeling, and consulting, with a strong background in Azure, Databricks, Azure Data Factory (ADF), SQL, and SSIS / SSRS. This role requires a self-starter who can work independently with minimal guidance and collaborate effectively with cross-functional teams to deliver high-impact data solutions in the healthcare domain
- Role: Cloud Data Platform Engineer
- Location: All Persistent Locations
- Experience: 6+ Years
- Job Type: Full Time Employment
What You'll Do:
- Design, implement, and manage Cloud Data Platform infrastructure for Azure Data Factory (ADF), Self-hosted Integration Runtime (SHIR), Logic Apps, Azure Data Lake Storage Gen2 (ADLS Gen2), and Blob Storage.
- Configure and optimize Snowflake data warehouse solutions to meet business requirements.
- Develop and maintain CI/CD pipelines using GitActions to ensure seamless integration and deployment processes.
- Utilize Terraform for infrastructure as code (IaC) to automate and manage cloud resources.
- Collaborate with cross-functional teams to understand data requirements and deliver scalable, reliable cloud solutions.
- Monitor and troubleshoot Cloud Data Platform infrastructure to ensure high availability and performance.
- Implement security best practices to safeguard data and cloud resources.
Expertise You'll Bring:
- 3+ years of hands-on experience in setting up and managing Cloud Data Platform infrastructure for Azure Cloud Data Stacks (ADF, SHIR, Logic Apps, ADLS Gen2, Blob Storage).
- Strong expertise in Snowflake configuration and optimization.
- Hands-on experience with DevOps tools and practices, including GitActions CI/CD and Terraform.
- Solid understanding of cloud architecture and infrastructure as code (IaC) principles.
- Excellent problem-solving skills and ability to work in a fast-paced environment.
- Strong communication and collaboration skills.
Benefits:
- Competitive salary and benefits package
- Culture focused on talent development with quarterly growth opportunities and company-sponsored higher education and certifications
- Opportunity to work with cutting-edge technologies
- Employee engagement initiatives such as project parties, flexible work hours, and Long Service awards
- Annual health check-ups
- Insurance coverage: group term life, personal accident, and Mediclaim hospitalization for self, spouse, two children, and parents
Values-Driven, People-Centric & Inclusive Work Environment:
Persistent Ltd. is dedicated to fostering diversity and inclusion in the workplace. We invite applications from all qualified individuals, including those with disabilities, and regardless of gender or gender preference. We welcome diverse candidates from all backgrounds.
- We support hybrid work and flexible hours to fit diverse lifestyles.
- Our office is accessibility-friendly, with ergonomic setups and assistive technologies to support employees with physical disabilities.
- If you are a person with disabilities and have specific requirements, please inform us during the application process or at any time during your employment
Let’s unleash your full potential at Persistent - persistent.com/careers
“Persistent is an Equal Opportunity Employer and prohibits discrimination and harassment of any kind.”
Senior Database Infrastructure Engineer- Cassandra, DataStax, Big Data Pipelines
Posted 5 days ago
Job Viewed
Job Description
HEROIC Cybersecurity ( HEROIC.com ) is seeking a Senior Data Infrastructure Engineer with deep expertise in DataStax Enterprise (DSE) and Apache Cassandra to help architect, scale, and maintain the data infrastructure that powers our cybersecurity intelligence platforms.
You will be responsible for designing and managing fully automated, big data pipelines that ingest, process, and serve hundreds of billions of breached and leaked records sourced from the surface, deep, and dark web. You'll work with DSE Cassandra, Solr, and Spark, helping us move toward a 99% automated pipeline for data ingestion, enrichment, deduplication, and indexing — all built for scale, speed, and reliability.
This position is critical in ensuring our systems are fast, reliable, and resilient as we ingest thousands of unique datasets daily from global threat intelligence sources.
What you will do:
Design, deploy, and maintain high-performance Cassandra clusters using DataStax Enterprise (DSE) Architect and optimize automated data pipelines to ingest, clean, enrich, and store billions of records daily Configure and manage DSE Solr and Spark to support search and distributed processing at scale Automate dataset ingestion workflows from unstructured surface, deep, and dark web sources Cluster management, replication strategy, capacity planning, and performance tuning Ensure data integrity, availability, and security across all distributed systems Write and manage ETL processes, scripts, and APIs to support data flow automation Monitor systems for bottlenecks, optimize queries and indexes, and resolve production issues Research and integrate third-party data tools or AI-based enhancements (e.g., smart data parsing, deduplication, ML-based classification) Collaborate with engineering, data science, and product teams to support HEROIC’s AI-powered cybersecurity platform Requirements Minimum 5 years experience with Cassandra / DataStax Enterprise in production environments Hands-on experience with DSE Cassandra, Solr, Apache Spark, CQL, and data modeling at scale Strong understanding of NoSQL architecture, sharding, replication, and high availability Advanced knowledge of Linux/Unix, shell scripting, and automation tools (e.g., Ansible, Terraform) Proficient in at least one programming language: Python, Java, or Scala Experience building large-scale automated data ingestion systems or ETL workflows Solid grasp of AI-enhanced data processing, including smart cleaning, deduplication, and classification Excellent written and spoken English communication skills Prior experience with cybersecurity or dark web data (preferred but not required) Benefits Position Type: Full-time Location: Pune, India (Remote – Work from anywhere) Compensation: Competitive salary depending on experience Benefits: Paid Time Off + Public Holidays Professional Growth: Amazing upward mobility in a rapidly expanding company. Innovative Culture: Fast-paced, innovative, and mission-driven. Be part of a team that leverages AI and cutting-edge technologies.About Us: HEROIC Cybersecurity ( HEROIC.com ) is building the future of cybersecurity. Unlike traditional cybersecurity solutions, HEROIC takes a predictive and proactive approach to intelligently secure our users before an attack or threat occurs. Our work environment is fast-paced, challenging and exciting. At HEROIC, you’ll work with a team of passionate, engaged individuals dedicated to intelligently securing the technology of people all over the world.
Position Keywords: DataStax Enterprise (DSE), Apache Cassandra, Apache Spark, Apache Solr, AWS, Jira, NoSQL, CQL (Cassandra Query Language), Data Modeling, Data Replication, ETL Pipelines, Data Deduplication, Data Lake, Linux/Unix Administration, Bash, Docker, Kubernetes, CI/CD, Python, Java, Distributed Systems, Cluster Management, Performance Tuning, High Availability, Disaster Recovery, AI-based Automation, Artificial Intelligence, Big Data, Dark Web Data
Be The First To Know
About the latest Cloud data engineer Jobs in India !
Senior Cloud Engineer - Data Hosting
Posted 2 days ago
Job Viewed
Job Description
2
About the position:
Chevron invites applications for the role of Senior Cloud Engineer - Data Hosting within our team in India. This position supports Chevron's data hosting environment by delivering modern digital data hosting capabilities in a cost competitive, reliable, and secure manner. This position will provide broad exposure to the application of technology to enable business with many opportunities for growth and professional development for the candidate.
Key Responsibilities:
- Design, implement, and manage scalable and secure data hosting solutions on Azure
- Develop and maintain data architectures, including data models, data warehouses, and data lakes
- Refine data storage and extraction procedures to enhance performance and cost-effectiveness
- Uphold stringent data security measures and ensure adherence to relevant industry standards and regulatory requirements. - Collaborate with data scientists, analysts, and other stakeholders to understand and address their data needs
- Monitor and troubleshoot data hosting environments to ensure high availability and reliability
- Streamline data workflows and operations through the automation capabilities of Azure Data Factory and comparable technologies
- Design, develop, and deploy modular cloud-based systems
- Develop and maintain cloud solutions in accordance with best practices
Required Qualifications:
- Must have bachelor's degree in computer science engineering or related discipline
- A Minimum of 8+ years of experience. 5 - 10 years of experience
- Over 5+ years of experience in data hosting for both on-premises and azure environments
- At least 5 years of expertise in data engineering and management across on-premises and azure environments
- Microsoft AZ900 Certification & AZ104 Certification
- Proficient in utilizing Azure data services, including Azure SQL Database, Azure Data Lake Storage, and Azure Data Factory
- In-depth understanding of cloud infrastructure, encompassing virtual networks, storage solutions, and compute resources within Azure
- Extensive hands-on experience with Azure services such as Azure SQL Database, Azure Blob Storage, Azure Data Lake, and Azure Synapse Analytics
- Well-versed in on-premises storage systems from vendors like NetApp, Dell, and others
- Skilled proficiency in scripting languages like Ansible, PowerShell, Python, and Azure CLI for automation and management tasks
- Comprehensive knowledge of Azure security best practices, including identity and access management, encryption, and compliance standards
Preferred Qualifications:
- Demonstrated proficiency in architecting, deploying, and managing secure and scalable data hosting solutions on the Azure platform
- Extensive experience in developing and maintaining robust data architectures, including data models, data warehouses, and data lakes, utilizing Azure services
- Expertise in optimizing data storage and retrieval processes for superior performance and cost efficiency within Azure environments
- In-depth knowledge of data security protocols and compliance with industry standards and regulations, with a focus on Azure cloud compliance
- Proven ability to collaborate effectively with data scientists, analysts, and other stakeholders to address their data needs using Azure's capabilities
- Strong track record of monitoring and troubleshooting Azure data hosting environments to ensure high availability and system reliability
- Skilled in automating data workflows and processes using Azure Data Factory and other Azure-based automation tools
- Experience in designing, developing, and deploying modular, cloud-based systems, with a particular emphasis on Azure solutions
- Commitment to maintaining cloud solutions in alignment with Azure best practices and continuously integrating Azure's latest updates and features
- Possession of Azure certifications, such as the Azure Data Engineer Associate or Azure Database Administrator Associate, with a preference for candidates holding the Azure Solutions Architect Expert certification or equivalent advanced credentials
Chevron ENGINE supports global operations, supporting business requirements across the world. Accordingly, the work hours for employees will be aligned to support business requirements. The standard work week will be Monday to Friday. Working hours are 8:00am to 5:00pm or 1.30pm to 10.30pm.
Chevron participates in E-Verify in certain locations as required by law.
Chevron Corporation is one of the world's leading integrated energy companies. Through its subsidiaries that conduct business worldwide, the company is involved in virtually every facet of the energy industry. Chevron explores for, produces and transports crude oil and natural gas; refines, markets and distributes transportation fuels and lubricants; manufactures and sells petrochemicals and additives; generates power; and develops and deploys technologies that enhance business value in every aspect of the company's operations. Chevron is based in Houston, Texas. More information about Chevron is available at .
Chevron is an Equal Opportunity / Affirmative Action employer. Qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability or protected veteran status, or other status protected by law or regulation.
Senior Data Engineer (Cloud)
Posted 23 days ago
Job Viewed
Job Description
Lead Data Engineer - Big Data & Cloud
Posted 9 days ago
Job Viewed
Job Description
The Lead Data Engineer will be responsible for building and optimizing scalable data pipelines using big data technologies and cloud platforms. You will work closely with data scientists, analysts, and business stakeholders to understand data requirements and deliver high-quality data solutions. Key responsibilities include designing data warehousing solutions, implementing ETL/ELT processes, and ensuring data quality, integrity, and security. You will evaluate and adopt new technologies to enhance data processing capabilities and improve performance. This role involves mentoring junior engineers, setting technical standards, and fostering a collaborative team environment. Expertise in cloud data services (e.g., AWS, Azure, GCP), distributed computing frameworks (e.g., Spark, Hadoop), and database technologies is essential. The ideal candidate possesses strong architectural skills, excellent problem-solving abilities, and a proven track record of delivering complex data engineering projects on time and within scope. You will play a key role in enabling data-driven decision-making across the business.
Responsibilities:
- Lead the design, development, and implementation of scalable data pipelines and architectures.
- Architect and build data warehouses and data lakes on cloud platforms.
- Develop and optimize ETL/ELT processes for data ingestion and transformation.
- Ensure data quality, consistency, and reliability across all data systems.
- Evaluate and implement new data technologies and tools to improve efficiency.
- Mentor and guide junior data engineers and data analysts.
- Collaborate with data scientists and business stakeholders to define data needs.
- Manage and maintain the performance and security of data infrastructure.
- Develop and enforce data governance policies and best practices.
- Troubleshoot and resolve data-related issues promptly.
- Bachelor's or Master's degree in Computer Science, Engineering, or a related quantitative field.
- Minimum of 7 years of experience in data engineering, with at least 2 years in a lead or supervisory role.
- Proficiency with big data technologies such as Apache Spark, Hadoop, Kafka.
- Extensive experience with cloud data services (AWS Redshift, S3, Glue; Azure Data Factory, Synapse; GCP BigQuery, Dataflow).
- Strong SQL skills and experience with various database systems (relational and NoSQL).
- Proficiency in programming languages like Python or Java.
- Experience with data warehousing concepts and best practices.
- Excellent problem-solving, analytical, and communication skills.
- Ability to lead and mentor a team effectively.
- Experience with data visualization tools is a plus.