2652 Data Engineer jobs in Bengaluru
GCP Data Engineer-Architecture
Posted today
Job Viewed
Job Description
Role:-GCP Data Architect
Exp:-10-12 Yrs
Locations: Hyderabad, Bangalore and Chennai
Primary Skills: GCP Data Engineer, BigQuery, Kafka, Airflow and Architect Exp
Please share your resumes to ,
Job Description:-
What we're looking for.
- You are curious and passionate about Data and truly believe in the high impact it can create for the business.
- People count on you for your expertise in data management in all phases of the software development cycle.
- You enjoy the challenge of solving complex data management problems and challenging priorities in a multifaceted, complex and deadline-oriented environment.
- Building effective working relationships and collaborating with other technical teams across the organization comes naturally to you.
You'll need to have:
- Bachelor's degree or four or more years of work experience.
- Six or more years of relevant work experience.
- Knowledge of Information Systems and their applications to data management processes.
- Experience performing detailed analysis of business problems and technical environments and designing the solution.
- Experience working with
Google Cloud Platform & BigQuery. - Experience working with
Bigdata Technologies & Utilities - Hadoop/ Spark/ Scala/ Kafka/ NiFi. - Experience with relational SQL and NoSQL databases.
- Experience with data pipeline and workflow management & Governance tools.
- Experience with stream-processing systems.
- Experience with object-oriented/object function scripting languages.
- Experience building data solutions for Machine learning and Artificial Intelligence.
- Knowledge of Data Analytics and modeling tools.
Even better if you have any one or more of the following:
- Knowledge of Telecom and Network
- Master's degree in computer science or a related field.
- Contributed to Open-Source Data Warehousing.
- Certifications in any Data Warehousing/Analytical solutioning.
- Certifications in GCP.
- Ability to clearly articulate the pros and cons of various technologies and platforms.
- Experience collaborating with multi-functional teams and managing partner expectations.
- Written and verbal communication skills.
- Ability to work in a fast-paced agile development environment.
Data Engineer- Lead Data Engineer
Posted today
Job Viewed
Job Description
Role Overview
We are seeking an experienced Lead Data Engineer to join our Data Engineering team at Paytm, India's leading digital payments and financial services platform. This is a critical role responsible for designing, building, and maintaining large-scale, real-time data streams that process billions of transactions and user interactions daily. Data accuracy and stream reliability are essential to our operations, as data quality issues can result in financial losses and impact customer
a Lead Data Engineer at Paytm, you will be responsible for building robust data systems that support India's largest digital payments ecosystem. You'll architect and implement reliable, real-time data streaming solutions where precision and data correctness are fundamental requirements. Your work will directly support millions of users across merchant payments, peer-to-peer transfers, bill payments, and financial services, where data accuracy is crucial for maintaining customer confidence and operational excellence.
This role requires expertise in designing fault-tolerant, scalable data architectures that maintain high uptime standards while processing peak transaction loads during festivals and high-traffic events. We place the highest priority on data quality and system reliability, as our customers depend on accurate, timely information for their financial decisions. You'll collaborate with cross-functional teams including data scientists, product managers, and risk engineers to deliver data solutions that enable real-time fraud detection, personalized recommendations, credit scoring, and regulatory compliance reporting.
Key technical challenges include maintaining data consistency across distributed systems with demanding performance requirements, implementing comprehensive data quality frameworks with real-time validation, optimizing query performance on large datasets, and ensuring complete data lineage and governance across multiple business domains. At Paytm, reliable data streams are fundamental to our operations and our commitment to protecting customers' financial security and maintaining India's digital payments
Responsibilities
Data Stream Architecture & DevelopmentDesign and implement reliable, scalable data streams handling high-volume transaction data with strong data integrity controlsBuild real-time processing systems using modern data engineering frameworks (Java/Python stack) with excellent performance characteristicsDevelop robust data ingestion systems from multiple sources with built-in redundancy and monitoring capabilitiesImplement comprehensive data quality frameworks, ensuring the 4 C's: Completeness, Consistency, Conformity, and Correctness - ensuring data reliability that supports sound business decisionsDesign automated data validation, profiling, and quality monitoring systems with proactive alerting capabilitiesInfrastructure & Platform ManagementManage and optimize distributed data processing platforms with high availability requirements to ensure consistent service deliveryDesign data lake and data warehouse architectures with appropriate partitioning and indexing strategies for optimal query performanceImplement CI/CD processes for data engineering workflows with comprehensive testing and reliable deployment proceduresEnsure high availability and disaster recovery for critical data systems to maintain business continuity
Performance & OptimizationMonitor and optimize streaming performance with focus on latency reduction and operational efficiencyImplement efficient data storage strategies including compression, partitioning, and lifecycle management with cost considerationsTroubleshoot and resolve complex data streaming issues in production environments with effective response protocolsConduct proactive capacity planning and performance tuning to support business growth and data volume increases
Collaboration & Leadership Work closely with data scientists, analysts, and product teams to understand important data requirements and service level expectationsMentor junior data engineers with emphasis on data quality best practices and customer-focused approachParticipate in architectural reviews and help establish data engineering standards that prioritize reliability and accuracyDocument technical designs, processes, and operational procedures with focus on maintainability and knowledge sharing
Required Qualifications
Experience & EducationBachelor's or Master's degree in Computer Science, Engineering, or related technical field
7+ years (Senior) of hands-on data engineering experience
Proven experience with large-scale data processing systems (preferably in fintech/payments domain)
Experience building and maintaining production data streams processing TB/PB scale data with strong performance and reliability standards
Technical Skills & RequirementsProgramming Languages:
Expert-level proficiency in both Python and Java; experience with Scala preferred
Big Data Technologies: Apache Spark (PySpark, Spark SQL, Spark with Java), Apache Kafka, Apache Airflow
Cloud Platforms: AWS (EMR, Glue, Redshift, S3, Lambda) or equivalent Azure/GCP services
Databases: Strong SQL skills, experience with both relational (PostgreSQL, MySQL) and NoSQL databases (MongoDB, Cassandra, Redis)
Data Quality Management: Deep understanding of the 4 C's framework - Completeness, Consistency, Conformity, and Correctness
Data Governance: Experience with data lineage tracking, metadata management, and data cataloging
Data Formats & Protocols: Parquet, Avro, JSON, REST APIs, GraphQLContainerization & DevOps: Docker, Kubernetes, Git, GitLab/GitHub with CI/CD pipeline experience
Monitoring & Observability: Experience with Prometheus, Grafana, or similar monitoring tools
Data Modeling: Dimensional modeling, data vault, or similar methodologies
Streaming Technologies: Apache Flink, Kinesis, or Pulsar experience is a plus
Infrastructure as Code: Terraform, CloudFormation (preferred)
Java-specific: Spring Boot, Maven/Gradle, JUnit for building robust data services
Preferred Qualifications
Domain Expertise
Previous experience in fintech, payments, or banking industry with solid understanding of regulatory compliance and financial data requirementsUnderstanding of financial data standards, PCI DSS compliance, and data privacy regulations where compliance is essential for business operationsExperience with real-time fraud detection or risk management systems where data accuracy is crucial for customer protection
Advanced Technical Skills (Preferred)
Experience building automated data quality frameworks covering all 4 C's dimensionsKnowledge of machine learning stream orchestration (MLflow, Kubeflow)Familiarity with data mesh or federated data architecture patternsExperience with change data capture (CDC) tools and techniques
Leadership & Soft SkillsStrong problem-solving abilities with experience debugging complex distributed systems in production environmentsExcellent communication skills with ability to explain technical concepts to diverse stakeholders while highlighting business valueExperience mentoring team members and leading technical initiatives with focus on building a quality-oriented cultureProven track record of delivering projects successfully in dynamic, fast-paced financial technology environments
Data Engineer- Senior Data Engineer
Posted 1 day ago
Job Viewed
Job Description
The Role
We're looking for a senior AI engineer who can build production-grade agentic AI systems. You'll be working at the intersection of cutting-edge AI research and scalable engineering, creating autonomous agents that can reason, plan, and execute complex tasks reliably at scale.
What We Need
Agentic AI & LLM Engineering
You should have hands-on experience with:
Multi-agent systems: Building agents that coordinate, communicate, and work together on complex workflows
Agent orchestration: Designing systems where AI agents can plan multi-step tasks, use tools, and make autonomous decisions
LLMOps Experience: End-to-End LLM Lifecycle Management - hands-on experience managing the complete LLM workflow from prompt engineering and dataset curation through model fine-tuning, evaluation, and deployment. This includes versioning prompts, managing training datasets, orchestrating distributed training jobs, and implementing automated model validation pipelines. Production LLM Infrastructure - experience building and maintaining production LLM serving infrastructure including model registries, A/B testing frameworks for comparing model versions, automated rollback mechanisms, and monitoring systems that track model performance, latency, and cost metrics in real-time.
AI Observability: Experience implementing comprehensive monitoring and tracing for AI systems, including prompt tracking, model output analysis, cost monitoring, and agent decision-making visibility across complex workflows.
Evaluation frameworks: Creating comprehensive testing for agent performance, safety, and goal achievement
LLM inference optimization: Scaling model serving with techniques like batching, caching, and efficient frameworks (vLLM, TensorRT-LLM)
Systems Engineering
Strong backend development skills including:
Python expertise: FastAPI, Django, or Flask for building robust APIs that handle agent workflows
Distributed systems: Microservices, event-driven architectures, and message queues (Kafka, RabbitMQ) for agent coordination
Database strategy: Vector databases, traditional SQL/NoSQL, and caching layers optimized for agent state management
Web-scale design: Systems handling millions of requests with proper load balancing and fault tolerance
DevOps (Non-negotiable)
Kubernetes: Working knowledge required - deployments, services, cluster management
Containerization: Docker with production optimization and security best practices
CI/CD: Automated testing and deployment pipelines
Infrastructure as Code: Terraform, Helm charts
Monitoring: Prometheus, Grafana for tracking complex agent behaviors
Programing Language : Java , Python
What You'll Build
You'll architect the infrastructure that powers our autonomous AI systems:
Agent Orchestration Platform: Multi-agent coordination systems that handle complex, long-running workflows with proper state management and failure recovery.
Evaluation Infrastructure: Comprehensive frameworks that assess agent performance across goal achievement, efficiency, safety, and decision-making quality.
Production AI Services: High-throughput systems serving millions of users with intelligent resource management and robust fallback mechanisms.
Training Systems: Scalable pipelines for SFT and DPO that continuously improve agent capabilities based on real-world performance and human feedback.
Who You Are
You've spent serious time in production environments building AI systems that actually work. You understand the unique challenges of agentic AI - managing state across long conversations, handling partial failures in multi-step processes, and ensuring agents stay aligned with their intended goals.
You've dealt with the reality that the hardest problems aren't always algorithmic. Sometimes it's about making an agent retry gracefully when an API call fails, or designing an observability layer that catches when an agent starts behaving unexpectedly, or building systems that can scale from handling dozens of agent interactions to millions.
You're excited about the potential of AI agents but pragmatic about the engineering work required to make them reliable in production.
Data Engineer- Lead Data Engineer
Posted today
Job Viewed
Job Description
Role Overview
We are seeking an experienced Lead Data Engineer to join our Data Engineering team at Paytm, India's leading digital payments and financial services platform. This is a critical role responsible for designing, building, and maintaining large-scale, real-time data streams that process billions of transactions and user interactions daily. Data accuracy and stream reliability are essential to our operations, as data quality issues can result in financial losses and impact customer trust.
As a Lead Data Engineer at Paytm, you will be responsible for building robust data systems that support India's largest digital payments ecosystem. You'll architect and implement reliable, real-time data streaming solutions where precision and data correctness are fundamental requirements . Your work will directly support millions of users across merchant payments, peer-to-peer transfers, bill payments, and financial services, where data accuracy is crucial for maintaining customer confidence and operational excellence.
This role requires expertise in designing fault-tolerant, scalable data architectures that maintain high uptime standards while processing peak transaction loads during festivals and high-traffic events. We place the highest priority on data quality and system reliability, as our customers depend on accurate, timely information for their financial decisions. You'll collaborate with cross-functional teams including data scientists, product managers, and risk engineers to deliver data solutions that enable real-time fraud detection, personalized recommendations, credit scoring, and regulatory compliance reporting.
Key technical challenges include maintaining data consistency across distributed systems with demanding performance requirements, implementing comprehensive data quality frameworks with real-time validation, optimizing query performance on large datasets, and ensuring complete data lineage and governance across multiple business domains. At Paytm, reliable data streams are fundamental to our operations and our commitment to protecting customers' financial security and maintaining India's digital payments infrastructure.
Key Responsibilities
Data Stream Architecture & Development Design and implement reliable, scalable data streams handling high-volume transaction data with strong data integrity controlsBuild real-time processing systems using modern data engineering frameworks (Java/Python stack) with excellent performance characteristicsDevelop robust data ingestion systems from multiple sources with built-in redundancy and monitoring capabilitiesImplement comprehensive data quality frameworks, ensuring the 4 C's: Completeness, Consistency, Conformity, and Correctness - ensuring data reliability that supports sound business decisionsDesign automated data validation, profiling, and quality monitoring systems with proactive alerting capabilities Infrastructure & Platform Management Manage and optimize distributed data processing platforms with high availability requirements to ensure consistent service deliveryDesign data lake and data warehouse architectures with appropriate partitioning and indexing strategies for optimal query performanceImplement CI/CD processes for data engineering workflows with comprehensive testing and reliable deployment proceduresEnsure high availability and disaster recovery for critical data systems to maintain business continuity
Performance & Optimization Monitor and optimize streaming performance with focus on latency reduction and operational efficiencyImplement efficient data storage strategies including compression, partitioning, and lifecycle management with cost considerationsTroubleshoot and resolve complex data streaming issues in production environments with effective response protocolsConduct proactive capacity planning and performance tuning to support business growth and data volume increases
Collaboration & Leadership Work closely with data scientists, analysts, and product teams to understand important data requirements and service level expectationsMentor junior data engineers with emphasis on data quality best practices and customer-focused approachParticipate in architectural reviews and help establish data engineering standards that prioritize reliability and accuracyDocument technical designs, processes, and operational procedures with focus on maintainability and knowledge sharing
Required Qualifications
Experience & Education Bachelor's or Master's degree in Computer Science, Engineering, or related technical field
7+ years (Senior) of hands-on data engineering experience
Proven experience with large-scale data processing systems (preferably in fintech/payments domain)
Experience building and maintaining production data streams processing TB/PB scale data with strong performance and reliability standards
Technical Skills & RequirementsProgramming Languages:
Expert-level proficiency in both Python and Java; experience with Scala preferred
Big Data Technologies: Apache Spark (PySpark, Spark SQL, Spark with Java), Apache Kafka, Apache Airflow
Cloud Platforms: AWS (EMR, Glue, Redshift, S3, Lambda) or equivalent Azure/GCP services
Databases: Strong SQL skills, experience with both relational (PostgreSQL, MySQL) and NoSQL databases (MongoDB, Cassandra, Redis)
Data Quality Management: Deep understanding of the 4 C's framework - Completeness, Consistency, Conformity, and Correctness
Data Governance: Experience with data lineage tracking, metadata management, and data cataloging
Data Formats & Protocols: Parquet, Avro, JSON, REST APIs, GraphQL Containerization & DevOps: Docker, Kubernetes, Git, GitLab/GitHub with CI/CD pipeline experience
Monitoring & Observability: Experience with Prometheus, Grafana, or similar monitoring tools
Data Modeling: Dimensional modeling, data vault, or similar methodologies
Streaming Technologies: Apache Flink, Kinesis, or Pulsar experience is a plus
Infrastructure as Code: Terraform, CloudFormation (preferred)
Java-specific: Spring Boot, Maven/Gradle, JUnit for building robust data services
Preferred Qualifications
Domain Expertise
Previous experience in fintech, payments, or banking industry with solid understanding of regulatory compliance and financial data requirementsUnderstanding of financial data standards, PCI DSS compliance, and data privacy regulations where compliance is essential for business operationsExperience with real-time fraud detection or risk management systems where data accuracy is crucial for customer protection
Advanced Technical Skills (Preferred)
Experience building automated data quality frameworks covering all 4 C's dimensionsKnowledge of machine learning stream orchestration (MLflow, Kubeflow)Familiarity with data mesh or federated data architecture patternsExperience with change data capture (CDC) tools and techniques
Leadership & Soft Skills Strong problem-solving abilities with experience debugging complex distributed systems in production environmentsExcellent communication skills with ability to explain technical concepts to diverse stakeholders while highlighting business valueExperience mentoring team members and leading technical initiatives with focus on building a quality-oriented cultureProven track record of delivering projects successfully in dynamic, fast-paced financial technology environments
What We Offer
Opportunity to work with cutting-edge technology at scaleCompetitive salary and equity compensation
Comprehensive health and wellness benefits
Professional development opportunities and conference attendanceFlexible working arrangements
Chance to impact millions of users across India's digital payments ecosystem
Application Process
Interested candidates should submit:
Updated resume highlighting relevant data engineering experience with emphasis on real-time systems and data quality
Portfolio or GitHub profile showcasing data engineering projects, particularly those involving high-throughput streaming systems
Cover letter explaining interest in fintech/payments domain and understanding of data criticality in financial services
References from previous technical managers or senior colleagues who can attest to your data quality standards
PIfc5a5d46cf
Data Engineer- Senior Data Engineer
Posted today
Job Viewed
Job Description
Data Engineer- Senior Data Engineer
Posted today
Job Viewed
Job Description
The Role
We're looking for a senior AI engineer who can build production-grade agentic AI systems. You'll be working at the intersection of cutting-edge AI research and scalable engineering, creating autonomous agents that can reason, plan, and execute complex tasks reliably at scale.
What We Need
Agentic AI & LLM Engineering
You should have hands-on experience with:
Multi-agent systems : Building agents that coordinate, communicate, and work together on complex workflows
Agent orchestration : Designing systems where AI agents can plan multi-step tasks, use tools, and make autonomous decisions
LLMOps Experience : End-to-End LLM Lifecycle Management - hands-on experience managing the complete LLM workflow from prompt engineering and dataset curation through model fine-tuning, evaluation, and deployment. This includes versioning prompts, managing training datasets, orchestrating distributed training jobs, and implementing automated model validation pipelines. Production LLM Infrastructure - experience building and maintaining production LLM serving infrastructure including model registries, A/B testing frameworks for comparing model versions, automated rollback mechanisms, and monitoring systems that track model performance, latency, and cost metrics in real-time.
AI Observability : Experience implementing comprehensive monitoring and tracing for AI systems, including prompt tracking, model output analysis, cost monitoring, and agent decision-making visibility across complex workflows.
Evaluation frameworks : Creating comprehensive testing for agent performance, safety, and goal achievement
LLM inference optimization : Scaling model serving with techniques like batching, caching, and efficient frameworks (vLLM, TensorRT-LLM)
Systems Engineering
Strong backend development skills including:
Python expertise : FastAPI, Django, or Flask for building robust APIs that handle agent workflows
Distributed systems : Microservices, event-driven architectures, and message queues (Kafka, RabbitMQ) for agent coordination
Database strategy : Vector databases, traditional SQL/NoSQL, and caching layers optimized for agent state management
Web-scale design : Systems handling millions of requests with proper load balancing and fault tolerance
DevOps (Non-negotiable)
Kubernetes : Working knowledge required - deployments, services, cluster management
Containerization : Docker with production optimization and security best practices
CI/CD : Automated testing and deployment pipelines
Infrastructure as Code : Terraform, Helm charts
Monitoring : Prometheus, Grafana for tracking complex agent behaviors
Programing Language : Java , Python
What You'll Build
You'll architect the infrastructure that powers our autonomous AI systems:
Agent Orchestration Platform : Multi-agent coordination systems that handle complex, long-running workflows with proper state management and failure recovery.
Evaluation Infrastructure : Comprehensive frameworks that assess agent performance across goal achievement, efficiency, safety, and decision-making quality.
Production AI Services : High-throughput systems serving millions of users with intelligent resource management and robust fallback mechanisms.
Training Systems : Scalable pipelines for SFT and DPO that continuously improve agent capabilities based on real-world performance and human feedback.
Who You Are
You've spent serious time in production environments building AI systems that actually work. You understand the unique challenges of agentic AI - managing state across long conversations, handling partial failures in multi-step processes, and ensuring agents stay aligned with their intended goals.
You've dealt with the reality that the hardest problems aren't always algorithmic. Sometimes it's about making an agent retry gracefully when an API call fails, or designing an observability layer that catches when an agent starts behaving unexpectedly, or building systems that can scale from handling dozens of agent interactions to millions.
You're excited about the potential of AI agents but pragmatic about the engineering work required to make them reliable in production.
PI503be25532c
Data Engineer

Posted 4 days ago
Job Viewed
Job Description
**The Team**
You will be joining the Catalog team within the Business Apps department. Our mission is to build end-to-end solutions on the Celonis platform, including data models and end-user applications, to accelerate time to value for our customers and partners. The Catalog team within Business Apps specializes in three aspects: defining the data ontology of the most common business processes, building prebuilt transformations for such ontologies for major source systems like SAP, Oracle etc, and lastly, collaborating with various teams in both the Product and Go-to-market organizations to drive adoption at scale.
As a **Data Engineer** , you will own and focus on primarily two aspects: On the one hand, refining prebuilt transformations for existing ontologies (for processes like Order to Cash, Procure to Pay, Inventory Management) for SAP, Oracle etc and validating them across our early adopters in our customer base. On the other, defining and extending the existing ontologies with additional processes and extending to additional systems. In addition, you would also be responsible for maintaining the quality of content that we produce, and write documentation on the ontology definitions. This will ensure both internal and external application developers will be able to leverage the data foundation to develop their solutions
**The work you'll do:**
+ Build data models for the defined ontologies and mappings using the object-centric process mining methodologies with performant SQL transformations.
+ Design and implement business objects, process events, and data models in the Celonis platform.
+ Research and design:
+ ontologies for new business processes, improve and extend capabilities of existing ones
+ the source system transformations to map them with the defined ontologies.
+ Facilitate cross-functional interactions with product managers, domain experts, engineers, and consultants.
+ Test and validate the models in development environments and customer environments to gather early feedback
+ Document the data model governing principles and development.
**The qualifications you need:**
+ You have that rare combination-a strong technical expertise and business acumen. You'll use this to build a system-agnostic data model for various business processes.
+ 3-6+ years of experience working in the data field as a Data Engineer, Data Analyst or similar.
+ **Must-have:**Experience working with data from at least one of the following system types:
+ ERP (e.g. SAP ECC or S/4, Oracle EBS or Fusion)
+ Supply Chain Management (e.g. BlueYonder, SAP or Oracle Transportation Management)
+ CRM (e.g. Salesforce, Microsoft Dynamics)
+ IT (e.g. ServiceNow)
+ Strong solution designing skills with solid understanding of business processes (supply chain, financial, CRM or IT-related processes) and data beneath the IT systems that run these processes.
+ Experience with databases and data modeling, and hands-on experience with SQL.
+ Ability to work independently and own a part of the team's goals
+ Very good knowledge of spoken and written English
+ Ability to communicate effectively and build a good rapport with team members.
**What Celonis Can Offer You:**
+ **Pioneer Innovation:** Work with the leading, award-winning process mining technology, shaping the future of business.
+ **Accelerate Your Growth:** Benefit from clear career paths, internal mobility, a dedicated learning program, and mentorship opportunities.
+ **Receive Exceptional Benefits:** Including generous PTO, hybrid working options, company equity (RSUs), comprehensive benefits, extensive parental leave, dedicated volunteer days, and much more ( . Interns and working students explore your benefits here ( .
+ **Prioritize Your Well-being:** Access to resources such as gym subsidies, counseling, and well-being programs.
+ **Connect and Belong:** Find community and support through dedicated inclusion and belonging programs.
+ **Make Meaningful Impact:** Be part of a company driven by strong values ( that guide everything we do: Live for Customer Value, The Best Team Wins, We Own It, and Earth Is Our Future.
+ **Collaborate Globally:** Join a dynamic, international team of talented individuals.
+ **Empowered Environment:** Contribute your ideas in an open culture with autonomous teams.
**About Us:**
Celonis makes processes work for people, companies and the planet. The Celonis Process Intelligence Platform uses industry-leading process mining and AI technology and augments it with business context to give customers a living digital twin of their business operation. It's system-agnostic and without bias, and provides everyone with a common language for understanding and improving businesses. Celonis enables its customers to continuously realize significant value across the top, bottom, and green line. Celonis is headquartered in Munich, Germany, and New York City, USA, with more than 20 offices worldwide.
Get familiar with the Celonis Process Intelligence Platform by watching this video ( .
**Celonis Inclusion Statement:**
At Celonis, we believe our people make us who we are and that "The Best Team Wins". We know that the best teams are made up of people who bring different perspectives to the table. And when everyone feels included, able to speak up and knows their voice is heard - that's when creativity and innovation happen.
**Your Privacy:**
Any information you submit to Celonis as part of your application will be processed in accordance with Celonis' Accessibility and Candidate Notices ( submitting this application, you confirm that you agree to the storing and processing of your personal data by Celonis as described in our Privacy Notice for the Application and Hiring Process ( .
Please be aware of common job offer scams, impersonators and frauds. Learn more here ( .
Be The First To Know
About the latest Data engineer Jobs in Bengaluru !
Data Engineer
Posted 9 days ago
Job Viewed
Job Description
NTT DATA strives to hire exceptional, innovative and passionate individuals who want to grow with us. If you want to be part of an inclusive, adaptable, and forward-thinking organization, apply now.
We are currently seeking a Data Engineer to join our team in Bangalore, Karnātaka (IN-KA), India (IN).
+ **6+ years** of experience in Data Engineering or related role.
+ Hands-on experience with **Snowflake** (data modelling, performance tuning, query optimization, Snowpipe, Time Travel, Streams & Tasks).
+ Strong expertise in **AWS Glue** for ETL job development, orchestration, and optimization.
+ Proficiency with AWS services such as **S3, Lambda, Redshift, Athena, Step Functions, Kinesis** , and **CloudWatch** .
+ Proficiency in Pulling data from API
+ Strong programming skills in **Python** and/or **PySpark** .
+ Knowledge of SQL and experience with performance tuning for large datasets.
+ Experience with **data warehouse** and **data lake** architectures.
+ Familiarity with CI/CD pipelines and infrastructure-as-code tools (Terraform, CloudFormation).
**Preferred Qualifications:**
+ AWS Certified Solutions Architect / Data Analytics Specialty certification.
+ Experience integrating **Snowflake** with third-party tools like Tableau or Power BI
+ Experience with streaming data ingestion using **Kinesis** or Kafka.
**About NTT DATA**
NTT DATA is a $30 billion trusted global innovator of business and technology services. We serve 75% of the Fortune Global 100 and are committed to helping clients innovate, optimize and transform for long term success. As a Global Top Employer, we have diverse experts in more than 50 countries and a robust partner ecosystem of established and start-up companies. Our services include business and technology consulting, data and artificial intelligence, industry solutions, as well as the development, implementation and management of applications, infrastructure and connectivity. We are one of the leading providers of digital and AI infrastructure in the world. NTT DATA is a part of NTT Group, which invests over $3.6 billion each year in R&D to help organizations and society move confidently and sustainably into the digital future. Visit us at us.nttdata.com ( possible, we hire locally to NTT DATA offices or client sites. This ensures we can provide timely and effective support tailored to each client's needs. While many positions offer remote or hybrid work options, these arrangements are subject to change based on client requirements. For employees near an NTT DATA office or client site, in-office attendance may be required for meetings or events, depending on business needs. At NTT DATA, we are committed to staying flexible and meeting the evolving needs of both our clients and employees. NTT DATA recruiters will never ask for payment or banking information and will only use @nttdata.com and @talent.nttdataservices.com email addresses. If you are requested to provide payment or disclose banking information, please submit a contact us form, .
**_NTT DATA endeavors to make_** **_ **_accessible to any and all users. If you would like to contact us regarding the accessibility of our website or need assistance completing the application process, please contact us at_** **_ **_._** **_This contact information is for accommodation requests only and cannot be used to inquire about the status of applications. NTT DATA is an equal opportunity employer. Qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability or protected veteran status. For our EEO Policy Statement, please click here ( . If you'd like more information on your EEO rights under the law, please click here ( . For Pay Transparency information, please click here ( ._**
Data Engineer
Posted 9 days ago
Job Viewed
Job Description
NTT DATA strives to hire exceptional, innovative and passionate individuals who want to grow with us. If you want to be part of an inclusive, adaptable, and forward-thinking organization, apply now.
We are currently seeking a Data Engineer to join our team in Bangalore, Karnātaka (IN-KA), India (IN).
+ **6+ years** of experience in Data Engineering or related role.
+ Hands-on experience with **Snowflake** (data modelling, performance tuning, query optimization, Snowpipe, Time Travel, Streams & Tasks).
+ Strong expertise in **AWS Glue** for ETL job development, orchestration, and optimization.
+ Proficiency with AWS services such as **S3, Lambda, Redshift, Athena, Step Functions, Kinesis** , and **CloudWatch** .
+ Proficiency in Pulling data from API
+ Strong programming skills in **Python** and/or **PySpark** .
+ Knowledge of SQL and experience with performance tuning for large datasets.
+ Experience with **data warehouse** and **data lake** architectures.
+ Familiarity with CI/CD pipelines and infrastructure-as-code tools (Terraform, CloudFormation).
**Preferred Qualifications:**
+ AWS Certified Solutions Architect / Data Analytics Specialty certification.
+ Experience integrating **Snowflake** with third-party tools like Tableau or Power BI
+ Experience with streaming data ingestion using **Kinesis** or Kafka.
**About NTT DATA**
NTT DATA is a $30 billion trusted global innovator of business and technology services. We serve 75% of the Fortune Global 100 and are committed to helping clients innovate, optimize and transform for long term success. As a Global Top Employer, we have diverse experts in more than 50 countries and a robust partner ecosystem of established and start-up companies. Our services include business and technology consulting, data and artificial intelligence, industry solutions, as well as the development, implementation and management of applications, infrastructure and connectivity. We are one of the leading providers of digital and AI infrastructure in the world. NTT DATA is a part of NTT Group, which invests over $3.6 billion each year in R&D to help organizations and society move confidently and sustainably into the digital future. Visit us at us.nttdata.com ( possible, we hire locally to NTT DATA offices or client sites. This ensures we can provide timely and effective support tailored to each client's needs. While many positions offer remote or hybrid work options, these arrangements are subject to change based on client requirements. For employees near an NTT DATA office or client site, in-office attendance may be required for meetings or events, depending on business needs. At NTT DATA, we are committed to staying flexible and meeting the evolving needs of both our clients and employees. NTT DATA recruiters will never ask for payment or banking information and will only use @nttdata.com and @talent.nttdataservices.com email addresses. If you are requested to provide payment or disclose banking information, please submit a contact us form, .
**_NTT DATA endeavors to make_** **_ **_accessible to any and all users. If you would like to contact us regarding the accessibility of our website or need assistance completing the application process, please contact us at_** **_ **_._** **_This contact information is for accommodation requests only and cannot be used to inquire about the status of applications. NTT DATA is an equal opportunity employer. Qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability or protected veteran status. For our EEO Policy Statement, please click here ( . If you'd like more information on your EEO rights under the law, please click here ( . For Pay Transparency information, please click here ( ._**
Data Engineer
Posted 9 days ago
Job Viewed
Job Description
+ Bachelor's degree in Computer Science, Engineering, Mathematics, a related field, or equivalent practical experience.
+ Experience in developing and troubleshooting data processing algorithms and software using Python, Java, Scala, Spark and hadoop frameworks.
+ Experience in data processing frameworks and Google Cloud Platform with investigative and transactional data stores like BigQuery, CloudSQL, AlloyDB, etc.
+ Experience in Google Cloud Platform.
**Preferred qualifications:**
+ Experience with encryption techniques like symmetric, asymmetric, Hardware Security Module (HSMs) and envelop with ability to implement secure key storage using Key Management System.
+ Experience in working with data warehouses, including technical architectures, infrastructure components, ETL/ELT and reporting tools, environments, and data structures.
+ Experience in building applications with technologies like NoSQL, MongoDB, SparkML, and TensorFlow.
+ Experience with Infrastructure as Code (IaC) and CI/CD tools like Terraform, Ansible, Jenkins, etc.
+ Experience in Big Data, information retrieval, data mining, or Machine Learning.
+ Experience with architecting, developing software, or internet production-grade Big Data solutions in virtualized environments.
The Google Cloud Consulting Professional Services team guides customers through the moments that matter most in their cloud journey to help businesses thrive. We help customers transform and evolve their business through the use of Google's global network, web-scale data centers, and software infrastructure. As part of an innovative team in this rapidly growing business, you will help shape the future of businesses of all sizes and use technology to connect with customers, employees, and partners.
In this role, you will guide customers on how to store, process, analyze, and explore/visualize data on the Google Cloud Platform (GCP). You will work on data migrations and modernization projects, and with customers to design data processing systems, develop data pipelines and troubleshoot platform/product tests. You will have an understanding of data governance and security controls. You will travel to customer sites to deploy solutions and deliver workshops to educate and empower customers. You will work with Product Management and Product Engineering teams to build and drive excellence in products.
Google Cloud accelerates every organization's ability to digitally transform its business and industry. We deliver enterprise-grade solutions that leverage Google's cutting-edge technology, and tools that help developers build more sustainably. Customers in more than 200 countries and territories turn to Google Cloud as their trusted partner to enable growth and solve their most critical business problems.
**Responsibilities:**
+ Interact with stakeholders to translate customer requirements into recommendations for solution architectures and advisory services.
+ Engage with technical leads, and partners to lead migration and modernization to Google Cloud Platform (GCP).
+ Design, migrate/build, and operationalize data storage and processing infrastructure using Cloud native products.
+ Develop and implement data quality and governance procedures to ensure the reliability of data.
+ Gather project requirements and organize them into goals and objectives, and create a work breakdown structure to manage internal and external stakeholders.
Google is proud to be an equal opportunity workplace and is an affirmative action employer. We are committed to equal employment opportunity regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability, gender identity or Veteran status. We also consider qualified applicants regardless of criminal histories, consistent with legal requirements. See also and If you have a need that requires accommodation, please let us know by completing our Accommodations for Applicants form: