508 Data Scientists jobs in Chennai
Data Scientists
Posted today
Job Viewed
Job Description
Thrive in a vibrant, collaborative environment where your ideas matter. Work on impactful projects, grow fast, and learn continuously.
Why You Should Join UsLearning Opportunity
Unlock continuous learning opportunities to enhance your skills and knowledge, with access to resources, training programs, and mentorship that support your professional growth and keep you at the forefront of the industry.
InnovationExperience a culture that embraces innovation, where your ideas are valued and nurtured to drive groundbreaking solutions and push the boundaries of what's possible.
Transparent CultureBe part of a transparent culture where open communication, trust, and respect are valued. Embrace an environment where diverse perspectives are encouraged, collaboration thrives, and decisions are made collectively.
Fun at WorkEnjoy a vibrant and enjoyable work atmosphere that promotes a healthy work-life balance. Engage in team-building activities, celebrations, and events that foster camaraderie and create memorable experiences with your colleagues.
What We OfferWe believe in taking care of our team members and providing an environment where they can thrive both professionally and personally.
Competitive salary
Health insurance coverage
Flexible working hours
Hybrid work options
Professional development budget
Modern office space
Team outings
Mentorship programs
Work at ImpeloxCollaborative Team
Work with passionate AI enthusiasts
Innovation FirstExperiment with cutting-edge technology
Work-Life BalanceFlexible schedules and remote options
Data ScientistsFull-time
Chennai, India
3-6 years
Analyze complex datasets and build predictive models to drive business insights.
Required Skills:Python
Deep Learning
Machine Learning
SQL
RAG
LLM
Data Scientists
Posted today
Job Viewed
Job Description
Description
Key Responsibilities
AI/ML Development & Research
• Design, develop, and deploy advanced machine learning and deep learning models for complex business problems
• Implement and optimize Large Language Models (LLMs) and Generative AI solutions
• Build agentic AI systems with autonomous decision-making capabilities
• Conduct research on emerging AI technologies and their practical applications
• Perform model evaluation, validation, and continuous improvement
Cloud Infrastructure & Full-Stack Development
• Architect and implement scalable cloud-native ML/AI solutions on AWS, Azure, or GCP
• Develop full-stack applications integrating AI models with modern web technologies
• Build and maintain ML pipelines using cloud services (SageMaker, ML Engine, etc.)
• Implement CI/CD pipelines for ML model deployment and monitoring
• Design and optimize cloud infrastructure for high-performance computing workloads
Data Engineering & Database Management
• Design and implement data pipelines for large-scale data processing
• Work with both SQL and NoSQL databases (PostgreSQL, MongoDB, Cassandra, etc.)
• Optimize database performance for ML workloads and real-time applications
• Implement data governance and quality assurance frameworks
• Handle streaming data processing and real-time analytics
Leadership & Collaboration
• Mentor junior data scientists and guide technical decision-making
• Collaborate with cross-functional teams including product, engineering, and business stakeholders
• Present findings and recommendations to technical and non-technical audiences
• Lead proof-of-concept projects and innovation initiatives
Required Qualifications
Education & Experience
• Master's or PhD in Computer Science, Data Science, Statistics, Mathematics, or related field
• 5+ years of hands-on experience in data science and machine learning
• 3+ years of experience with deep learning frameworks and neural networks
• 2+ years of experience with cloud platforms and full-stack development
Technical Skills - Core AI/ML
• Machine Learning: Scikit-learn, XGBoost, LightGBM, advanced ML algorithms
• Deep Learning: TensorFlow, PyTorch, Keras, CNN, RNN, LSTM, Transformers
• Large Language Models: GPT, BERT, T5, fine-tuning, prompt engineering
• Generative AI: Stable Diffusion, DALL-E, text-to-image, text generation
• Agentic AI: Multi-agent systems, reinforcement learning, autonomous agents
Technical Skills - Development & Infrastructure
• Programming: Python (expert), R, Java/Scala, JavaScript/TypeScript
• Cloud Platforms: AWS (SageMaker, EC2, S3, Lambda), Azure ML, or Google Cloud AI
• Databases: SQL (PostgreSQL, MySQL), NoSQL (MongoDB, Cassandra, DynamoDB)
• Full-Stack Development: React/Vue.js, Node.js, FastAPI, Flask, Docker, Kubernetes
• MLOps: MLflow, Kubeflow, Model versioning, A/B testing frameworks
• Big Data: Spark, Hadoop, Kafka, streaming data processing
Preferred Qualifications
• Experience with vector databases and embeddings (Pinecone, Weaviate, Chroma)
• Knowledge of LangChain, LlamaIndex, or similar LLM frameworks
• Experience with model compression and edge deployment
• Familiarity with distributed computing and parallel processing
• Experience with computer vision and NLP applications
• Knowledge of federated learning and privacy-preserving ML
• Experience with quantum machine learning
• Expertise in MLOps and production ML system design
Key Competencies
Technical Excellence
• Strong mathematical foundation in statistics, linear algebra, and optimization
• Ability to implement algorithms from research papers
• Experience with model interpretability and explainable AI
• Knowledge of ethical AI and bias detection/mitigation
Problem-Solving & Innovation
• Strong analytical and critical thinking skills
• Ability to translate business requirements into technical solutions
• Creative approach to solving complex, ambiguous problems
• Experience with rapid prototyping and experimentation
Communication & Leadership
• Excellent written and verbal communication skills
• Ability to explain complex technical concepts to diverse audiences
• Strong project management and organizational skills
• Experience mentoring and leading technical teams
How We Partner To Protect You: TaskUs will neither solicit money from you during your application process nor require any form of payment in order to proceed with your application. Kindly ensure that you are always in communication with only authorized recruiters of TaskUs.DEI: In TaskUs we believe that innovation and higher performance are brought by people from all walks of life. We welcome applicants of different backgrounds, demographics, and circumstances. Inclusive and equitable practices are our responsibility as a business. TaskUs is committed to providing equal access to opportunities. If you need reasonable accommodations in any part of the hiring process, please let us know.We invite you to explore all TaskUs career opportunities and apply through the provided URL.
Data Scientists - Nlp
Posted today
Job Viewed
Job Description
- Acquire deep understanding of the business problems and translate them into appropriate mathematical representations
- Work with large, complex data sets using tools such as SQL, Google Cloud Services, Hadoop, Alteryx, Python
- Develop and deliver business solutions using skills such as data acquisition, data engineering, programming, and visualization.
- Design intuitive visual interfaces for users to interact with the data using dashboard and programming software
- Develop and deliver analytic products using skills such as statistical analysis, machine learning, algorithm design, and interface development
- Interpret modeling results and communicate them to technical and non-technical audiences, cross-functional teams and leadership
- Interact and work cross-functionally with a wide variety of teams
- Develop trust with stakeholders and peers by delivering results on time
- Ensure overall quality of the data & solutions throughout the analytic development process
- Work with business teams on change management
- Provide training and maintenance of implemented tools to business partners
- Collect feedback from business users and continuously improve analytic products
**The minimum requirements we seek**:
- Bachelor’s degree in Statistics, Data Science, Computer Science or a related quantitative discipline
- One year of experience or other significant, independent project experience using at least one statistical software package such as R or Python,
- 3+ months of experience using SQL for acquiring and transforming data
- 3+ months of experience with real-world data, data cleaning, data collection or other data wrangling challenges
**Our preferred requirements**:
- Masters or PhD in quantitative field, such as (but not limited to) Statistics, Computer Science, Economics, Mathematics, Data Science, Operations Research
- 2+ years of work experience with Python, SQL, and/or visualization/dashboard tools (such as Tableau, PowerBI or Qliksense)
- 2+ years of work experience in data mining, statistical analysis, modeling, optimization or similar analytics
- Comfortable working in an environment where problems are not always well-defined
- Inquisitive, proactive, and interested in learning new tools and techniques
- Proven quantitative modeling and statistical analysis skills
- Strong oral, written and interpersonal communication skills
- Ability to effectively advocate technical solutions to peer, management and business audiences
- Well-organized, a self-starter, independent and ready to work with mínimal supervision
- A respectful and committed teammate, willing to excel and work with talented people
- Work experience in automotive industry is a big plus, as is experience in a procurement, logistics or program management function
- 3-4 years of industry experience
- Bachelor's in Engineering/ M.S. or Ph.D. in Computer Science, Operational research, Statistics, Applied mathematics, or in a related field
- Experience with SQL database
- Knowledge in fundamental text data processing (like use of regex, token/word analysis, spelling correction/noise reduction in text, segmenting noisy unfamiliar sentences/phrases at right places, deriving insights from clustering, etc.,)
- Excellent problem solving, communication, and data presentation skills
- Flexibility to work on multiple projects/domains/tools
- Collaborative skills to work with business teams.
- Comfort level to code in Tensorflow and/or Pytorch, Numpy and Pandas, Scikit-learn
- Comfort level to use two/more of open source NLP modules like SpaCy, HuggingFace, TorchText, fastai.text, farm-haystack, and others
- Understanding the Business problem and formulating the analytical problem
- Develop NLP models to process text in different languages (German, Chinese, Spanish, Portuguese and more)
- Speech to text translation for Video/Audio NLP Analysis
- Train, tune, validate, and monitor predictive models
- Analyze and extract relevant information from large amounts of Ford’s historical business data, both in structured and unstructured formats
- Establish scalable, efficient, automated processes for large scale data analyses
- Develop and deploy Data Science models on cloud platforms like GCP, Azure, AWS.
Big Data Engineer
Posted 4 days ago
Job Viewed
Job Description
**Responsibilities:**
+ Design, development of BigData applications/ pipelines using Spark, Scala, SQL, Pyspark, Python, Java
+ Consult with users, clients, and other technology groups on issues, and recommend programming solutions, install, and support customer exposure systems
+ Appropriately assess risk when business decisions are made, demonstrating particular consideration for the firm's reputation and safeguarding Citigroup, its clients and assets, by driving compliance with applicable laws, rules and regulations, adhering to Policy, applying sound ethical judgment regarding personal behavior, conduct and business practices, and escalating, managing and reporting control issues with transparency.
**Qualifications:**
+ 4-8 years of experience in software development, building large scale distributed data processing systems or large-scale applications
+ Designing & developing Big Data solutions with at least one end to end implementation.
+ Strong Hands-on experience in following technical skills: Apache Spark, Scala/ Java, XML/ JSON/ Parquet/ Avro, SQL, Linux, Hadoop Ecosystem (HDFS, Spark, Impala, HIVE, HBASE etc.), Kafka.
+ Performance analysis, troubleshooting and issue resolution and Exposure to latest Cloudera offerings like Ozone, Iceberg.
+ Intermediate level experience in Applications Development role
+ Consistently demonstrates clear and concise written and verbal communication
+ Demonstrated problem-solving and decision-making skills
+ Ability to work under pressure and manage deadlines or unexpected changes in expectations or requirements
**Education:**
+ Bachelor's degree/University degree or equivalent experience
This job description provides a high-level review of the types of work performed. Other job-related duties may be assigned as required.
---
**Job Family Group:**
Technology
---
**Job Family:**
Applications Development
---
**Time Type:**
Full time
---
**Most Relevant Skills**
Please see the requirements listed above.
---
**Other Relevant Skills**
For complementary skills, please see above and/or contact the recruiter.
---
_Citi is an equal opportunity employer, and qualified candidates will receive consideration without regard to their race, color, religion, sex, sexual orientation, gender identity, national origin, disability, status as a protected veteran, or any other characteristic protected by law._
_If you are a person with a disability and need a reasonable accommodation to use our search tools and/or apply for a career opportunity review_ _Accessibility at Citi ( _._
_View Citi's_ _EEO Policy Statement ( _and the_ _Know Your Rights ( _poster._
Citi is an equal opportunity and affirmative action employer.
Minority/Female/Veteran/Individuals with Disabilities/Sexual Orientation/Gender Identity.
Big Data Developer
Posted today
Job Viewed
Job Description
Job Title : Senior Big Data Engineer (Python)
Experience : 5+ Years
Locations : Chennai, Bangalore, Gurugram
Employment Type : Full-Time
Notice Period : Immediate to 30 Days Preferred
Job Summary
We are seeking an experienced Big Data Engineer with strong expertise in Python to design, build, and manage large-scale data pipelines and analytics solutions. The ideal candidate will have hands-on experience working with Big Data technologies and cloud platforms, and a passion for writing efficient, scalable, and maintainable code.
Key Responsibilities
- Design, develop, and maintain scalable big data pipelines using Python and other data processing tools.
- Work with distributed data processing frameworks like Spark, Hadoop, Hive, or similar.
- Implement ETL processes for structured and unstructured data from various sources.
- Collaborate with data scientists, analysts, and other engineering teams to understand data needs.
- Optimize data workflows and ensure data quality, integrity, and security.
- Deploy solutions on cloud platforms (AWS/GCP/Azure) and automate data workflows.
- Monitor data pipelines, troubleshoot issues, and ensure high availability and performance.
Required Skills & Qualifications
- Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field.
- Minimum 5 years of hands-on experience in Big Data engineering roles.
- Strong proficiency in Python for data processing and scripting.
- Experience with Apache Spark , Hadoop ecosystem (Hive, HDFS, HBase), Kafka.
- Solid understanding of data warehousing concepts and data modeling.
- Hands-on experience with SQL and NoSQL databases.
- Familiarity with cloud data platforms like AWS (EMR, S3, Glue) , Azure , or Google Cloud .
- Good understanding of CI/CD, containerization (Docker, Kubernetes), and version control (Git).
Preferred Skills (Nice to Have)
- Experience with Airflow or other workflow orchestration tools.
- Knowledge of real-time data processing (e.g., Spark Streaming, Flink).
- Exposure to data governance, data lineage, and catalog tools.
Why Join Us?
- Opportunity to work on cutting-edge Big Data solutions with global clients.
- Collaborative and innovation-driven work culture.
- Competitive compensation and career growth opportunities.
- Flexible work environment with hybrid (based on role).
Big Data Developer
Posted 2 days ago
Job Viewed
Job Description
Job Title: Big Data Developer (Java/Python)
Location: Chennai, Bangalore, Gurugram
Experience Required: 5+ years (5 to 15 yrs)
Joining: Immediate or Early Joiners Preferred
Employment Type: Full-time
Job Summary:
We are looking for a passionate and experienced Big Data Developer with expertise in either Java or Python to join our dynamic team. The ideal candidate will have a strong background in designing and implementing large-scale data processing systems and a solid understanding of modern data technologies. Candidates who are available to join immediately or at short notice will be given preference.
Key Responsibilities:
- Design, develop, and maintain scalable Big Data solutions using Hadoop ecosystem, Spark, and other distributed frameworks.
- Build and optimize data pipelines for batch and real-time data processing.
- Collaborate with data scientists, analysts, and other developers to integrate data-driven solutions into production.
- Write efficient, testable, and reusable code using Java or Python.
- Work closely with DevOps teams to deploy and monitor applications on cloud/on-prem infrastructure.
- Ensure data integrity, security, and performance tuning of large-scale data systems.
Technical Skills:
- Strong programming skills in Java or Python (both is a plus).
- Hands-on experience with Big Data technologies such as Hadoop , Hive , HDFS , Spark , Kafka , etc.
- Familiarity with data modeling, ETL pipelines, and data warehousing concepts.
- Good understanding of SQL and NoSQL databases.
- Experience with cloud platforms such as AWS, GCP, or Azure is a plus.
- Knowledge of CI/CD tools and containerization (Docker/Kubernetes) is desirable.
Required Qualifications:
- Bachelor's or Master’s degree in Computer Science, Engineering, or a related field.
- Minimum 5 years of experience in Big Data development with strong coding background in Java or Python.
- Strong problem-solving skills and ability to work independently or in a team.
- Excellent communication and collaboration skills.
Nice to Have:
- Experience with data lakes, lakehouses, or real-time analytics.
- Exposure to tools like Airflow, NiFi, or similar workflow orchestration tools.
Why Join Us?
- Opportunity to work with cutting-edge Big Data technologies.
- Collaborative and innovative work environment.
- Competitive compensation and benefits.
- Immediate onboarding for early joiners.
Big Data Developer
Posted 2 days ago
Job Viewed
Job Description
- 4+ years of hands on development experience in programming languages such as JAVA,SCALA using Maven, Apache Spark Frameworks and Unix Shell scripting
- Should be comfortable with Unix File system as well as HDFS commands
- Should have worked on query languages such as Oracle SQL, Hive SQL, Spark SQL, Impala, HBase DB Should be flexible
- Should have good communication and customer management skills
- Should have knowledge on Big data Data Ingestion tools such as SQOOP and KAFKA. Should be aware of the components in Big Data ecosystem. Should have worked on building projects using Eclipse IDE, Tectia Client, Oracle SQL Developer
Be The First To Know
About the latest Data scientists Jobs in Chennai !
Big Data Developer
Posted today
Job Viewed
Job Description
Job Title: Developer
Work Location: Chennai TN
Skill Required: XDigital : BigData and Hadoop Ecosystems
Experience Range in Required Skills: 4-6 years
Job Description: BigData and Hadoop
Essential Skills:
BigData and Hadoop
Big Data Engineer_C
Posted today
Job Viewed
Job Description
Hi All,
Skill: Bigdata Engineer
Exp: 6-9 Years
Location: Pune, Chennai
F2F Interview on 19th Jul 2025. Who are interested please send me your updated resume.
Mandatory Skills: PySpark, spark, python , GCP, SCALA, SQL, Hadoop, Hive, AWS, GCP
Key Responsibilities:
- Design, develop, and maintain scalable data pipelines and ETL workflows using PySpark, Hadoop, and Hive.
- Deploy and manage big data workloads on cloud platforms like GCP and AWS.
- Work closely with cross-functional teams to understand data requirements and deliver high-quality solutions.
- Optimize data processing jobs for performance and cost-efficiency on cloud infrastructure.
- Implement automation and CI/CD pipelines to streamline deployment and monitoring of data workflows.
- Ensure data security, governance, and compliance in cloud environments.
- Troubleshoot and resolve data issues, monitoring job executions and system health.
Mandatory Skills:
- PySpark: Strong experience in developing data processing jobs and ETL pipelines.
- Google Cloud Platform (GCP): Hands-on experience with BigQuery, Dataflow, Dataproc, or similar services.
- Hadoop Ecosystem: Expertise with Hadoop, Hive, and related big data tools.
- AWS: Familiarity with AWS data services like S3, EMR, Glue, or Redshift.
- Strong SQL and data modeling skills.
Good to Have:
- Experience with CI/CD tools and DevOps practices (Jenkins, GitLab, Terraform, etc.).
- Containerization and orchestration knowledge (Docker, Kubernetes).
- Experience with Infrastructure as Code (IaC).
- Knowledge of data governance and data security best practices.
Big Data Lead
Posted today
Job Viewed
Job Description
Investor Services, a leading business line in the Citi Services and offers the full spectrum of capabilities to clients including Custody, Fund Accounting, Investment Accounting, Fund Administration, Middle Office Services, Performance and Risk Analytics, Transfer Agency and Securities lending across multiple jurisdictions.
Investor Services has embarked on significant business growth through digital transformation and technology investments. To meet the objectives of the business, Investor Services Technology is undergoing an exciting platform modernization journey focused on improving agility, scalability, and simplifying the architecture. Program Execution is underway since 2022, with transformational leaders dedicated to this multiyear program and moving to a new next generation architecture on a cloud native platform.
The Applications Development Technology Senior Data Lead Analyst is a senior level position responsible for establishing and implementing canonical data architecture and drive data governance in coordination with the Technology Team. The overall objective of this role is to lead data analysis and programming activities for suite of applications across Investor Services and also drive standardization and modernization of Data strategy and architecture across Services.
Responsibilities:
- Design & implement Spark, Hive, Scala pipelines using Medallion model
- Architect data integration across custody platforms
- Embed automation in metadata, data dictionary, data catalogues, and quality checks
- Ensure reliability, cost optimization, and data governance
- Mentor engineers and stay hands-on with architecture reviews
- Code Quality and Standards: Ensure application design adheres to the overall architecture blueprint. Develop and enforce standards for coding, testing, debugging, and implementation. Conduct code reviews to ensure code quality and compliance with standards.
- Collaboration and Communication: Collaborate with cross-functional teams, including architects, infrastructure engineers, and business analysts, to deliver integrated solutions. Consistently demonstrate clear and concise written and verbal communication.
- Mentoring and Coaching: Serve as an advisor or coach to mid-level developers and analysts, allocating work as necessary and providing guidance on technical best practices.
- Risk Management: Appropriately assess risk when business decisions are made, demonstrating particular consideration for the firm's reputation and safeguarding Citigroup, its clients, and assets. Drive compliance with applicable laws, rules, and regulations, adhering to Policy, applying sound ethical judgment regarding personal behavior, conduct, and business practices, and escalating, managing, and reporting control issues with transparency.
- DevOps Practices: Implement and promote DevOps practices, including continuous integration, continuous delivery, and automated testing.
- Containerization and Orchestration: Utilize Openshift for container orchestration, ensuring applications are scalable, resilient, and easily deployed.
- Version Control: Manage source code using GitHub, following established branching strategies and code review processes.
Skills:
- Big Data Engineering: Hive, Spark, Scala, Delta Lake, performance tuning
- Data Architecture: Medallion, Data Mesh, multi-zone Data Lake
- GenAI for Data: metadata, test data, code gen, lineage
- Cloud & DevOps: AWS/Azure, GitHub Actions, Airflow
- Data Governance: schema evolution, contracts, observability
Education:
- Bachelor's or Master's in CS, Data Engineering, or related field
- 12+ years in data engineering, 5+ in financial services (Custody preferred)
-
Job Family Group:
Technology
-
Job Family:
Applications Development
-
Time Type:
Full time
-
Most Relevant Skills
Please see the requirements listed above.
-
Other Relevant Skills
For complementary skills, please see above and/or contact the recruiter.
-
Citi is an equal opportunity employer, and qualified candidates will receive consideration without regard to their race, color, religion, sex, sexual orientation, gender identity, national origin, disability, status as a protected veteran, or any other characteristic protected by law.
If you are a person with a disability and need a reasonable accommodation to use our search tools and/or apply for a career opportunity review Accessibility at Citi .
View Citi's EEO Policy Statement and the Know Your Rights poster.