273 Data Mining jobs in Noida
Data Mining Analyst
Posted 3 days ago
Job Viewed
Job Description
Position: Data Mining Analyst
Please Note: Only Candidates with 2-3 years of experience should apply.
This is a Delhi-based position and work from office only!
Work Location: Sector 23 Dwarka, Delhi
We are seeking a skilled Data Mining Analyst with expertise in automating data extraction processes from web platforms. The ideal candidate will be experienced in Python, Selenium, Pandas, SQL, and APIs, with the ability to design and implement efficient and scalable data scraping systems. If you have a passion for working with data and a solid understanding of web technologies, we want to hear from you!
Key Responsibilities:
- Design, develop, and maintain robust web scraping solutions to extract structured and unstructured data from various websites and APIs.
- Use tools like Python, Selenium, BeautifulSoup, Scrapy, and Pandas for data scraping and processing.
- Build and manage automated scripts to scrape dynamic websites, including handling JavaScript-driven content.
- Optimize scraping workflows to ensure data extraction is efficient, accurate, and scalable.
- Work with APIs to gather and integrate data, ensuring proper rate limits and authentication handling.
- Clean, preprocess, and store extracted data in databases (SQL) or cloud-based systems.
- Collaborate with data analysts and other stakeholders to provide required data for further analysis and reporting.
- Debug and troubleshoot issues in scraping pipelines and scripts.
- Ensure compliance with ethical data scraping standards, including legal considerations like website terms of use and robots.txt policies.
Required Skills & Qualifications:
- Experience : 2-3 years of hands-on experience in web scraping and data extraction.
Technical Skills :
- Strong proficiency in Python.
- Experience with web scraping frameworks and libraries like Selenium, Scrapy, BeautifulSoup, and Requests.
- Experience with data manipulation libraries like Pandas.
- Familiarity with API integration (REST, GraphQL, etc.).
- Proficiency in SQL for data querying, database design, and managing large datasets.
- Knowledge of JavaScript and front-end technologies to work with dynamic web pages.
- Experience with version control (Git) and collaborative development environments.
Other Skills :
- Problem-solving skills with attention to detail.
- Ability to write clean, maintainable code and automate workflows.
- Good understanding of HTTP, HTML, CSS, and JavaScript.
- Familiarity with cloud services (AWS, Azure, GCP) is a plus.
Python Nice to Have:
- Experience with cloud-based scraping tools or services (e.g., AWS Lambda, Google Cloud Functions).
- Familiarity with distributed scraping and data pipeline management.
- Experience with large-scale data collection and storage systems.
- Knowledge of ethical and legal issues related to web scraping.
About Nuvoretail (
Nuvoretail Enlytical Technologies Private Limited is an e-commerce analytics and automation company. Our proprietary digital shelf analytics and automation platform called Enlytical.ai helps e-commerce brands solve the complexities in today’s e-commerce landscape by offering a unified and all- encompassing business view on the various aspects of e-commerce business. Our platform leverages insights drawn from multiple data points that help our clients win in e-commerce by gaining a competitive edge with data-driven insights for sharper decision-making. The insights cover all aspects of e-commerce such as digital product portfolio analysis, supply chain analytics, e-commerce operations automation, pricing, and competitor benchmarking, and Amazon advertising automation using our proprietary algorithms.
As a leading e-commerce service provider, we offer the most comprehensive end-to-end
e-commerce solutions to brands, both in India and abroad. Right from preparing a road map to writing our client’s e- commerce success story to assisting them In increasing their online sales, we do everything via our diverse e-commerce services and bespoke strategies and technology. Our services span across the brand’s e-commerce enablement including content and digital asset creation for product listing, On Platform, and Off Platform marketing services with deep expertise in Amazon Marketing Services (AMS), Amazon SEO through keyword research, e-Commerce operations across various e-commerce platforms, website development, social media marketing, and AI-enabled e-Commerce MIS Dashboards.
Awards & Recognition:
Thanks to the faith reposed on us by our clients, NuvoRetail has been featured as "The Most Promising Ecommerce Technology Service Providers in India 2020” by CIOReviewIndia Magazine. Our leadership is often acknowledged by leading e-commerce services, digital marketing, consulting, and other e- commerce programs around the world. We are now one of the very few companies in India that have become an Amazon Ads Advanced partner.
Remote Senior Data Mining Engineer
Posted 10 days ago
Job Viewed
Job Description
Responsibilities:
- Design, develop, and implement advanced data mining algorithms and models.
- Explore and analyze large, complex datasets to identify patterns, trends, and insights.
- Develop predictive models and machine learning solutions.
- Collaborate with data scientists and stakeholders to define data mining requirements.
- Clean, transform, and prepare data for analysis.
- Develop and maintain data pipelines and ETL processes.
- Utilize big data technologies and platforms for data processing and analysis.
- Evaluate and improve the performance of data mining models.
- Communicate findings and recommendations through clear visualizations and reports.
- Stay current with the latest advancements in data mining, machine learning, and AI.
- Master's or Ph.D. in Computer Science, Data Science, Statistics, or a related quantitative field.
- Minimum of 5 years of experience in data mining, machine learning, or data science.
- Proficiency in programming languages such as Python or R.
- Strong experience with SQL and NoSQL databases.
- Hands-on experience with big data technologies (e.g., Hadoop, Spark, Hive).
- Expertise in statistical modeling and machine learning techniques.
- Experience with data visualization tools (e.g., Tableau, Power BI).
- Excellent analytical, problem-solving, and critical thinking skills.
- Strong communication and collaboration abilities for remote work.
- Ability to work independently and manage project timelines effectively.
- This role is fully remote, supporting teams globally and potentially interacting with operations in Ghaziabad, Uttar Pradesh, IN .
Remote Senior Data Scientist - Mining Analytics
Posted 15 days ago
Job Viewed
Job Description
Responsibilities:
- Design, develop, and implement machine learning models for applications such as predictive maintenance, resource estimation, ore grade optimization, and operational efficiency.
- Analyze large and complex datasets from various sources (e.g., geological surveys, sensor data, production logs, financial records).
- Develop data pipelines for cleaning, transforming, and integrating data from disparate systems.
- Conduct exploratory data analysis to identify trends, patterns, and anomalies relevant to mining operations.
- Build and maintain robust data visualizations and dashboards to communicate findings to stakeholders.
- Collaborate with subject matter experts to understand business needs and translate them into data science solutions.
- Evaluate the performance of models and algorithms, and iterate for improvement.
- Stay current with the latest advancements in data science, machine learning, and relevant industry technologies.
- Mentor junior data scientists and contribute to the team's technical growth.
- Present complex analytical results and recommendations clearly and concisely to both technical and non-technical audiences.
- Ensure data quality and integrity throughout the analytical process.
- Contribute to the strategic direction of data analytics initiatives within the company.
- Master's or Ph.D. in Data Science, Statistics, Computer Science, Engineering, or a related quantitative field.
- 5+ years of professional experience in data science and machine learning.
- Proven experience in developing and deploying machine learning models in production.
- Strong proficiency in programming languages such as Python or R, and relevant libraries (e.g., scikit-learn, TensorFlow, PyTorch).
- Expertise in SQL and experience with big data technologies (e.g., Spark, Hadoop).
- Familiarity with cloud platforms (AWS, Azure, GCP) and their data science services.
- Experience in the mining or natural resources industry, with knowledge of geological or operational data, is highly preferred.
- Excellent analytical, problem-solving, and critical thinking skills.
- Strong communication and collaboration skills, with the ability to explain technical concepts effectively.
Data Science Intern
Posted today
Job Viewed
Job Description
Did you notice a shortage of food at supermarkets during covid? Have you heard about the recent issues in the global shipping industry? or perhaps you’ve heard about the shortages of microchips? These problems are called supply chain disruptions. They have been increasing in frequency and severity. Supply chain disruptions are threatening our very way of life.
Our vision is to advance society’s capacity to withstand shocks and stresses. Kavida.ai believes the only way to ensure security is through supply chain resiliency. We are on a mission to help companies proactively manage disruption supply chain disruption risks using integrated data.
Our Story
In March 2020 over 35 academics, data scientists, students, and software engineering volunteers came together to address the food shortage issues caused by the pandemic - Covid19foodsupply.com. A core team of 9 was formed and spun off into a startup and the rest is history.
Our investors include one of the world's largest supply chain quality & compliance monitoring companies, a £1.25bn apparel manufacturer, and some very impressive angel investors.
Social Impact:
Social impact is in our DNA. We believe private sector innovation is the only way to address social problems at scale. If we achieve our mission, humanity will always have access to its essential goods for sustenance. No more shortages of food, PPE, medicine, etc.
Our Culture:
Idea Meritocracy:
The best ideas win. We only care about what is right, not who is right. We know arriving at the best answer requires constructive tension. Sometimes it can get heated but it's never personal. Everyone contributes to better ideas knowing they will be heard but also challenged.
Drivers Not Passengers:
We think as owners who drive the bus, not as passengers. We are self-starters and never wait for instructions. We are hungry for autonomy, trust, and responsibility. Everyone is a leader because we know leadership is a trait, not a title. Leaders drive growth and navigate the chaos.
We Figure Out The Answers:
We trust our ability to figure stuff out. We do not need all the information to start answering the question. We can connect the dots and answer difficult questions with logic.
Customer & Mission Obsessed:
Our customers are our heroes and we are obsessed with helping them. We are obsessed with; understanding their supply chains better, resolving their biggest headaches, and advancing their competitiveness.
Learning and growth
We all take personal responsibility for becoming smarter, wiser, more skilled, happier. We are obsessed with learning about our industry and improving our own skills. We are obsessed with our personal growth; to become more.
Job Description:
As a member of our Research team, you will be responsible for researching, developing, and coding Agents using state-of-the-art LLM's with automated pipelines.
- Write code for the development of our ML engines and micro-services pipelines.
- use, optimize, train, and evaluate state-of-the-art GPT models.
- research and Develop Agentic pipelines using LLM's.
- research and develop RAG based pipeline using vector DB's .
Essential Requirements:
- prompt engineering and Agentic LLm frameworks like langchain/llama index
- good enough undersanding of vectors/tensors and RAG pipelines
- Knowledge of building NLP systems using transfer learning or building custom NLP systems from scratch using TensorFlow or PyTorch.
- In-depth knowledge of DSA, async, python, and containers.
- Knowledge of transformers and NLP techniques is essential, and deployment experience is a significant advantage.
Salary Range: ₹15000 - ₹25000
We are offering a full-time internship position to final-year students. The internship will last for an initial period of 6-12 months before converting to a full-time job, depending on suitability for both parties. If the applicant is a student who needs to return to university, they can continue with the program on a part-time basis.
Data Science Intern
Posted today
Job Viewed
Job Description
ZeTheta Algorithms Private Limited is a FinTech start-up which has been recently set up and is developing innovative AI tools.
the Role
As a Data Scientist intern, you will work on cutting-edge projects involving financial data analysis, investment research, and risk modelling. You will have the opportunity to engage in multiple mini-projects or take up a focused innovation-based research project. The project experience is designed to provide practical exposure to data science in the context of asset management, trading, and financial technology. We provide problem statements, methodology and after you submit your solution to develop the solutions/ model, we also showcase to you sample solution. You can use our sample solution to modify your project submission and expand further based on suggestions given in our sample solution. You can opt for your own research based data science solution to develop/ model.
Responsibilities
- Conduct data cleaning, wrangling, and pre-processing for financial datasets.
- Assist investment teams in equity research, fixed income research, portfolio management, and economic analysis.
- Apply statistical techniques to financial problems such as credit risk modelling, probability of default, and value-at-risk estimation.
- Work with big data sources including financial reports, macroeconomic datasets, and alternative investment data.
- Use either one – Python, Excel or R to analyse, visualize, and model financial data.
- Participate in research projects related to quantitative trading, financial derivatives, and portfolio optimization.
Who Should Apply?
- Any student even without coding skills can upskill (self learning) to develop Data Science Solutions. Some basic knowledge of Excel or Python or R script can help complete the projects quicker. We permit the use of all LLMs/ NLPs to help students to develop the solutions.
- Strong problem-solving and analytical skills.
- Able to self-learn and work independently in a remote, flexible environment.
Internship Details
- Duration: Option of 1 month, 2 month, 3 month, 4 month or 6 months
- Timing: Self-paced.
- Type: Unpaid
Data Science Intern
Posted today
Job Viewed
Job Description
NLP Data Science Intern
Did you notice a shortage of food at supermarkets during covid? Have you heard about the recent issues in the global shipping industry? or perhaps you've heard about the shortages of microchips? These problems are called supply chain disruptions. They have been increasing in frequency and severity. Supply chain disruptions are threatening our very way of life.
Our vision is to advance society's capacity to withstand shocks and stresses. Kavida.ai believes the only way to ensure security is through supply chain resiliency. We are on a mission to help companies proactively manage disruption supply chain disruption risks using integrated data.
Our Story
In March 2020 over 35 academics, data scientists, students, and software engineering volunteers came together to address the food shortage issues caused by the pandemic - Covid19foodsupply.com. A core team of 9 was formed and spun off into a startup and the rest is history.
Our investors include one of the world's largest supply chain quality & compliance monitoring companies, a £1.25bn apparel manufacturer, and some very impressive angel investors.
Social Impact:
Social impact is in our DNA. We believe private sector innovation is the only way to address social problems at scale. If we achieve our mission, humanity will always have access to its essential goods for sustenance. No more shortages of food, PPE, medicine, etc.
Our Culture:
Idea Meritocracy:
The best ideas win. We only care about what is right, not who is right. We know arriving at the best answer requires constructive tension. Sometimes it can get heated but it's never personal. Everyone contributes to better ideas knowing they will be heard but also challenged.
Drivers Not Passengers:
We think as owners who drive the bus, not as passengers. We are self-starters and never wait for instructions. We are hungry for autonomy, trust, and responsibility. Everyone is a leader because we know leadership is a trait, not a title. Leaders drive growth and navigate the chaos.
We Figure Out The Answers:
We trust our ability to figure stuff out. We do not need all the information to start answering the question. We can connect the dots and answer difficult questions with logic.
Customer & Mission Obsessed:
Our customers are our heroes and we are obsessed with helping them. We are obsessed with; understanding their supply chains better, resolving their biggest headaches, and advancing their competitiveness.
Learning and growth
We all take personal responsibility for becoming smarter, wiser, more skilled, happier. We are obsessed with learning about our industry and improving our own skills. We are obsessed with our personal growth; to become more.
Job Description:
As a member of our Research team, you will be responsible for researching, developing, and coding Agents using state-of-the-art LLM's with automated pipelines.
- Write code for the development of our ML engines and micro-services pipelines.
- use, optimize, train, and evaluate state-of-the-art GPT models.
- research and Develop Agentic pipelines using LLM's.
- research and develop RAG based pipeline using vector DB's .
Essential Requirements:
- prompt engineering and Agentic LLm frameworks like langchain/llama index
- good enough undersanding of vectors/tensors and RAG pipelines
- Knowledge of building NLP systems using transfer learning or building custom NLP systems from scratch using TensorFlow or PyTorch.
- In-depth knowledge of DSA, async, python, and containers.
- Knowledge of transformers and NLP techniques is essential, and deployment experience is a significant advantage.
Salary Range: ₹15000 - ₹25000
We are offering a full-time internship position to final-year students. The internship will last for an initial period of 6-12 months before converting to a full-time job, depending on suitability for both parties. If the applicant is a student who needs to return to university, they can continue with the program on a part-time basis.
Data Science Specialist
Posted today
Job Viewed
Job Description
Job Title: Data Scientist-Customer Analytics About the Role:
We are seeking a skilled Data Scientist to join our team. The ideal candidate will have expertise in data analysis, machine learning, and statistical techniques.
Responsibilities:- Develop and implement data analytics solutions to drive business growth.
- Analyze customer behavior data to identify trends and opportunities.
- Design and deploy machine learning models to improve customer engagement and retention.
To be successful in this role, you will need:
- Advanced knowledge of programming languages such as SQL, Python, and R.
- Experience with machine learning algorithms and statistical techniques.
- Strong analytical and problem-solving skills.
- Excellent communication and collaboration skills.
We offer a competitive salary, comprehensive benefits package, and opportunities for career growth and development.
EEO Employer:We are an equal opportunity employer and welcome applications from all qualified candidates.
Be The First To Know
About the latest Data mining Jobs in Noida !
Data Science Intern
Posted today
Job Viewed
Job Description
About the Company
ZeTheta Algorithms Private Limited is a FinTech start-up which has been recently set up and is developing innovative AI tools.
About the Role
As a Data Scientist intern, you will work on cutting-edge projects involving financial data analysis, investment research, and risk modelling. You will have the opportunity to engage in multiple mini-projects or take up a focused innovation-based research project. The project experience is designed to provide practical exposure to data science in the context of asset management, trading, and financial technology. We provide problem statements, methodology and after you submit your solution to develop the solutions/ model, we also showcase to you sample solution. You can use our sample solution to modify your project submission and expand further based on suggestions given in our sample solution. You can opt for your own research based data science solution to develop/ model.
Responsibilities
- Conduct data cleaning, wrangling, and pre-processing for financial datasets.
- Assist investment teams in equity research, fixed income research, portfolio management, and economic analysis.
- Apply statistical techniques to financial problems such as credit risk modelling, probability of default, and value-at-risk estimation.
- Work with big data sources including financial reports, macroeconomic datasets, and alternative investment data.
- Use either one – Python, Excel or R to analyse, visualize, and model financial data.
- Participate in research projects related to quantitative trading, financial derivatives, and portfolio optimization.
Who Should Apply?
- Any student even without coding skills can upskill (self learning) to develop Data Science Solutions. Some basic knowledge of Excel or Python or R script can help complete the projects quicker. We permit the use of all LLMs/ NLPs to help students to develop the solutions.
- Strong problem-solving and analytical skills.
- Able to self-learn and work independently in a remote, flexible environment.
Internship Details
- Duration: Option of 1 month, 2 month, 3 month, 4 month or 6 months
- Timing: Self-paced.
- Type: Unpaid
Data Science Intern
Posted today
Job Viewed
Job Description
NLP Data Science Intern
Did you notice a shortage of food at supermarkets during covid? Have you heard about the recent issues in the global shipping industry? or perhaps you've heard about the shortages of microchips? These problems are called supply chain disruptions. They have been increasing in frequency and severity. Supply chain disruptions are threatening our very way of life.
Our vision is to advance society's capacity to withstand shocks and stresses. Kavida.ai believes the only way to ensure security is through supply chain resiliency. We are on a mission to help companies proactively manage disruption supply chain disruption risks using integrated data.
Our Story
In March 2020 over 35 academics, data scientists, students, and software engineering volunteers came together to address the food shortage issues caused by the pandemic - A core team of 9 was formed and spun off into a startup and the rest is history.
Our investors include one of the world's largest supply chain quality & compliance monitoring companies, a £1.25bn apparel manufacturer, and some very impressive angel investors.
Social Impact:
Social impact is in our DNA. We believe private sector innovation is the only way to address social problems at scale. If we achieve our mission, humanity will always have access to its essential goods for sustenance. No more shortages of food, PPE, medicine, etc.
Our Culture :
Idea Meritocracy:
The best ideas win. We only care about what is right, not who is right. We know arriving at the best answer requires constructive tension. Sometimes it can get heated but it's never personal. Everyone contributes to better ideas knowing they will be heard but also challenged.
Drivers Not Passengers:
We think as owners who drive the bus, not as passengers. We are self-starters and never wait for instructions. We are hungry for autonomy, trust, and responsibility. Everyone is a leader because we know leadership is a trait, not a title. Leaders drive growth and navigate the chaos.
We Figure Out The Answers:
We trust our ability to figure stuff out. We do not need all the information to start answering the question. We can connect the dots and answer difficult questions with logic.
Customer & Mission Obsessed:
Our customers are our heroes and we are obsessed with helping them. We are obsessed with; understanding their supply chains better, resolving their biggest headaches, and advancing their competitiveness.
Learning and growth
We all take personal responsibility for becoming smarter, wiser, more skilled, happier. We are obsessed with learning about our industry and improving our own skills. We are obsessed with our personal growth; to become more.
Job Description:
As a member of our Research team, you will be responsible for researching, developing, and coding Agents using state-of-the-art LLM's with automated pipelines.
- Write code for the development of our ML engines and micro-services pipelines.
- use, optimize, train, and evaluate state-of-the-art GPT models.
- research and Develop Agentic pipelines using LLM's.
- research and develop RAG based pipeline using vector DB's .
Essential Requirements:
- prompt engineering and Agentic LLm frameworks like langchain/llama index
- good enough undersanding of vectors/tensors and RAG pipelines
- Knowledge of building NLP systems using transfer learning or building custom NLP systems from scratch using TensorFlow or PyTorch.
- In-depth knowledge of DSA, async, python, and containers.
- Knowledge of transformers and NLP techniques is essential, and deployment experience is a significant advantage.
Salary Range: 15000 - 25000
We are offering a full-time internship position to final-year students. The internship will last for an initial period of 6-12 months before converting to a full-time job, depending on suitability for both parties. If the applicant is a student who needs to return to university, they can continue with the program on a part-time basis.