Found 165 repositories(showing 30)
Exercises answers to the book "machine-learning" written by Zhou Zhihua。周志华《机器学习》课后习题,个人解答。各算法都拿numpy和pandas实现了一遍
shreyasharma04
🤖 HealthCare ChatBot Major -1 (4th year - 7th semester) Health Care Chat-Bot is a Healthcare Domain Chatbot to simulate the predictions of a General Physician. ChatBot can be described as software that can chat with people using artificial intelligence. These software are used to perform tasks such as quickly responding to users, informing them, helping to purchase products and providing better service to customers. We have made a healthcare based chatbot. The three main areas where chatbots can be used are diagnostics, patient engagement outside medical facilities, and mental health. In our major we are working on diagnostic. 📃 Brief A chatbot is an artificially intelligent creature which can converse with humans. This could be text-based, or a spoken conversation. In our project we will be using Python as it is currently the most popular language for creating an AI chatbot. In the middle of AI chatbot, architecture is the Natural Language Processing (NLP) layer. This project aims to build an user-friendly healthcare chatbot which facilitates the job of a healthcare provider and helps improve their performance by interacting with users in a human-like way. Through chatbots one can communicate with text or voice interface and get reply through artificial intelligence Typically, a chat bot will communicate with a real person. Chat bots are used in applications such as E-commerce customer service, Call centres, Internet gaming,etc. Chatbots are programs built to automatically engage with received messages. Chatbots can be programmed to respond the same way each time, to respond differently to messages containing certain keywords and even to use machine learning to adapt their responses to fit the situation. A developing number of hospitals, nursing homes, and even private centres, presently utilize online Chatbots for human services on their sites. These bots connect with potential patients visiting the site, helping them discover specialists, booking their appointments, and getting them access to the correct treatment. In any case, the utilization of artificial intelligence in an industry where individuals’ lives could be in question, still starts misgivings in individuals. It brings up issues about whether the task mentioned above ought to be assigned to human staff. This healthcare chatbot system will help hospitals to provide healthcare support online 24 x 7, it answers deep as well as general questions. It also helps to generate leads and automatically delivers the information of leads to sales. By asking the questions in series it helps patients by guiding what exactly he/she is looking for. 📜 Problem Statement During the pandemic, it is more important than ever to get your regular check-ups and to continue to take prescription medications. The healthier you are, the more likely you are to recover quickly from an illness. In this time patients or health care workers within their practice, providers are deferring elective and preventive visits, such as annual physicals. For some, it is not possible to consult online. In this case, to avoid false information, our project can be of help. 📇 Features Register Screen. Sign-in Screen. Generates database for user login system. Offers you a GUI Based Chatbot for patients for diagnosing. [A pragmatic Approach for Diagnosis] Reccomends an appropriate doctor to you for the following symptom. 📜 Modules Used Our program uses a number of python modules to work properly: tkinter os webbrowser numpy pandas matplotlib 📃 Algorithm We have used Decision tree for our health care based chat bot. Decision Tree is a Supervised learning technique that can be used for both classification and Regression problems, but mostly it is preferred for solving Classification problems. It is a tree-structured classifier, where internal nodes represent the features of a dataset, branches represent the decision rules and each leaf node represents the outcome.It usually mimic human thinking ability while making a decision, so it is easy to understand. :suspect: Project Members Anushka Bansal - 500067844 - R164218014 Shreya Sharma - 500068573 - R164218070 Silvi - 500069092 - R164218072 Ishika Agrawal - 500071154 - R164218097
mudigosa
Image Classifier Going forward, AI algorithms will be incorporated into more and more everyday applications. For example, you might want to include an image classifier in a smartphone app. To do this, you'd use a deep learning model trained on hundreds of thousands of images as part of the overall application architecture. A large part of software development in the future will be using these types of models as common parts of applications. In this project, you'll train an image classifier to recognize different species of flowers. You can imagine using something like this in a phone app that tells you the name of the flower your camera is looking at. In practice, you'd train this classifier, then export it for use in your application. We'll be using this dataset of 102 flower categories. When you've completed this project, you'll have an application that can be trained on any set of labelled images. Here your network will be learning about flowers and end up as a command line application. But, what you do with your new skills depends on your imagination and effort in building a dataset. This is the final Project of the Udacity AI with Python Nanodegree Prerequisites The Code is written in Python 3.6.5 . If you don't have Python installed you can find it here. If you are using a lower version of Python you can upgrade using the pip package, ensuring you have the latest version of pip. To install pip run in the command Line python -m ensurepip -- default-pip to upgrade it python -m pip install -- upgrade pip setuptools wheel to upgrade Python pip install python -- upgrade Additional Packages that are required are: Numpy, Pandas, MatplotLib, Pytorch, PIL and json. You can donwload them using pip pip install numpy pandas matplotlib pil or conda conda install numpy pandas matplotlib pil In order to intall Pytorch head over to the Pytorch site select your specs and follow the instructions given. Viewing the Jyputer Notebook In order to better view and work on the jupyter Notebook I encourage you to use nbviewer . You can simply copy and paste the link to this website and you will be able to edit it without any problem. Alternatively you can clone the repository using git clone https://github.com/fotisk07/Image-Classifier/ then in the command Line type, after you have downloaded jupyter notebook type jupyter notebook locate the notebook and run it. Command Line Application Train a new network on a data set with train.py Basic Usage : python train.py data_directory Prints out current epoch, training loss, validation loss, and validation accuracy as the netowrk trains Options: Set direcotry to save checkpoints: python train.py data_dor --save_dir save_directory Choose arcitecture (alexnet, densenet121 or vgg16 available): pytnon train.py data_dir --arch "vgg16" Set hyperparameters: python train.py data_dir --learning_rate 0.001 --hidden_layer1 120 --epochs 20 Use GPU for training: python train.py data_dir --gpu gpu Predict flower name from an image with predict.py along with the probability of that name. That is you'll pass in a single image /path/to/image and return the flower name and class probability Basic usage: python predict.py /path/to/image checkpoint Options: Return top K most likely classes: python predict.py input checkpoint ---top_k 3 Use a mapping of categories to real names: python predict.py input checkpoint --category_names cat_To_name.json Use GPU for inference: python predict.py input checkpoint --gpu Json file In order for the network to print out the name of the flower a .json file is required. If you aren't familiar with json you can find information here. By using a .json file the data can be sorted into folders with numbers and those numbers will correspond to specific names specified in the .json file. Data and the json file The data used specifically for this assignemnt are a flower database are not provided in the repository as it's larger than what github allows. Nevertheless, feel free to create your own databases and train the model on them to use with your own projects. The structure of your data should be the following: The data need to comprised of 3 folders, test, train and validate. Generally the proportions should be 70% training 10% validate and 20% test. Inside the train, test and validate folders there should be folders bearing a specific number which corresponds to a specific category, clarified in the json file. For example if we have the image a.jpj and it is a rose it could be in a path like this /test/5/a.jpg and json file would be like this {...5:"rose",...}. Make sure to include a lot of photos of your catagories (more than 10) with different angles and different lighting conditions in order for the network to generalize better. GPU As the network makes use of a sophisticated deep convolutional neural network the training process is impossible to be done by a common laptop. In order to train your models to your local machine you have three options Cuda -- If you have an NVIDIA GPU then you can install CUDA from here. With Cuda you will be able to train your model however the process will still be time consuming Cloud Services -- There are many paid cloud services that let you train your models like AWS or Google Cloud Coogle Colab -- Google Colab gives you free access to a tesla K80 GPU for 12 hours at a time. Once 12 hours have ellapsed you can just reload and continue! The only limitation is that you have to upload the data to Google Drive and if the dataset is massive you may run out of space. However, once a model is trained then a normal CPU can be used for the predict.py file and you will have an answer within some seconds. Hyperparameters As you can see you have a wide selection of hyperparameters available and you can get even more by making small modifications to the code. Thus it may seem overly complicated to choose the right ones especially if the training needs at least 15 minutes to be completed. So here are some hints: By increasing the number of epochs the accuracy of the network on the training set gets better and better however be careful because if you pick a large number of epochs the network won't generalize well, that is to say it will have high accuracy on the training image and low accuracy on the test images. Eg: training for 12 epochs training accuracy: 85% Test accuracy: 82%. Training for 30 epochs training accuracy 95% test accuracy 50%. A big learning rate guarantees that the network will converge fast to a small error but it will constantly overshot A small learning rate guarantees that the network will reach greater accuracies but the learning process will take longer Densenet121 works best for images but the training process takes significantly longer than alexnet or vgg16 *My settings were lr=0.001, dropoup=0.5, epochs= 15 and my test accuracy was 86% with densenet121 as my feature extraction model. Pre-Trained Network The checkpoint.pth file contains the information of a network trained to recognise 102 different species of flowers. I has been trained with specific hyperparameters thus if you don't set them right the network will fail. In order to have a prediction for an image located in the path /path/to/image using my pretrained model you can simply type python predict.py /path/to/image checkpoint.pth Contributing Please read CONTRIBUTING.md for the process for submitting pull requests. Authors Shanmukha Mudigonda - Initial work Udacity - Final Project of the AI with Python Nanodegree
Devtown-India
Over the past decade, bicycle-sharing systems have been growing in number and popularity in cities across the world. Bicycle-sharing systems allow users to rent bicycles on a very short-term basis for a price. This allows people to borrow a bike from point A and return it at point B, though they can also return it to the same location if they'd like to just go for a ride. Regardless, each bike can serve several users per day. Thanks to the rise in information technologies, it is easy for a user of the system to access a dock within the system to unlock or return bicycles. These technologies also provide a wealth of data that can be used to explore how these bike-sharing systems are used. In this project, you will use data provided by Motivate, a bike share system provider for many major cities in the United States, to uncover bike share usage patterns. You will compare the system usage between three large cities: Chicago, New York City, and Washington, DC. Day:1 In this project, Students will make use of Python to explore data related to bike share systems for three major cities in the United States—Chicago, New York City, and Washington. You will write code to import the data and answer interesting questions about it by computing descriptive statistics. They will also write a script that takes in raw input to create an interactive experience in the terminal to present these statistics. Technologies that will be covered are Numpy, Pandas, Matplotlib, Seaborn, Jupyter notebook. We will be giving the students a deep dive into the Data Analytical process Day:2 We will be giving the students an insight into one of the major fields of Machine Learning ie. Time Series forcasting we will be taking them through the relevant theory and make them understand of the importance and different techniques that are available to deal with it. After that we will be working hands on the bike share data set implementing different algorithms and understanding them to the core We aim to provide students an insight into what exactly is the job of a data analyst and get them familiarise to how does the entire data analysis process work. The session will be hosted by Shaurya Sinha a data analyst at Jio and Parag Mittal Software engineer at Microsoft.
swami-hai-ham
This project focuses on analyzing Spotify track data using data analytics techniques and Python libraries such as Matplotlib, Seaborn, Pandas, and NumPy. The goal is to derive meaningful insights and answer various questions about the tracks based on the available dataset.
Predicted if students will answer incorrectly or correctly on questions with an Artificial Intelligent tutor developed at RIIID Labs. Utilized Postgres to store the AI tutor data into a SQL database. Used Tableau and Seaborn for data visualization and understanding, Engineered new features from data using Pandas and Numpy to improve model predictions, Predicted student answers with Scikit-learn models, the best model being a Random Forest Classifier with 68% accuracy.
kristinvmartin
This is a dimensional data warehouse that seeks to provide insights into the raw data that FEMA provides publicly for its Individual and Housing Program. I used Jupyter Notebook, Python (Pandas, NumPy, Pyodbc), and SQL to perform ETL on the dataset, loading the warehouse based on the schema I designed. I created visualizations using Tableau from the data warehouse to provide targeted insights that answered the key business questions of the project (see README file). Note: If etl_IHP.ipynb is throwing an error on load, it can be viewed using nbviewer by following this link: https://nbviewer.jupyter.org/github/kristinvmartin/datawarehouse-fema-bu/blob/main/etl_IHP.ipynb, or you can view the CODEONLY file, which has the scripts without the output.
vnsgamer
Introduction : This data set is a masked data set which is similar to what data analysts at Uber handle. Solving this assignment will give you an idea about how problems are systematically solved using EDA and data visualisation. Business Understanding : You may have some experience of travelling to and from the airport. Have you ever used Uber or any other cab service for this travel? Did you at any time face the problem of cancellation by the driver or non-availability of cars? Well, if these are the problems faced by customers, these very issues also impact the business of Uber. If drivers cancel the request of riders or if cars are unavailable, Uber loses out on its revenue. As an analyst, you decide to address the problem Uber is facing - driver cancellation and non-availability of cars leading to loss of potential revenue. Business Objectives : The aim of analysis is to identify the root cause of the problem (i.e. cancellation and non-availability of cars) and recommend ways to improve the situation. As a result of your analysis, you should be able to present to the client the root cause(s) and possible hypotheses of the problem(s) and recommend ways to improve them. There are six attributes associated with each request made by a customer: 1. Request id: A unique identifier of the request 2. Time of request: The date and time at which the customer made the trip request 3. Drop-off time: The drop-off date and time, in case the trip was completed 4. Pick-up point: The point from which the request was made 5. Driver id: The unique identification number of the driver 6. Status of the request: The final status of the trip, that can be either completed, cancelled by the driver or no cars available Note: For this assignment, only the trips to and from the airport are being considered. Results Expected : 1. Visually identify the most pressing problems for Uber. Hint: Create plots to visualise the frequency of requests that get cancelled or show 'no cars available'; identify the most problematic types of requests (city to airport / airport to city etc.) and the time slots (early mornings, late evenings etc.) using plots. 2. Find out the gap between supply and demand and show the same using plots. a. Find the time slots when the highest gap exists b. Find the types of requests (city-airport or airport-city) for which the gap is the most severe in the identified time slots 3. What do you think is the reason for this issue for the supply-demand gap? Write the answer in less than 100 words. You may accompany the write-up with plot(s). 4. Recommend some ways to resolve the supply-demand gap. IDE : jupyter notebook Language : Python Libraries : Numpy, Pandas, Matplotlib, Seaborn Please do explore the dataset further to your own and see what kind of other insights you can get across various other columns.
Mayank-Bhatt22
This project analyzes the Titanic dataset using Pandas, NumPy, Seaborn, and Matplotlib. It answers 20 data analysis questions covering data quality, univariate, bivariate, and multivariate analysis to explore survival patterns based on gender, age, class, fare, family status, and embarkation.
sanehkr08
This is a repository of my data analysis project on Startup funding data sets. I have used python and its libraries like NumPy, Pandas, Matplotlib etc. to analyze the data and find out answers for the provided questions. I am uploading my Jupyter notebook files and pdf of explanation of solution with insights extracted from the data.
Divyanshu960
In this project, I will examine a dataset on the pay for several data science professions that I downloaded from Kaggle. Using this dataset, I’ll provide answers to a number of questions, including: Does a person’s work role have an impact on their salary? Does the sort of employment affect the wage earned? Does the size and location of the company have an impact on the salary paid? e.t.c,. To answer these questions, I’ll be utilizing Python libraries such as Numpy and Pandas for computations, and Matplotlib and Seaborn for visualizations.
Python is point of fact the Next Big Thing to investigate. There is no need to be worried about its worth, profession possibilities, or accessible positions. Python's commitment to the advancement of your calling is huge, as its notoriety among designers and different areas is step by step waning. Python is "the one" for an assortment of reasons. It's a straightforward pre-arranged language that is not difficult to get. Subsequently, the general improvement time for the task code is diminished. It accompanies an assortment of structures and APIs that assistance with information examination, perception, and control. Employment opportunities in Python While India has a critical interest for Python engineers, the stock is very restricted. We'll utilize a HR master articulation to validate this. For both Java and Python, the expert was relied upon to employ ten developers. For Java, they got over 100 fantastic resumes, however just eight for Python. In this way, while they needed to go through an extensive method to get rid of resilient people, they had no real option except to acknowledge those eight individuals with Python. What does this say about the circumstance to you? Regardless of Python's straightforward language structure, we desperately need more individuals in India to update their abilities. This is the reason learning Python is a particularly colossal opportunity for Indians. With regards to work openings, there may not be numerous for Python in India. Notwithstanding, we have countless assignments accessible per Python developer. In the relatively recent past, one of India's unicorn programming organizations was stood up to with an issue. It had gotten a $200 million (Rs. 1200 crore) arrangement to develop an application store for a significant US bank. Be that as it may, the organization required talented Python developers. Since Python was the best language for the undertaking, it wound up paying a gathering of independent Python developers in the United States multiple times the charging sum. For sure and Naukri, for instance, have 20,000 to 50,000 Python work postings, showing that Python vocation openings in India are copious. It is an insightful choice to seek after a profession in Python. The diagrams underneath show the absolute number of occupation advertisements for the most well known programming dialects. Python Job Descriptions Anyway, what sorts of work would you be able to get in the event that you know Python? Python's degree is broad in information science and investigation, first off. Customers regularly demand that secret examples be separated from their informational indexes. In AI and man-made reasoning, it is additionally suggested. Python is a top choice among information researchers. Furthermore, we figured out how Python is used in web advancement, work area applications, information examination, and organization programming in our article on Python applications. Python Job Profiles With Python on your resume, you might wind up with one of the accompanying situations in a presumed organization: 1. Programmer Investigate client necessities Compose and test code Compose functional documentation Counsel customers and work intimately with other staff Foster existing projects 2. Senior Software Engineer Foster excellent programming engineering Mechanize assignments by means of prearranging and different apparatuses Survey and troubleshoot code Perform approval and confirmation testing Carry out form control and configuration designs 3. DevOps Engineer Send refreshes and fixes Break down and resolve specialized issues Plan systems for support and investigating Foster contents to mechanize representation Convey Level 2 specialized help 4. Information Scientist Recognize information sources and mechanize the assortment Preprocess information and dissect it to find patterns Plan prescient models and ML calculations Perform information representation Propose answers for business challenges 5. Senior Data Scientist Manage junior information experts Construct logical devices to create knowledge, find designs, and foresee conduct Execute ML and measurements based calculations Propose thoughts for utilizing had information Impart discoveries to colleagues While many significant firms are as yet utilizing Java, Python is a more seasoned yet at the same time well known innovation. Python's future is splendid, on account of: 1.Artificial Intelligence (AI): Machine knowledge is alluded to as man-made consciousness. This is as a conspicuous difference to the regular astuteness that people and different creatures have. It is one of the most up to date advances that is clearing the globe. With regards to AI, Python is one of the main dialects that rings a bell; truth be told, it is probably the most ideally equipped language for the work. We have different structures, libraries, and devices devoted to permitting AI to swap human work for this objective. It supports this, however it additionally further develops productivity and precision. Discourse acknowledgment frameworks, self-driving vehicles, and other AI-based advancements are models. The accompanying devices and libraries transport for these parts of AI: AI – PyML, PyBrain, scikit-learn, MDP Toolkit, GraphLab Create, MIPy General AI – pyDatalog, AIMA, EasyAI, SimpleAI Neural Networks – PyAnn, pyrenn, ffnet, neurolab Normal Language and Text Processing – Quepy, NLTK, genism 2. Enormous Data Enormous Data is the term for informational collections so voluminous and complex that conventional information handling application programming is insufficient in managing them. Python has assisted Big Data with developing, its libraries permit us to break down and work with a lot of information across groups: Pandas scikit-learn NumPy SciPy GraphLab Create IPython Bokeh Agate PySpark Dask 3. Systems administration Python additionally allows us to design switches and switches, and perform other organization mechanization undertakings cost-viably. For this, we have the accompanying Python libraries: Ansible Netmiko NAPALM(Network Automation and Programmability Abstraction Layer with Multivendor Support) Pyeapi JunosPyEZ PySNM Paramiko SSH Python Course
atharvjairath
Uses Zomato Data to Answer some Questions Using Numpy,Pandas and Matplot
vincentmuiruri
- Analyzing Penda Health medical data using Python Pandas and Numpy to answer the following questions:
samlexrod
Posed a question about a dataset, then used NumPy and Pandas to answer that question based on the data and created a report to share the results.
gileiva
A trivia game written in Python by Gi Leiva. The game generates questions and answers from different datasets with general information, using the Pandas and Numpy libraries.
AbhishekRatho-2004
This is project on EDA on the Netflix dataset from the kaggle ,It gives answers to many questions and technologies used are numpy,pandas ,matplolib and seaborn
madhura0106
An application that provides accurate natural language answers to the questions asked by a user about the given image. Technology : Deep Learning Tools/Libraries : keras, numpy, pandas,sklearn Models : ResNet152, LSTM
S13G
SurveyPieChart is a project made to take the answers submitted by users on a survey form, analyze it using numpy, pandas etc and then visualise the data using libraries like matplotlib
Using Pandas, NumPy, Matplotlib, Seaborn, Plotly and Cufflinks, the answers for financial crisis of 2007/08 & the most crime affected US regions are visualized with detailed description leading to those events.
JonWeber0328
This repo creates a SQL database from six .csv files (data engineering) then answers questions about the data (data analysis) using: SQL, Python (SQLAlchemy, pandas, NumPy, and Matplotlib), pgAdmin 4, and PostgreSQL.
This project is an assignment given in my internship in which I have to analyse the summer olympics csv file and answer the questions of assignment using numpy and pandas library of Python.
AI-powered developer documentation search engine using RAG (Retrieval-Augmented Generation) with FAISS, Sentence Transformers, and local LLM (Ollama). Enables fast, context-aware answers from Python, Django, Flask, FastAPI, NumPy, and Pandas docs.
saeedfalana
# Udacity--Project-Investigate-TMDB-Movies-Dataset Hello Everyone! I am saeed falana from Palestine with specializations in computer information Systems . My ultimate aim is to derive some great results by combining the knowledge and Experince. I am passionate about data and insights. I love Data science and Analytics. As one of the important steps I have joined Data Analyst Nanodegree. #### "Udacity-DA_Nanodegree" repositories, I will be showing my projects in the Udacity's Data Analyst Nanodegree. # Udacity--Project-Investigate-TMDB-Movies-Dataset Project Overview In this project, we have to analyze a dataset and then communicate our findings about it. We will use the Python libraries NumPy, pandas, and Matplotlib to make your analysis easier. What do I need to install? You will need an installation of Python, plus the following libraries: pandas NumPy Matplotlib csv It will be recommend to installing Anaconda, which comes with all of the necessary packages, as well as IPython notebook. Why this Project? In this project, we have to go through the data analysis process and see how everything fits together. I have also use the Python libraries NumPy, pandas, and Matplotlib, which make writing data analysis code in Python a lot easier! What I have learn? After completing the project, I have learned following : Know all the steps involved in a typical data analysis process Be comfortable posing questions that can be answered with a given dataset and then answering those questions Know how to investigate problems in a dataset and wrangle the data into a format you can use Have practice communicating the results of your analysis Be able to use vectorized operations in NumPy and pandas to speed up your data analysis code Be familiar with pandas' Series and DataFrame objects, which let you access your data more conveniently Know how to use Matplotlib to produce plots showing your findings
yogendra-08
A Python Quiz Application powered by the Open Trivia Database API. Built with Tkinter GUI, it fetches live questions, tracks user answers, calculates scores, and stores quiz history in CSV. Includes detailed statistics using pandas and numpy for performance analysis
Ahad-Al-Seraihi
I performed basic exploratory data analysis (EDA) using Pandas, NumPy and Matplotlib libraries on a Kaggle dataset looking at COVID-19 world vaccination progress in the period (12/2020 - 04/2021). This included data cleaning, wrangling and visualization to answer some hypotheses based on data.
SarthakAgase
Django Quiz Application: The application provides an API interface for managing quizzes, questions, and answers. It utilizes a SQLite3 database for storing the data. The user interface (UI) is developed using HTML, CSS Bootstrap, and JavaScript and python libraries such as pandas, matplotlib, and numpy to provide data analysis functionalities.
ganeshkavhar
Many real-world data sets contain strings, integers, time-stamps and unstructured data. How do you store data like this so that you can manipulate it and easily retrieve important information? The answer is in a pandas DataFrame! In this course, you'll learn how to use the industry-standard pandas library to import, build, and manipulate DataFrames. With pandas, you'll always be able to convert your data into a form that permits you to analyze it easily. You'll also learn more about NumPy, how it stores data, and its relation to the pandas DataFrame.
Libardo1
Udacity Data Analyst Nanodegree: Project 2-Investigate a Dataset For the final project, you will conduct your own data analysis and create a file to share that documents your findings. You should start by taking a look at your dataset and brainstorming what questions you could answer using it. Then you should use Pandas and NumPy to answer the questions you are most interested in, and create a report sharing the answers. You will not be required to use statistics or machine learning to complete this project, but you should make it clear in your communications that your findings are tentative. This project is open-ended in that we are not looking for one right answer.
Detailed AGENDA of the workshop 1. Discuss the needs and applications of data analytics in the healthcare industry and give a description of the most recent trends in this field; Introduce Python, its features and use for data preprocessing and analyses. Explain Python libraries (Pandas, Numpy, Scipy, Statsmodels, Scikit-Learn, Matplotlib, Seaborn, Scrapy etc.) and their applications (1 hour). 2. Break (10 minutes). 3. Demonstration: Installing Python, Loading/importing data (CSV, Excel), selecting and filtering data, deleting columns, data cleaning, sorting, merging etc.(1.15 hours). 4. Break (10 minutes) 5. Demonstration continues Basic visualizations (Bar chart, Line chart, Scatter plot etc.)., and basic analysis: measures of central tendency(mean, median, mode)Correlation, Chi-Square and t-test. (1.15 hour). 7. Questions and Answers (10 minutes).