Database Management System Using Visual C# .NET: MySQL, SQL Server, and Access

Database Management System Using Visual C# .NET: MySQL, SQL Server, and Access

Author: Vivian Siahaan

Publisher: BALIGE PUBLISHING

Published: 2020-09-14

Total Pages: 740

ISBN-13:

DOWNLOAD EBOOK

Book 1: VISUAL C# .NET WITH MYSQL: A Definitive Guide to Develop Database-Oriented Desktop Applications In chapter one, you will learn to know the properties and events of each control in a Windows Visual C# application. You need to learn and know in order to be more familiar when applying them to some applications in this book. In chapter two, you will go through step by step to build a SALES database using MySQL. You will build each table and add associated data fields (along with the necessary keys and indexes). The first field in the Client table is ClientID. Enter the clien ID in the Name Field and select AutoNumber in the Data Type. You define primary key and other indexes which are useful for quick searching. ClientID is a primary field. You will define FamilyName as an index. You then will create Ordering table with three fields: OrderID, ClientID, and OrderDate. You then will create Purchase table with three fields: OrderID, ProductID, and Quantity. And you will create Product table with four fields: ProductID, Description, Price, and QtySold. Before designing Visual C# interface, you will build the relationships between four tables. The interface will be used to enter new orders into the database. The order form will be used to enter the following information into the database: order ID, order date, client ID, client’s first name and family name, client’s address, product information ordered. The form will have the ability to add new orders, find clients, add new clients. The completed order invoice will be provided in a printed report. In chapter three, you will build a database management system where you can store information about valuables in your warehouse. The table will have seven fields: Item (description of the item), Location (where the item was placed), Shop (where the item was purchased), DatePurchased (when the item was purchased), Cost (how much the item cost), SerialNumber (serial number of the item), PhotoFile (path of the photo file of the item), and Fragile (indicates whether a particular item is fragile or not). The development of this Warehouse Inventory Project will be performed, as usual, in a step-by-step manner. You will first create the database. Furthermore, the interface will be built so that the user can view, edit, add, or add data records from the database. Finally, you add code to create a printable list of information from the database. In chapter four, you will build an application that can be used to track daily high and low pollutant PM2.5 and air quality level. The steps that need to be taken in building Siantar Air Quality Index (SAQI) database project are: Build and test a Visual C# interface; Create an empty database using code; and Report database. The designed interface will allow the user to enter max pollutant, min pollutant, and air quality for any date that the user chooses in a particular year. This information will be stored in a database. Graphical result of the data will be provided, along with summary information relating to the maximum value, minimum value, and mean value. You will use a tab control as the main component of the interface. The control has three tabs: one for viewing and editing data, one for viewing graph of pollutant data, and another for viewing graph of air quality data. Each tab on this control operates like a Visual C# control panel. In chapter five, you will perform the steps necessary to build a MySQL book inventory database that contains 4 tables. You will build each table and add the associated fields as needed. You will have four tables in the database and define the relationship between the primary key and foreign key. You will associate AuthorID (foreign key) field in the Title_Author table with AuthorID (primary key) in the Author table. Then, you want to associate the ISBN (foreign key) field in Title_Author table with ISBN (primary key) in the Title table. Book 2: Visual C# .NET For Programmers: A Progressive Tutorial to Develop Desktop Applications In chapter one, you will learn to know the properties and events of each control in a Windows Visual C# application. You need to learn and know in order to be more familiar when applying them to some applications in this book. In chapter two, you will go through step by step to build a SALES database using Microsoft Access and SQL Server. You will build each table and add associated data fields (along with the necessary keys and indexes). The first field in the Client table is ClientID. Enter the clien ID in the Name Field and select AutoNumber in the Data Type. You define primary key and other indexes which are useful for quick searching. ClientID is a primary field. If the small lock symbol is not displayed next to the ClientID row, then you need to place it there. Right click on ClientID row and select Primary Key. A small key is now displayed next to the entry indicating it is the primary key. You will define FamilyName as an index. Select the FamilyName line. On the General tab, set the Indexed property to Yes (Duplicates OK). You then will create Ordering table with three fields: OrderID, ClientID, and OrderDate. You then will create Purchase table with three fields: OrderID, ProductID, and Quantity. And you will create Product table with four fields: ProductID, Description, Price, and QtySold. Before designing Visual C# interface, you will build the relationships between four tables. In chapter three, you will build a Visual C# interface for the database. The interface will be used to enter new orders into the database. The order form will be used to enter the following information into the database: order ID, order date, client ID, client’s first name and family name, client’s address, product information ordered. The form will have the ability to add new orders, find clients, add new clients. The completed order invoice will be provided in a printed report. In chapter four, you will build a database management system where you can store information about valuables in your warehouse. The table will have seven fields: Item (description of the item), Location (where the item was placed), Shop (where the item was purchased), DatePurchased (when the item was purchased), Cost (how much the item cost), SerialNumber (serial number of the item), PhotoFile (path of the photo file of the item), and Fragile (indicates whether a particular item is fragile or not). The development of this Warehouse Inventory Project will be performed, as usual, in a step-by-step manner. You will first create the database. Furthermore, the interface will be built so that the user can view, edit, add, or add data records from the database. Finally, you add code to create a printable list of information from the database. In chapter five, you will build an application that can be used to track daily high and low pollutant PM2.5 and air quality level. You will do this in stages, from database development to creation of distribution packages. These steps are the same as those used in developing a commercial database application. The steps that need to be taken in building Siantar Air Quality Index (SAQI) database project are: Build and test a Visual C# interface; Create an empty database using code; and Report database. The designed interface will allow the user to enter max pollutant, min pollutant, and air quality for any date that the user chooses in a particular year. This information will be stored in a database. Graphical result of the data will be provided, along with summary information relating to the maximum value, minimum value, and mean value. You will use a tab control as the main component of the interface. The control has three tabs: one for viewing and editing data, one for viewing graph of pollutant data, and another for viewing graph of air quality data. Each tab on this control operates like a Visual C# control panel. In chapter six, you will perform the steps necessary to build a SQL Server book inventory database that contains 4 tables using Microsoft Visual Studio 2019. You will build each table and add the associated fields as needed. You will have four tables in the database and define the relationship between the primary key and foreign key. You will associate AuthorID (foreign key) field in the Title_Author table with AuthorID (primary key) in the Author table. Then, you want to associate the ISBN (foreign key) field in Title_Author table with ISBN (primary key) in the Title table.


VISUAL C# .NET WITH MYSQL

VISUAL C# .NET WITH MYSQL

Author: Vivian Siahaan

Publisher: BALIGE PUBLISHING

Published: 2020-09-13

Total Pages: 348

ISBN-13:

DOWNLOAD EBOOK

In chapter one, you will learn to know the properties and events of each control in a Windows Visual C# application. You need to learn and know in order to be more familiar when applying them to some applications in this book. In chapter two, you will go through step by step to build a SALES database using MySQL. You will build each table and add associated data fields (along with the necessary keys and indexes). The first field in the Client table is ClientID. Enter the clien ID in the Name Field and select AutoNumber in the Data Type. You define primary key and other indexes which are useful for quick searching. ClientID is a primary field. You will define FamilyName as an index. You then will create Ordering table with three fields: OrderID, ClientID, and OrderDate. You then will create Purchase table with three fields: OrderID, ProductID, and Quantity. And you will create Product table with four fields: ProductID, Description, Price, and QtySold. Before designing Visual C# interface, you will build the relationships between four tables. The interface will be used to enter new orders into the database. The order form will be used to enter the following information into the database: order ID, order date, client ID, client’s first name and family name, client’s address, product information ordered. The form will have the ability to add new orders, find clients, add new clients. The completed order invoice will be provided in a printed report. In chapter three, you will build a database management system where you can store information about valuables in your warehouse. The table will have seven fields: Item (description of the item), Location (where the item was placed), Shop (where the item was purchased), DatePurchased (when the item was purchased), Cost (how much the item cost), SerialNumber (serial number of the item), PhotoFile (path of the photo file of the item), and Fragile (indicates whether a particular item is fragile or not). The development of this Warehouse Inventory Project will be performed, as usual, in a step-by-step manner. You will first create the database. Furthermore, the interface will be built so that the user can view, edit, add, or add data records from the database. Finally, you add code to create a printable list of information from the database. In chapter four, you will build an application that can be used to track daily high and low pollutant PM2.5 and air quality level. The steps that need to be taken in building Siantar Air Quality Index (SAQI) database project are: Build and test a Visual C# interface; Create an empty database using code; and Report database. The designed interface will allow the user to enter max pollutant, min pollutant, and air quality for any date that the user chooses in a particular year. This information will be stored in a database. Graphical result of the data will be provided, along with summary information relating to the maximum value, minimum value, and mean value. You will use a tab control as the main component of the interface. The control has three tabs: one for viewing and editing data, one for viewing graph of pollutant data, and another for viewing graph of air quality data. Each tab on this control operates like a Visual C# control panel. In chapter five, you will perform the steps necessary to build a MySQL book inventory database that contains 4 tables. You will build each table and add the associated fields as needed. You will have four tables in the database and define the relationship between the primary key and foreign key. You will associate AuthorID (foreign key) field in the Title_Author table with AuthorID (primary key) in the Author table. Then, you want to associate the ISBN (foreign key) field in Title_Author table with ISBN (primary key) in the Title table.


LEARN FROM SCRATCH VISUAL C# .NET WITH SQL SERVER To Develop Database-Driven Desktop Applications

LEARN FROM SCRATCH VISUAL C# .NET WITH SQL SERVER To Develop Database-Driven Desktop Applications

Author: VIVIAN SIAHAAN

Publisher: BALIGE PUBLISHING

Published: 2020-10-10

Total Pages: 686

ISBN-13:

DOWNLOAD EBOOK

In Tutorial 1, you will start building a Visual C# interface for database management system project with SQL Server. The database, named DBMS, is created. The designed interface in this tutorial will used as the main terminal in accessing other forms. This tutorial will also discuss how to create login form and login table. In Tutorial 2, you will build a project, as part of database management system, where you can store information about valuables in school. In Tutorial 3 up to Tutorial 4, you will perform the steps necessary to add 6 tables into DBMS database. You will build each table and add the associated fields as needed. In this tutorials, you will create a library database project, as part of database management system, where you can store all information about library including author, title, and publisher. In Tutorial 5 up to Tutorial 7, you will perform the steps necessary to add 6 more tables into DBMS database. You will build each table and add the associated fields as needed. In this tutorials, you will create a high school database project, as part of database management system, where you can store all information about school including parent, teacher, student, subject, and, title, and grade.


LEARN FROM SCRATCH VISUAL C# .NET WITH MYSQL

LEARN FROM SCRATCH VISUAL C# .NET WITH MYSQL

Author: Vivian Siahaan

Publisher: BALIGE PUBLISHING

Published: 2020-10-05

Total Pages: 684

ISBN-13:

DOWNLOAD EBOOK

In Tutorial 1, you will start building a Visual C# interface for database management system project using MySQL. The database, named DBMS, is created. The designed interface in this tutorial will used as the main terminal in accessing other forms. This tutorial will also discuss how to create login form and login table. In Tutorial 2, you will build a project, as part of database management system, where you can store information about valuables in school. The table will have seven fields: Item (description of the item), Location (where the item was placed), Shop (where the item was purchased), DatePurchased (when the item was purchased), Cost (how much the item cost), SerialNumber (serial number of the item), PhotoFile (path of the photo file of the item), and Fragile (indicates whether a particular item is fragile or not). In Tutorial 3 up to Tutorial 4, you will perform the steps necessary to add 6 tables using phpMyAdmin into DBMS database. You will build each table and add the associated fields as needed. In this tutorials, you will create a library database project, as part of database management system, where you can store all information about library including author, title, and publisher. In Tutorial 5 up to Tutorial 7, you will perform the steps necessary to add 8 more tables using phpMyAdmin into DBMS database. You will build each table and add the associated fields as needed. In this tutorials, you will create a high school database project, as part of database management system, where you can store all information about school including parent, teacher, student, subject, and, title, and grade.


VISUAL C# .NET AND DATABASE

VISUAL C# .NET AND DATABASE

Author: Vivian Siahaan

Publisher: BALIGE PUBLISHING

Published: 2020-10-21

Total Pages: 402

ISBN-13:

DOWNLOAD EBOOK

This book aims to develop a database-driven desktop application that readers can develop for their own purposes to implement database-oriented digital image processing, machine learning, and image retrieval applications. In Tutorial 1, you will perform the steps necessary to add 6 tables using Visual C# into ImageProc database. You will build each table and add the associated fields as needed. In this tutorial, you will also build such a form for Officer table. This table has sixteen fields: OfficerID, FirstName, LastName, RegNumber, BirthDate, AppDate, Gender, Status, Rank, Address, Mobile, Phone, Email, Description, PhotoFile, and FingerFile). You need seventeen label controls, two picture boxes, ten text boxes, two comboxes, one check box, two date time pickers, one openfiledialog, and one printpreviewdialog. You also need four buttons for navigation, eight buttons for utilites, one button for searching officer’s name, one button to upload officer’s photo, and one button to upload officer’s fingerprint. In Tutorial 2, you will perform the steps necessary to create and implement police station form. In this tutorial, you will build such a form for PoliceStation table. This table has seven fields: PSID, OfficerID, PSName, City, Address, Phone, and Description. You need an input form so that user can edit existing records, delete records, or add new records. The form will also have the capability of navigating from one record to another. You need eight label controls, six text boxes, two comboxes, one check box, and one printpreviewdialog. You also need four buttons for navigation, eight buttons for utilites, and one button for searching officer’s name. Place these controls on the form. In Tutorial 3, you will build such a form for Accused table. This table has thirteen fields: AccusedID, FullName, MotherName, CrimeCase, BirthDate, Gender, Address, Mobile, Phone, Email, Description, PhotoFile, and FingerFile). You need an input form so that user can edit existing records, delete records, or add new records. The form will also have the capability of navigating from one record to another. You need fourteen label controls, two picture boxes, nine text boxes, two comboxes, one date time picker, one openfiledialog, and one printpreviewdialog. You also need four buttons for navigation, eight buttons for utilites, one button for searching accused’s name, one button to upload accused’s photo, and one button to upload accused’s fingerprint. In Tutorial 4, you will build such a form for Witness table. This table has thirteen fields: WitnessID, FullName, MotherName, CrimeCase, BirthDate, Gender, Address, Mobile, Phone, Email, Description, PhotoFile, and FingerFile). You need an input form so that user can edit existing records, delete records, or add new records. The form will also have the capability of navigating from one record to another. You need fourteen label controls, two picture boxes, nine text boxes, two comboxes, one date time picker, one openfiledialog, and one printpreviewdialog. You also need four buttons for navigation, eight buttons for utilites, one button for searching witness’s name, one button to upload witness’s photo, and one button to upload witness’s fingerprint. In Tutorial 5, you will build such a form for Victim table. This table has thirteen fields: VictimID, FullName, MotherName, CrimeCase, BirthDate, Gender, Address, Mobile, Phone, Email, Description, PhotoFile, and FingerFile). You need an input form so that user can edit existing records, delete records, or add new records. The form will also have the capability of navigating from one record to another. You need fourteen label controls, two picture boxes, nine text boxes, two comboxes, one date time picker, one openfiledialog, and one printpreviewdialog. You also need four buttons for navigation, eight buttons for utilites, one button for searching victim’s name, one button to upload victim’s photo, and one button to upload victim’s fingerprint. In Tutorial 6, you will build such a form for CrimeReg table. This table has fourteen fields: CRID, CRNumber, PSID, VictimID, AccusedID, DateReport, DateCrime, Arrested, CaseStatus, Description, Feature1, Feature2, Feature3, and Feature4. You need an input form so that user can edit existing records, delete records, or add new records. The form will also have the capability of navigating from one record to another. You need thirty two label controls, seven text boxes, ten comboxes, one check box, two date time pickers, six picture boxes, and one printpreviewdialog. You then need four buttons for navigation, eight buttons for utilites, and one button for searching crime register number. You also need button to save every feature.


MYSQL with Visual Basic and Visual C#

MYSQL with Visual Basic and Visual C#

Author: Vivian Siahaan

Publisher: BALIGE PUBLISHING

Published: 2020-11-20

Total Pages: 645

ISBN-13:

DOWNLOAD EBOOK

Book 1: This book aims to develop a MySQL-driven desktop application that readers can develop for their own purposes to implement library project using Visual Basic .NET. In Tutorial 1, you will build a Visual Basic interface for the database. This interface will used as the main terminal in accessing other forms. This tutorial will also discuss how to create login form and login table. You will create login form. Place on the form one picture box, two labels, one combo box, one text box, and two buttons. In Tutorial 2, you will build a school inventory project where you can store information about valuables in school. The table will have nine fields: Item (description of the item), Quantity, Location (where the item was placed), Shop (where the item was purchased), DatePurchased (when the item was purchased), Cost (how much the item cost), SerialNumber (serial number of the item), PhotoFile (path of the photo file of the item), and Fragile (indicates whether a particular item is fragile or not). In Tutorial 3, you will perform the steps necessary to add 5 new tables using phpMyAdmin into Academy database. You will build each table and add the associated fields as needed. Every table in the database will need input form. In this tutorial, you will build such a form for Author table. Although this table is quite simple (only four fields: AuthorID, Name, BirthDate, and PhotoFile), it provides a basis for illustrating the many steps in interface design. SQL statement is required by the Command object to read fields (sorted by Name). Then, you will build an interface so that the user can maintain the Publisher table in the database (Academy). The Publisher table interface is more or less the same as Author table interface. This Publisher table interface only requires more input fields. So you will use the interface for the Author table and modify it for the Publisher table. Book 2: In chapter one, you will learn to know the properties and events of each control in a Windows Visual C# application. You need to learn and know in order to be more familiar when applying them to some applications in this book. In chapter two, you will go through step by step to build a SALES database using MySQL. You will build each table and add associated data fields (along with the necessary keys and indexes). The first field in the Client table is ClientID. Enter the clien ID in the Name Field and select AutoNumber in the Data Type. You define primary key and other indexes which are useful for quick searching. ClientID is a primary field. You will define FamilyName as an index. You then will create Ordering table with three fields: OrderID, ClientID, and OrderDate. You then will create Purchase table with three fields: OrderID, ProductID, and Quantity. And you will create Product table with four fields: ProductID, Description, Price, and QtySold. Before designing Visual C# interface, you will build the relationships between four tables. The interface will be used to enter new orders into the database. The order form will be used to enter the following information into the database: order ID, order date, client ID, client’s first name and family name, client’s address, product information ordered. The form will have the ability to add new orders, find clients, add new clients. The completed order invoice will be provided in a printed report. In chapter three, you will build a database management system where you can store information about valuables in your warehouse. The table will have seven fields: Item (description of the item), Location (where the item was placed), Shop (where the item was purchased), DatePurchased (when the item was purchased), Cost (how much the item cost), SerialNumber (serial number of the item), PhotoFile (path of the photo file of the item), and Fragile (indicates whether a particular item is fragile or not). The development of this Warehouse Inventory Project will be performed, as usual, in a step-by-step manner. You will first create the database. Furthermore, the interface will be built so that the user can view, edit, add, or add data records from the database. Finally, you add code to create a printable list of information from the database. In chapter four, you will build an application that can be used to track daily high and low pollutant PM2.5 and air quality level. The steps that need to be taken in building Siantar Air Quality Index (SAQI) database project are: Build and test a Visual C# interface; Create an empty database using code; and Report database. The designed interface will allow the user to enter max pollutant, min pollutant, and air quality for any date that the user chooses in a particular year. This information will be stored in a database. Graphical result of the data will be provided, along with summary information relating to the maximum value, minimum value, and mean value. You will use a tab control as the main component of the interface. The control has three tabs: one for viewing and editing data, one for viewing graph of pollutant data, and another for viewing graph of air quality data. Each tab on this control operates like a Visual C# control panel. In chapter five, you will perform the steps necessary to build a MySQL book inventory database that contains 4 tables. You will build each table and add the associated fields as needed. You will have four tables in the database and define the relationship between the primary key and foreign key. You will associate AuthorID (foreign key) field in the Title_Author table with AuthorID (primary key) in the Author table. Then, you want to associate the ISBN (foreign key) field in Title_Author table with ISBN (primary key) in the Title table.


ANALYSIS AND PREDICTION PROJECTS USING MACHINE LEARNING AND DEEP LEARNING WITH PYTHON

ANALYSIS AND PREDICTION PROJECTS USING MACHINE LEARNING AND DEEP LEARNING WITH PYTHON

Author: Vivian Siahaan

Publisher: BALIGE PUBLISHING

Published: 2022-02-17

Total Pages: 860

ISBN-13:

DOWNLOAD EBOOK

PROJECT 1: DEFAULT LOAN PREDICTION BASED ON CUSTOMER BEHAVIOR Using Machine Learning and Deep Learning with Python In finance, default is failure to meet the legal obligations (or conditions) of a loan, for example when a home buyer fails to make a mortgage payment, or when a corporation or government fails to pay a bond which has reached maturity. A national or sovereign default is the failure or refusal of a government to repay its national debt. The dataset used in this project belongs to a Hackathon organized by "Univ.AI". All values were provided at the time of the loan application. Following are the features in the dataset: Income, Age, Experience, Married/Single, House_Ownership, Car_Ownership, Profession, CITY, STATE, CURRENT_JOB_YRS, CURRENT_HOUSE_YRS, and Risk_Flag. The Risk_Flag indicates whether there has been a default in the past or not. The machine learning models used in this project are K-Nearest Neighbor, Random Forest, Naive Bayes, Logistic Regression, Decision Tree, Support Vector Machine, Adaboost, LGBM classifier, Gradient Boosting, XGB classifier, MLP classifier, and CNN 1D. Finally, you will plot boundary decision, ROC, distribution of features, feature importance, cross validation score, and predicted values versus true values, confusion matrix, learning curve, performance of the model, scalability of the model, training loss, and training accuracy. PROJECT 2: AIRLINE PASSENGER SATISFACTION Analysis and Prediction Using Machine Learning and Deep Learning with Python The dataset used in this project contains an airline passenger satisfaction survey. In this case, you will determine what factors are highly correlated to a satisfied (or dissatisfied) passenger and predict passenger satisfaction. Below are the features in the dataset: Gender: Gender of the passengers (Female, Male); Customer Type: The customer type (Loyal customer, disloyal customer); Age: The actual age of the passengers; Type of Travel: Purpose of the flight of the passengers (Personal Travel, Business Travel); Class: Travel class in the plane of the passengers (Business, Eco, Eco Plus); Flight distance: The flight distance of this journey; Inflight wifi service: Satisfaction level of the inflight wifi service (0:Not Applicable;1-5); Departure/Arrival time convenient: Satisfaction level of Departure/Arrival time convenient; Ease of Online booking: Satisfaction level of online booking; Gate location: Satisfaction level of Gate location; Food and drink: Satisfaction level of Food and drink; Online boarding: Satisfaction level of online boarding; Seat comfort: Satisfaction level of Seat comfort; Inflight entertainment: Satisfaction level of inflight entertainment; On-board service: Satisfaction level of On-board service; Leg room service: Satisfaction level of Leg room service; Baggage handling: Satisfaction level of baggage handling; Check-in service: Satisfaction level of Check-in service; Inflight service: Satisfaction level of inflight service; Cleanliness: Satisfaction level of Cleanliness; Departure Delay in Minutes: Minutes delayed when departure; Arrival Delay in Minutes: Minutes delayed when Arrival; and Satisfaction: Airline satisfaction level (Satisfaction, neutral or dissatisfaction) The machine learning models used in this project are K-Nearest Neighbor, Random Forest, Naive Bayes, Logistic Regression, Decision Tree, Support Vector Machine, LGBM classifier, Gradient Boosting, XGB classifier, MLP classifier, and CNN 1D. Finally, you will plot boundary decision, ROC, distribution of features, feature importance, cross validation score, and predicted values versus true values, confusion matrix, learning curve, performance of the model, scalability of the model, training loss, and training accuracy. PROJECT 3: CREDIT CARD CHURNING CUSTOMER ANALYSIS AND PREDICTION USING MACHINE LEARNING AND DEEP LEARNING WITH PYTHON The dataset used in this project consists of more than 10,000 customers mentioning their age, salary, marital_status, credit card limit, credit card category, etc. There are 20 features in the dataset. In the dataset, there are only 16.07% of customers who have churned. Thus, it's a bit difficult to train our model to predict churning customers. Following are the features in the dataset: 'Attrition_Flag', 'Customer_Age', 'Gender', 'Dependent_count', 'Education_Level', 'Marital_Status', 'Income_Category', 'Card_Category', 'Months_on_book', 'Total_Relationship_Count', 'Months_Inactive_12_mon', 'Contacts_Count_12_mon', 'Credit_Limit', 'Total_Revolving_Bal', 'Avg_Open_To_Buy', 'Total_Amt_Chng_Q4_Q1', 'Total_Trans_Amt', 'Total_Trans_Ct', 'Total_Ct_Chng_Q4_Q1', and 'Avg_Utilization_Ratio',. The target variable is 'Attrition_Flag'. The machine learning models used in this project are K-Nearest Neighbor, Random Forest, Naive Bayes, Logistic Regression, Decision Tree, Support Vector Machine, LGBM classifier, Gradient Boosting, XGB classifier, MLP classifier, and CNN 1D. Finally, you will plot boundary decision, ROC, distribution of features, feature importance, cross validation score, and predicted values versus true values, confusion matrix, learning curve, performance of the model, scalability of the model, training loss, and training accuracy. PROJECT 4: MARKETING ANALYSIS AND PREDICTION USING MACHINE LEARNING AND DEEP LEARNING WITH PYTHON This data set was provided to students for their final project in order to test their statistical analysis skills as part of a MSc. in Business Analytics. It can be utilized for EDA, Statistical Analysis, and Visualizations. Following are the features in the dataset: ID = Customer's unique identifier; Year_Birth = Customer's birth year; Education = Customer's education level; Marital_Status = Customer's marital status; Income = Customer's yearly household income; Kidhome = Number of children in customer's household; Teenhome = Number of teenagers in customer's household; Dt_Customer = Date of customer's enrollment with the company; Recency = Number of days since customer's last purchase; MntWines = Amount spent on wine in the last 2 years; MntFruits = Amount spent on fruits in the last 2 years; MntMeatProducts = Amount spent on meat in the last 2 years; MntFishProducts = Amount spent on fish in the last 2 years; MntSweetProducts = Amount spent on sweets in the last 2 years; MntGoldProds = Amount spent on gold in the last 2 years; NumDealsPurchases = Number of purchases made with a discount; NumWebPurchases = Number of purchases made through the company's web site; NumCatalogPurchases = Number of purchases made using a catalogue; NumStorePurchases = Number of purchases made directly in stores; NumWebVisitsMonth = Number of visits to company's web site in the last month; AcceptedCmp3 = 1 if customer accepted the offer in the 3rd campaign, 0 otherwise; AcceptedCmp4 = 1 if customer accepted the offer in the 4th campaign, 0 otherwise; AcceptedCmp5 = 1 if customer accepted the offer in the 5th campaign, 0 otherwise; AcceptedCmp1 = 1 if customer accepted the offer in the 1st campaign, 0 otherwise; AcceptedCmp2 = 1 if customer accepted the offer in the 2nd campaign, 0 otherwise; Response = 1 if customer accepted the offer in the last campaign, 0 otherwise; Complain = 1 if customer complained in the last 2 years, 0 otherwise; and Country = Customer's location. The machine and deep learning models used in this project are K-Nearest Neighbor, Random Forest, Naive Bayes, Logistic Regression, Decision Tree, Support Vector Machine, LGBM classifier, Gradient Boosting, XGB classifier, MLP classifier, and CNN 1D. Finally, you will plot boundary decision, ROC, distribution of features, feature importance, cross validation score, and predicted values versus true values, confusion matrix, learning curve, performance of the model, scalability of the model, training loss, and training accuracy. PROJECT 5: METEOROLOGICAL DATA ANALYSIS AND PREDICTION USING MACHINE LEARNING WITH PYTHON Meteorological phenomena are described and quantified by the variables of Earth's atmosphere: temperature, air pressure, water vapour, mass flow, and the variations and interactions of these variables, and how they change over time. Different spatial scales are used to describe and predict weather on local, regional, and global levels. The dataset used in this project consists of meteorological data with 96453 total number of data points and with 11 attributes/columns. Following are the columns in the dataset: Formatted Date; Summary; Precip Type; Temperature (C); Apparent Temperature (C); Humidity; Wind Speed (km/h); Wind Bearing (degrees); Visibility (km); Pressure (millibars); and Daily Summary. The machine learning models used in this project are K-Nearest Neighbor, Random Forest, Naive Bayes, Logistic Regression, Decision Tree, Support Vector Machine, LGBM classifier, Gradient Boosting, XGB classifier, and MLP classifier. Finally, you will plot boundary decision, distribution of features, feature importance, cross validation score, and predicted values versus true values, confusion matrix, learning curve, performance of the model, scalability of the model, training loss, and training accuracy.


Hands-On Guide To IMAGE CLASSIFICATION Using Scikit-Learn, Keras, And TensorFlow with PYTHON GUI

Hands-On Guide To IMAGE CLASSIFICATION Using Scikit-Learn, Keras, And TensorFlow with PYTHON GUI

Author: Vivian Siahaan

Publisher: BALIGE PUBLISHING

Published: 2023-06-20

Total Pages: 210

ISBN-13:

DOWNLOAD EBOOK

In this book, implement deep learning on detecting face mask, classifying weather, and recognizing flower using TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries. In chapter 1, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform detecting face mask using Face Mask Detection Dataset provided by Kaggle (https://www.kaggle.com/omkargurav/face-mask-dataset/download). Here's an overview of the steps involved in detecting face masks using the Face Mask Detection Dataset: Import the necessary libraries: Import the required libraries like TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, and NumPy.; Load and preprocess the dataset: Load the dataset and perform any necessary preprocessing steps, such as resizing images and converting labels into numeric representations.; Split the dataset: Split the dataset into training and testing sets using the train_test_split function from Scikit-Learn. This will allow us to evaluate the model's performance on unseen data.; Data augmentation (optional): Apply data augmentation techniques to artificially increase the size and diversity of the training set. Techniques like rotation, zooming, and flipping can help improve the model's generalization.; Build the model: Create a Convolutional Neural Network (CNN) model using TensorFlow and Keras. Design the architecture of the model, including the number and type of layers.; Compile the model: Compile the model by specifying the loss function, optimizer, and evaluation metrics. This prepares the model for training. Train the model: Train the model on the training dataset. Adjust the hyperparameters, such as the learning rate and number of epochs, to achieve optimal performance.; Evaluate the model: Evaluate the trained model on the testing dataset to assess its performance. Calculate metrics such as accuracy, precision, recall, and F1 score.; Make predictions: Use the trained model to make predictions on new images or video streams. Apply the face mask detection algorithm to identify whether a person is wearing a mask or not.; Visualize the results: Visualize the predictions by overlaying bounding boxes or markers on the images or video frames to indicate the presence or absence of face masks. In chapter 2, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform how to classify weather using Multi-class Weather Dataset provided by Kaggle (https://www.kaggle.com/pratik2901/multiclass-weather-dataset/download). To classify weather using the Multi-class Weather Dataset from Kaggle, you can follow these general steps: Load the dataset: Use libraries like Pandas or NumPy to load the dataset into memory. Explore the dataset to understand its structure and the available features.; Preprocess the data: Perform necessary preprocessing steps such as data cleaning, handling missing values, and feature engineering. This may include resizing images (if the dataset contains images) or encoding categorical variables.; Split the data: Split the dataset into training and testing sets. The training set will be used to train the model, and the testing set will be used for evaluating its performance.; Build a model: Utilize TensorFlow and Keras to define a suitable model architecture for weather classification. The choice of model depends on the type of data you have. For image data, convolutional neural networks (CNNs) often work well.; Train the model: Train the model using the training data. Use appropriate training techniques like gradient descent and backpropagation to optimize the model's weights.; Evaluate the model: Evaluate the trained model's performance using the testing data. Calculate metrics such as accuracy, precision, recall, or F1-score to assess how well the model performs.; Fine-tune the model: If the model's performance is not satisfactory, you can experiment with different hyperparameters, architectures, or regularization techniques to improve its performance. This process is called model tuning.; Make predictions: Once you are satisfied with the model's performance, you can use it to make predictions on new, unseen data. Provide the necessary input (e.g., an image or weather features) to the trained model, and it will predict the corresponding weather class. In chapter 3, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform how to recognize flower using Flowers Recognition dataset provided by Kaggle (https://www.kaggle.com/alxmamaev/flowers-recognition/download). Here are the general steps involved in recognizing flowers: Data Preparation: Download the Flowers Recognition dataset from Kaggle and extract the contents. Import the required libraries and define the dataset path and image dimensions.; Loading and Preprocessing the Data: Load the images and their corresponding labels from the dataset. Resize the images to a specific dimension. Perform label encoding on the flower labels and split the data into training and testing sets. Normalize the pixel values of the images.; Building the Model: Define the architecture of your model using TensorFlow's Keras API. You can choose from various neural network architectures such as CNNs, ResNet, or InceptionNet. The model architecture should be designed to handle image inputs and output the predicted flower class..; Compiling and Training the Model: Compile the model by specifying the loss function, optimizer, and evaluation metrics. Common choices include categorical cross-entropy loss and the Adam optimizer. Train the model using the training set and validate it using the testing set. Adjust the hyperparameters, such as the learning rate and number of epochs, to improve performance.; Model Evaluation: Evaluate the trained model on the testing set to measure its performance. Calculate metrics such as accuracy, precision, recall, and F1-score to assess how well the model is recognizing flower classes.; Prediction: Use the trained model to predict the flower class for new images. Load and preprocess the new images in a similar way to the training data. Pass the preprocessed images through the trained model and obtain the predicted flower class labels.; Further Improvements: If the model's performance is not satisfactory, consider experimenting with different architectures, hyperparameters, or techniques such as data augmentation or transfer learning. Fine-tuning the model or using ensembles of models can also improve accuracy.


In-Depth Tutorials: Deep Learning Using Scikit-Learn, Keras, and TensorFlow with Python GUI

In-Depth Tutorials: Deep Learning Using Scikit-Learn, Keras, and TensorFlow with Python GUI

Author: Vivian Siahaan

Publisher: BALIGE PUBLISHING

Published: 2021-06-05

Total Pages: 1459

ISBN-13:

DOWNLOAD EBOOK

BOOK 1: LEARN FROM SCRATCH MACHINE LEARNING WITH PYTHON GUI In this book, you will learn how to use NumPy, Pandas, OpenCV, Scikit-Learn and other libraries to how to plot graph and to process digital image. Then, you will learn how to classify features using Perceptron, Adaline, Logistic Regression (LR), Support Vector Machine (SVM), Decision Tree (DT), Random Forest (RF), and K-Nearest Neighbor (KNN) models. You will also learn how to extract features using Principal Component Analysis (PCA), Linear Discriminant Analysis (LDA), Kernel Principal Component Analysis (KPCA) algorithms and use them in machine learning. In Chapter 1, you will learn: Tutorial Steps To Create A Simple GUI Application, Tutorial Steps to Use Radio Button, Tutorial Steps to Group Radio Buttons, Tutorial Steps to Use CheckBox Widget, Tutorial Steps to Use Two CheckBox Groups, Tutorial Steps to Understand Signals and Slots, Tutorial Steps to Convert Data Types, Tutorial Steps to Use Spin Box Widget, Tutorial Steps to Use ScrollBar and Slider, Tutorial Steps to Use List Widget, Tutorial Steps to Select Multiple List Items in One List Widget and Display It in Another List Widget, Tutorial Steps to Insert Item into List Widget, Tutorial Steps to Use Operations on Widget List, Tutorial Steps to Use Combo Box, Tutorial Steps to Use Calendar Widget and Date Edit, and Tutorial Steps to Use Table Widget. In Chapter 2, you will learn: Tutorial Steps To Create A Simple Line Graph, Tutorial Steps To Create A Simple Line Graph in Python GUI, Tutorial Steps To Create A Simple Line Graph in Python GUI: Part 2, Tutorial Steps To Create Two or More Graphs in the Same Axis, Tutorial Steps To Create Two Axes in One Canvas, Tutorial Steps To Use Two Widgets, Tutorial Steps To Use Two Widgets, Each of Which Has Two Axes, Tutorial Steps To Use Axes With Certain Opacity Levels, Tutorial Steps To Choose Line Color From Combo Box, Tutorial Steps To Calculate Fast Fourier Transform, Tutorial Steps To Create GUI For FFT, Tutorial Steps To Create GUI For FFT With Some Other Input Signals, Tutorial Steps To Create GUI For Noisy Signal, Tutorial Steps To Create GUI For Noisy Signal Filtering, and Tutorial Steps To Create GUI For Wav Signal Filtering. In Chapter 3, you will learn: Tutorial Steps To Convert RGB Image Into Grayscale, Tutorial Steps To Convert RGB Image Into YUV Image, Tutorial Steps To Convert RGB Image Into HSV Image, Tutorial Steps To Filter Image, Tutorial Steps To Display Image Histogram, Tutorial Steps To Display Filtered Image Histogram, Tutorial Steps To Filter Image With CheckBoxes, Tutorial Steps To Implement Image Thresholding, and Tutorial Steps To Implement Adaptive Image Thresholding. You will also learn: Tutorial Steps To Generate And Display Noisy Image, Tutorial Steps To Implement Edge Detection On Image, Tutorial Steps To Implement Image Segmentation Using Multiple Thresholding and K-Means Algorithm, Tutorial Steps To Implement Image Denoising, Tutorial Steps To Detect Face, Eye, and Mouth Using Haar Cascades, Tutorial Steps To Detect Face Using Haar Cascades with PyQt, Tutorial Steps To Detect Eye, and Mouth Using Haar Cascades with PyQt, Tutorial Steps To Extract Detected Objects, Tutorial Steps To Detect Image Features Using Harris Corner Detection, Tutorial Steps To Detect Image Features Using Shi-Tomasi Corner Detection, Tutorial Steps To Detect Features Using Scale-Invariant Feature Transform (SIFT), and Tutorial Steps To Detect Features Using Features from Accelerated Segment Test (FAST). In Chapter 4, In this tutorial, you will learn how to use Pandas, NumPy and other libraries to perform simple classification using perceptron and Adaline (adaptive linear neuron). The dataset used is Iris dataset directly from the UCI Machine Learning Repository. You will learn: Tutorial Steps To Implement Perceptron, Tutorial Steps To Implement Perceptron with PyQt, Tutorial Steps To Implement Adaline (ADAptive LInear NEuron), and Tutorial Steps To Implement Adaline with PyQt. In Chapter 5, you will learn how to use the scikit-learn machine learning library, which provides a wide variety of machine learning algorithms via a user-friendly Python API and to perform classification using perceptron, Adaline (adaptive linear neuron), and other models. The dataset used is Iris dataset directly from the UCI Machine Learning Repository. You will learn: Tutorial Steps To Implement Perceptron Using Scikit-Learn, Tutorial Steps To Implement Perceptron Using Scikit-Learn with PyQt, Tutorial Steps To Implement Logistic Regression Model, Tutorial Steps To Implement Logistic Regression Model with PyQt, Tutorial Steps To Implement Logistic Regression Model Using Scikit-Learn with PyQt, Tutorial Steps To Implement Support Vector Machine (SVM) Using Scikit-Learn, Tutorial Steps To Implement Decision Tree (DT) Using Scikit-Learn, Tutorial Steps To Implement Random Forest (RF) Using Scikit-Learn, and Tutorial Steps To Implement K-Nearest Neighbor (KNN) Using Scikit-Learn. In Chapter 6, you will learn how to use Pandas, NumPy, Scikit-Learn, and other libraries to implement different approaches for reducing the dimensionality of a dataset using different feature selection techniques. You will learn about three fundamental techniques that will help us to summarize the information content of a dataset by transforming it onto a new feature subspace of lower dimensionality than the original one. Data compression is an important topic in machine learning, and it helps us to store and analyze the increasing amounts of data that are produced and collected in the modern age of technology. You will learn the following topics: Principal Component Analysis (PCA) for unsupervised data compression, Linear Discriminant Analysis (LDA) as a supervised dimensionality reduction technique for maximizing class separability, Nonlinear dimensionality reduction via Kernel Principal Component Analysis (KPCA). You will learn: Tutorial Steps To Implement Principal Component Analysis (PCA), Tutorial Steps To Implement Principal Component Analysis (PCA) Using Scikit-Learn, Tutorial Steps To Implement Principal Component Analysis (PCA) Using Scikit-Learn with PyQt, Tutorial Steps To Implement Linear Discriminant Analysis (LDA), Tutorial Steps To Implement Linear Discriminant Analysis (LDA) with Scikit-Learn, Tutorial Steps To Implement Linear Discriminant Analysis (LDA) Using Scikit-Learn with PyQt, Tutorial Steps To Implement Kernel Principal Component Analysis (KPCA) Using Scikit-Learn, and Tutorial Steps To Implement Kernel Principal Component Analysis (KPCA) Using Scikit-Learn with PyQt. In Chapter 7, you will learn how to use Keras, Scikit-Learn, Pandas, NumPy and other libraries to perform prediction on handwritten digits using MNIST dataset. You will learn: Tutorial Steps To Load MNIST Dataset, Tutorial Steps To Load MNIST Dataset with PyQt, Tutorial Steps To Implement Perceptron With PCA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement Perceptron With LDA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement Perceptron With KPCA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement Logistic Regression (LR) Model With PCA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement Logistic Regression (LR) Model With LDA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement Logistic Regression (LR) Model With KPCA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement , Tutorial Steps To Implement Support Vector Machine (SVM) Model With LDA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement Support Vector Machine (SVM) Model With KPCA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement Decision Tree (DT) Model With PCA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement Decision Tree (DT) Model With LDA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement Decision Tree (DT) Model With KPCA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement Random Forest (RF) Model With PCA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement Random Forest (RF) Model With LDA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement Random Forest (RF) Model With KPCA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement K-Nearest Neighbor (KNN) Model With PCA Feature Extractor on MNIST Dataset Using PyQt, Tutorial Steps To Implement K-Nearest Neighbor (KNN) Model With LDA Feature Extractor on MNIST Dataset Using PyQt, and Tutorial Steps To Implement K-Nearest Neighbor (KNN) Model With KPCA Feature Extractor on MNIST Dataset Using PyQt. BOOK 2: THE PRACTICAL GUIDES ON DEEP LEARNING USING SCIKIT-LEARN, KERAS, AND TENSORFLOW WITH PYTHON GUI In this book, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to implement deep learning on recognizing traffic signs using GTSRB dataset, detecting brain tumor using Brain Image MRI dataset, classifying gender, and recognizing facial expression using FER2013 dataset In Chapter 1, you will learn to create GUI applications to display line graph using PyQt. You will also learn how to display image and its histogram. In Chapter 2, you will learn how to use TensorFlow, Keras, Scikit-Learn, Pandas, NumPy and other libraries to perform prediction on handwritten digits using MNIST dataset with PyQt. You will build a GUI application for this purpose. In Chapter 3, you will learn how to perform recognizing traffic signs using GTSRB dataset from Kaggle. There are several different types of traffic signs like speed limits, no entry, traffic signals, turn left or right, children crossing, no passing of heavy vehicles, etc. Traffic signs classification is the process of identifying which class a traffic sign belongs to. In this Python project, you will build a deep neural network model that can classify traffic signs in image into different categories. With this model, you will be able to read and understand traffic signs which are a very important task for all autonomous vehicles. You will build a GUI application for this purpose. In Chapter 4, you will learn how to perform detecting brain tumor using Brain Image MRI dataset provided by Kaggle (https://www.kaggle.com/navoneel/brain-mri-images-for-brain-tumor-detection) using CNN model. You will build a GUI application for this purpose. In Chapter 5, you will learn how to perform classifying gender using dataset provided by Kaggle (https://www.kaggle.com/cashutosh/gender-classification-dataset) using MobileNetV2 and CNN models. You will build a GUI application for this purpose. In Chapter 6, you will learn how to perform recognizing facial expression using FER2013 dataset provided by Kaggle (https://www.kaggle.com/nicolejyt/facialexpressionrecognition) using CNN model. You will also build a GUI application for this purpose. BOOK 3: STEP BY STEP TUTORIALS ON DEEP LEARNING USING SCIKIT-LEARN, KERAS, AND TENSORFLOW WITH PYTHON GUI In this book, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to implement deep learning on classifying fruits, classifying cats/dogs, detecting furnitures, and classifying fashion. In Chapter 1, you will learn to create GUI applications to display line graph using PyQt. You will also learn how to display image and its histogram. Then, you will learn how to use OpenCV, NumPy, and other libraries to perform feature extraction with Python GUI (PyQt). The feature detection techniques used in this chapter are Harris Corner Detection, Shi-Tomasi Corner Detector, and Scale-Invariant Feature Transform (SIFT). In Chapter 2, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform classifying fruits using Fruits 360 dataset provided by Kaggle (https://www.kaggle.com/moltean/fruits/code) using Transfer Learning and CNN models. You will build a GUI application for this purpose. In Chapter 3, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform classifying cats/dogs using dataset provided by Kaggle (https://www.kaggle.com/chetankv/dogs-cats-images) using Using CNN with Data Generator. You will build a GUI application for this purpose. In Chapter 4, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform detecting furnitures using Furniture Detector dataset provided by Kaggle (https://www.kaggle.com/akkithetechie/furniture-detector) using VGG16 model. You will build a GUI application for this purpose. In Chapter 5, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform classifying fashion using Fashion MNIST dataset provided by Kaggle (https://www.kaggle.com/zalando-research/fashionmnist/code) using CNN model. You will build a GUI application for this purpose. BOOK 4: Project-Based Approach On DEEP LEARNING Using Scikit-Learn, Keras, And TensorFlow with Python GUI In this book, implement deep learning on detecting vehicle license plates, recognizing sign language, and detecting surface crack using TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries. In Chapter 1, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform detecting vehicle license plates using Car License Plate Detection dataset provided by Kaggle (https://www.kaggle.com/andrewmvd/car-plate-detection/download). In Chapter 2, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform sign language recognition using Sign Language Digits Dataset provided by Kaggle (https://www.kaggle.com/ardamavi/sign-language-digits-dataset/download). In Chapter 3, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform detecting surface crack using Surface Crack Detection provided by Kaggle (https://www.kaggle.com/arunrk7/surface-crack-detection/download). BOOK 5: Hands-On Guide To IMAGE CLASSIFICATION Using Scikit-Learn, Keras, And TensorFlow with PYTHON GUI In this book, implement deep learning-based image classification on detecting face mask, classifying weather, and recognizing flower using TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries. In Chapter 1, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform detecting face mask using Face Mask Detection Dataset provided by Kaggle (https://www.kaggle.com/omkargurav/face-mask-dataset/download). In Chapter 2, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform how to classify weather using Multi-class Weather Dataset provided by Kaggle (https://www.kaggle.com/pratik2901/multiclass-weather-dataset/download). In Chapter 3, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform how to recognize flower using Flowers Recognition dataset provided by Kaggle (https://www.kaggle.com/alxmamaev/flowers-recognition/download). BOOK 6: Step by Step Tutorial IMAGE CLASSIFICATION Using Scikit-Learn, Keras, And TensorFlow with PYTHON GUI In this book, implement deep learning-based image classification on classifying monkey species, recognizing rock, paper, and scissor, and classify airplane, car, and ship using TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries. In Chapter 1, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform how to classify monkey species using 10 Monkey Species dataset provided by Kaggle (https://www.kaggle.com/slothkong/10-monkey-species/download). In Chapter 2, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform how to recognize rock, paper, and scissor using 10 Monkey Species dataset provided by Kaggle (https://www.kaggle.com/sanikamal/rock-paper-scissors-dataset/download). In Chapter 3, you will learn how to use TensorFlow, Keras, Scikit-Learn, OpenCV, Pandas, NumPy and other libraries to perform how to classify airplane, car, and ship using Multiclass-image-dataset-airplane-car-ship dataset provided by Kaggle (https://www.kaggle.com/abtabm/multiclassimagedatasetairplanecar).


HIGHER EDUCATION STUDENT ACADEMIC PERFORMANCE ANALYSIS AND PREDICTION USING MACHINE LEARNING WITH PYTHON GUI

HIGHER EDUCATION STUDENT ACADEMIC PERFORMANCE ANALYSIS AND PREDICTION USING MACHINE LEARNING WITH PYTHON GUI

Author: Vivian Siahaan

Publisher: BALIGE PUBLISHING

Published: 2022-04-24

Total Pages: 222

ISBN-13:

DOWNLOAD EBOOK

The dataset used in this project was collected from the Faculty of Engineering and Faculty of Educational Sciences students in 2019. The purpose is to predict students' end-of-term performances using ML techniques. Attribute information in the dataset are as follows: Student ID; Student Age (1: 18-21, 2: 22-25, 3: above 26); Sex (1: female, 2: male); Graduated high-school type: (1: private, 2: state, 3: other); Scholarship type: (1: None, 2: 25%, 3: 50%, 4: 75%, 5: Full); Additional work: (1: Yes, 2: No); Regular artistic or sports activity: (1: Yes, 2: No); Do you have a partner: (1: Yes, 2: No); Total salary if available (1: USD 135-200, 2: USD 201-270, 3: USD 271-340, 4: USD 341-410, 5: above 410); Transportation to the university: (1: Bus, 2: Private car/taxi, 3: bicycle, 4: Other); Accommodation type in Cyprus: (1: rental, 2: dormitory, 3: with family, 4: Other); Mother's education: (1: primary school, 2: secondary school, 3: high school, 4: university, 5: MSc., 6: Ph.D.); Father's education: (1: primary school, 2: secondary school, 3: high school, 4: university, 5: MSc., 6: Ph.D.); Number of sisters/brothers (if available): (1: 1, 2:, 2, 3: 3, 4: 4, 5: 5 or above); Parental status: (1: married, 2: divorced, 3: died - one of them or both); Mother's occupation: (1: retired, 2: housewife, 3: government officer, 4: private sector employee, 5: self-employment, 6: other); Father's occupation: (1: retired, 2: government officer, 3: private sector employee, 4: self-employment, 5: other); Weekly study hours: (1: None, 2: <5 hours, 3: 6-10 hours, 4: 11-20 hours, 5: more than 20 hours); Reading frequency (non-scientific books/journals): (1: None, 2: Sometimes, 3: Often); Reading frequency (scientific books/journals): (1: None, 2: Sometimes, 3: Often); Attendance to the seminars/conferences related to the department: (1: Yes, 2: No); Impact of your projects/activities on your success: (1: positive, 2: negative, 3: neutral); Attendance to classes (1: always, 2: sometimes, 3: never); Preparation to midterm exams 1: (1: alone, 2: with friends, 3: not applicable); Preparation to midterm exams 2: (1: closest date to the exam, 2: regularly during the semester, 3: never); Taking notes in classes: (1: never, 2: sometimes, 3: always); Listening in classes: (1: never, 2: sometimes, 3: always); Discussion improves my interest and success in the course: (1: never, 2: sometimes, 3: always); Flip-classroom: (1: not useful, 2: useful, 3: not applicable); Cumulative grade point average in the last semester (/4.00): (1: <2.00, 2: 2.00-2.49, 3: 2.50-2.99, 4: 3.00-3.49, 5: above 3.49); Expected Cumulative grade point average in the graduation (/4.00): (1: <2.00, 2: 2.00-2.49, 3: 2.50-2.99, 4: 3.00-3.49, 5: above 3.49); Course ID; and OUTPUT: Grade (0: Fail, 1: DD, 2: DC, 3: CC, 4: CB, 5: BB, 6: BA, 7: AA). The models used in this project are K-Nearest Neighbor, Random Forest, Naive Bayes, Logistic Regression, Decision Tree, Support Vector Machine, Adaboost, LGBM classifier, Gradient Boosting, and XGB classifier. Three feature scaling used in machine learning are raw, minmax scaler, and standard scaler. Finally, you will develop a GUI using PyQt5 to plot cross validation score, predicted values versus true values, confusion matrix, learning curve, decision boundaries, performance of the model, scalability of the model, training loss, and training accuracy.