SlideShare a Scribd company logo
1 of 5
Download to read offline
ISSN (online) 2583-455X
BOHR International Journal of Computer Science
2021, Vol. 1, No. 1, pp. 6–10
https://doi.org/10.54646/bijcs.002
www.bohrpub.com
Implementation of Web Application for Disease Prediction Using AI
Manasvi Srivastava, Vikas Yadav and Swati Singh∗
IILM, Academy of Higher Learning, College of Engineering and Technology Greater Noida,
Uttar Pradesh, India
∗Corresponding author: swati.singh@iilm.edu
Abstract. The Internet is the largest source of information created by humanity. It contains a variety of materials
available in various formats, such as text, audio, video, and much more. In all, web scraping is one way. There is a
set of strategies here in which we get information from the website instead of copying the data manually. Many web-
based data extraction methods are designed to solve specific problems and work on ad hoc domains. Various tools
and technologies have been developed to facilitate web scraping. Unfortunately, the appropriateness and ethics of
using these web scraping tools are often overlooked. There are hundreds of web scraping software available today,
most of them designed for Java, Python, and Ruby. There is also open-source software and commercial software.
Web-based software such as YahooPipes, Google Web Scrapers, and Firefox extensions for Outwit are the best tools
for beginners in web cutting. Web extraction is basically used to cut this manual extraction and editing process and
provide an easy and better way to collect data from a web page and convert it into the desired format and save it to
a local or archive directory. In this study, among other kinds of scrub, we focus on those techniques that extract the
content of a web page. In particular, we use scrubbing techniques for a variety of diseases with their own symptoms
and precautions.
Keywords: Web Scraping, Disease, Legality, Software, Symptoms.
INTRODUCTION
Web scraper is a process for downloading and extract-
ing important data by scanning a web page. Web scrap-
ers work best when page content is either transferred,
searched, or modified. The collected information is then
copied to a spreadsheet or stored in a database for further
analysis. For the ultimate purpose of analysis, data need
to be categorized by progressively different developments,
for example, by starting with its specification collection,
editing process, cleaning process, remodeling, and using
different models and various algorithms and end result.
There are two ways to extract data from websites: the
first is the manual extraction process and the second is
the automatic extraction process. Web scrapers compile
site information in the same way that a person can do
that by removing access to a web page of the site, finding
relevant information, and moving on to the next web
page. Each website has a different structure that is why
web scrapers are usually designed to search through a
website. Web deletion can help in finding any kind of
targeted information. We will then have the opportunity
to find, analyze, and use information in the way we
need it. Web logging therefore paves the way for data
acquisition, speeds up automation, and makes it easier to
access extracted data by rendering it in comma-separated
values (CSV) pattern. Web publishing often removes a lot
of data from websites, for example, monitoring consumer
interests, price monitoring (e.g., price checking), advancing
AI models, data collection, tracking issues, and so on. So,
there is no doubt that web removal is a systematic way to
get more data from websites. It requires two stages, mainly
crawling and removal. A search engine is an algorithm
designed by a person who goes through the web to look
for specific information needed by following online links.
Deleter is a specific tool designed to extract data from sites.
Web scraper will work that way; if the patient is suffering
from any kind of illness or illness, he will add his symp-
toms and problems and when the crawl work starts, he
will start scrolling and look for a disease from the database
provided on the website and it will show the best disease
like patient symptoms. When those specific diseases show
6
Implementation of Web Application for Disease Prediction Using AI 7
up, they will also show the precautionary measures that
the patient needs to take care in order to overcome them
and treat the infection.
OVERVIEW OF WEB SCRAPING
Web scraping is a great way to extract random data from
websites and convert that data into organized data that can
be stored and analyzed in a database. Web scraping is also
known as web data extraction, web data removal, web har-
vesting, or screen scanning. Web scraping is a form of data
mining. The whole purpose of the web crawling process is
to extract information from websites and convert it into an
understandable format such as spreadsheets, a database, or
a CSV, as shown in Figure 1. Data such as item prices, stock
prices, various reports, market prices, and product details
can be collected with web termination. Extracting website-
based information helps to make effective decisions for
your business.
Figure 1. Web scraping structure.
PRACTICES OF WEB SCRAPING
• Data scraping
• Research
• Web mash up—integrate data from multiple sources
• Extract business details from business directory web-
sites, such as Yelp and Yellow pages
• Collect government data
• Market analysis
The web data scraper process, a software agent, also
known as a Web robot, mimics browsing communication
between web servers and a person using a normal web
browser. Step by step, the robot enters as many websites as
it needs, transfers its content to find and extract interesting
data, and builds that content as desired. AP scraping
APIs and frameworks address the most common web data
scrapers involved in achieving specific recovery goals, as
described in the following text.
Hypertext Transfer Protocol (HTTP)
This method is used to extract data from static and
dynamic web pages. Data can be retrieved by sending
HTTP requests to a remote web server using a socket
system.
Hyper Text Markup Language (HTML)
Exploring languages for query data, such as XQuery and
Hyper Text Query Language (HTQL), can be used to scan
HTML pages and retrieve and modify content on the page.
Release Structure
The main purpose is to convert the published content into
a formal representation for further analysis and retention.
Although this last step is on the side of web scraping, some
tools are aware of post-results, providing memory data
formats and text-based solutions, such as cables or files
(XML or CSV files).
LITERATURE SURVEY
Python has a rich set of libraries available for downloading
digital content online. Among the libraries available, the
following three are the most popular ones: BeautifulSoup,
LXml, and RegEx. Statistical research performed on the
available data sets indicated that RegEx was able to deliver
the requested information at an average rate of 153.6 ms.
However, RegEx has limitations of data extraction of web
pages with internal HTML tags. Because of this demerit,
RegEx is used to perform complex data extraction only.
Some libraries, such as BeautifulSoup and LXml, are able to
extract content from web pages in a complex environment
that has yielded a response rate of 457.66 and 203 ms,
respectively.
The main purpose of data analysis is to get useful
information from data and make decisions based on that
analysis. Web deletion refers to the collection of data on
the web. Web scraping is also known as data scraping. Data
analysis can be divided into several steps such as cleaning
and editing. Scrapy is the most widely used source of
information needed by the user. The main purpose of using
scrapy is to extract data from its sources. Scrapy, which
crawls on the web and is based on python programming
language, is very helpful in finding the data we need by
using the URLs needed to clear the data from its sources.
Web scraper is a useful API to retrieve data from a website.
Scrapy provides all the necessary tools to extract data from
a website, process data according to user needs, and store
data in a specific format as defined by users.
The Internet is very much looking at web pages that
include a large number of descriptive elements including
text, audio, graphics, video, etc. This process, called web
scraping, is mainly responsible for the collection of raw
data from the website. It is a process in which you extract
data automation very quickly. The process enables us to
extract specific data requested by the user. The most popu-
lar method used is to create individual web data structures
using any known language.
8 Manasvi Srivastava et al.
EXPERIMENTAL WORK
Technology Used
Firebase
For database, we have used Cloud Firestore from firebase.
It is a real-time NoSQL database that stores two key-value
data in the form of collections and documents.
TensorFlow
TensorFlow is used to train the database model and to
make predictions. There are various algorithms for mod-
eling training or using line format in our project.
JavaScript Frameworks
• Node.js
Node.js is an open-source, cross-platform,
JavaScript runtime environment running back to
V8 engine and extracting JavaScript code without a
web browser.
Our rewriting code is written for Node.js as it is a
fast and platform language.
• ElectronJS
Electron is a framework for building native appli-
cations with web technologies such as JavaScript,
HTML, and CSS. As Electron is used to create a short
web application, it helps us write our code and thus
reduces the development time.
• ReactJS
React makes it less painful to create interactive
UIs. Design a simple view of each state in your app
and React will carefully review and provide relevant
sections as your data changes.
React can also render to the server using Node
and has the power of mobile applications using React
Native.
PYTHON
Python is a high-level programming language translated
into high-level translations.
In this project, various libraries, such as pandas, NumPy,
good soup, etc., are used to create our database. Pandas
and NumPy are used to filter and process data needed to
train our model by extracting and removing them from a
separate data source.
COMPATIBILITY
OS X
Only 64-bit binaries are provided for OS X, and the lower
version of OS X is supported by OS X 10.9.
Windows
Electron supports Windows 7 and later, but older versions
of the OS are not supported.
Both x86 and amd64 (x64) binary are provided for
Windows and are not supported in the ARM version of
Windows.
Software Used
– VSCode
Visual Studio Code is a freeware source code editor
developed by Microsoft for Windows, Linux, and
macOS. Features include debugging support, syntax
highlighting, intelligent coding, captions, reuse code,
and embedded Git.
– Google colab notebook
Colaboratory, or Colab for short, is a product of
Google Research that allows developers to write and
use Python code through their browsers. Google
Colab is an excellent tool for in-depth learning activi-
ties. It is a compact Jupyter notebook that needs no
setup and has an excellent free version, providing
free access to Google computer resources such as
GPUs and TPUs.
– PyCharm
PyCharm is an integrated development platform
used for computer programs, especially in the
Python language, developed by the Czech company
JetBrains.
Data Source
As we did not get more than 40 diseases, to get a dataset
we have created our own dataset. The dataset that we have
used for our training and testing process has taken from
various sources. One of them is added below.
– https://github.com/DiseaseOntology/HumanDise
aseOntology
Use of Scrapy
Scrapy is a framework for crawling and retrieving nonfic-
tion data that can be used for the size of a supportive appli-
cation, such as data mining, managed, or actual reported
data. Apart from the way it was originally expected for
Scrapy to be removed from the web, it could be used in the
same way to extract data using APIs, for example, Amazon
AWS, or as a very important web browser. Scrappy is
written in Python. Let us take a Wiki example related to
one of these problems. A simple online photo gallery can
provide three options to users as defined by HTTP GET
parameters at URL. If there are four ways to filter images
with three thumbnail-sized options, two file formats, and
Implementation of Web Application for Disease Prediction Using AI 9
a user-provided disabling option, then the same content
set can be accessed with different URLs, all of which can
be linked to the site. This carefully crafted combination
creates a problem for the pages as they have to plan with
an endless combination of subtitle changes to get different
content.
Methodology
The method used by the project to collect all the required
data is extracted from various sources such as the CDC’s
database and Kaggle resources. Then, analyze the extracted
data using texts written in the Python language according
to project requirements. Pandas and NumPy are widely
used to perform various functions on the database.
After sorting the data according to each need, it is then
uploaded to the database. In the database, we have used
Cloud Firestore as it is a real-time NoSQL database with
extensive API support.
Furthermore, the TensorFlow project is used to train our
model according to needs.
In this project, we predict the disease because of the
given symptoms.
Training data set: 70%
Setting test data: 30%
TensorFlow supports Linear Regression, which is used
to predict diseases based on the given indicators.
Coding
Project Frontend is written using ReactJS & TypeScript.
However, we have used the MaterialUI kit from Google
ReactJS to speed up our development process.
To provide our app, Electron is used. Our web system
supports macOS and Windows. Most of today’s web app
are written with the help of ElectronJS.
Testing
The project is tested using an Electron built-in test frame-
work called Spectron.
The project is being implemented in the browser. The
output generated turns out to be completely consistent and
the generated analysis is approximate.
Electron’s standard workflow with Spectron can involve
engineers who write unit tests in the standard TDD format
and then write integration tests to ensure that acceptance
criteria are met before approving a feature to be used.
Continuous integration servers can ensure that all these
tests are passed before they are incorporated into the
production.
Algorithm Used
Linear Regression is a standard mathematical method that
allows us to study a function or relationship in a given set
of continuous data. For example, we are given some of the
corresponding x and y data points, and we need to study
the relationship between them called a hypothesis.
In the event of a line reversal, the hypothesis is a straight
line, i.e.,
Where the vector is called Weights and b is a scale called
bias. Weights and bias are called model parameters.
All we need to do is estimate the values of w and b
from the set of data, given that the result of the assumption
has produced the minimum cost J defined by the next cost
function
where m is the number of data points in the data pro-
vided. This cost function is also called the mean squared
error.
To find the optimized value of the J’s minimum param-
eters, we will be using a widely used optimizer algorithm
called Gradient Descent. The following is a fake Gradient
Descent code:
RESULT DISCUSSION
The overall results of the project are useful in predicting
diseases with the given symptoms. The script that was
written to extract data can be used later to compile and
format it according to needs.
Users can pick up symbols by typing them themselves
or by selecting them from the given options. The training
model will predict the disease according to it. Users are
able to create their own medical profile, where they can
submit their medical records and prescribed medication;
this greatly helps us to feed our database and better predict
disease over time, as some of these diseases occur directly
during the season.
Moreover, the analysis performed showed a very sim-
ilar disease, but the training model lacks the size of the
database.
CONCLUSIONS AND FUTURE SCOPE
The use of the Python program also emphasizes under-
standing the use of pattern matching and general expres-
sions for web releases. Database data are compiled from
factual reports and sent directly to government media out-
lets for local media where it is considered reliable. A team
of experts and analysts who validate the information from
a continuous list of more than 5,000 items is likely to be the
site that collects data effectively. User-provided inputs are
analyzed and deleted from the website, and the output is
extracted as the user enters the user interface encounters.
Output is generated in the form of text. This method is
simple and straightforward to eradicate the disease from
companies and provides vigilance against that disease.
10 Manasvi Srivastava et al.
For future work, we plan tests that aim to show the med-
ication that a patient can take for treatment. In addition,
we are looking to link this website to various hospitals and
pharmacies for easy use.
REFERENCES
[1] Thivaharan. S, Srivatsun. G and Sarathambekai. S, “A Survey on
Python Libraries Used for Social Media Content Scraping”, Proceed-
ings of the International Conference on Smart Electronics and Com-
munication (ICOSEC 2020) IEEE Xplore Part Number: CFP20V90-
ART; ISBN: 978-1-7281-5461-9, PP: 361–366.
[2] Shreya Upadhyay, Vishal Pant, Shivansh Bhasin and Mahantesh K
Pattanshetti “Articulating the Construction of a Web Scraper for
Massive Data Extraction”, 2017 IEEE.
[3] Amruta Kulkarni, Deepa Kalburgi and Poonam Ghuli, “Design of
Predictive Model for Healthcare Assistance Using Voice Recogni-
tion”, 2nd IEEE International Conference on Computational Sys-
tems and Information Technology for Sustainable Solutions 2017,
PP: 61–64.
[4] Dimitri Dojchinovski, Andrej Ilievski, Marjan Gusev Interactive home
healthcare system with integrated voice assistant MIPRO 2019,
PP: 284–288 Posted: 2019.
[5] Mohammad Shahnawaz, Prashant Singh, Prabhat Kumar and Dr.
Anuradha Konidena, “Grievance Redressal System”, International
Journal of Data Mining and Big Data, 2020, Vol. 1, No. 1,
PP. 1–4.
Hyperlink of Research Paper
1. https://www.sciencedirect.com/science/article/pii/S2352914817302253

More Related Content

Similar to Implementation ofWeb Application for Disease Prediction Using AI

IRJET - Review on Search Engine Optimization
IRJET - Review on Search Engine OptimizationIRJET - Review on Search Engine Optimization
IRJET - Review on Search Engine OptimizationIRJET Journal
 
A Web Extraction Using Soft Algorithm for Trinity Structure
A Web Extraction Using Soft Algorithm for Trinity StructureA Web Extraction Using Soft Algorithm for Trinity Structure
A Web Extraction Using Soft Algorithm for Trinity Structureiosrjce
 
a novel technique to pre-process web log data using sql server management studio
a novel technique to pre-process web log data using sql server management studioa novel technique to pre-process web log data using sql server management studio
a novel technique to pre-process web log data using sql server management studioINFOGAIN PUBLICATION
 
STRATEGY AND IMPLEMENTATION OF WEB MINING TOOLS
STRATEGY AND IMPLEMENTATION OF WEB MINING TOOLSSTRATEGY AND IMPLEMENTATION OF WEB MINING TOOLS
STRATEGY AND IMPLEMENTATION OF WEB MINING TOOLSAM Publications
 
A Novel Method for Data Cleaning and User- Session Identification for Web Mining
A Novel Method for Data Cleaning and User- Session Identification for Web MiningA Novel Method for Data Cleaning and User- Session Identification for Web Mining
A Novel Method for Data Cleaning and User- Session Identification for Web MiningIJMER
 
Search Engine Scrapper
Search Engine ScrapperSearch Engine Scrapper
Search Engine ScrapperIRJET Journal
 
SPEEDING UP THE WEB CRAWLING PROCESS ON A MULTI-CORE PROCESSOR USING VIRTUALI...
SPEEDING UP THE WEB CRAWLING PROCESS ON A MULTI-CORE PROCESSOR USING VIRTUALI...SPEEDING UP THE WEB CRAWLING PROCESS ON A MULTI-CORE PROCESSOR USING VIRTUALI...
SPEEDING UP THE WEB CRAWLING PROCESS ON A MULTI-CORE PROCESSOR USING VIRTUALI...ijwscjournal
 
Odam an optimized distributed association rule mining algorithm (synopsis)
Odam an optimized distributed association rule mining algorithm (synopsis)Odam an optimized distributed association rule mining algorithm (synopsis)
Odam an optimized distributed association rule mining algorithm (synopsis)Mumbai Academisc
 
DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...
DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...
DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...ijmech
 
DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...
DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...
DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...ijmech
 
Design and Implementation of Carpool Data Acquisition Program Based on Web Cr...
Design and Implementation of Carpool Data Acquisition Program Based on Web Cr...Design and Implementation of Carpool Data Acquisition Program Based on Web Cr...
Design and Implementation of Carpool Data Acquisition Program Based on Web Cr...ijmech
 
International conference On Computer Science And technology
International conference On Computer Science And technologyInternational conference On Computer Science And technology
International conference On Computer Science And technologyanchalsinghdm
 
ANALYTICAL IMPLEMENTATION OF WEB STRUCTURE MINING USING DATA ANALYSIS IN ONLI...
ANALYTICAL IMPLEMENTATION OF WEB STRUCTURE MINING USING DATA ANALYSIS IN ONLI...ANALYTICAL IMPLEMENTATION OF WEB STRUCTURE MINING USING DATA ANALYSIS IN ONLI...
ANALYTICAL IMPLEMENTATION OF WEB STRUCTURE MINING USING DATA ANALYSIS IN ONLI...IAEME Publication
 
Sekhon final 1_ppt
Sekhon final 1_pptSekhon final 1_ppt
Sekhon final 1_pptManant Sweet
 

Similar to Implementation ofWeb Application for Disease Prediction Using AI (20)

IRJET - Review on Search Engine Optimization
IRJET - Review on Search Engine OptimizationIRJET - Review on Search Engine Optimization
IRJET - Review on Search Engine Optimization
 
A Web Extraction Using Soft Algorithm for Trinity Structure
A Web Extraction Using Soft Algorithm for Trinity StructureA Web Extraction Using Soft Algorithm for Trinity Structure
A Web Extraction Using Soft Algorithm for Trinity Structure
 
G017334248
G017334248G017334248
G017334248
 
a novel technique to pre-process web log data using sql server management studio
a novel technique to pre-process web log data using sql server management studioa novel technique to pre-process web log data using sql server management studio
a novel technique to pre-process web log data using sql server management studio
 
Large-Scale Web Scraping: An Ultimate Guide
Large-Scale Web Scraping: An Ultimate GuideLarge-Scale Web Scraping: An Ultimate Guide
Large-Scale Web Scraping: An Ultimate Guide
 
STRATEGY AND IMPLEMENTATION OF WEB MINING TOOLS
STRATEGY AND IMPLEMENTATION OF WEB MINING TOOLSSTRATEGY AND IMPLEMENTATION OF WEB MINING TOOLS
STRATEGY AND IMPLEMENTATION OF WEB MINING TOOLS
 
Scrappy
ScrappyScrappy
Scrappy
 
A Novel Method for Data Cleaning and User- Session Identification for Web Mining
A Novel Method for Data Cleaning and User- Session Identification for Web MiningA Novel Method for Data Cleaning and User- Session Identification for Web Mining
A Novel Method for Data Cleaning and User- Session Identification for Web Mining
 
Search Engine Scrapper
Search Engine ScrapperSearch Engine Scrapper
Search Engine Scrapper
 
SPEEDING UP THE WEB CRAWLING PROCESS ON A MULTI-CORE PROCESSOR USING VIRTUALI...
SPEEDING UP THE WEB CRAWLING PROCESS ON A MULTI-CORE PROCESSOR USING VIRTUALI...SPEEDING UP THE WEB CRAWLING PROCESS ON A MULTI-CORE PROCESSOR USING VIRTUALI...
SPEEDING UP THE WEB CRAWLING PROCESS ON A MULTI-CORE PROCESSOR USING VIRTUALI...
 
Odam an optimized distributed association rule mining algorithm (synopsis)
Odam an optimized distributed association rule mining algorithm (synopsis)Odam an optimized distributed association rule mining algorithm (synopsis)
Odam an optimized distributed association rule mining algorithm (synopsis)
 
DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...
DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...
DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...
 
DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...
DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...
DESIGN AND IMPLEMENTATION OF CARPOOL DATA ACQUISITION PROGRAM BASED ON WEB CR...
 
Design and Implementation of Carpool Data Acquisition Program Based on Web Cr...
Design and Implementation of Carpool Data Acquisition Program Based on Web Cr...Design and Implementation of Carpool Data Acquisition Program Based on Web Cr...
Design and Implementation of Carpool Data Acquisition Program Based on Web Cr...
 
International conference On Computer Science And technology
International conference On Computer Science And technologyInternational conference On Computer Science And technology
International conference On Computer Science And technology
 
Pxc3893553
Pxc3893553Pxc3893553
Pxc3893553
 
ANALYTICAL IMPLEMENTATION OF WEB STRUCTURE MINING USING DATA ANALYSIS IN ONLI...
ANALYTICAL IMPLEMENTATION OF WEB STRUCTURE MINING USING DATA ANALYSIS IN ONLI...ANALYTICAL IMPLEMENTATION OF WEB STRUCTURE MINING USING DATA ANALYSIS IN ONLI...
ANALYTICAL IMPLEMENTATION OF WEB STRUCTURE MINING USING DATA ANALYSIS IN ONLI...
 
Pf3426712675
Pf3426712675Pf3426712675
Pf3426712675
 
Sekhon final 1_ppt
Sekhon final 1_pptSekhon final 1_ppt
Sekhon final 1_ppt
 
Web Scraping Services.pptx
Web Scraping Services.pptxWeb Scraping Services.pptx
Web Scraping Services.pptx
 

More from BOHR International Journal of Computer Science (BIJCS)

More from BOHR International Journal of Computer Science (BIJCS) (8)

Object Detection and Ship Classification Using YOLOv5
Object Detection and Ship Classification Using YOLOv5Object Detection and Ship Classification Using YOLOv5
Object Detection and Ship Classification Using YOLOv5
 
A Comprehensive Study on Online Teaching–Learning (OTL) System and Platforms
A Comprehensive Study on Online Teaching–Learning (OTL) System and PlatformsA Comprehensive Study on Online Teaching–Learning (OTL) System and Platforms
A Comprehensive Study on Online Teaching–Learning (OTL) System and Platforms
 
Proposed Arrangements and Cyber Security Challenges in Nuclear Domain: An Exp...
Proposed Arrangements and Cyber Security Challenges in Nuclear Domain: An Exp...Proposed Arrangements and Cyber Security Challenges in Nuclear Domain: An Exp...
Proposed Arrangements and Cyber Security Challenges in Nuclear Domain: An Exp...
 
Disease Prediction Using Machine Learning
Disease Prediction Using Machine LearningDisease Prediction Using Machine Learning
Disease Prediction Using Machine Learning
 
Do US Government and Commercial Media Concern Similar Topics? A Text-mining (...
Do US Government and Commercial Media Concern Similar Topics? A Text-mining (...Do US Government and Commercial Media Concern Similar Topics? A Text-mining (...
Do US Government and Commercial Media Concern Similar Topics? A Text-mining (...
 
Weed Detection Using Convolutional Neural Network
Weed Detection Using Convolutional Neural NetworkWeed Detection Using Convolutional Neural Network
Weed Detection Using Convolutional Neural Network
 
Virtualization in Distributed System: A Brief Overview
Virtualization in Distributed System: A Brief OverviewVirtualization in Distributed System: A Brief Overview
Virtualization in Distributed System: A Brief Overview
 
Reframing the Possibilities in Healthcare Using Blue Brain Technology
Reframing the Possibilities in Healthcare Using Blue Brain TechnologyReframing the Possibilities in Healthcare Using Blue Brain Technology
Reframing the Possibilities in Healthcare Using Blue Brain Technology
 

Recently uploaded

CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdfCCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdfAsst.prof M.Gokilavani
 
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfCCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfAsst.prof M.Gokilavani
 
GDSC ASEB Gen AI study jams presentation
GDSC ASEB Gen AI study jams presentationGDSC ASEB Gen AI study jams presentation
GDSC ASEB Gen AI study jams presentationGDSCAESB
 
Heart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptxHeart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptxPoojaBan
 
Introduction to Microprocesso programming and interfacing.pptx
Introduction to Microprocesso programming and interfacing.pptxIntroduction to Microprocesso programming and interfacing.pptx
Introduction to Microprocesso programming and interfacing.pptxvipinkmenon1
 
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur High Profile
 
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130Suhani Kapoor
 
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...srsj9000
 
chaitra-1.pptx fake news detection using machine learning
chaitra-1.pptx  fake news detection using machine learningchaitra-1.pptx  fake news detection using machine learning
chaitra-1.pptx fake news detection using machine learningmisbanausheenparvam
 
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ
 
College Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service NashikCollege Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service NashikCall Girls in Nagpur High Profile
 
Artificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxArtificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxbritheesh05
 
What are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxWhat are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxwendy cai
 
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escortsranjana rawat
 
Past, Present and Future of Generative AI
Past, Present and Future of Generative AIPast, Present and Future of Generative AI
Past, Present and Future of Generative AIabhishek36461
 
OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...
OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...
OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...Soham Mondal
 
Call Girls Narol 7397865700 Independent Call Girls
Call Girls Narol 7397865700 Independent Call GirlsCall Girls Narol 7397865700 Independent Call Girls
Call Girls Narol 7397865700 Independent Call Girlsssuser7cb4ff
 
main PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidmain PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidNikhilNagaraju
 

Recently uploaded (20)

CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdfCCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
 
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfCCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
 
GDSC ASEB Gen AI study jams presentation
GDSC ASEB Gen AI study jams presentationGDSC ASEB Gen AI study jams presentation
GDSC ASEB Gen AI study jams presentation
 
Heart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptxHeart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptx
 
Introduction to Microprocesso programming and interfacing.pptx
Introduction to Microprocesso programming and interfacing.pptxIntroduction to Microprocesso programming and interfacing.pptx
Introduction to Microprocesso programming and interfacing.pptx
 
Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCRCall Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
 
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
 
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
 
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
 
chaitra-1.pptx fake news detection using machine learning
chaitra-1.pptx  fake news detection using machine learningchaitra-1.pptx  fake news detection using machine learning
chaitra-1.pptx fake news detection using machine learning
 
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
 
College Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service NashikCollege Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
 
Artificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxArtificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptx
 
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptxExploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
 
What are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxWhat are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptx
 
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
 
Past, Present and Future of Generative AI
Past, Present and Future of Generative AIPast, Present and Future of Generative AI
Past, Present and Future of Generative AI
 
OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...
OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...
OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...
 
Call Girls Narol 7397865700 Independent Call Girls
Call Girls Narol 7397865700 Independent Call GirlsCall Girls Narol 7397865700 Independent Call Girls
Call Girls Narol 7397865700 Independent Call Girls
 
main PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidmain PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfid
 

Implementation ofWeb Application for Disease Prediction Using AI

  • 1. ISSN (online) 2583-455X BOHR International Journal of Computer Science 2021, Vol. 1, No. 1, pp. 6–10 https://doi.org/10.54646/bijcs.002 www.bohrpub.com Implementation of Web Application for Disease Prediction Using AI Manasvi Srivastava, Vikas Yadav and Swati Singh∗ IILM, Academy of Higher Learning, College of Engineering and Technology Greater Noida, Uttar Pradesh, India ∗Corresponding author: swati.singh@iilm.edu Abstract. The Internet is the largest source of information created by humanity. It contains a variety of materials available in various formats, such as text, audio, video, and much more. In all, web scraping is one way. There is a set of strategies here in which we get information from the website instead of copying the data manually. Many web- based data extraction methods are designed to solve specific problems and work on ad hoc domains. Various tools and technologies have been developed to facilitate web scraping. Unfortunately, the appropriateness and ethics of using these web scraping tools are often overlooked. There are hundreds of web scraping software available today, most of them designed for Java, Python, and Ruby. There is also open-source software and commercial software. Web-based software such as YahooPipes, Google Web Scrapers, and Firefox extensions for Outwit are the best tools for beginners in web cutting. Web extraction is basically used to cut this manual extraction and editing process and provide an easy and better way to collect data from a web page and convert it into the desired format and save it to a local or archive directory. In this study, among other kinds of scrub, we focus on those techniques that extract the content of a web page. In particular, we use scrubbing techniques for a variety of diseases with their own symptoms and precautions. Keywords: Web Scraping, Disease, Legality, Software, Symptoms. INTRODUCTION Web scraper is a process for downloading and extract- ing important data by scanning a web page. Web scrap- ers work best when page content is either transferred, searched, or modified. The collected information is then copied to a spreadsheet or stored in a database for further analysis. For the ultimate purpose of analysis, data need to be categorized by progressively different developments, for example, by starting with its specification collection, editing process, cleaning process, remodeling, and using different models and various algorithms and end result. There are two ways to extract data from websites: the first is the manual extraction process and the second is the automatic extraction process. Web scrapers compile site information in the same way that a person can do that by removing access to a web page of the site, finding relevant information, and moving on to the next web page. Each website has a different structure that is why web scrapers are usually designed to search through a website. Web deletion can help in finding any kind of targeted information. We will then have the opportunity to find, analyze, and use information in the way we need it. Web logging therefore paves the way for data acquisition, speeds up automation, and makes it easier to access extracted data by rendering it in comma-separated values (CSV) pattern. Web publishing often removes a lot of data from websites, for example, monitoring consumer interests, price monitoring (e.g., price checking), advancing AI models, data collection, tracking issues, and so on. So, there is no doubt that web removal is a systematic way to get more data from websites. It requires two stages, mainly crawling and removal. A search engine is an algorithm designed by a person who goes through the web to look for specific information needed by following online links. Deleter is a specific tool designed to extract data from sites. Web scraper will work that way; if the patient is suffering from any kind of illness or illness, he will add his symp- toms and problems and when the crawl work starts, he will start scrolling and look for a disease from the database provided on the website and it will show the best disease like patient symptoms. When those specific diseases show 6
  • 2. Implementation of Web Application for Disease Prediction Using AI 7 up, they will also show the precautionary measures that the patient needs to take care in order to overcome them and treat the infection. OVERVIEW OF WEB SCRAPING Web scraping is a great way to extract random data from websites and convert that data into organized data that can be stored and analyzed in a database. Web scraping is also known as web data extraction, web data removal, web har- vesting, or screen scanning. Web scraping is a form of data mining. The whole purpose of the web crawling process is to extract information from websites and convert it into an understandable format such as spreadsheets, a database, or a CSV, as shown in Figure 1. Data such as item prices, stock prices, various reports, market prices, and product details can be collected with web termination. Extracting website- based information helps to make effective decisions for your business. Figure 1. Web scraping structure. PRACTICES OF WEB SCRAPING • Data scraping • Research • Web mash up—integrate data from multiple sources • Extract business details from business directory web- sites, such as Yelp and Yellow pages • Collect government data • Market analysis The web data scraper process, a software agent, also known as a Web robot, mimics browsing communication between web servers and a person using a normal web browser. Step by step, the robot enters as many websites as it needs, transfers its content to find and extract interesting data, and builds that content as desired. AP scraping APIs and frameworks address the most common web data scrapers involved in achieving specific recovery goals, as described in the following text. Hypertext Transfer Protocol (HTTP) This method is used to extract data from static and dynamic web pages. Data can be retrieved by sending HTTP requests to a remote web server using a socket system. Hyper Text Markup Language (HTML) Exploring languages for query data, such as XQuery and Hyper Text Query Language (HTQL), can be used to scan HTML pages and retrieve and modify content on the page. Release Structure The main purpose is to convert the published content into a formal representation for further analysis and retention. Although this last step is on the side of web scraping, some tools are aware of post-results, providing memory data formats and text-based solutions, such as cables or files (XML or CSV files). LITERATURE SURVEY Python has a rich set of libraries available for downloading digital content online. Among the libraries available, the following three are the most popular ones: BeautifulSoup, LXml, and RegEx. Statistical research performed on the available data sets indicated that RegEx was able to deliver the requested information at an average rate of 153.6 ms. However, RegEx has limitations of data extraction of web pages with internal HTML tags. Because of this demerit, RegEx is used to perform complex data extraction only. Some libraries, such as BeautifulSoup and LXml, are able to extract content from web pages in a complex environment that has yielded a response rate of 457.66 and 203 ms, respectively. The main purpose of data analysis is to get useful information from data and make decisions based on that analysis. Web deletion refers to the collection of data on the web. Web scraping is also known as data scraping. Data analysis can be divided into several steps such as cleaning and editing. Scrapy is the most widely used source of information needed by the user. The main purpose of using scrapy is to extract data from its sources. Scrapy, which crawls on the web and is based on python programming language, is very helpful in finding the data we need by using the URLs needed to clear the data from its sources. Web scraper is a useful API to retrieve data from a website. Scrapy provides all the necessary tools to extract data from a website, process data according to user needs, and store data in a specific format as defined by users. The Internet is very much looking at web pages that include a large number of descriptive elements including text, audio, graphics, video, etc. This process, called web scraping, is mainly responsible for the collection of raw data from the website. It is a process in which you extract data automation very quickly. The process enables us to extract specific data requested by the user. The most popu- lar method used is to create individual web data structures using any known language.
  • 3. 8 Manasvi Srivastava et al. EXPERIMENTAL WORK Technology Used Firebase For database, we have used Cloud Firestore from firebase. It is a real-time NoSQL database that stores two key-value data in the form of collections and documents. TensorFlow TensorFlow is used to train the database model and to make predictions. There are various algorithms for mod- eling training or using line format in our project. JavaScript Frameworks • Node.js Node.js is an open-source, cross-platform, JavaScript runtime environment running back to V8 engine and extracting JavaScript code without a web browser. Our rewriting code is written for Node.js as it is a fast and platform language. • ElectronJS Electron is a framework for building native appli- cations with web technologies such as JavaScript, HTML, and CSS. As Electron is used to create a short web application, it helps us write our code and thus reduces the development time. • ReactJS React makes it less painful to create interactive UIs. Design a simple view of each state in your app and React will carefully review and provide relevant sections as your data changes. React can also render to the server using Node and has the power of mobile applications using React Native. PYTHON Python is a high-level programming language translated into high-level translations. In this project, various libraries, such as pandas, NumPy, good soup, etc., are used to create our database. Pandas and NumPy are used to filter and process data needed to train our model by extracting and removing them from a separate data source. COMPATIBILITY OS X Only 64-bit binaries are provided for OS X, and the lower version of OS X is supported by OS X 10.9. Windows Electron supports Windows 7 and later, but older versions of the OS are not supported. Both x86 and amd64 (x64) binary are provided for Windows and are not supported in the ARM version of Windows. Software Used – VSCode Visual Studio Code is a freeware source code editor developed by Microsoft for Windows, Linux, and macOS. Features include debugging support, syntax highlighting, intelligent coding, captions, reuse code, and embedded Git. – Google colab notebook Colaboratory, or Colab for short, is a product of Google Research that allows developers to write and use Python code through their browsers. Google Colab is an excellent tool for in-depth learning activi- ties. It is a compact Jupyter notebook that needs no setup and has an excellent free version, providing free access to Google computer resources such as GPUs and TPUs. – PyCharm PyCharm is an integrated development platform used for computer programs, especially in the Python language, developed by the Czech company JetBrains. Data Source As we did not get more than 40 diseases, to get a dataset we have created our own dataset. The dataset that we have used for our training and testing process has taken from various sources. One of them is added below. – https://github.com/DiseaseOntology/HumanDise aseOntology Use of Scrapy Scrapy is a framework for crawling and retrieving nonfic- tion data that can be used for the size of a supportive appli- cation, such as data mining, managed, or actual reported data. Apart from the way it was originally expected for Scrapy to be removed from the web, it could be used in the same way to extract data using APIs, for example, Amazon AWS, or as a very important web browser. Scrappy is written in Python. Let us take a Wiki example related to one of these problems. A simple online photo gallery can provide three options to users as defined by HTTP GET parameters at URL. If there are four ways to filter images with three thumbnail-sized options, two file formats, and
  • 4. Implementation of Web Application for Disease Prediction Using AI 9 a user-provided disabling option, then the same content set can be accessed with different URLs, all of which can be linked to the site. This carefully crafted combination creates a problem for the pages as they have to plan with an endless combination of subtitle changes to get different content. Methodology The method used by the project to collect all the required data is extracted from various sources such as the CDC’s database and Kaggle resources. Then, analyze the extracted data using texts written in the Python language according to project requirements. Pandas and NumPy are widely used to perform various functions on the database. After sorting the data according to each need, it is then uploaded to the database. In the database, we have used Cloud Firestore as it is a real-time NoSQL database with extensive API support. Furthermore, the TensorFlow project is used to train our model according to needs. In this project, we predict the disease because of the given symptoms. Training data set: 70% Setting test data: 30% TensorFlow supports Linear Regression, which is used to predict diseases based on the given indicators. Coding Project Frontend is written using ReactJS & TypeScript. However, we have used the MaterialUI kit from Google ReactJS to speed up our development process. To provide our app, Electron is used. Our web system supports macOS and Windows. Most of today’s web app are written with the help of ElectronJS. Testing The project is tested using an Electron built-in test frame- work called Spectron. The project is being implemented in the browser. The output generated turns out to be completely consistent and the generated analysis is approximate. Electron’s standard workflow with Spectron can involve engineers who write unit tests in the standard TDD format and then write integration tests to ensure that acceptance criteria are met before approving a feature to be used. Continuous integration servers can ensure that all these tests are passed before they are incorporated into the production. Algorithm Used Linear Regression is a standard mathematical method that allows us to study a function or relationship in a given set of continuous data. For example, we are given some of the corresponding x and y data points, and we need to study the relationship between them called a hypothesis. In the event of a line reversal, the hypothesis is a straight line, i.e., Where the vector is called Weights and b is a scale called bias. Weights and bias are called model parameters. All we need to do is estimate the values of w and b from the set of data, given that the result of the assumption has produced the minimum cost J defined by the next cost function where m is the number of data points in the data pro- vided. This cost function is also called the mean squared error. To find the optimized value of the J’s minimum param- eters, we will be using a widely used optimizer algorithm called Gradient Descent. The following is a fake Gradient Descent code: RESULT DISCUSSION The overall results of the project are useful in predicting diseases with the given symptoms. The script that was written to extract data can be used later to compile and format it according to needs. Users can pick up symbols by typing them themselves or by selecting them from the given options. The training model will predict the disease according to it. Users are able to create their own medical profile, where they can submit their medical records and prescribed medication; this greatly helps us to feed our database and better predict disease over time, as some of these diseases occur directly during the season. Moreover, the analysis performed showed a very sim- ilar disease, but the training model lacks the size of the database. CONCLUSIONS AND FUTURE SCOPE The use of the Python program also emphasizes under- standing the use of pattern matching and general expres- sions for web releases. Database data are compiled from factual reports and sent directly to government media out- lets for local media where it is considered reliable. A team of experts and analysts who validate the information from a continuous list of more than 5,000 items is likely to be the site that collects data effectively. User-provided inputs are analyzed and deleted from the website, and the output is extracted as the user enters the user interface encounters. Output is generated in the form of text. This method is simple and straightforward to eradicate the disease from companies and provides vigilance against that disease.
  • 5. 10 Manasvi Srivastava et al. For future work, we plan tests that aim to show the med- ication that a patient can take for treatment. In addition, we are looking to link this website to various hospitals and pharmacies for easy use. REFERENCES [1] Thivaharan. S, Srivatsun. G and Sarathambekai. S, “A Survey on Python Libraries Used for Social Media Content Scraping”, Proceed- ings of the International Conference on Smart Electronics and Com- munication (ICOSEC 2020) IEEE Xplore Part Number: CFP20V90- ART; ISBN: 978-1-7281-5461-9, PP: 361–366. [2] Shreya Upadhyay, Vishal Pant, Shivansh Bhasin and Mahantesh K Pattanshetti “Articulating the Construction of a Web Scraper for Massive Data Extraction”, 2017 IEEE. [3] Amruta Kulkarni, Deepa Kalburgi and Poonam Ghuli, “Design of Predictive Model for Healthcare Assistance Using Voice Recogni- tion”, 2nd IEEE International Conference on Computational Sys- tems and Information Technology for Sustainable Solutions 2017, PP: 61–64. [4] Dimitri Dojchinovski, Andrej Ilievski, Marjan Gusev Interactive home healthcare system with integrated voice assistant MIPRO 2019, PP: 284–288 Posted: 2019. [5] Mohammad Shahnawaz, Prashant Singh, Prabhat Kumar and Dr. Anuradha Konidena, “Grievance Redressal System”, International Journal of Data Mining and Big Data, 2020, Vol. 1, No. 1, PP. 1–4. Hyperlink of Research Paper 1. https://www.sciencedirect.com/science/article/pii/S2352914817302253