Tourism Enhancement Using LLMs & Neural Network_Report (1)
Tourism Enhancement Using LLMs & Neural Network_Report (1)
Tourism Enhancement Using LLMs & Neural Network_Report (1)
Network
By
Name of Student Roll Number
Anurag Verma 2020UCO1580
Gaurav Bansal 2020UCO1585
Rohit Kumar 2020UCO1597
Of
May 2024
Certificate
This is to certify that the work embodied in project thesis titled, "Tourism
enhancement using LLMs and Neural Networks” by Anurag Verma
(2020UCO1580), Gaurav Bansal (2020UCO1585) and Rohit Kumar
(2020UCO1597) is the bonafide work of the group submitted to Netaji Subhas
University of Technology for consideration in 8th Semester B.Tech. Project
Evaluation.
The original Research work was carried out by the team under my/our guidance
and supervision in the academic year 2023-2024. This work has not been
submitted for any other diploma or degree of any university. On the basis of
declaration made by the group, we recommend the project report for evaluation.
2
Candidate(s) Declaration
3
Certificate of Declaration
This is to certify that the Project-Thesis titled "Tourism enhancement using LLMs
and Neural Networks" which is being submitted by Anurag Verma
(2020UCO1580), Gaurav Bansal (2020UCO1585) and Rohit Kumar
(2020UCO1597) to the Department of Computer Science & Engineering, Netaji
Subhas University of Technology (NSUT) Dwarka, New Delhi in partial fulfilment of
the requirement for the award of the degree of Bachelor of Technology, is a record of
the thesis work carried out by the students under my supervision and guidance. The
content of this thesis, in full or in parts, has not been submitted for any other degree or
diploma.
Place:
Date:
4
Acknowledgement
We would like to express our gratitude and appreciation to all those who made
it possible for us to complete this project. Special thanks to our project supervisor(s)
Dr. Vandana Bhatia whose help, stimulating suggestions and encouragement helped
us in writing this report. We also sincerely thank our colleagues for the time spent
proofreading and correcting our mistakes.
We would also like to acknowledge with much appreciation the crucial role of
the staff in the Department of Computer Science & Engineering, who gave us the
permission to use the lab and the systems and gave permission to use all necessary
things related to the project.
5
Plagiarism Report
6
Abstract
Traveller happiness and destination management have grown more dependent on the
utilisation of data-driven approaches in the fast changing tourism market. In order to
maximise the complete travel experience, this thesis investigates the integration of
modern data analytics to transform recommendation systems and tourist analysis.
Through the utilisation of various datasets, such as user preferences, past travel
trends, and up-to-date contextual data, this study aims to establish a sturdy structure
for all-encompassing tourism examination.
The goal of the project is to create intelligent recommendation systems that employ
machine learning algorithms to customise recommendations according to user
preferences, behavioural trends, and the ever-changing nature of travel locations. It
also explores the privacy ramifications and ethical issues surrounding the processing
of sensitive travel data.
7
List of Contents
Introduction ………………………………………………………………… 10
Background ………………………………………………………… 10
Motivation ………………………………………………………… 10
Problem Statement ………………………………………………… 11
Objectives ………………………………………………………… 11
Literature Survey ………………………………………………… 12
Conclusion …………………………………………………………………… 33
Accomplishments ………………………………………………… 33
Future Scope ………………………………………………………… 34
References …………………………………………………………………… 36
8
List of Figures
2 Feature Matrix 20
9
Chapter - 1
Introduction
1.1 Background
India's economy greatly benefits from the tourism sector, which is vital to the country.
India has a rich history, varied landscapes, rich cultural legacy and emerging trends
draws millions of people there every year. But choosing the best places to go and
things to do that fit their interests and preferences may be difficult for travellers. We
suggest creating a recommendation model for Indian tourism destinations in order to
address this problem.
The goal of this research is to create a recommendation model that is especially suited
for Indian tourism. In order to offer customised suggestions, it will take into account
variables including user preferences, geographic location, and cultural interests and
provide the best alternatives in the near vicinity of their preferences.
1.2 Motivation
India's tourism economy is based mostly on cultural visits, which boosts GDP and
creates a large number of jobs. India's incredible diversity—which includes beautiful
landscapes, a rich cultural tapestry, historical riches, and a lively tapestry of
traditions—is what makes it so alluring. Due to its diversity, India is one of the most
popular travel destinations in the world, drawing millions of visitors each year. Even
though there are a lot of other places that do not necessarily have cultural
significance, they offer peace, joy and thrill to those seeking adventure and want to
take a relaxing break from their daily lives. There are certainly a lot of available
choices to make. But even with all of India's resources, travellers frequently encounter
a difficult obstacle: sorting through the plethora of choices and choosing locations and
activities that suit their own interests, limitations, and preferences.
Information Overload
When planning a vacation, tourists mostly encounter information overload because
there are a lot of websites, blogs, and guidebooks available that give contradicting
recommendations.
10
Technological Advancements
Developments in data analytics and machine learning offer a chance to build complex
recommendation systems. In the tourism industry, employing these technologies may
greatly improve the calibre of suggestions.
Our project aims to create a personalised travel recommendation system for India that
takes into account the varied interests of users as well as their financial limitations. In
order to provide a more personalised and improved travel experience, we try to take
into account user feedback and trends in destination popularity. In addition, we'll
concentrate on ethical data handling, scalability, and real-time updates to develop a
complete solution for tourists visiting India.
1.4 Objectives
11
6. Privacy and Data Security: When creating a recommendation system, it is
important to guarantee the confidentiality and integrity of user information.
12
hidden representations of tourism preferences, this study introduces an autoencoder
and then uses it for clustering. With the help of this deep learning method, complex
patterns of tourist behavior can be discovered and highly accurate and personalized
recommendations can be made. This work is an example of using intelligent machine
learning techniques to extract different insights from tourism data.
Based on these findings, NWPY Praditya, AE Permanasari[6] present a unique hybrid
tagging system that combines clustering, content-based analysis, and collaborative
analysis to provide travel recommendations. This comprehensive strategy uses each
method and its capabilities to provide highly accurate and context-sensitive
recommendations. This study shows how this hybrid system can improve accessibility
and user satisfaction, demonstrating its potential as a recommender system.
In a related work, CC Yu and H Chang[7] investigate how on-site services can be
combined with aggregation methods to provide tourism advice. Based on the tourist
and their current geographic location, the system uses location data and algorithms to
provide trending information. This method ensures that it is not only personalised but
more relevant to the user and their surroundings, improving the overall travel
experience.
In addition, deep neural networks such as General Adversarial Networks (GANs)
have great potential for use in tourism recommendation systems. Research by EE
Stephy and M Rajeswari shows that GANs can generate synthetic data that matches
the needs and behaviours of real tourists. By using various datasets of tourist
interactions and interests during training, the GAN can generate accurate information
about virtual visitors. These documents can be used to rank algorithms to increase the
accuracy and depth of testing. This method provides a new opportunity to generate
highly personalised and contextual tourism recommendations, thus improving the user
experience.
13
Chapter - 2
Data Collection and Cleaning
We used web scraping techniques on different travel websites to collect data for
training a machine learning model to recommend tourist attractions in India. In order
to scrape data from Lonely Planet, we used Beautiful Soup[9] with the requests[10]
library, and we used Selenium[11] to extract data from the dynamic websites Trip.com
and TripAdvisor.in. We also incorporated google maps to enrich our data with
location coordinates to enhance the recommendations based on the location
preferences. We explain our methodology in more detail and give the rationale behind
our decisions below.
1. Data Availability: Trip.com and TripAdvisor.in are known for their extensive
databases, offering detailed information on attractions, user reviews, and
ratings, providing our model with a rich knowledge base.
We employed tailored web scraping techniques for each website. Beautiful Soup was
used for Lonely Planet due to its static structure, while Selenium was chosen for
Trip.com and TripAdvisor.in, which feature dynamic content.
14
2.1.3 Reasons for choosing Beautiful Soap:
We used Beautiful Soap to extract data from the Lonely Planet[12] website.
● The HTML source code of Lonely Planet's website contains easily accessible
data, and its structure is comparatively static.
● When it comes to effectively extracting information from static web pages,
Beautiful Soup excels.
● The website is a good option because it permits scraping in accordance with
its terms of service.
We used Beautiful Soap to extract data from the Trip Advisor[13] and Trip.com[13]
websites.
The Dataset contains Names of Tourist destinations, Rating out of 5 for the given
destination, Amount spent to visit these places, Reviews count, City in which the
Place is located at and Tags associated with the destination. The tags serve the most
important role in our recommendation algorithm and we have assigned clusters
according to the number of tags in the entire dataset.
The missing values are filled with the average of the entire dataset as of now to
suggest relatively similar places. ‘
Feature engineering was pivotal in enhancing our dataset's quality and relevance. We
extracted geographical coordinates, enabling location-based recommendations.
Textual features, including sentiment analysis scores and keyword extraction, were
15
derived from descriptions and user-generated reviews. Features related to ratings,
review counts, and popularity quantified attraction appeal. We also categorised
attractions based on attributes like historical significance, natural beauty, adventure,
and cultural relevance, providing additional features for personalised
recommendations.
In summary, our data collection and feature engineering efforts have produced a
robust dataset, empowering our machine learning model to offer personalised and
accurate recommendations for travellers exploring India's diverse tourist attractions.
16
Chapter - 3
Exploring Fuzzy Clustering Algorithm
One of our initial choices for the recommendation algorithm is Fuzzy C-Means
Clustering. It offers more flexibility than its counterparts like K-means Clustering. It
does not strictly adhere to the principle that all the clusters need to be spherical in
shape. Given that a destination might be a part of many clusters, it produces superior
results in datasets that overlap. In this case, fuzzy clustering would be a preferable
option.
1. Soft Assignments:
Soft assignments are supported by FCM, each data point might have various
degrees of membership in several clusters. Item features and user preferences
frequently show some degree of resemblance. By considering intrinsic
ambiguity in user-item relationships, it often portrays complex preferences.
2. Flexibility in Membership:
FCM permits data points to partially belong to multiple clusters whereas hard
clustering states that a data point strictly belongs to one cluster.
Recommendation systems benefit from this flexibility because it represents the
heterogeneous and overlapping nature of user preferences.
3. Robustness to Noise:
It is well known that FCM is resistant to data noise and outliers. The
dependability of clustering findings in recommendation systems is improved
by the capacity to handle noisy data, as user behaviour might occasionally be
irregular or prone to abnormalities. The recommendation system's accuracy
and stability are enhanced by this resilience.
17
a variety of data sources, including user ratings, demographic data, and textual
evaluations. FCM's capacity to combine data from several sources makes for a
more thorough clustering strategy.
There were several challenges in implementing the naïve Fuzzy C-Means (FCM)
clustering method. Which resulted in inconsistent results and made it difficult to
handle noise in the data. These problems obstructed the clustering process and called
for more research and adjustments.
The volatility of the clustering assignments produced by the FCM method was one of
the main causes for worry. The model frequently assigned distinct destinations to
separate clusters throughout successive iterations. The algorithm's stochastic structure
selects the initial cluster centroids randomly. As a result, slight variations in the
starting circumstances or the sequence of the data may provide varied clustering
results. Repeating the procedure with several initializations lessened the randomness
of assigned clusters.
Another noteworthy obstacle was the vulnerability of FCM to noise present in the
dataset. Because FCM is intrinsically sensitive to noise and outliers, the clustering
results may be skewed. Frequently, a significant portion of the destinations were first
assigned to no cluster, leaving an imperfect clustering solution. One popular solution
to this problem is to use the dataset's average values in place of the null entries. This
method, however, may result in a bias in the clustering, assigning the bulk of the
destinations to one conspicuous cluster.
The overall clustering result might be greatly distorted by the dominance of a single
cluster brought about by noise reduction measures. Other approaches to treating noise
should be investigated in order to get around this problem. For example, robust
clustering algorithms that can manage outliers or preprocessing techniques that can
recognise and remove noisy data points should be investigated. Additionally, by
specifying suitable noise thresholds or adding weighted distance measures to lessen
the impact of noisy data, domain expertise or expert input can assist optimize the
clustering process.
To conclude, the implementation of naïve FCM clustering has brought to light some
significant issues, such as noise intolerance and unstable cluster assignments.
Strategies like limiting randomization in initialization and implementing more
advanced noise-handling techniques should be taken into consideration in order to
improve the resilience and reliability of the clustering findings. More consistent and
significant clustering findings for destination analysis or related applications can be
achieved by carefully weighing these difficulties and investigating suitable solutions.
18
Similar to fuzzy logic, an example in a fuzzy cluster has a degree of membership to
the cluster but does not belong to any particular cluster. We have a coefficient that
indicates the degree uk(t) of belonging to the k-th cluster for each sample t. Since the
total of those coefficients is typically 1, uk(t) may represent the likelihood that an
example t falls into a certain cluster:
(3.1)
A fuzzy version of the k-means partitional method is called the fuzzy c-means. The
centroid of a cluster, also known as a prototype, is calculated using fuzzy c-means as
follows: the mean of all samples is weighted according to their degree (uk) of
belonging to the cluster Ck:
(3.2)
To ensure that the total of the coefficients is equal to 1, they are normalised and
fuzzified with a real parameter m > 1.
(3.3)
19
Fig 2. Feature Matrix
1. Sensitivity to initial conditions: The original cluster centre location can have
a significant impact on the fuzzy clustering result. Different initial cluster
assignments lead to distinct cluster outcomes. Since the initial cluster is not
that the algorithm allows to assign it is difficult to conclude meaningful
results.
2. Computational Complexity: For a high number of data points computations
are also considerably large, it requires computational resources and a lot of
time to derive results.
3. Need for pre-specification of cluster number: We need to define the number
of clusters beforehand which is difficult to predict since there are various
factors affecting the cluster assignment.
4. Difficulty in interpreting results: Rather of rigid assignments, membership
degrees are produced using fuzzy clustering. In contrast to rigorous clustering
techniques like K-means, this may make it more difficult to comprehend the
groupings.
5. Not suitable for all types of data: It might not function effectively on data
including clusters that are complexly geometrized or have irregular shapes.
6. Assumption of fuzzy membership: It is assumed that every data point has
some degree of affiliation with every cluster. For some datasets, this
assumption might not always be true.
20
7. Vulnerable to noise and outliers: Fuzzy C-means, like many clustering
techniques, is susceptible to noise and outliers, which might result in less
significant clusters.
8. Lack of clear stopping criterion: Convergence of this algorithm is not
guaranteed. We need to specify the interactions it can perform which can be
tricky to estimate.
9. Difficulties with high-Dimensional data: The "curse of dimensionality,"
which can reduce the significance of distance-based computations, may
prevent fuzzy C-means from performing effectively with high-dimensional
data.
10. Limited Scalability: The computing needs of very big datasets may cause it
to suffer.
In order to address these drawbacks, we have used other techniques, like hierarchical
clustering that are more suited for handling huge datasets and structured results.
Furthermore we will investigate unsupervised RNN like LSTM, Deep Neural network
models, such as GAN, in addition to these techniques, using the insights we have
gathered from related research.
21
Chapter - 4
Applying Hierarchical Clustering
This was motivated by the fact that most work on hierarchical clustering was focused
on providing algorithms rather than optimizing a specific objective. He demonstrated
the following desired characteristics of this cost function: Higher layers of the
hierarchy must be used to separate unconnected components, or dissimilar pieces, in
order to attain the best cost. When the similarity between data items is the same, all
clusterings result in the same cost.
Fuzzy C-Means (FCM) clustering and hierarchical clustering are two separate
methods with varying advantages. Here are some benefits of hierarchical clustering
over FCM, albeit the decision between them will rely on the particulars of the data
and the analysis's objectives:
3. Interpretability of Results:
In many cases, hierarchical clustering's hierarchical clustering structure
closely mimics the hierarchical patterns seen in real data. This can help with
22
the results' interpretability, particularly in cases when there is a hierarchical
structure in the interactions between the clusters. Although fuzzy clustering
methods like FCM offer membership degrees, their interpretation of clusters
may not be as clear-cut.
4. Handling of Outliers:
Because the effects of an individual outlier are usually contained inside a tiny
subtree of the dendrogram, hierarchical clustering is comparatively resistant to
outliers. However, because FCM gives membership degrees to individual data
points, it may be susceptible to outliers, which might affect the degree of
membership across several clusters.
A simple technique used in data analysis to show the hierarchical structure of a data
set is hierarchical classification. Data preparation is the first of many important steps
in planning. What is compiled indicates that the data set under investigation was
generated or collected. The next important step is to calculate the connection. In other
words, it determines how similar or different the data points are from each other.
Different clustering methods, such as neighbourhoods, sums, or averages, affect the
clustering process by determining whether clusters should be merged or split.
23
logical groupings of data.
In addition, the use of hierarchical classification allows for deeper exploration of data
relationships. Using a variety of correlation methods, analysts can tailor the
aggregation method to the specific characteristics of the data set. This method is a
systematic way of looking for patterns in the data, depending on whether the focus is
on the neighbourhood correlation, the total correlation, or the mean. Ward correlations
emphasize cohesion within clusters, while mean correlations capture the diversity of
individuals. Due to its useful and detailed representation of dendrograms, hierarchical
clustering is a very useful and intelligent tool for understanding complex data
structures at a deep level.
24
4.4 2D Scatter Plot Visualisation using Hierarchical Clustering:
2D scatterplots are a useful tool to visualize clustering results clearly and concisely in
the context of hierarchical classification. Scatterplots can be used to show the two
components when the power is increased, so that both sets of data have significant
features.
● Dendrogram Overlay:
A dendrogram is the first graphical representation of a hierarchical
classification, but overlapping cluster assignments on a two-dimensional
scatter plot provides a quick and informative way to collect data points. The
value of each point in the two selected categories corresponds to a point in the
scatter, and the color and shape of the point indicate its relationship to a
specific group.
Average Linkage:
daverage(Ci, Cj) = (1 / |Ci| * |Cj|) 𝚺pϵCi 𝚺qϵCj d(Xp, Xq) (4.1)
● Identification of outliers:
2D scatter plots make it easy to see what's in and what's in your data. Visual
cues are provided to help identify and deal with outliers, such as deviations
from the default build template.
● Pattern Recognition:
The scatter plot you create by viewing a hierarchical ranking will help you
identify patterns and trends in your data. Analysts can determine whether
clusters expand or form tight groupings, providing insight into the underlying
structure.
25
Fig 5. 2D Scatter plot using t-sne
26
shown in a three-dimensional scatterplot. To facilitate visual segmentation,
data points in the same cluster are often assigned the same symbol color or
shape.
● Interactive Exploration:
Enhance your browsing experience with tools and frameworks that provide
interactive 3D visualizations. Panning, zooming, and rotating allow analysts to
dynamically visualize the spatial distribution of clusters. This interactive
feature allows you to understand the hierarchical structure of your data.
27
Chapter - 5
Sentimental Analysis using LLMs
There are two main stages to the LLM training procedure. They go through
pre-training first, when the model learns to anticipate the word that will come after a
particular phrase using large datasets. This first stage gives the model a general
knowledge of semantics and language structure. The model is then fine-tuned for
certain activities or domains, allowing its vast knowledge to be applied in more
specialized settings.
The ability of LLMs to comprehend plain language is what sets them apart. Their
comprehension of written text's context, sentiment, and intent is unmatched. This skill
has real-world uses in sentiment analysis, chatbots, and virtual assistants. The model's
ability to understand subtleties in language improves user engagement and makes for
a more interesting and natural-feeling discussion.
28
5.2 Sentimental Analysis using LLMs
Using Large Language Models (LLMs) for sentiment analysis offers a sophisticated
and context-aware method of deciphering the sentiment represented in textual data,
which is a significant leap in natural language processing. The use of LLMs, most
notably models like Gemini, which are excellent at capturing the subtleties of
language patterns and contextual nuances, is the fundamental component of this
technique.
The creation of a properly labeled dataset is the first stage in the sentiment analysis
process using LLMs. This dataset, which includes a variety of text samples matched
with appropriate sentiment labels—including positive, negative, and neutral
sentiments—is essential for training the model. The model can effectively generalize
to the heterogeneity of linguistic expressions in real-world circumstances thanks to
the dataset's diversity.
The trained model goes through its layers to process the input text and provide a
sentiment prediction during the inference and prediction phase. The model's
confidence in each sentiment category is often reflected in the output, which is a
probability distribution across the various sentiment classes. In order to provide a
final sentiment label, post-processing stages may entail applying thresholds to these
probabilities.
The ability of LLMs to manage contextual sentiment is one of its most prominent
strengths in sentiment analysis. These algorithms are able to understand subtle
differences in sentiment depending on the context of a sentence or paragraph. Since
natural language utterances can contain layers of complexity and ambiguity,
contextual awareness is essential for effectively reading sentiment in natural
language.
29
Sentiment analysis with LLMs has many useful and profound real-world uses.
LLM-based sentiment analysis has proven beneficial in a variety of contexts,
including market research to detect sentiment patterns, customer feedback analysis to
assess product evaluations, and social media monitoring to measure public opinion.
Furthermore, incorporating these models into chatbots and virtual assistants improves
their capacity to connect with people in an emotionally intelligent manner.
When sentiment analysis is applied using Large Language Models (LLMs) and
hierarchical clustering, textual data is first thoroughly understood and categorized
according to sentiment, and then hierarchical clustering is used to further refine the
results. Data preparation is the first stage in the process, and it's important to do this to
make sure the textual data is ready for further analysis. To construct a clean and
uniform dataset, this requires tokenization, stop word removal, and sometimes
stemming or lemmatization.
The emphasis then shifts to using a pre-trained LLM, such GPT-3, for sentiment
analysis on the dataset that has already been preprocessed. Because of their innate
talents, LLMs are able to recognize the subtleties of language and give each text entry
a sentiment label, usually categorizing the entries as positive, negative, or neutral.
Through the creation of sentiment-labeled clusters, this sentiment analysis stage offers
a fundamental knowledge of the emotional tone portrayed within the textual data.
Hierarchical clustering is the next step after the dataset has been given sentiment
labels. A useful method for organizing related data points into hierarchical structures
or trees is hierarchical clustering. This clustering stage attempts to arrange the dataset
according to the discovered sentiments in the context of sentiment-labeled data. Text
30
items that share sentiment labels are clustered together to form cohesive groups that
encapsulate the shared emotional context of the related sentences.
Potential subcategories under broad sentiment labels can be found through a nuanced
analysis of sentiment patterns made possible by the hierarchical clustering method. A
more detailed picture of the sentiment landscape is given by this hierarchical
structure, which identifies groups that share both general sentiment and maybe more
nuanced emotional aspects.
This two-step method provides a comprehensive way to find and arrange patterns in
textual data by integrating sentiment analysis utilizing LLMs with hierarchical
clustering. It makes it easier to identify desirable clusters based on emotional tones
and allows for a more comprehensive comprehension of sentiment differences.
Applying this technique to customer reviews, social media data, or any other text-rich
dataset offers a strong foundation for gaining insights and honing the research through
hierarchical organization. The accuracy and significance of the detected clusters may
be further improved by ongoing refinement and validation against ground truth data,
which can provide important insights into the emotional terrain captured in the textual
material.
31
Sentiment analysis applied to hierarchical clustering clusters provides a potent way to
investigate the subtle emotional aspects in large, intricate datasets. It goes beyond a
general sentiment analysis by offering a hierarchical organized framework for
sentiment analysis. A more focused investigation of sentiment dynamics within
certain settings or topics is made possible by the representation of clusters, which are
collections of data points with comparable attributes.
32
Chapter - 6
Conclusion
6.1 Accomplishments:
The way we view and traverse tourist attractions has been revolutionised by the
subsequent visualisation of these clusters on Google Maps. With the help of this
interactive mapping function, visitors and destination managers may make
well-informed selections by simplifying complicated data into cohesive clusters. The
easily navigable website helps with strategic tourism management as well as effective
itinerary planning, enabling local authorities to pinpoint and attend to particular
requirements within each cluster.
● Scraped data from different websites and compiled it down to more than 17k
locations within India.
● Used LLMs to extract the sentiment of user from the prompts he passes to our
application, providing him with relevant tourism recommendations with ease
of use.
● Used Hierarchical Clustering for dividing the locations into different clusters
using multiple features, including reviews, review count, distance, tags,
latitude and longitude.
● Visualised the clusters in 2D as well as 3D scatter plots using Principal
Component Analysis (PCA) as well as t-sne methods.
● Visualised the clusters on Google Maps along with the inter connecting lines
between different locations in a cluster, thus aiding in clear visualisation of
distance between similar locations with a locality.
● Made a website where users can get recommendations similar to a location
within their city, or within complete India, if they wish so.
33
Fig 9. Visualisation of Clusters in Delhi
34
We also want to create customised assessment criteria in order to measure the efficacy
of our updated clustering strategy. With the use of these indicators, we will be able to
evaluate the effectiveness of our recommendations statistically and make any
necessary corrections.
We also intend to incorporate user feedback systems and dynamic updates. This will
guarantee that, over time, our recommendations stay relevant and in line with the
tastes of specific users. We are dedicated to providing a tourism destination
recommendation system that constantly raises user happiness and provides
high-quality suggestions by listening to customer comments and improving our
algorithms.
35
References
[1] Z. Wang, B. Liu, and B. Liu, “Tourism recommendation system based on data
mining,” J. Phys. Conf. Ser., vol. 1345, no. 2, p. 022027, Nov. 2019, doi:
10.1088/1742-6596/1345/2/022027.
[3] Z. Jia, Y. Yang, W. Gao, and X. Chen, “User-Based Collaborative Filtering for
Tourist Attraction Recommendations,” in 2015 IEEE International Conference on
Computational Intelligence & Communication Technology, Feb. 2015, pp. 22–25.
doi: 10.1109/CICT.2015.20.
[9] “Beautiful Soup Documentation.” Accessed: Feb. 08, 2023. [Online]. Available:
https://www.crummy.com/software/BeautifulSoup/bs4/doc/#module-bs4
36
[10] K. Reitz, “requests: Python.” Accessed: Feb. 08, 2023. [OS Independent].
Available: https://requests.readthedocs.io
[14] “Trip.com Official Site.” Accessed: Feb. 08, 2023. [Online]. Available:
https://us.trip.com/?locale=en-us
37