1. Introduction
Next POI recommendation, as a research hotspot in location-based social networks, not only allows users to find their desired POIs quickly, but also provides opportunities for merchants to cater to user preferences. However, in commonly used datasets for next POI recommendation, the user-POI interaction matrix is relatively sparse, making accurate recommendation challenging based on limited information. This is often referred to as cold start problem in recommendation systems. One feasible solution to address the cold start problem is to leverage additional auxiliary information such as POI information, user profiles, social relationships [
1], spatial information [
2], social network detection [
3]. However, due to privacy concerns, the obtained auxiliary information related to users is usually scare, which makes it extremely difficult to accurately capture users’ preferences based on limited information. Moreover, it is not easy to effectively simulate dynamic changes in user preferences for cold start users.
In real life, there are a large number of users who have some check-in activities in the past period, but have few or almost no check-in activities in the recent period. We define these users as recent check-in cold start users. Currently, there are methods [
4,
5] that study how to capture dynamic preferences of users changing with spatio-temporal factors. However, these methods mainly focus on active users with frequent check-ins, filtering out cold start users with fewer check-in activities and cold start POIs with fewer user check-ins. Therefore, there is a lack of a next POI recommendation model for recent check-in cold-start users, which integrates multi-dimensional auxiliary information from user check-in activities to effectively simulate user dynamic preferences.
Recently, research shows that meta-learning [
6] can endow models with the ability to “learn how to learn”, enabling them to quickly adapt the global model to new tasks (users) with little or no interaction information, so as to perform accurately recommendation for new tasks (users). Consequently, faced with cold start data in user check-ins, researchers began to integrate meta-learning into next POI recommendation systems, transforming the recommendation task into a few-shot learning task. For example, Zhang et al. [
7] captured the transition information between two adjacent POIs in a user’s long term historical check-in sequence as transferable generalized knowledge, and then simulated the transition preferences of cold start users to visit next POI. Vinayak et al. [
8] conducted lightweight unsupervised cluster-based information propagation based on long term check-in sequence information between users with similar preferences and locations in data-rich regions, thus improving the quality of recommendation in data-scarce regions. Although existing next POI recommendation approaches based on meta-learning has achieved inspiring results, there are still two limitations.
(1) Existing methods mainly learn users’ long term static preferences according to their historical check-in sequences, and ignore the impact of multi-dimensional dynamic preferences of recent check-in cold start users on next POI recommendation.
(2) Ignoring auxiliary information such as user check-in frequency and POI category, there is a lack of an integrated recommendation model for next POI to alleviate user cold start problem.
To address the issues as described above, we propose a POI Recommendation model based on Meta-Learning (ML-POIRec) for recent check-in cold start users. This model takes into account multi-dimensional auxiliary information in user check-in activities. Firstly, it utilizes a static LSTM module to capture users’ multi-dimensional long term static preferences based on their check-in activities within historical time periods. Then it employs a dynamic meta-learning module to capture users’ multi-dimensional dynamic preferences based on their check-in activities within the current time period. In addition, we incorporate a POI category filter into our model to filter candidate POIs, and calculate the preference score for each POI by taking dot product between filtered POIs and the user comprehensive preference obtained by integrating the above two user preferences, and rank preference scores for next POI recommendation. Finally, to validate the effectiveness of the proposed model, we conduct a series of experiments on two well-known LBSN datasets. The results indicate that ML-POIRec outperforms seven baseline methods on both datasets.
To summarize, the main contributions of this paper are listed as follows:
(1) We propose a novel approach for next POI recommendation by modeling long and short term user preferences with multi-dimensional auxiliary information, which greatly improves the recommendation performance.
(2) We deeply mine users’ multi-dimensional long term static preferences and dynamic preferences by static LSTM module and dynamic meta-learning module, respectively. Moreover, we incorporate a POI category filter into our model. These factors are considered comprehensively to more accurately simulate users’ preferences.
(3) Extensive experiments on two real-world datasets are conducted to evaluate the performance of the proposed model. The results show that our ML-POIRec model achieves significant performance improvements over the state-of-the-art baseline methods.
The remainder of this paper is organized as follows.
Section 2 reviews the related work. In
Section 3, the problem definition and data analysis are presented, along with an explanation of key symbols used in this paper. In
Section 4, we introduce the details of the proposed ML-POIRec model. Next, the effectiveness of ML-POIRec is evaluated in
Section 5.
Section 6 discusses the practical implications of ML-POIRec. Finally,
Section 7 concludes this paper.
4. Proposed Method
In this section, we elaborate the overall framework of our ML-POIRec model. As shown in
Figure 3, our model mainly includes the following four parts:
(1) Multi-dimensional check-in information embedding layer. It is used to obtain a dense representation of a user check-in sequence information, including POI ID, POI category information, time interval information, spatial interval information and user check-in frequency.
(2) Multi-dimensional static preference LSTM module. In order to obtain user’s multi-dimensional long term static preference , this module is leveraged to update LSTM model parameters according to the information of user’s interaction with POIs in the past time period.
(3) Multi-dimensional dynamic preference meta-learning module. It aims to capture user’s multi-dimensional short term dynamic preference by considering the information of user’s interaction with POIs in the current time period. Training the meta-learning module with multiple users (tasks) enables the model to have strong generalization ability, thus making it applicable for testing new users.
(4) Next POI Recommendation. The obtained user’s multi-dimensional static preference representation and dynamic preference representation are integrated to obtain the comprehensive preference representation of the user, which is then matched with the POI embedding after category filtering in the candidate pool, the preference score of next POI is calculated, and POIs with top-n scores are recommended.
4.1. Multi-Dimensional Check-In Information Embedding Representation
The multi-dimensional check-in information embedding layer is leveraged to encode the multi-dimensional auxiliary information of user and POI interactions into a latent representation. According to the analysis in
Section 3.2, the check-in time of a user visiting a POI, the geographical distance between two visited POIs, the check-in frequency of POIs, and POI categories all have an impact on user’s selection of next POI. Therefore, the multi-dimensional auxiliary information such as POI ID information, POI category information, spatio-temporal label of the user visiting POIs, and user check-in frequency are taken as inputs, and the embedding vectors are obtained according to the inputs and then concatenated.
To take full advantage of the multi-dimensional auxiliary information, we use a similar approach in [
25] to generate the initial representation of POI using the embedding matrix
, where
d is the latent dimension and
m is the dimension of POI feature. First, POI
is encoded as a binary vector
, and then
is used to transform to obtain
. Thus, the obtained embedding vectors of POIs can be stacked as:
where L is the total number of POIs from a user check-in sequence. After learning user’s latent representation,
will be optimized as the model is trained, as described in
Section 4.5.
4.2. Multi-Dimensional Static Preference Modeling for Users
The historical check-in information of users is particularly important for modeling their long term stable (static) preferences. By capturing users’ multi-dimensional static preferences and combining them with their multi-dimensional dynamic preferences, it is helpful to more accurately simulate users’ comprehensive preferences. Considering LSTM is very effective for processing sequential data with rich contextual information, and can alleviate the gradient vanishing problem of RNN. Therefore, we use LSTM network to encode user’s historical check-in information to simulate user’s long term stable latent preference representation
, defined as:
where
represents the LSTM model,
denotes user’s multi-dimensional static latent preference factor in the previous time period,
represents the set of check-in activities of user
u in the time period
t, and
is the network parameter. We update
using the stochastic gradient descent method, shown as:
where
is the step size hyperparameter and
denotes the regularization parameter. For
and
, see
Section 4.3.
4.3. Multi-Dimensional Dynamic Preference Modeling for Users
The main task of user’s multi-dimensional dynamic preference meta-learning module is to capture user’s multi-dimensional dynamic preference by considering the interaction information between users and POIs in the current time period. Here, we adopt meta-learning because it is a popular few-shot learning method and can alleviate the cold start problem, as described in
Section 2.2. However, different from existing meta-learning based POI recommendations [
26,
27], our model takes input from user’s check-in activity in the current time period. In this way, it can capture user’s potential preferences for visiting POIs in the current time period, so as to provide more accurate next POI recommendation. We consider each user as a learning task to learn a meta-learning parameter
that represents the global user model parameter within time period
t in the meta-training users set including users with enough check-in activities. We follow settings of few-shot learning [
32], where tasks’ distribution is represented as
. The model is iteratively trained by sampling task
from
, and user’s multi-dimensional dynamic latent preference within time period
t generated by this module is shown as:
where
represents the meta-learning model, and
indicates the task of user
u in time period
t, that is, the check-in activity set
of the user at current time period
t. While each task includes support set
and query set
, where
includes the first
k check-in activities of a user in current time period
t, and
includes the remaining check-in activities of a user in current time period
t.
Specifically, we adopt optimization-based meta-learning methods to learn users’ multi-dimensional dynamic preferences. The input to the meta-learning module is the embeddings of POIs visited by the user during current time period t. The training process of the meta-learning module includes both inner and outer loops as described below.
Inner loop: we pass support set
of a task
to
and adjust the model parameter
of user
u based on the global user model parameter
to perform local update, as shown in Formula (5).
The update is achieved by one or more gradients with respect to the global parameter
. Where
is the step size hyperparameter,
represents the gradient error with respect to the initial parameter
. While
is the loss function of model propagation sample, which is calculated based on support set
. Our ML-POIRec model mainly calculates such loss using mean square error (MSE), defined as:
where
and
represent the learnable model parameters, and
denotes the predicted score of user
u for POI
, which will be described in
Section 4.4. The initial
represents check-in frequencies of users visiting POI
, indicating the degree of users preferences for visiting that POI.
After performing batch-size to train , the parameter of LSTM module is updated through user multi-dimensional static preference module based on the historical check-in activity set of batch-size users before the time period t. It is trained iteratively according to different tasks of different users. From here on, the inner loop ends.
Outer loop: only after the completion of the inner loop, we can continue to update the initial parameter
of the original meta-learning model based on feedback from all tasks
. That is, we need to obtain the total loss function for all tasks and perform a second gradient descent on the meta-learning model. We use query set
of each
in batch-size to test the effect of the meta-learning model with parameter
, and calculate the total loss, defined as:
The total loss function is expressed as the sum of loss
in the meta-learning model on the corresponding parameters of the query set of each task in a batch size. Therefore, the global update is to update the parameter of meta-learning module at the current time period according to query set
in the inner loop, and the weight update of the global model parameter is defined as:
where
is the step size hyperparameter. In this way, continuous inner and outer loop training can obtain model parameters with good generalization ability on the dataset, thereby obtaining global knowledge shared by users, so as to accurately simulate dynamic visiting preferences of cold start users with only a small number of check-in activities in the current time period in the meta-test users set.
We provide the pseudo code for the model training process in Algorithm 1. Algorithm 1 summarizes the process of updating user multi-dimensional dynamic preference model parameter through meta-learning and updating user multi-dimensional static preference model parameter through LSTM. Among them, lines 3–8 represent the process of locally updating the meta-learning module and updating LSTM model parameters through the support set in user’s check-in activities in the current time period and check-in activities in the past time period, respectively. Line 9 represents the global update of the meta-learning module through the query set in user’s check-in activities in the current time period.
4.4. Next POI Recommendation by Integrating User Static and Dynamic Preferences
Considering that user’s preference for POI category will affect user’s selection of next POI recommendation [
33], therefore, before returning the recommendation result, we add a POI category filter to remove those POIs belonging to the POI category that has never been visited in user’s check-in sequence in the POI candidate pool, so as to further improve the recommendation accuracy.
After filtering, we calculate the preference score for each POI by taking dot product between the captured user comprehensive preference and the filtered POIs using Formula (9) and recommend the user a top-
n list of POIs with high scores.
where user’s comprehensive preference representation is the vector sum of user’s multi-dimension-al static preference
in the past time period and user’s multi-dimensional dynamic preference
from the current time period.
represents the embedding of POI
filtered by the POI category filter.
Algorithm 1 The training process of the model |
Input: Meta training task set , initialization parameter Output: The optimal parameter of the model 1: while not converge do 2: Task sampling ~ 3: for all do 4: Calculate the gradient error of the initial parameter 5: Sampling support set for local updates 6: Calculate the descent gradient to update model parameter through Formula (6) 7: Use Formula (3) to update user’s multi-dimensional static preference module parameter based on the check-in activity set in the past time period 8: end for 9: Use Formula (8) to perform meta updates on the multi-dimensional dynamic preference module based on the query set 10: end while |
4.5. Model Optimization
This subsection is to optimize the model we proposed. Based on user’s historical check-in sequence , after learning the latent preferences of the user multi-dimensional static and dynamic representations and , the matrix is optimized with the process of model training.
We represent
as a differentiable loss function for training the embedding matrix
, represented as:
where
is the original embedding representation of POI
,
k represents the number of meta-learning samples, and
is the embedded representation of the POI after updating, which is obtained by:
where
represents the decoding model and
represents the intermediate representation after POI
decoding, which is converted by the sigmoid function of attribute-wise. While
is the parameter of the sigmoid transformation.
In order to jointly train the POI embedding matrix, user multi-dimensional dynamic preference meta-learning module, and user multi-dimensional static preference LSTM module, we optimize the total loss function defined as:
where
represents the weight and
is the regularization parameter.
5. Experiments
5.1. Datasets and Preprocessing
We evaluate our ML-POIRec model on two real-world LBSNs datasets [
34], i.e., Foursquare and Gowalla. The Foursquare dataset contains check-in records of users collected from April 2012 to September 2013, and Gowalla contains users’check-ins extracted from February 2009 to October 2010. Due to the fact that neither dataset (Foursquare and Gowalla) has been processed and already includes many cold start users with less than 5 check-in activities, so both datasets can be considered as cold start datasets (The density of user-POI interaction matrix is 0.13% and 0.22%, respectively).
Next, we sorted check-in activities of each user in both datasets based on the corresponding timestamps of check-in activities. For the check-in time information of users, we divide each dataset into different check-in time periods based on timestamps of users’ visit POIs. As a result, Foursquare dataset is divided into six check-in time periods, and each check-in time period is three months. While Gowalla dataset is divided into ten check-in time periods, each of which is two months.
Similarly, we divide the geographic information of user’s visit POIs by geographic distance. Specifically, we use to calculate the spatial (geographic) distance between two POIs according to the corresponding latitude and longitude. The geographic information in user’s current check-in activity is the geographic distance between the POI visited in the current check-in activity and the previous POI visited (the geographic information in user’s first check-in activity is zero). Furthermore, we classify the spatial distance into different geographic segments, each with a 10 km interval, as the geographic interval label for users to visit the POI.
We follow the setting [
25] of meta-learning model and dynamically divide users into meta-training users sets and meta-test users sets. Specifically, we consider users with fewer than five check-in activities in the current time period as meta-test users, and the remaining users as meta-training users. We consider the distribution of task
, and each user is represented as a few-shot regression task
sampled from a given task distribution. Each task includes a support set
and a query set
. Here,
includes user’s first
k check-in activities in time period
t, where
k is represented as the number of meta-learning samples. While user’s remaining check-in activities within the time period
t is considered as query set.
For the non-meta-learning comparison methods, we first collect user’s check-in activities in the current time period, and then obtain the first k check-in activities in current time period as training set, and select the remaining check-in activities in current time period as test set, so as to simulate the few-shot problem of users visiting POIs in the current time period.
5.2. Baselines
We adopt the following baselines to demonstrate the effectiveness of our ML-POIRec model.
FM [
35]: a standard model of factor factorization, which considers second-order interactions between users and items.
NeuMF [
36]: a state-of-the-art approach for item recommendation which combines traditional MF and MLP in neural networks to predict user-item interactions.
GRURec [
37]: it is the first to use RNN in recommendation systems, optimizing GRU units on the basis of traditional RNN, which can effectively alleviate the long-standing problem of traditional RNN in obtaining users’ visiting preferences.
Caser [
38]: a sequential recommendation approach with convolutional sequence embedding to capture preferences and sequential patterns by modeling recommendation as a uniform and flexible structure.
SASRec [
39]: a Transformer-based sequential recommendation model that captures users’ long term visiting preferences. At each moment, the model aggregates information related to the current check-in activity and uses this information for next-step prediction.
MeLU [
25]: a recommendation method based on meta-learning is to estimate preferences of a new user according to a small number of user-item interaction records. In addition, a candidate item selection strategy is provided for personalized preference estimation.
MAMO [
26]: in order to solve the problem of generating simple initialization vectors of users or POIs based on MAML methods, and quickly convergence through a small amount of training to obtain a good global knowledge, resulting in local convergence of some users and poor generalization performance, MAMO introduces several additional memory modules to improve the generalization ability of the model.
These methods can be divided into four categories: one is Matrix Factorization based Approach (MFbA), such as FM, NeuMF; the second is Neural Network based Approach (NNbA), including GRU-based method (GRURec) and CNN-based method (Caser); the third is Self-Attention Network based Approach (SANbA), e.g., SASREC; the fourth is Meta-Learning based Approach (MLbA), such as MeLU, MAMO and ML-POIRec.
5.3. Parameter Settings and Evaluation Metrics
We use Torch 1.2.0 as a machine learning framework and NumPy 1.19.5 for the experiments. In our proposed model, we conducted grid search for all parameters and selected the best values of the corresponding parameters to make the model perform best. Specially, the first and second hidden layers in the meta-learning module have 128 and 64 hidden units with ReLU activation function, respectively, and the size of both the input layer and the output layer is 128. We set the number of samples for meta-learning
k = 5 as a limited number of user check-in activities for the support set. The latent dimension
d for both datasets is set to 128. We optimize the model through Adam optimizer [
40] using L2 regularization, and set the regularization parameter to 0.0001. The learning rate is 0.001 and dropout is 0.2 to avoid overfitting. The epoch is 50 and batch-size is 128. The source code of our proposed model is publicly available at
https://github.com/glpppp/ML-POIRec (accessed on 24 August 2023).
To evaluate the performance of ML-POIRec, we adopt two widely used evaluation metrics [
41]:
and
, where
n is the number of recommended POIs.
evaluates the quality of ranking, which indicates that whether POIs that users actually check-in rank at the top of next POI recommendation list. While
measures the presence of POIs that users actually visit among the top-
n recommended POIs. The higher the
and
, the better the model performance.
and
are defined as follows.
where
denotes the graded correlation of POI ranked
i. The value of
can be obtained by binary relevance, i.e.,
= 1 if a POI actually visited by the user is in the POI recommendation list, and 0 otherwise.
indicates the value of
at ideal ranking.
where
represents positive POIs that user
u has visited. Note that
= 1 in our work.
5.4. Sensitive Analysis of Parameters
In this section, we investigate whether the performance of ML-POIRec is sensitive to the latent dimension d, value of epochs, and number of meta-learning samples k. We adopt to analyze the effects of these parameters. Note that the results on are similar to that on .
5.4.1. Effect of Latent Dimension
In order to study how sensitive ML-POIRec is to latent dimension
d, we conduct the experiments while keeping other hyperparameters unchanged.
Figure 4 shows the results. It can be seen that the performance first increases and then decreases slowly as
d increases. The reason is that
d denotes the model complexity, and a large
d is too complicated to describe the datasets, while a small
d is not enough to model the datasets. Therefore, the optimal value of
d is 128 on both datasets.
5.4.2. Effect of Epochs
In general, in neural networks, we need to transfer all datasets in the same neural network for several times to train the model. Too many times of epochs will easily result in overfitting, while too few times will easily make the training parameters fail to reach the optimal. Based on this, we take model performance convergence as the basis for setting epochs. As shown in
Figure 5, the model performance first increases as epochs increase, then grows slowly and gradually becomes stable roughly after 40–50 epochs in two datasets. Therefore, we set epochs to 50.
5.4.3. Effect of the Number of Meta-Learning Samples
In this section, we demonstrate the effect of the number of samples
k on the model performance. The meta-learning model with too small number of samples is not enough to simulate user’s comprehensive preference. While the number of samples is too large, it does not conform to the concept of few-shot learning. Therefore, we turned
k from {1, 2, 3, 4, 5}. In order to illustrate the effect of
k and the effectiveness of ML-POIRec, we compare two methods (MELU and MAMO) based on meta-learning.
Figure 6 shows the results. We can observe that the performance of three models improves with the increase of
k, and the performance of ML-POIRec is the best. This is because the more check-in activities included in the support set for each task, the more it reflects the known visiting behavior of the target cold start user, which in turn can more accurately simulate user’s preferences, so as to achieve more improvements. Therefore, we set
k to 5.
Next, we will use the selected parameter values for subsequent experiments.
5.5. Recommendation Performance
The performance comparisons of eight methods on two datasets are shown in
Table 2 and
Table 3, in the form of mean and standard deviation of
and
. The numbers shown in bold in
Table 2 and
Table 3 represent the best performance of each column in the corresponding tables. From the experimental results shown in
Table 2 and
Table 3, we can make the following observations.
First, compared with the other seven models, the performance of ML-POIRec model has improved significantly on both two datasets. Taking FourSquare dataset as an example, compared with the best baseline, ML-POIRec improves and by 5.77% and 4.64%, respectively. The performance improvements on Gowalla are also great. These results clearly demonstrate the effectiveness of ML-POIRec.
Second, in most cases, deep learning-based methods (GRU4REC, Caser and SASREC) perform better than traditional matrix factorization methods (FM and NFM). The reason is that those deep learning methods simulate user’s preferences through training a large amount of data, which can more accurately simulate the trend of user preferences over time, space and other factors, so as to more accurately predict user’s selection of next POI.
Third, when considering the impact of user check-in sequences, the attention network-based method (SASRec) performs better than RNN based method (GRU4Rec). The reason is that attention network-based methods can effectively alleviate the defects of RNN models, which can effectively simulate users’ long term preferences, and perform parallel operations to improve model efficiency.
In addition, when faced with cold start data, the performance of deep learning based models appears to be quite limited. This is because deep learning methods require sufficient training data to simulate the comprehensive preferences of users as much as possible. Therefore, the performance in few-shot recommendation tasks has been greatly affected. Among them, SASRec model depends on the self-attention network and can obtain the long term dependence of users in visiting POIs. However, when users have fewer or no check-in activities in the current time period, SASRec does not perform well for simulating the short term preferences of users. While meta-learning based methods (MeLU, MAMO) can alleviate the problem of incomplete user preference acquisition caused by sparse user check-in activities by utilizing the shared knowledge among users in new user recommendation tasks. Therefore, the methods based on meta-learning can mine POI visiting preferences of individual users in limited data, so as to make next POI recommendation.
Finally, the performance of ML-POIRec is better than MeLU and MAMO on both datasets, mainly because it not only considers the sequential effects between user check-in activities, but also considers some auxiliary information such as geographic distance between two POIs in user check-in activities, check-in frequency, and POI category. Furthermore, in order to comprehensively simulate users’ visiting preferences, we divide users’ historical check-in activities according to time periods. Moreover, we conduct multiple iterative training through the meta-learning model to obtain the global knowledge of users’ check-in activities in the current time period, so that it can be applied to those recent check-in cold start users, and capture users’ long term static preferences in the past time period by combining with LSTM network. In addition, we leverage a POI category filter to filter POIs in the candidate pool to further improve the recommendation performance.
5.6. Ablation Study
In this section, we report the effectiveness of different components in ML-POIRec, including sequence impacts of user check-in sequence (ML-SE), user preferences at different time periods (ML-TP), geographic distance information (ML-GD) and POI category information (ML-POIRec). To this end, we conduct experiments on the following variants of ML-POIRec.
ML-SE: this model only considers the sequence influence, and does not distinguish users’ visiting preferences in different time periods.
ML-TP: the model divides user’s historical check-in sequence according to different time periods, and jointly considers user’s past(long) check-in preference and current (short) check-in preference.
ML-GD: this model leverages the geographical distance between POIs in user check-in activities, and divides geographical intervals according to different geographical distances, as a geographical interval label for users visiting POIs.
Table 4 shows the characteristics of different ML-POIRec variants. We use
and
to illustrate the effectiveness of different variants on two datasets. The results are presented in
Table 5 and
Table 6.
Take
as an example. From
Table 5 and
Table 6, it can be seen that in the three variants of ML-POIRec, ML-SE has poor performance compared with ML-TP and ML-GD. This is because ML-SE only considers the influence of the sequence of user’s check-in activities and cannot obtain users’ visiting preferences in different time periods, so it may cause outdated recommendation.
While ML-GD achieves better performance than ML-TP because ML-GD can obtain users’ geographical distance preferences for visiting POIs from their historical check-in sequences. It can assign larger weights to POIs with large distance interval and high user check-in frequency, so it is more reliable than ML-TP that only considers time information.
In contrast, the performance of complete ML-POIRec is superior to ML-SE, ML-TP and ML-GD. This is because when we simulate user preferences, we incorporate multi-dimensional information such as spatio-temporal interval information, user check-in frequency to POI, and POI categories. We comprehensively simulate users’ long term static preferences and recent dynamic preferences, and add POI category filter to filter POIs in the candidate pool, so as to further improve the model performance.
5.7. Complexity Analysis
In this subsection, we analyze the approximate time complexity of the baseline approaches. Considering that it is difficult to directly analyze the time complexity of each approach, we estimate the approximate complexity of calculating user preference representation in each method to evaluate its efficiency.
Assume that all baseline approaches have the same sample size (denoted by n), the same hidden layer dimension (denoted by m) and the same check-in activities (denoted by l). For matrix factorization based approach, since FM method does not involve neural networks, its time complexity is , while the time complexity of NeuFM is . For neural network based approach, the time complexity of GRURec is mainly related to the number of GRU layers and the size of hidden units in the GUR layer, its time complexity is , where t is the number of GRU layers. Caser used a convolutional neural network with a time complexity of .
For the self-attention network based approach, SASREC used multi-head attention to capture sequence information, with a time complexity of , where h is the number of attention heads. For meta-learning based approach, MeLU used multiple fully connected layers for feature extraction and combination, with a time complexity of , where L represents the model consisting of L fully connected layers with 2 layers. MAMO designed two memory matrices to provide personalized initialization for the recommendation model, with a time complexity of . Since our model includes two modules: static preference LSTM module and dynamic preference meta-learning module, with 1 and 2 hidden layers respectively, its time complexity is .
6. Discussion
According to our statistics, user-POI interaction matrix in LBSN datasets is relatively sparse compared with user-item interaction matrix in the traditional e-commerce recommendation, and there are a large number of cold start users. Therefore, how to accurately model user preferences in the limited check-in records is an urgent problem to be solved. For recent check-in cold start users, we proposed the ML-POIRec model, which captures multi-dimensional static preferences and dynamic preferences of users by LSTM module and meta-learning module, and adds a POI category filter to further improve the accuracy of recommendation and effectively alleviate the cold start problem. It is of great significance to ensure the quality of next POI recommendation and provides a wide range of practical implications. More specifically, on the one hand, for social platforms and merchants, it is beneficial for LBSN service providers to recommend new POIs to users based on the captured long and short term visiting preferences, which can not only improve user experience, but also help them find more potential customers and bring potential benefits. On the other hand, for users, high-quality next POI recommendation can help users quickly find POIs that well meet their personalized needs, which is the key to attracting users and improving user satisfaction.