Abstract
Due to the personalized needs for specific aspect evaluation on product quality, these years have witnessed a boom of researches on aspect rating prediction, whose goal is to extract ad hoc aspects from online reviews and predict rating or opinion on each aspect. Most of the existing works on aspect rating prediction have a basic assumption that the overall rating is the average score of aspect ratings or the overall rating is very close to aspect ratings. However, after analyzing real datasets, we have an insightful observation: there is an obvious rating bias between overall rating and aspect ratings. Motivated by this observation, we study the problem of aspect mining with rating bias, and design a novel RAting-center model with BIas (RABI). Different from the widely used review-center models, RABI adopts the overall rating as the center of the probabilistic model, which generates reviews and topics. In addition, a novel aspect rating variable in RABI is designed to effectively integrate the rating bias priori information. Experiments on two real datasets (Dianping and TripAdvisor) validate that RABI significantly improves the prediction accuracy over existing state-of-the-art methods.
You have full access to this open access chapter, Download conference paper PDF
1 Introduction
With the rapid development of the Internet, the information which people can gain from the Internet grows exponentially. Nowadays, people are used to viewing online reviews before making decisions. For example, if a user wants to go out for dinner, he or she may look at the reviews of restaurants around on the Internet and choose one according to his or her taste. These reviews contain mainly overall ratings which evaluate restaurants from a general view. However, people may expect more subtle aspect ratings, such as the taste, environment, service, and so on. This problem has inspired the research on aspect-level opinion mining. The goal of the aspect-level opinion mining (i.e., aspect identification and aspect rating prediction) is to extract ad hoc aspects from online reviews and predict rating or opinion on each aspect.
Because of its great practical significance, there is a surge of researches on aspect identification and aspect rating prediction in recent years. Some works generate ratable aspects for reviews with whole overall ratings [7] or scarce overall ratings [6], and some works consider to integrate external knowledge [9]. Most of the existing works predict aspect ratings with the help of overall ratings, and they all have a basic assumption. That is, the overall rating is the average score of aspect ratings or the overall rating is close to aspect ratings.
However, the analysis on real datasets shows an insightful phenomenon: there is an obvious and systemic rating bias between overall ratings and aspect ratings. Figure 1 illustrates the rating distributions on two real datasets: DianpingFootnote 1 (a well-known social media platform in China, which contains the information and reviews of restaurant, hotel, entertainment, movie, etc.) and TripAdvisorFootnote 2 (a widely used dataset in this field, which is a social media platform about travel, hotel, scenic spot, etc.). The datasets we use are the restaurant data in Dianping and the hotel data in TripAdvisor. Note that the overall ratings of restaurants/hotels are sorted in an ascending order in Fig. 1. We can find that the overall ratings in TripAdvisor are obviously lower than two aspect ratings, while the overall ratings in Dianping are significantly larger than aspect ratings. The interesting observation implies that the previous aspect rating prediction approaches may achieve poor performance, if ignoring the rating bias between overall ratings and aspect ratings.
Motivated by the observed rating bias, we try to study the problem of aspect mining with rating bias. That is, the goal is to decompose the reviews into different aspects and predict the rating of different aspects on each entity, with the help of the overall rating and the rating bias priori information. However, aspect mining with rating bias may face two challenges. First, the rating process of users may conform to some behaviour patterns, which determine the dependency relationship among the variables in the topic model. Most of the existing works on aspect rating prediction are based on probabilistic graphical model. Inspired by the word generation process, these works usually consider ratings are finally generated by reviews, topics or aspects. However, does it really comply to user behaviour? We have a different view. We believe that users form an intuitive impression (good or bad) as soon as they experienced the product, which is reflected by rating. Only after the impression (rating) is formed will the user write a review (or words) to express his/her feeling. So we think the previous models may not conform to user behaviour properly, and thus we need to mine the authentic rating behaviour of users. Second, how to effectively utilize the rating bias information? As we mentioned above, there is an obvious bias between overall rating and aspect ratings. The rating bias may cause the inaccuracy of aspect rating prediction, and influence the results tremendously. Luo et al. [6] have discovered the rating bias, but nobody has considered it in the model until now. So how to use the rating bias priori information properly to improve the prediction accuracy is also a challenge.
To solve the challenges mentioned above, we design a novel RAting-center model with BIas (RABI). Different from traditional rating generating process [6, 7, 9], RABI considers rating as the center of the model, which generates the reviews and topics. This idea stems from users’ real experiences. When users decide to write a review, they usually have intuitional opinions (i.e., overall ratings) on the products, and then they will use proper phrases to represent their opinions. In addition, RABI introduces a novel latent aspect rating variable which can effectively learn the correlation of the overall rating, aspects, and rating bias. Experimental results on two real datasets (i.e., Dianping and TripAdvisor) validate the effectiveness of RABI on both Chinese and English reviews, compared to existing state-of-the-art methods. The results also show that RABI can accurately decompose the reviews into different aspects.
Our contributions are summarized as follows:
-
We first analyze the rating bias between overall rating and aspect ratings in real data, and put forward the problem of aspect mining with rating bias.
-
We propose a novel RABI model for aspect mining with rating bias. Different from existing models, RABI considers rating as the center of the model, which simulates the generation of the review better. In addition, an aspect rating variable is proposed to effectively utilize the rating bias information.
-
Experiments on real datasets have shown the effectiveness of our algorithm over existing state-of-the-art methods.
2 Data Analysis
In order to show the rating bias phenomenon, we analyze two real datasets. The first dataset is crawled from Dianping website, a well-known social media platform in China, which provides a review platform for businesses and entertainments. In Dianping website, a user can give a review to a business after enjoying a service in this business. Besides an overall rating, the review information includes Chinese comments and three aspect ratings on Taste, Service, and Environment, respectively. In addition, we also employ the widely used TripAdvisor dataset [10]. Accompanying with English comments, reviews in this dataset are not only associated with overall ratings, but also with ground truth aspect ratings on 7 aspects: Value, Room, Location, Cleanliness, Front desk/staff, Service, and Business. All the ratings in the datasets are in the range from 1 to 5. The statistic information of these datasets is shown in Table 1.
We first intuitively show the distributions of overall and aspect ratings on these two datasets in Fig. 1. Note that, we only show the distributions of some aspect ratings due to the space limitation. Moreover, we sort products according to their overall ratings for clarity. From Fig. 1, we can find that there are obvious rating biases between overall rating and aspect rating on both datasets. In Dianping dataset, the overall rating is far above the aspect ratings in all three aspects, while the overall rating is smaller than two aspect ratings in TripAdvisor.
Furthermore, we calculate the rating bias on each aspect on both datasets. The calculating process can be seen in Eq. (1) and the results are listed in Table 2. The rating biases in Dianping are huge on most aspects, especially +0.48 for Service and +0.54 for Environment, which are pretty huge values. So the rating biases in Dianping should be well considered. The rating biases in TripAdvisor are small on some aspects (e.g., +0.01 for Value and −0.01 for Room), but huge on other aspects (e.g., −0.33 for Location and −0.26 for Cleanliness). Although the rating biases in TripAdvisor are not as much as those in Dianping, they all truly exist. The interesting observation implies that the previous aspect rating prediction approaches may achieve poor performance, if ignoring the rating bias. As shown in Table 2, the rating biases are different in different datasets and aspects, which can influence the results to varying degrees and cause the inaccuracy of aspect rating prediction. So the proper consideration of the rating bias can improve the prediction accuracy.
3 Preliminary Notations and Problem Definition
In this section, we first introduce the notations and concepts used in this paper, and then formally propose the problem of aspect mining with rating bias.
Entity: An entity e indicates a product which belongs to the product set E (e.g., a restaurant in Dianping dataset or a hotel in TripAdvisor dataset). \(N_e\) indicates the number of entities in E.
Review: A review d is the user’s opinion about the entity e. An entity e can have many reviews from different users. A review consists of the text content, the overall rating and many aspect ratings. There are \(N_d\) reviews in total.
Phrase: A phrase \(f=(h,m)\) consists of a pair of words, which are extracted from the review’s text content. h denotes the head term, and m is the modifier term which modifies h. A review d contains several phrases f.
Head term: The head term h is used to describe the aspect information. It decides which aspect the phrase f is expressing. For instance, “attitude” is a head term, and it belongs to the aspect “Service”.
Modifier term: The modifier term m is used to describe the sentiment information. It is used to describe the aspect, which is decided by h, is good or bad. For instance, for the head term “attitude”, “cold” or “passionate” may be used as the modifier term.
Overall rating: An overall rating r of a review d is a numerical rating, which indicates the user’s overall sentiment tendency on the entity e. The number of the values of rating is \(N_r\) and it is usually 5, which means the values of rating r are from 1 to 5.
Aspect: An aspect \(A_i\) is a specific side of the entity e, e.g., the taste of the restaurant. It is a set of many similar characteristic of the entity e. \(N_A\) indicates the number of aspects.
Aspect rating: An aspect rating \(r_{A_i}\) is a numerical rating, which indicates the user’s sentiment tendency on the aspect \(A_i\) of the entity e, and is also from 1 to 5. And a review d has \(N_A\) aspect ratings, which corresponds to \(N_A\) aspect.
Rating bias: The rating bias is the gap between the average of overall ratings and the average of aspect ratings. There are \(N_A\) biases on \(N_A\) aspects, and they are in connection with the current aspect \(A_i\). The rating bias \(b_{A_i}\) on aspect \(A_i\) can be calculated as follows:
Aspect mining with rating bias: The problem of aspect mining with rating bias is to predict the rating on each aspect with the rating bias prior information. Specifically, given a set of reviews \(D=\{d_1, d_2, \cdots , d_{N_d}\}\) about entities \(E=\{e_1,e_2, \cdots , e_{N_e}\}\), we know that each review \(d_i\in D\) contains text content (Chinese or English) and overall rating r on an entity \(e_j\in E\), as well as the rating bias \(b_{A_i}\) between the overall rating and the aspect rating on \(N_A\) aspects for all reviews. The goal is to decompose the phrases f, which are extracted from texts in D, into \(N_A\) aspects \(\{A_1, A_2, \cdots , A_{N_A}\}\), and rate the aspects of each entity e with \(\{r_{A_1}, r_{A_2}, \cdots , r_{A_{N_A}}\}\).
In fact, our goal includes two sub-tasks. (1) The first sub-task is aspect identification, which is to correctly identify the aspect label \(A_i\) given phrase f. (2) The second sub-task is aspect rating prediction, which is to predict the aspect rating \(r_{A_i}\) given the entity e and aspect \(A_i\).
The problem of aspect mining with rating bias is very important in real applications. The problem is also the base of many tasks, such as overall rating prediction and aspect-level product recommendation. Compared to overall ratings, the aspect ratings are always missing and more unreliable. The aspect rating prediction is an effective way to repair the missing ratings and correct the unreliable ratings. However, the existent rating bias may make current methods on aspect rating prediction not effective anymore, so it is desired to consider rating bias for aspect rating prediction. Please note that the rating bias is known in our problem setting. Moreover, the rating bias can be easily obtained through limited reliable aspect ratings or a small quantity of manual labeling in real applications. So we can use the information of rating bias to correct the aspect rating prediction.
4 Rating-Center Model with Bias
The simplest way to handle rating bias is to subtract rating bias from the rating prediction results of existing models. However, it does not consider the correlations of ratings, aspects, and rating bias, so it may result in poor performances. In this section, we propose a novel RABI to handle the problem of the existent rating bias. Furthermore, we derive an iterative optimization solution with the EM algorithm.
4.1 Model Description
Existing models on aspect rating prediction usually consider reviews as the center to generate ratings and topics [6, 7, 9]. However, it does not conform to the authentic rating behaviour of users. In daily life, we form an intuitive impression as soon as we experienced a product. Only after we form an intuitive opinion (like or dislike, quantitatively represented by a rating) on a product, will we write a review to express our opinion. In addition, our opinion may involve multiple aspects of the product, such as taste, service and environment. So in the generative process of a product review, we will choose proper head terms to represent the aspect we want to express, and proper modifier terms to express sentiments on corresponding head terms. Finally, we organize these terms and other words to form a review. Therefore, we believe it is more reasonable to consider rating (overall rating) as the center to generate topics and reviews, which conforms to the authentic rating behaviour of users. Following this idea, we design the probabilistic model of RABI, shown in Fig. 2.
In Fig. 2, d indicates the reviews, r indicates the overall rating, h indicates the head term and m indicates the modifier term. These four variables are represented as the shaded circles, which means these four variables are observable. z indicates the aspect \(A_i\). In order to keep consistent with the topic model, the aspect \(A_i\) is expressed as the topic z. And \(r_b\) indicates aspect rating, which will be introduced in the following. These two variables are represented as the open circles, which means these two variables are latent variables. Furthermore, N indicates the number of phrases in a review. And M indicates the number of reviews, which is equal to \(N_d\).
To utilize the rating bias information effectively, we bring in a new latent aspect rating variable \(r_b\). The modifier term m is used to modify the head term h to express the opinion (like or dislike) on aspect \(A_i\) (represented with z in the model), so m is actually influenced by the corresponding aspect rating \(r_{A_i}\). As we mentioned above, there is an obvious rating bias between overall rating and aspect ratings. This observation causes that we cannot use the overall rating r to influence the modifier term m directly. So we bring in a new variable \(r_b\) between r and m to eliminate the influence of rating bias. \(r_b\) indicates an unknown aspect rating, so it is a latent variable. For a certain aspect \(A_i\), the value of \(r_b\) is set as the overall rating r minus the rating bias \(b_{A_i}\). Note that \(r_b\) can take \(N_r\) values in \(A_i\), since the variable r can take \(N_r\) values. By bringing in the latent variable \(r_b\), the association between r and m is modeled more reasonably in RABI.
According to the RABI model shown in Fig. 2, as the origin of the model, the overall rating r generates the review d and the latent topic z. The latent aspect rating \(r_b\) depends on the topic z and the overall rating r. And the head term h and the modifier term m are influenced by the topic z and the aspect rating \(r_b\), respectively. So the joint probability over all variables is as follows:
All the parameters can be iteratively calculated using the EM algorithm [4], which is a common method to solve the problem with latent variable. The detail derivation is given in next section.
4.2 EM Solution
In the E-step, we need to maximize the lower bound function \(\mathcal {L}_0\) (i.e., Jensens inequality [2]),
Here, as usual, \(q(z,r_b)\) is set as follows:
Then we simplify Eq. (3), we can get
So the second part is a const, which can be ignored. Then we ignore the const, and only consider the \(\mathcal {L}\).
The function for the posterior probabilities of the latent variables is as follows:
where \(\varLambda \) includes all parameters, i.e., \(p(m|r_b)\), \(p(r_b|r,z)\), p(h|z), p(z|r), p(d|r) and p(r), which are mentioned in Eq. (2). Besides, n(h, m, r, d) is the number of co-occurrences of h, m, r and d.
The function \(q(z,r_b)\) and \(p(h,m,r,d,z,r_b|\varLambda )\) in Eq. (3) are expanded as follows:
In the M-step, the Lagrangian Multiplier method is used to maximize \(\mathcal {L}\) and calculate the parameters.
For \(p(m|r_b)\), there is a basic constraint as follows:
Applying the Lagrangian Multiplier method, we can get a function for \(p(m|r_b)\) as follows:
After calculation, we have
Then the update function for \(p(m|r_b)\) is as follows:
Similarly, the update functions for other parameters are as follows:
Through these functions above, we can iteratively calculate the parameters until the model has converged.
4.3 Aspect Rating Prior
To verify our model’s effectiveness, we need to compare the predicted aspect ratings with the real aspect ratings. So the aspects should correspond to the real aspects which are set by the e-commerce review sites. To make the predicted aspects similar to the real aspects, we need to assign some seed words to each aspect. For instance, the aspect “Taste” may include a few prior words, such as “taste” and “flavor”.
In our model, we inject the prior knowledge for the aspect z. The function is as follows:
where \(\tau (h,z)\) indicates the prior knowledge of the prior words. Only when there is a relationship between the head term h and the topic z, in other words, h belongs to z, does \(\tau (h,z)\) have a value \(\delta \), otherwise 0.
Note that, in the real applications, we can set aspects manually or generate aspects by the model directly. Moreover, manual aspect setting usually has better performances.
4.4 Aspect Identification and Aspect Rating Prediction
We can get \(p(z,r_b|h,m)\) from the model by the following function,
The goal of aspect identification is to find the mapping function \(\mathcal {G}\) that correctly assigns the aspect label for given phrase f.
The goal of aspect rating prediction is to predict the aspect rating \(r_{A_i}\) of the entity e given all the phrases f from all reviews and aspect \(A_i\)(z). The aspect rating function is as follows:
where \(r_{e,A_i}\) indicates the aspect rating on the aspect \(A_i\) of the entity e.
In this way, RABI learns the joint probability distribution of phrases, aspects and ratings, and predicts aspect ratings with bias.
5 Evaluation
In this section, we introduce experimental preparation, evaluation metric and baselines. Then we conduct extensive experiments to evaluate the effectiveness of RABI on two real datasets.
5.1 Experimental Preparation
Experiments are conducted on two real datasets (i.e., Dianping and TripAdvisor), which are introduced in Sect. 2. The preprocessing of TripAdvisor is similar to that in [6]. But the preprocessing of Dianping is slightly different. Since Dianping is a Chinese website, the Word SegmenterFootnote 3 and the rules from [8] are adopted for preprocessing. To inject the prior knowledge for the aspect, we select some words as prior for each aspect, and Table 3 lists some of the prior words (not all of the prior words due to the space limitation). For better understanding, we translate the Chinese words in Dianping into English.
Besides, all of the initial parameters (\(p(m|r_b)\), \(p(r_b|r,z)\), p(h|z), p(z|r), p(d|r) and p(r) in Eq. (2)) are assigned uniformly and randomly. \(\delta \) in the Sect. 4.3 is set as 1 after some preliminary tests. The number of aspects or topics K is set as 3 for Dianping and 7 for TripAdvisor. The experiments are done on different-size of datasets (i.e., 25 %, 50 %, 75 %, and 100 % of review data) from Dianping and TripAdvisor, respectively. The maximum number of iterations is set as 500.
5.2 Evaluation Metric
RMSE (Root Mean Square Error) is one of the most common metrics for rating prediction. RMSE can measure the difference between the real values and the predicted values. For every entity e, we have the real aspect rating vector \(r_{e,A_i}\) and the predicted aspect rating vector \(\hat{r}_{e,A_i}\). The function of RMSE is as follows:
Smaller value of RMSE indicates a stronger predictor, which means the real values and the predicted values are nearer.
Besides, we use Pearson Correlation Coefficient \(\rho \) [10] to measure the relative ordering of products based on the predicted aspect rating and the real aspect rating. The correlation is stronger when the absolute value of \(\rho \) is closer to 1, and weaker when the absolute value of \(\rho \) is closer to 0. The function is as follows:
where N indicates the total amount, which is \(N_e*N_A\).
5.3 Baseline Methods
We compare the proposed model with three representative methods and one variation of RABI. Since all of these baselines do not consider the rating bias, we adjust the results of these baselines through subtracting the rating bias for fair comparison. The adjusted method is marked with “\(^*\)” to distinguish from the original method.
-
QPLSA/QPLSA\(^*\) [7] uses quad-tuples information to build a model based on PLSA framework. The model not only can generate fine-granularity aspects of products, but also capture the relationship between words and ratings.
-
GRAOS/GRAOS\(^*\) [6] is a semi-supervised model based on LDA framework. It also uses the quad-tuples information to capture the relationship between words and ratings. The model considers the rating distribution as a Gaussian distribution.
-
SATM/SATM\(^*\) [9] is a sentiment-aligned model based on LDA framework. The model uses two kinds of external knowledge: productlevel overall rating distribution and wordlevel sentiment lexicon.
-
RA/RA\(^*\) is a simplified model which removes the latent aspect rating variable \(r_b\) from our model RABI. It only considers the rating-center assumption. Through comparing RA\(^*\) and RABI, we can testify the importance of the good mechanism to utilize rating bias information.
5.4 Results Evaluation
We firstly validate the effectiveness of aspect identification of RABI through a case study, and then compare the results of different methods on the accuracy of aspect rating prediction with two criteria mentioned above.
Aspect Identification. RABI extracts a set of rated phrases to describe the product for each aspect. We list the top 20 automatically mined phrases for each aspect, from which we select several meaningful phrases to be shown in Table 4. The phrases are ranked by their ratings for every aspect.
Generally, the extracted phrases properly describe the corresponding aspects and accurately embody the opinion in both English and Chinese reviews. On one hand, the head terms can indicate the aspects well, such as “attitude” for service, “fitment” for environment, “setting” for room, and “area” for location. When a user sees the head term, he can understand which aspect is talked about. On the other hand, a positive modifier term indicates a positive attitude and is likely to obtain a higher rating, and a negative modifier term indicates a negative attitude and is likely to obtain a lower rating. For example, in the Service aspect of Dianning, the phrase “cold attitude” is rated as 1.67 because “cold” is a negative modifier term, while the phrase “smart waiter” has a score of 4.51 because “smart” is a positive modifier term. In addition, the phrases and their ratings are also able to reflect the different rating styles in Chinese and English. That is, users tend to give relatively lower ratings in Chinese reviews. The distribution of the predicted ratings on phrases also conforms to that of aspect-level ratings on these two datasets in Table 2. It also confirms the effectiveness of RABI on Chinese and English datasets.
Accuracy Experiment. Then we validate the performances of different methods through comparing predicted aspect ratings with real aspect ratings using the RMSE criterion by Eq. (22).
From the results shown in Table 5, we can clearly find that the integration of the rating bias information can significantly improve the prediction accuracy for all methods (e.g., QPLSA* has better performances than QPLSA), and RABI always performs best on both datasets. The improvement is particularly obvious for Dianping, because this dataset has large rating biases. Although the rating bias is small in TripAdvisor, the methods considering rating bias all achieve better performances than original methods. It illustrates that it is necessary to consider the rating bias for aspect rating prediction.
Besides, the rating-center model (i.e., RA) also achieves good performances among four baselines, which confirms the correctness of the rating-center assumption. Compared to simply subtracting the rating bias in four baselines, the best performances of RABI imply that the good mechanism to utilize rating bias information is also necessary. We think the rating-center and the latent aspect rating variable contribute to the good performances of RABI.
In addition, with the increment of review data, the accuracy of RABI increases steadily and slowly, which reflects that RABI is a steady method.
Relative Order Experiment. Furthermore, we verify the ability of different methods to maintain the relative order among products with the Pearson Correlation Coefficient \(\rho \). The results are shown in Table 6. Note that the rating bias has slight effect on the order of products, so we only display the results of original methods and ignore the adjusted methods. We can see that RABI obtains much higher \(\rho \) than other methods in all datasets. It once again shows that RABI is more effective to model the correlations between aspects and ratings, and thus better maintains aspect ranking orders compared to other methods. The results also imply that RABI is very promising for aspect-level recommender system, since it can generate very similar product order to the real order.
6 Related Work
In recent years, sentiment analysis on reviews becomes a research hotspot. Reviews focus on the products in each aspect, so sentiment analysis on reviews usually involves aspect. This situation leads to the aspect rating prediction. Aspect rating prediction usually contains two subtasks, aspect identification and aspect rating prediction.
Topic model is widely used to solve aspect identification. It mainly contains LSI [3], PLSA [5] and LDA [1]. Xu et al. [12] centered on implicit feature identification in Chinese product reviews via LDA and SVM. An AEP-based Latent Dirichlet Allocation (AEP-LDA) [13] model was also proposed to extract product and service aspect words automatically from reviews. Fu et al. [11] proposed an approach to automatically discover the aspects discussed in Chinese social reviews and classified the polarity of the associated sentiment by HowNet lexicon. Our model RABI is designed based on the PLSA framework.
To solve aspect identification and aspect rating prediction simultaneously, many researches adopted the topic-sentiment mixture models. QPLSA [7] adopted the quad-tuples, which consist of head, modifier, rating and entity. It can generate fine-granularity aspects and capture the correlations between words and ratings. SATM [9] used external knowledge, product-level overall rating distribution and word-level sentiment lexicon, to extract the product aspects and predict aspect ratings simultaneously. Luo et al. [6] proposed a model based on LDA to predict aspect ratings and overall ratings for unrated reviews and made two assumptions for the rating distribution. However, all of these works did not consider the existing rating bias, which is firstly studied in this paper.
7 Conclusion
Aspect rating prediction for reviews is a hot research issue nowadays. Most of researches base on such a basic assumption, the overall rating is the average score of aspect ratings or the overall rating is close to aspect ratings. However in the real world, there may be rating biases between overall rating and aspect ratings, and existing works did not consider these rating biases.
In this paper, we study the problem of aspect mining with rating bias and propose a novel probabilistic model RABI based on PLSA framework. The RABI model makes rating as the center to generate ratings and topics, and introduces a latent aspect rating variable to integrate the rating bias information. Experiments on two real datasets validate the effectiveness of RABI. In the future, we can import the Dirichlet prior and redesign our model based on LDA framework. The effectiveness will be enhanced further.
References
Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent Dirichlet allocation. J. Mach. Learn. Res. 3, 993–1022 (2003)
Chandler, D.: Introduction to modern statistical mechanics. Phys. Today 1, 288 (1987)
Deerwester, S.C., Dumais, S.T., Landauer, T.K., Furnas, G.W., Harshman, R.A.: Indexing by latent semantic analysis. J. Am. Soc. Inform. Sci. 41(6), 391–407 (1990)
Dempster, A.P., Laird, N.M., Rubin, D.B.: Maximum likelihood from incomplete data via the EM algorithm. J. Roy. Stat. Soc. Ser. B (Methodol.) 39, 1–38 (1977)
Hofmann, T.: Probabilistic latent semantic indexing. In: Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 50–57. Association for Computing Machinery, Berkeley, August 1999
Luo, W., Zhuang, F., Cheng, X., He, Q., Shi, Z.: Ratable aspects over sentiments: predicting ratings for unrated reviews. In: 2014 IEEE International Conference on Data Mining (ICDM), pp. 380–389. Institute of Electrical and Electronics Engineers, Shenzhen, December 2014
Luo, W., Zhuang, F., He, Q., Shi, Z.: Quad-tuple PLSA: incorporating entity and its rating in aspect identification. In: Tan, P.-N., Chawla, S., Ho, C.K., Bailey, J. (eds.) PAKDD 2012. LNCS (LNAI), pp. 392–404. Springer, Heidelberg (2012). doi:10.1007/978-3-642-30217-6_33
Moghaddam, S., Ester, M.: On the design of LDA models for aspect-based opinion mining. In: Proceedings of the 21st ACM International Conference on Information and Knowledge Management, pp. 803–812. Association for Computing Machinery, Sheraton, October 2012
Wang, H., Ester, M.: A sentiment-aligned topic model for product aspect rating prediction. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, pp. 1192–1202. The Association for Computational Linguistics, Doha, October 2014
Wang, H., Lu, Y., Zhai, C.: Latent aspect rating analysis on review text data: a rating regression approach. In: Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 783–792. Association for Computing Machinery, Washington DC, July 2010
Fu, X., Liu, G., Guo, Y., Wang, Z.: Multi-aspect sentiment analysis for Chinese online social reviews based on topic modeling and HowNet lexicon. Knowl.-Based Syst. 37, 186–195 (2013)
Xu, H., Zhang, F., Wang, W.: Implicit feature identification in Chinese reviews using explicit topic mining model. Knowl.-Based Syst. 76, 166–175 (2015)
Zheng, X., Lin, Z., Wang, X., Lin, K.J., Song, M.: Incorporating appraisal expression patterns into topic modeling for aspect and sentiment word identification. Knowl.-Based Syst. 61, 29–47 (2014)
Acknowledgements
This work is supported in part by the National Key Basic Research and Department (973) Program of China (No. 2013CB329606), and the National Natural Science Foundation of China (No. 61375058, 61473273), and the Co-construction Project of Beijing Municipal Commission of Education, and the CCF-Tencent Open Fund, and 2015 Microsoft Research Asia Collaborative Research Program.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this paper
Cite this paper
Li, Y., Shi, C., Zhao, H., Zhuang, F., Wu, B. (2016). Aspect Mining with Rating Bias. In: Frasconi, P., Landwehr, N., Manco, G., Vreeken, J. (eds) Machine Learning and Knowledge Discovery in Databases. ECML PKDD 2016. Lecture Notes in Computer Science(), vol 9852. Springer, Cham. https://doi.org/10.1007/978-3-319-46227-1_29
Download citation
DOI: https://doi.org/10.1007/978-3-319-46227-1_29
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-46226-4
Online ISBN: 978-3-319-46227-1
eBook Packages: Computer ScienceComputer Science (R0)