1. Introduction
Livestock is widely known as an important part of agriculture. According to the Food and Agriculture Organization (Food and Agriculture Organization of the United Nations (Available online:
In previous research, efforts have been made to predict future prices based on various historical factors, such as historical prices, climate change, seasonal factors, agricultural calamities, and other economic effects. However, many of these factors, such as capital operation, policy, and disease, are difficult to quantify to make a prediction, making it hard to choose the influencing factors.
To address this issue, we explore the influence of forum discussions on hog price prediction. As forum discussions contain people’s analysis and reflect their expectations towards this topic, we assume that they include and interpret many factors, like the influence of consumer preferences, political events and other factors that are difficult to quantify. In fact, including textual information such as news articles for classification or prediction problems is not rare, especially for stock price prediction [7] and many NLP tasks. It is very likely that the forum discussions can enhance hog price prediction. Besides this, hog prices follow a non-linear and non-stationary time-series. For this time-series prediction task, researchers seek statistical methods and later machine learning methods. To extend this research line of applying deep neural networks to extract the necessary features, we further construct heterogeneous graphs to capture the representations of online discussions, enhancing hog price prediction.
In this study, we explore the influence of forum discussions on hog price prediction and propose a method that predicts the weekly hog prices. We extract historical prices of hog, maize and bean, as well as forum discussions for hog price prediction. It has been proved that bean and maize prices can largely influence hog price [8]. More importantly, the historical prices of hog and maize are easy to quantify and acquire. After obtaining representations of forum discussions and price series, our HGLSTM will combine price features and discussion information to forecast hog prices.
Our contributions are summarized as follows:
As far as we are concerned, this is the first study to make use of discussion information, acquired from the online professional pig community, for hog price prediction, and prove it to be effective;
Due to our limited time and effort, we find no other research to deeply integrate discussion information and prices series based on heterogeneous graph for hog price forecast;
We propose a heterogeneous graph-enhanced LSTM network (HGLSTM) and conduct extensive experiments to prove its effectiveness. Our experiments show that it outperforms state-of-the-art models.
This paper is organized in the following manner: In Section 2, we introduce some important related works. In Section 3, we discuss how our model is constructed and give necessary explanations. In Section 4, we illustrate the experimental design and results. We also give a brief analysis of the results in this section. Finally, in Section 5, we present our conclusions and insight into this task’s possible future direction.
2. Related Work
In this section, we introduce some critical studies related to the price forecasting of agricultural commodities. Price forecasting is often regarded as a time-series prediction problem. Thus, traditional statistical and deep learning methods have been commonly used for this. Significant studies of natural language processing and deep neural network concerning our method will also be discussed.
2.1. Price Forecasting Using Statistical Methods
Regression methods, like the autoregressive integrated moving average (ARIMA), generalized ARIMA, and seasonal ARIMA, are often used to solve this type of task. They are usually classified as traditional statistical methods. The ARIMA model is exploited by researchers [9,10,11,12] for agricultural price prediction. When studying cocoa bean price forecasting, Assis and Remali [13] tried to figure out the best method in various time-series prediction models. Their experiments showed that the generalized ARIMA model achieved the best performance. Adanacioglu and Yercan [14] applied seasonal ARIMA to tomato price forecasting in Turkey. In an attempt to solve corn prices forecasting, Gu et al. [15] proposed a multivariate linear regression model. They tried to model the effect of supply and demand, but their model’s performance is still not very desirable due to drastic changes in the corn market. BV and Dakshayini [16] tried to predict the prices and demand of tomatoes. Their study compared the performance of Holt Winter’s model and other benchmark models, such as simple (multiple) linear regression. Their experiments presented huge variations between targets and predictions. They also concluded that seasonality was an influencing factor because Holt Winter’s model, which considers seasonality, achieved the best performance.
Statistical methods show good performance on linear price series, but their performance drops drastically when faced with non-linear and non-stationary price series.
2.2. Price Forecasting Using Machine Learning Methods
Thanks to the rapid development of machine learning and deep learning algorithms, many researchers have developed new approaches to solve time-series forecasting problems. These new methods can extract hidden features from price series and, as a result, show a much better performance than traditional statistical models.
A back-propagation neural network proposed by Minghua et al. [17] is applied to the price forecasting of agricultural products. They conducted extensive experiments and found their proposed artificial neural network’s superiority against a statistical method. Other researchers, such as Nasira and Hemageetha [18], also exploit back-propagation neural network (BPNN) to predict tomatoes’ prices. Trying to predict the non-linear garlic price series, Wang et al. [19] proposed a hybrid ARIMA support vector machine (SVM) model. Experimental results showed that this very model surpassed the performance of both single ARIMA and SVM. Hemageetha and Nasira [20] proposed a radial basis function neural network (RBF) to predict tomato prices. Their model achieved better accuracy than the BPNN model. Using a chaotic neural network, Li et al. [21] found it to be a superior algorithm for weekly egg price prediction than ARIMA.
Many researchers have also made an effort to combine multiple models into a hybrid model. Luo et al. [22] propose three models and a hybrid model to forecast Lentinus edodes mushroom prices of Beijing. Their integrated model combines BPNN, RBF neural network, and genetic-algorithm-based neural network to achieve the best performance. Zhang et al. [23] proposes a quantile regression-based RBF (QR-RBF) neural network model to predict soybean prices in China. In the process of model optimization, they apply a gradient descent with GA to improve performance. Their experimental results align with previous studies [24,25].
Other researchers seek to preprocess the price series before feeding them into the model. Xiong, Li and Bao [26] first use the STL-based method to decompose the price series to predict cabbage, hot pepper, cucumber, kidney bean, and tomato prices. They consider the seasonal characteristics of vegetables and preprocess the time-series price data based on these characteristics. Their experiments prove the effectiveness of their method. To forecast vegetable prices, Li and Zheng [27] proposes a model that integrates an H-P filter and a neural network. Their study’s main contribution is that they decompose trend and cyclical components in the price series and recombine prediction values using the H-P filter. Another study [28] aims to forecast five monthly crop prices in the Korean market. They propose the STL-LSTM model, which eliminates high seasonality in vegetable prices. Their model performance has improved a lot by doing so. Following this research line, Liu et al. [29] propose a model that divides hog price series into the trend and cyclical components. They use the most similar sub-series search method to predict them and recombine these components. Finally, with the help of support vector regression, they successfully forecast the hog prices.
Researchers also exploit other information to help forecast the price series. Yoo et al. [30] makes use of climate factors and production information along with trends and seasonality of price data for prediction. They aim to forecast the prices of Korean cabbage and achieve good results. Chen et al. [31] aims to predict cabbage prices in the Chinese market. They propose a wavelet analysis-based LSTM model. The wavelet method that removes noise from the price series, therefore, helps improve model performance.
As we can see, most researchers using deep neural networks include LSTM in their model, which is not surprising because LSTM has shown superiority in dealing with series data. With the development of the attention mechanism of Bahdanau et al. [32], researchers began to apply it to their model. The attention mechanism can assign weights for different input vectors, thus calculating each vector’s importance value. There are many variants of attention, suggesting that the structure is very flexible and can be combined with many existing models. Consequently, it has applications in various fields, such as classification, recommendation, regression, and price prediction.
Qin et al. [33] proposes a dual-stage attention-based recurrent neural network for stock price forecast. Feature attention and temporal attention are used in their model. Attention structure also helps to explain the correlations between input vectors and outputs. Ran et al. [34] addresses travel time prediction by an attention-based LSTM. The attention structure assigns different weights to different features, thus improving model efficiency. Li et al. [35] proposed evolutionary attention-based LSTM model explains the correlations between local features in time steps. Aiming to solve financial time series prediction, Zhang et al. [36] designs attention-based LSTM and addresses a long-term dependence issue.
We summarize the above literature review in the following table (Table 1):
2.3. LSTM
Long short-term memory (LSTM) is a type of recurrent neural network (RNN). It is proposed by Hochreiter and Schmidhuber [38] to solve long-term dependency and gradient vanishing problems. An LSTM cell usually consists of an input gate, an output gate, a forget gate and a cell state. The structure of a LSTM cell is shown below (Figure 1):
As shown in Figure 1, for each element in the input sequence, (the hidden state at time t) is computed via the following functions
(1)
where is the hidden state at time is the cell state at time is the input at time is the hidden state at time or the initial hidden state, and are the input, forget, cell, and output gates, respectively. is the sigmoid function and · denotes element-wise matrix multiplication.LSTM networks are well-suited to classifying, processing, and making predictions based on time series data. A lot of research [28,34,36,39,40] into agricultural price prediction have demonstrated the effectiveness of LSTM in dealing with prices series. Therefore, in this paper, we decide to follow this line of research by exploiting LSTM network to process price series.
3. Materials and Methods
3.1. Problem Statement
Let be a thread consisting of a set of posts, be a group of users participating in this thread uploading at least one post, where denotes the number of all posts involved and denotes the number of all users involved in this thread.
To make the best use of the discussion network and capture the user-enhanced semantic features, we construct the heterogeneous graph , where V denotes the node set and E denotes the edge set. is the adjacency matrix of graph G. An example of this heterogeneous graph is shown in Figure 2. Considering the graph’s heterogeneity, there are two types of nodes: users node and posts node . Therefore, there are two types of edge: post-user edges and post-post edges . The connections between users are not considered in this study because, in a discussion thread, users’ connections are rare, thus contributing little to our goal. Moreover, we treat G as an undirected graph.
For historical prices, let be the processed weekly prices, where each , and denotes hog, maize, and bean price, respectively.
We regard this price prediction task as a binary classification problem. denotes the label, where means hog price will increase next week and represents other situations. So our goal is to train a model to predict the label of given input (forum discussions and historical prices).
3.2. Overall Structure
In this paper, we propose a forecasting method for hog prices. The overall structure of our proposed method is shown in Figure 3. It will be explained in detail later. For clarity, all steps are presented below:
Necessary pre-processing of historical price data and discussion text;
Acquire hidden representation of price series via an LSTM network;
Construct a heterogeneous graph based on forum discussion network to capture semantic and network features.
Integrate the features extracted from the above process and make the prediction.
3.3. Pre-Processing of Data
After acquiring raw data from the Internet, we have to do some data cleaning and pre-processing before feeding them into our model. For forum discussions, We first remove stop words and irregular words or expressions. Then, we use the nltk [41] package for tokenization and transform words into vectors with GloVe [42]. For price data, we replace the price’s absolute value with the change of price relative to the previous week. As our price data is weekly, we choose the thread with the most comments every week to make graphs used in later steps. We assume that the more comments a thread contains, the more information we can extract from the discussion, thus helping the price prediction.
3.4. Acquiring Hidden Representation of Price Series
Let , where , is defined in Section 3.1. As shown in Figure 3, we feed into a one-layer LSTM network
(2)
and we use the representation of the last hidden state as the feature of historical prices.3.5. Constructing Heterogeneous Graph Based on Discussion Network
There are two types of relations in our constructed graph. To obtain a global representation combining semantics, propagation, and user information, we decompose the heterogeneous graph into a post-post subgraph and a post-user subgraph based on meta-path post-post and post-user. After decomposition, only one type of relationship is considered for each subgraph. This process is shown in Figure 3.
Then, we feed the subgraphs into GAT [43]. GATs have shown great capacity in capturing the graph structures. Therefore, we choose GAT in our work. We will describe the details here.
The propagation step from the l-th layer to the (l + 1)-th layer of GAT is
(3)
where is the representation of node in the l-th layer. is a trainable weight matrix, is the ReLU activation function. is the set of one-hop neighbors of node , itself is also included in the set. And the attention coefficients are computed as(4)
is the node embedding matrix. To extract the structure information of subgraphs, We reserve the matrix of activations in the l-th layer for later use.Now that we have the node embedding matrix of post-post subgraph and that of post-user subgraph , after feeding them into GAT, we can obtain node representations (output) and , respectively.
The decomposed subgraphs contain different information. The post-post subgraph contains the semantic information of text contents and propagation features, while the post-user subgraph primarily contains user features and relations between the user and its post. To acquire a global and complete representation of heterogeneous graphs, we design an attention mechanism to fuse the information in different subgraphs together.
For this part, we have and as input, we need to calculate the weights of each subgraph and
(5)
To learn the weights and , we first transform the representation of nodes in subgraphs into higher-level features by applying a linear transformation. Then, we compute an attention score for each node by doing dot product operations between the transformed node representations and a learnable weight vector . Next, we average the attention scores of all nodes in the subgraph and use it as the subgraph score. The score of the subgraph is computed as follows
(6)
where e is the score of subgraph, W is the learnable weight matrix. W together with attention vector are shared by all subgraphs.After above steps, we normalize the attention scores e ( or ) using softmax function
(7)
where and ( or ) denotes the weight of subgraphs.Finally, with the learned weight and , we fuse the node representations in subgraphs to form a global representation of the heterogeneous graph
(8)
(9)
contains rich global relation information of the discussion network. Therefore, after a necessary pooling layer, we attain the discussion network’s representation for price prediction in a later section.3.6. Intergrating Features and Making Prediction
As in Figure 3, after extracting from the discussion network and from price series, we finally combine those information and make the prediction, which is formulated as follows
(10)
where is the concatenation of and . We then feed into a simple feed-forward neural network with softmax function(11)
where W and b are learned parameters. is the predicted probability distributions.Finally, in order to train the parameters, the cross-entropy loss is used as the model’s objective function. The loss is computed as
(12)
where is [1,0] or [0,1]. N is the number of training data and c indicates the class label.4. Experiments and Results
4.1. Description of Data
In this study, we first collect historical prices of the hog, maize, and bean from 2013 to 2020. All these historical prices are available from
4.2. Experimental Setup
When we make the dataset, each input price series has a corresponding discussion network and a label c. means hog price will increase next week and vice versa.
We implement our model and other compared models using PyTorch [44] and PyTorch-geometric [45]. Our experiments have been conducted on Tesla P100-16GB. We use the cross-entropy loss function and the Adagrad optimizer to train our model and set the learning rate as .
4.3. Evaluation Metrics
Considering that we transform the price prediction task into a binary classification problem, we use four popular performance indices to evaluate the models. These evaluation metrics are Accuracy, F1-score, Precision, and Recall. They are calculated as follows
(13)
where TP denotes true positive, TN denotes true negative, FP denotes false positive and FN denotes false negative.4.4. Competing Models
We have conducted extensive experiments to compare our proposed method’s performance with several popular methods for classification problems, single LSTM, multilayer perceptron (MLP), and STL-ATTLSTM. We briefly discuss these competing methods here.
-
Single LSTM: Proposed by Hochreiter and Schmidhuber [38], LSTM networks have shown superiority in processing time-series data. Therefore, LSTM networks are usually exploited when dealing with time series classification problems. In this study, we build a one-layer LSTM network for comparison;
-
MLP: As a class of feedforward artificial neural network, multilayer perceptron usually consists of an input layer, an output layer, and several hidden layers. Researchers often make use of MLPs to solve regression problems. Since classification is a particular case of regression, MLPs also make good classifiers;
-
STL-ATTLSTM: Proposed by Yin et al. [37], STL-Attention-based LSTM is a state-of-the-art method to forecast the price of agricultural products. In their original paper, STL-ATTLSTM makes use of several types of information to forecast monthly vegetable prices, such as vegetable prices, weather information, and market trading volumes [37]. According to their paper, the STL algorithm decomposes the price series into three parts: trend, seasonality, and remainder components. Then, they feed the remainder components into an LSTM network with an attention layer by removing the trend and seasonality components. Their experiments have shown promising results;
-
BERTLSTM [46]: As BERT [47] has shown a great capacity to capture semantic information from text, Ko and Chang [46] exploited BERT to extract better representations of news article. After feeding the stock prices into LSTM module, they integrate price features and news features. Inspired by their study, we select BERTLSTM as one of the competing models;
-
GCNLSTM [48]: GCN [49] is a popular model to extract hidden representation on graph structure data. Li et al. [48] proposed GCNLSTM for traffic flow prediction. They employ GCN to mine the spatial relationships of traffic flow. Then, they use LSTM module to extract temporal features. Finally, they design a structure to make the final prediction.
Table 2 shows the performance of our proposed method and all competing methods on the dataset.
4.5. Discussion of Results
4.5.1. Performance Comparison
As is shown in Table 2, deep neural networks achieve much better results than MLP. This phenomenon is very reasonable because deep neural networks have much more powerful learning abilities and can extract better representations. In contrast, typical MLP architectures are not deep, and they do not have many hidden layers, resulting in their relatively poor performance. Our experiments again prove the effectiveness of deep neural networks in classification.
LSTMs are well known for their effectiveness in dealing with series data. Our experiments also demonstrate this. As in Table 2, Single LSTM, STL-ATTLSTM, and our HGLSTM all contain LSTM networks, accounting for their better performance than MLP. Yin et al. [37] integrates the STL method and attention mechanism into LSTM. Therefore, their STL-ATTLSTM outperforms Single LSTM.
According to Table 2, our proposed Heterogeneous Graph-enhanced LSTM (HGLSTM) outperforms every competing method in terms of all metrics, indicating the effectiveness of our model. When we compare our HGLSTM with Single LSTM, the main difference is that HGLSTM includes the information of online discussions. This alone proves that discussion networks contain helpful information for hog price prediction.
It is also worth noting that, although we do not deal with Seasonality or Trends like STL-ATTLSTM, our HGLSTM still outperforms STL-ATTLSTM. According to their paper, the STL algorithm decomposes the price series into trend, seasonality, and remainder components before feeding the remainder components into LSTM. Their experiments proved the effectiveness of their model. Although we do not decompose the price series using the STL algorithm, our model still outperforms STL-ATTLSTM. This is mainly due to the introduction of discussion information, showing the success of including such information.
4.5.2. Importance of Constructing Heterogeneous Graph
Now that we have proved the effectiveness of including discussion information, we still have various ways of capturing that representation. Thus, we further perform experiments to show that constructing the heterogeneous graph is the most effective way. We carefully choose two competing methods, GCNLSTM and BERTLSTM, which have been described in Section 4.4.
As Figure 2 shows, HGLSTM unquestionably outperforms both GCNLSTM and BERTLSTM. Here, we analyze the reasons for this. For BERTLSTM, it neglects both propagation structure and user information. Such information is indispensable for classification. GCNLSTM has a similar shortcoming. Its ability to model graph network enables it to exploit the propagation structure; however, it only allows one type of node and does not consider user information. Thus, GCNLSTM treats every node equally. This is definitely not good because, in a real discussion network, the credit of different users is not the same. High-credit users should influence the prediction far more than low-credit users. Adding user nodes into the graph, our proposed HGLSTM has two types of nodes, solving this problem.
5. Conclusions and Future Work
In this paper, we propose a heterogeneous graph-enhanced LSTM network for hog price prediction. We assume the online discussions can enhance hog price prediction and prove this through our experiments. To make the best use of discussions and user information, we resort to constructing the heterogeneous graphs. Our experiments demonstrate the effectiveness of incorporating online discussions and constructing heterogeneous graphs.
In the future, we plan to investigate how to combine discussion information and price series representations more efficiently and effectively.
Author Contributions
Conceptualization, K.Y., X.C., Y.P. and K.Z.; methodology, K.Y.; software, K.Y. and K.Z.; investigation, K.Y. and Y.P.; resources, X.C.; writing—original draft preparation, K.Y.; writing—review and editing, K.Y., X.C., Y.P. and K.Z. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by National Key Research and Development Program of China No.2018YFC1604000, Fundamental Research Funds for the Central Universities No.2042017gf0035.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
All the data used in this paper is available online and we have given the specific information about dataset in previous sections.
Acknowledgments
We would like to thank the support of National Key Research and Development Program of China No. 2018YFC1604000, Fundamental Research Funds for the Central Universities No. 2042017gf0035 as well as all the reviewers for their enlightening comments.
Conflicts of Interest
The authors declare no conflict of interest.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Figures and Tables
Figure 1. LSTM cell.
Figure 2. This is an example of the discussion network we make use of in this paper. Four users and five posts are engaged in this discussion network.
Figure 3. Overall structure of our proposed model.
Figure 4. Weekly prices of the three agricultural products from 2013 to 2020.
Figure 5. An example of the discussion content.
Summary of the literature review.
| Model | Purpose | Input |
|---|---|---|
| ARIMA [13] | Cocoa bean | Price |
| Seasonal ARIMA [14] | Tomato | Price |
| Multivariate linear regression [15] | Corn | Price, production |
| BPNN [18] | Tomato | Price |
| ARIMA-SVM [19] | Garlic | Price |
| RBF [20] | Tomato | Price |
| Hybrid model of BPNN, RBF and GA [22] | Mushroom | Price |
| QR-RBF [23] | Soybean | Price, import/Output, consumer index, money supply |
| STL-LSTM [28] | Crop | Price, climate, trading volumes |
| Similar Sub-Series Search and SVM [29] | Hog | Price |
| Wavelet analysis based LSTM [31] | Cabbage | Price |
| Dual-stage attention based RNN [33] | Stock price | Price |
| Attention-based LSTM [34] | Travel time | time |
| STL-ATTLSTM [37] | Vegetable prices | Price, weather, market trading volumes |
Experimental results on our dataset.
| Method | Accuracy | Precision | Recall | F1 |
|---|---|---|---|---|
| MLP | 0.552 | 0.591 | 0.433 | 0.501 |
| Single LSTM | 0.741 | 0.742 | 0.767 | 0.754 |
| STL-ATTLSTM | 0.792 | 0.763 | 0.799 | 0.781 |
| BERTLSTM | 0.809 | 0.783 | 0.809 | 0.796 |
| GCNLSTM | 0.814 | 0.812 | 0.804 | 0.808 |
| proposed HGLSTM | 0.832 | 0.838 | 0.812 | 0.825 |
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Forecasting the prices of hogs has always been a popular field of research. Such information has played an essential role in decision-making for farmers, consumers, corporations, and governments. It is hard to predict hog prices because too many factors can influence them. Some of the factors are easy to quantify, but some are not. Capturing the characteristics behind the price data is also tricky considering their non-linear and non-stationary nature. To address these difficulties, we propose Heterogeneous Graph-enhanced LSTM (HGLTSM), which is a method that predicts weekly hog price. In this paper, we first extract the historical prices of necessary agricultural products in recent years. Then, we utilize discussions from the online professional community to build heterogeneous graphs. These graphs have rich information of both discussions and the engaged users. Finally, we construct HGLSTM to make the prediction. The experimental results demonstrate that forum discussions are beneficial to hog price prediction. Moreover, our method exhibits a better performance than existing methods.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer





