This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Heterogeneous Network Embedding for Deep Semantic Relevance Match in E-commerce Search

Ziyang Liu,1 Zhaomeng Cheng, 1 Yunjiang Jiang 1 Yue Shang 1 Wei Xiong 1 Sulong Xu 1 Bo Long 1 Di Jin 2
Abstract

Result relevance prediction is an essential task of e-commerce search engines to boost the utility of search engines and ensure smooth user experience. The last few years eyewitnessed a flurry of research on the use of Transformer-style models and deep text-match models to improve relevance. However, these two types of models ignored the inherent bipartite network structures that are ubiquitous in e-commerce search logs, making these models ineffective. We propose in this paper a novel Second-order Relevance, which is fundamentally different from the previous First-order Relevance, to improve result relevance prediction. We design, for the first time, an end-to-end First-and-Second-order Relevance prediction model for e-commerce item relevance. The model is augmented by the neighborhood structures of bipartite networks that are built using the information of user behavioral feedback, including clicks and purchases. To ensure that edges accurately encode relevance information, we introduce external knowledge generated from BERT to refine the network of user behaviors. This allows the new model to integrate information from neighboring items and queries, which are highly relevant to the focus query-item pair under consideration. Results of offline experiments showed that the new model significantly improved the prediction accuracy in terms of human relevance judgment. An ablation study showed that the First-and-Second-order model gained a 4.3% average gain over the First-order model. Results of an online A/B test revealed that the new model derived more commercial benefits compared to the base model.

Introduction

Semantic Match is one of the basic tasks for natural language processing and has many real-world applications such as question answering, textual entailment, paraphrase identification, and information retrieval (Berger et al. 2000; Dagan, Glickman, and Magnini 2005; Dolan et al. 2004; Li 2011). Unlike simple text matching, semantic similarity match aims to infer the semantic similarity of two sentences rather than the extent of common words between the two. For example, while “mac pro” and “mac lipstick” look alike, they describe two different items, i.e. computer and lipstick; “iPad” and “apple tablet” have no common word at all but rather refer to the same item, i.e. tablet computer. While similarity match usually deals with two homogeneous sentences of comparable lengths and expects to match every position of both sentences, semantic relevance match deals with heterogeneous pieces of text such as query and document in ad-hoc information retrieval and expects to match some keywords in documents with queries. As a specific application of ad-hoc information retrieval, e-commerce search serves as a platform to fetch candidate items highly relevant to a given query to satisfy the user’s purchase requirement. If a search system returns too many semantically irrelevant items, it will render unpleasant user experience and erode the user’s trust and confidence toward the e-commerce platform. Therefore, the semantic relevance estimation is critically important for long-term user satisfaction.

The current research on deep semantic learning can be grouped into two camps: representation-focused and interaction-focused. The representation-focused methods (Huang et al. 2013; Wan et al. 2016a) typically embed two input sentences separately using two neural networks and computes a relevance score to measure the similarity between the two embeddings. The common similarity measures used include the cosine similarity and negative Jensen-Shannon divergence. On the other hand, the interaction-focused methods (Pang et al. 2016a; Chen et al. 2017) concatenate two input sentences as input to a single neural network which captures interactions between their features. These two types of methods exploit the semantic relationship between the query and the item text and are effective for static and context-free matching problems. However, the semantic match problem in many real applications is often not context-free, e.g. search logs store plenty of valuable context data111In this work, context information refers broadly to the neighbor features in a query/item bipartite graph, rather than the more restricted sense of co-occurring items in a search session. such as the query/item incidence network and the users’ historical behavior sequences (view, click, purchase, etc). These types of context information provide the actual semantic content of a query or item, which as ignored by the current semantic match methods.

Refer to caption
Figure 1: In the e-commerce search scene, the search log essentially provides a heterogeneous bipartite network. Similar to 1st/2nd order proximity in network representation learning, we define 1st/2nd order relevance in a semantic match under the e-commerce search scene.

When exploiting the graphical context information for semantic relevance match, we face the following challenges:

C1: Limited supervised information. Although a variety of user behavior signals were recorded in search logs, they are often noisy and misaligned with the search relevance objective: many factors other than relevance may affect a user’s final decision such as item popularity, title attractiveness (click-baitiness), and result set diversity. Human labels can provide accurate relevance information, but training an excellent deep model often requires millions or more examples, which are labor-intensive and costly to collect. In short, high quality annotated signals are scarce in our problem domain.

C2: Uncertainty in how to integrate context information. Since user behavioral feedback cannot naively substitute relevance signals, systematic utilization of the massive amounts of search log data has been a central research theme in e-commerce search. Incorporating context information on top of the logged user feedback presents new challenges. Any proposed method should be context-aware and corroborative of the final relevance objective, an under-explored area in the current research.

C3: Memory and latency constraints. A query phrase and an item’s title in an e-commerce search can often be represented in diverse forms. If a text is stored as a unique entity in an online key-value store, it may take 100 Gigabytes of memory (space complexity). Such a large memory usage will result in large models, which in turn will fail to respond quickly to online queries and make the run-time complexity severely limited.

To address these challenges, we propose to study the semantic match problem in dynamically evolving search scenarios. This problem is different from the existing context-free search problems. In particular, we consider that queries and items in the search log constitute a natural heterogeneous bipartite network structure (Fig. 1). In this network, there are two types of nodes (queries and items) and two types of edges (click behaviors and purchase behaviors). Traditional approaches estimate the relevance of two adjacent nodes in this network in isolation. We argue that contextual relevance can be significantly improved by taking into account their neighbors’ semantic information.

A central problem in constructing a bipartite network is edge refinement, specifically that in the query/item co-occurrence graph. Due to the noisy nature of user behaviors, we cannot rely on them exclusively to build the connection between two nodes. On the other hand, the pre-trained language representation model (e.g. BERT (Devlin et al. 2019)) is equipped with a good semantic understanding capacity. So we choose BERT as the teacher model and extract relevance knowledge which can be used as annotated information (C1). We then use this external knowledge to refine the user behavior network, via adding ignored edges and deleting noisy edges, to ensure a high-quality network structure.

Similar to the concept of first and second order proximity (Tang et al. 2015; Wang, Cui, and Zhu 2016), we propose alternative definitions of the first-order relevance and second-order relevance. Previous semantic match research (Nigam et al. 2019; Jiang et al. 2019) only considers the first-order relevance, which is reasonable for context-free semantic match problems, but will lose valuable context information when applied to real-world applications. We argued in C2 that both the first and second order relevance should be taken into consideration together for e-commerce search. Therefore, we propose a new model of Heterogeneous GNN for Semantic Match problem (HG4SM), which can be broadly applied to any search ranking problem that seeks to incorporate the context of real-world applications. The model captures the first-order relevance using a word interaction matrix attached with positional encoding and captures the second-order relevance using the metapath-guided embedding with graph attention scores. To our best knowledge, HG4SM is the first heterogeneous network embedding for the task of search relevance.

Although a query phrase and an item’s title may appear in many different forms, the words in these sentences have smaller representation space and are easy to embed. Thus, we use a word distributed representation to depict various queries and titles, which greatly reduces the model’s space complexity compared to a document embedding. The above design ensures that our model derives explicit interaction matching signals and reasonable node semantic embeddings so that we only need to employ shallow neural networks to combine all embeddings. Therefore, the whole model has a low time complexity which is suitable to deploy online (C3). We list some related works and compare them with our work in the appendices.

Problem Definition

Here we give: 1) three basic definitions about heterogeneous network and node proximity; 2) and two new definitions about the novel problem of Second-order Relevance.

Definition 1

Heterogeneous Network. Given a network or graph 𝒢=(𝒱,,𝒜,)\mathcal{G}=(\mathcal{V},\mathcal{E},\mathcal{A},\mathcal{R}) with nodes or vertices set 𝒱={v1,v2,,vn}\mathcal{V}=\left\{v_{1},v_{2},\cdots,v_{n}\right\} and edges set ={e1,,em}\mathcal{E}=\left\{e_{1},\cdots,e_{m}\right\}, if the node type’s mapping function 𝒱𝒜\mathcal{V}\rightarrow\mathcal{A} and the edge type’s mapping function \mathcal{E}\rightarrow\mathcal{R} satisfy the condition: |𝒜|+||3|\mathcal{A}|+|\mathcal{R}|\geqslant 3, then 𝒢\mathcal{G} is a heterogeneous network.

For the convenience of exploiting the heterogeneous information in a network, we only consider unsigned networks (i.e. no negative edges) with undirected and unweighted edges. In a heterogeneous network, metapath and its corresponding instances are universal concepts and defined as:

Definition 2

Metapath and Metapath Instance. A metapath P=a1r1rlal+1P=a_{1}\overset{r_{1}}{\rightarrow}\cdots\overset{r_{l}}{\rightarrow}a_{l+1} represents the path from a1a_{1} to al+1a_{l+1} successively through r1,,rlr_{1},\cdots,r_{l} (ai𝒜a_{i}\in\mathcal{A}, rir_{i}\in\mathcal{R}). A metapath instance p=(v1,,vl+1)p=\left(v_{1},\cdots,v_{l+1}\right) is a definite node sequence instantiated from metapath PP.

A good network embedding can well preserve the network’s structural information, e.g., local network structure and global network structure. The local and global network structures can be respectively represented by the first-order and second-order proximity. Proximity represents two nodes’ spatial closeness, where the first-order proximity is the local pairwise proximity and the second-order proximity is the neighbor structure proximity between two nodes.

Definition 3

First-order Proximity and Second-order Proximity. For two nodes viv_{i} and vjv_{j} in a network, their first-order proximity can be formalized as the function Fprx1(vi,vj)F_{prx}^{1}(v_{i},v_{j}) and their second-order proximity can be formalized as the function Fprx2(Nei(vi),Nei(vj))F_{prx}^{2}(Nei(v_{i}),Nei(v_{j})) where Nei(.)Nei(.) denotes a neighbor set. If there is an edge between viv_{i} and vjv_{j}, then the Fprx1(vi,vj)F_{prx}^{1}(v_{i},v_{j})’s value is bigger than its value when there is no such edge. If viv_{i} and vjv_{j} share more common neighbors, then the Fprx2(Nei(vi),Nei(vj))F_{prx}^{2}(Nei(v_{i}),Nei(v_{j}))’s value is bigger.

For the first-order proximity function, a common design is:

Fprx1(vi,vj)={1,if e=(vi,vj);0,otherwiseF_{prx}^{1}(v_{i},v_{j})=\begin{cases}1,&\mbox{if $e=(v_{i},v_{j})\in\mathcal{E}$;}\\ 0,&\mbox{otherwise}\end{cases} (1)

For the second-order proximity function, a possible choice can be:

Fprx2(Nei(vi),Nei(vj))=Nei(vi)Nei(vj)Nei(vi)Nei(vi)F_{prx}^{2}(Nei(v_{i}),Nei(v_{j}))=\frac{Nei(v_{i})\cap Nei(v_{j})}{Nei(v_{i})\cup Nei(v_{i})} (2)

Intuitively, even though there is no edge between v1v_{1} and v2v_{2}, if they share many common neighbors, they should also be very similar to each other. So the second-order proximity is an important supplement to the first-order proximity.

If we view proximity from the perspective of “path”, we can conclude that: 1) the first-order proximity reveals that the shortest path between v1v_{1} and v2v_{2} is a path whose length is 1, and 2) the second-order proximity reveals that the length of their shortest path is 2. Higher-order proximity reveals that the length of their shortest path is greater than 2. With the increase of the path’s length, the information intensity will gradually weaken along paths, so higher-order proximity is not considered here. Proximity is also suitable for the heterogeneous network, but the definition of neighbors in the homogeneous network and heterogeneous networks are different. In heterogeneous networks, neighbors are metapath-based and denote those nodes that have the same type as the central node. For example, in a citation network, the author citation relationship can be represented as “AuthorAuthor-PaperPaper-AuthorAuthor” and its corresponding metapath is “AA-PP-AA”. On this metapath, if two authors together cite the same paper, they establish a metapath-based neighbor relationship and tend to have similar research interests.

In a heterogeneous network, the relevance score represents the semantic closeness of two nodes. Similar to node proximity, we introduce the following new definitions.

Definition 4

First-order Relevance and Second-order Relevance. When mapping two sentences into two nodes viv_{i} and vjv_{j} in a heterogeneous network 𝒢\mathcal{G}, the first-order relevance and second-order relevance can be separately formalized as the function Frel1(vi,vj)F_{rel}^{1}(v_{i},v_{j}) and Frel2(Nei(vi),Nei(vj))F_{rel}^{2}(Nei(v_{i}),Nei(v_{j})). If viv_{i} is semantically more similar to vjv_{j}, then Frel1(vi,vj)F_{rel}^{1}(v_{i},v_{j}) is larger. If viv_{i}’s neighbors are semantically more similar to vjv_{j}’s neighbors, then Frel2(Nei(vi),Nei(vj))F_{rel}^{2}(Nei(v_{i}),Nei(v_{j})) is larger.

Instead of proximity, our relevance considers local and global node semantics. We argue that both the first-order relevance and second-order relevance are necessary for semantic match. They supplement each other in the same way as the first-order and second-order proximity. A good mechanism should fully consider these two types of relevance and make them cooperate with each other, i.e. the first-order relevance plays a dominant role and the second-order relevance provides important auxiliary information. In other words, when it is difficult to judge whether two nodes are semantically relevant from the textual information of themselves, the second-order relevance which is based on neighbor set can become useful. Finally, we give the definition of the research objective of this work below.

Definition 5

Heterogeneous Network Embedding for Relevance Estimation. For a heterogeneous network 𝒢=(𝒱,,𝒜,)\mathcal{G}=(\mathcal{V},\mathcal{E},\mathcal{A},\mathcal{R}), the network embedding for relevance estimation aims to learn each node’s low-dimension distributed representation which simultaneously satisfies the need of the first-order relevance and the second-order relevance, and thus preserves both the local and global node semantics.

Model Description

We propose a complete heterogeneous network embedding architecture for the e-commerce search relevance match problem (Fig. 2). Concretely, we first need to denoise a user behavior network. We introduce the fine-tuned BERT model of producing external knowledge in order to refine the click behavior edge in the original network. Based on this knowledge-enhanced network derived, we select two most important metapaths. We apply a node-level encoder and a metapath attention unit together to integrate these neighboring nodes’ context information into the central node. In addition, considering the e-commerce scene’s particularity, we give: 1) a special vocabulary formation rule to preserve the complete semantics of many products or brands, 2) the word-level interaction representation to capture the micro semantics matching signals between the query and title; and 3) the sentence-level semantic representation to capture the macro semantics matching signals between the query and title.

Refer to caption
Figure 2: The overall architecture of HG4SM. It mainly includes three steps: knowledge-enhanced network refinement, heterogeneous network analysis and relevance modeling.

Network Construction

In a real-world search application, queries and items and their relationships naturally form a heterogeneous bipartite network based on the multi-type user behaviors like view, click, and purchase. However, as mentioned earlier, user behaviors are typically biased and noisy. So if one directly conducted embedding learning on the original user behavior network, it would be difficult for the model to estimate explicit relevance relationships. To solve this problem, we introduce external knowledge to refine the user behavior network and then construct a knowledge-enhanced heterogeneous network. Here the external knowledge is provided by the BERT model that is pre-trained on a large text corpus and then fine-tuned on some in-house data. The whole network construction includes the following phrases.

Fine-tuning BERT. Transformer-based models such as BERT and ERNIE (Sun et al. 2019) have been preferably used as NLP benchmark in recent years. Here we use the BERT-Base model222https://github.com/google-research/bert composed of stacked Transformer units and fine-tune it on some in-house data. The positive and negative examples in the data are human-labeled and cover various categories of items. The fine-tuned BERT is equipped with a high relevance discrimination ability and thus can act as an expert on filtering noisy data.

Behavior network formation. The user behavior network in the left network of Fig. 2 is built on some user search log (over several months) which records user clicks and purchase behaviors as well as their frequencies. In this network, an edge represents an existing click behavior or purchase behavior between a query-node and an item-node. The edge of purchase behavior is sparse but important. The edge of click behavior is denser but highly noisy, making it difficult to learn a good model and predict reasonable results. Therefore, we introduce BERT to refine this network.

Knowledge-enhanced network refinement. Given the scarcity of user purchase feedback and noisiness of click feedback, we retain all the original purchase behavior edges while use the external knowledge from BERT to refine the click behaviors. Specifically, we set two different thresholds α\alpha and β\beta, i.e. α\alpha is used for judging whether a clicked item is truly relevant to its query, and β\beta is used for judging whether an unclicked item is truly irrelevant to its query. This strategy can help remove noises in user behaviors and at the same time extract the missing but crucial relevance signals not captured by user behaviors. To preserve the high-quality neighbor set, for each central node, we rank its neighboring nodes with the priority of “purchase\rightarrowhigh click\rightarrowlow click” and sample top-2 of them as the final neighbor set.

Heterogeneous Network Analysis

In HG4SM, the basic units are query nodes, item nodes and the refined user-behavior-oriented edges between them. Based on this schema, we employ two metapaths, “QQ-II-QQ” and “II-QQ-II”, where “QQ” and “II” stand for “QueryQuery” and “ItemItem”. These two metapaths correspond closely to the second-order relevance definition defined earlier. Compared to some more complex metapaths (such as “QQ-II-QQ-II-QQ” and “II-QQ-II-QQ-II”), the adopted metapaths in our model is both effective (with less information density loss) and computationally efficient. We further choose two instances for each metapath whenever they are available and pad with zero embeddings otherwise . Take “QQ-II-QQ” as an example, its instances can include “QQ-Itop1I_{\mathrm{top1}}-Qtop1Q_{\mathrm{{top1}}}” and “QQ-Itop1I_{\mathrm{{top1}}}-Qtop2Q_{\mathrm{{top2}}}”, as shown in Fig.2.

First-order Relevance Modeling

The whole framework of the new HG4SM model includes both the first-order relevance modeling and second-order relevance modeling. We first introduce the first-order relevance modeling which captures macro and micro semantic match signals by incorporating both the representation-focused and interaction-focused designs.

Word embedding in e-commerce scene. In e-commerce applications, it is infeasible to represent queries and items as individual embeddings since their entity space is effectively unbounded. Instead, we adopt word embedding in HG4SM, which dramatically reduces the representation space and deals well with the cold-start situation. Numerous product type names or attribute names (like “iphone11” and “256GB”) exist, but the basic word segmentation based on N-gram or WordPiece word segmentation splits these special names and thus cannot reveal their complete semantics. To adapt to this feature, we first treat single Chinese characters, contiguous numerals or English letters as single words (Fig. 3). We then retain only the high-frequency words in this list. Compared to potentially million queries and billion items, our vocabularies are only in the tens of thousands, which saves memory consumption and lookup time of id and embedding by a large margin. We represent each word with a dd-dimensional vector and train these word vectors or embeddings using Word2Vec (Mikolov et al. 2013). We use 𝑬Qi\bm{E}_{Q}^{i} (and 𝑬Ii\bm{E}_{I}^{i}) to denote the ii-th word’s embedding of query QQ (and the item’s title II).

Macro matching element. Most of the first-order relevance methods are either representation-focused (capturing macro matching signals) or interaction-focused (capturing micro matching signals). We employ their mixture in the HG4SM’s first-order relevance modeling. For the representation-focused part, take query QQ with lQl_{Q} words as an example, its sequence embedding 𝑬seqQ\bm{E}_{seq}^{Q} is obtained by calculating the element-wise mean value of {𝑬Qi}\left\{\bm{E}_{Q}^{i}\right\}:

𝑬seqQ=1lQi=1lQ𝑬Qi\bm{E}_{seq}^{Q}=\frac{1}{l_{Q}}\sum_{i=1}^{l_{Q}}\bm{E}_{Q}^{i} (3)

The above 𝑬seqQ\bm{E}_{seq}^{Q} depicts the whole semantic information of query QQ, so that it can be viewed as a simplified version of representation-focused embedding.

Micro matching element. To capture micro matching signals, we need to model the word-level interaction information (Pang et al. 2016a) between queries and titles. Suppose the sequence lengths of query QQ and title II are lQl_{Q} and lIl_{I}, respectively, we build an interaction matrix 𝑴int\bm{M}_{int}:

𝑴int=[𝑬Q1,,𝑬QlQ]T×[𝑬I1,,𝑬IlI]\bm{M}_{int}=[\bm{E}_{Q}^{1},...,\bm{E}_{Q}^{l_{Q}}]^{\mathrm{T}}\times[\bm{E}_{I}^{1},...,\bm{E}_{I}^{l_{I}}] (4)

The interaction embedding is derived by reshaping 𝑴int\bm{M}_{int} into a one-dimension vector:

𝑬int=reshape(𝑴int)\bm{E}_{int}=reshape(\bm{M}_{int}) (5)
Refer to caption
Figure 3: An example of the word-level matching signal. The matched and unmatched words are respectively represented by solid and dashed rectangles.

Position encoding. Besides, considering the sequential structure of texts, we further add a position embedding to each word embedding before calculating the correlation matrix. The position embedding is set as a trainable embedding vector and has the same dimension as the word embedding.

Second-order Relevance Modeling

Most semantic match methods focus on the first-order relevance, but ignore the second-order relevance (which integrates the neighbor information on metapaths into the central node and is essentially important in many real context-aware scenes). A complete semantic relevance estimation model should integrate them together. Here we consider how to generate second-order relevance embeddings so that it can incorporate context information in the network to enrich and improve the central nodes’ semantics. In general, the second-order relevance model consists of a node-level encoder and a metapath instance-level aggregator.

Node-level encoder. Metapath instance bridges the communication gap between heterogeneous nodes and thereby can infer the node’s global semantic embedding (rather than the local semantic embedding). For each metapath instance, to derive the global node semantics, we integrate the neighboring node embedding into the central node embedding with mean encoder. Take the instance “QQ-Itop1I_{\mathrm{{top1}}}-Qtop1Q_{\mathrm{{top1}}}” as an example, its corresponding embedding is:

𝑬QItop1Qtop1=MEAN{𝑬seqQ,𝑬seqItop1,𝑬seqQtop1}\bm{E}_{Q-I_{\mathrm{{top1}}}-Q_{\mathrm{{top1}}}}=\mathrm{MEAN}\left\{\bm{E}_{seq}^{Q},\bm{E}_{seq}^{I_{\mathrm{top1}}},\bm{E}_{seq}^{Q_{\mathrm{top1}}}\right\} (6)

Instance-level aggregator with graph attention. Different meatapath instances convey different information, so they should have different effects on the final metapath’s embedding. However, the mapping relationship between the metapath instance’s embedding and the metapath’s embedding is unknown. To learn their relationship automatically, we introduce a “graph attention” mechanism to the formation progress of the metapath’s embedding. The attention mechanism enables the model to learn a weight distribution and assign different weights to different components, which has been successfully applied in computer vision and natural language processing (Vaswani et al. 2017). Here we introduce graph attention to represent the mapping relationship between metapath and its instances. The final metapath’s embedding is then obtained by accumulating the embeddings of each metapath’s instances with attention scores. Take metapath “QQ-II-QQ” as an example, its corresponding embedding is defined as:

𝑬QIQ=\displaystyle\bm{E}_{Q-I-Q}=
σ(Att1𝑬QItop1Qtop1+Att2𝑬QItop1Qtop2)\displaystyle\sigma(\mathrm{Att}_{1}\cdot\bm{E}_{Q-I_{\mathrm{{top1}}}-Q_{\mathrm{{top1}}}}+\mathrm{Att}_{2}\cdot\bm{E}_{Q-I_{\mathrm{{top1}}}-Q_{\mathrm{{top2}}}}) (7)

where σ(.)\sigma(.) is the activation function. Though Atti\mathrm{Att}_{i} can be set as a fixed value, we adopt a more flexible way, i.e., we use a single-layer neural network to learn it automatically. Specifically, we feed the concatenation of embeddings of the central node and the metapath instances into a one-layer neural network and output an attention distribution in the softmax layer:

𝑬concat=CONCAT(𝑬seqQ,𝑬QItop1Qtop1)\bm{E}_{\mathrm{concat}}=\mathrm{CONCAT}(\bm{E}_{seq}^{Q},\bm{E}_{Q-I_{\mathrm{{top1}}}-Q_{\mathrm{{top1}}}}) (8)
Atti=softmax(σ(Watt𝑬concat+b))\mathrm{Att}_{i}=\mathrm{softmax}(\sigma(W_{att}*\bm{E}_{\mathrm{concat}}+b)) (9)

where WattW_{att} is a 1*2d trainable vector, shared across all metapaths.

Embedding fusion. Based on the first-order and second-order relevance modeling, three types of embeddings can be generated, including the representation- focused, interaction-focused and metapath-guided embeddings. To combine them, we concatenate these three embeddings together, feed it to the deep neural networks, and output a relevance score. To alleviate the model’s high time-cost, we use efficient three-layer DNNs without considering more complex neural network structures such as CNN and LSTM.

Experiments

We now present experimental results. We first introduce the baseline methods, performance metrics, final comparison results, ablation study, and online performance. We discuss the datasets used, implementation details, and parameters sensitivity experiment in the appendices.

The Baselines and Metrics

The HG4SM model is compared to several existing state-of-the-art semantic models. To facilitate their implementation, we use the open-source codebase packages which are as follows.
: Nine methods are implemented using MatchZoo (Fan et al. 2017), which is a toolkit for deep text matching, based on TensorFlow and Keras. We compare these typical methods (Guo et al. 2016; Pang et al. 2016b; Hu et al. 2014; Wan et al. 2016b; Mitra, Diaz, and Craswell 2017; Xiong et al. 2017) in MatchZoo with HG4SM. We use the default hyper-parameter setting for all the methods in MatchZoo333https://github.com/NTMC-Community/MatchZoo.
§: Two other baselines DSSM (Huang et al. 2013) and ESIM (Chen et al. 2017) are also included in the comparison. DSSM is a classical representation-focused method using pairwise examples to train the model. ESIM is a successful interaction-focused method which incorporates the syntactic parsing information into chain LSTM for natural language inference.

To comprehensively measure the performance of our HG4SM and these baseline methods, we use six metrics, including: 1) Area Under the receiver operating characteristic Curve, 2) Accuracy, 3) Precision, 4) Recall, 5) F1-score and 6) False Negative Rate (since it is more serious than False Positive Rate for e-commerce ranking). They are respectively denoted by AUC, Acc, Pre, Recall, F1 and FNR for short. For the first five metrics, the higher the metric value, the better the model’s performance. For FNR, the lower its value, the better the model’s performance. Note that AUC often serves as the most important metric while the others also provide auxiliary supports for our analysis.

Table 1: Comparison on the in-house data. The best results are bolded. ‘Rep’ denotes the representation-focused method, ‘Int’ the interaction-focused method, and ‘Both’ their mixture. ‘Triple’ denotes ‘Both’ with ‘HIN’. ‘\blacktriangleleft’ denotes the most important metric in the real application. ‘(-)’ represents that the lower value corresponds to better performance. Here the used all-categories data is the sampled 10 million data.
Types Models Mobile-phone All-categories (sampled)
AUC\blacktriangleleft Acc Prec Recall F1 FNR(-) AUC\blacktriangleleft Acc Prec Recall F1 FNR(-)
Rep DSSM§ 0.6246 0.5304 0.5300 0.9977 0.6923 0.9953 0.8219 0.7686 0.7686 1.0000 0.8691 1.0000
Rep MVLSTM 0.8602 0.7752 0.7433 0.8791 0.8055 0.3416 0.7877 0.8052 0.8090 0.9786 0.8857 0.7802
Rep ARC-I 0.8343 0.7580 0.7210 0.8858 0.7949 0.3857 0.6919 0.7809 0.7814 0.9941 0.8750 0.9388
Int DRMM 0.6720 0.5771 0.5642 0.8850 0.6891 0.7692 0.6781 0.7765 0.7803 0.9888 0.8722 0.9401
Int MatchPyramid 0.7826 0.6806 0.6422 0.8957 0.7481 0.5615 0.7859 0.7911 0.7961 0.9802 0.8786 0.8475
Int ARC-II 0.8128 0.7411 0.6982 0.9001 0.7864 0.4377 0.7606 0.7878 0.7871 0.9938 0.8784 0.9076
Int K-NRM 0.7462 0.6438 0.6102 0.9057 0.7291 0.6510 0.7314 0.7799 0.7853 0.9837 0.8733 0.9081
Int DRMM-TKS 0.7678 0.6738 0.6417 0.8692 0.7383 0.5462 0.7793 0.7943 0.8053 0.9672 0.8789 0.7893
Int Conv-KNRM 0.8369 0.7575 0.7339 0.8504 0.7879 0.3469 0.8029 0.7894 0.7896 0.991 0.8789 0.8913
Int ESIM§ 0.8056 0.7520 0.7526 0.8029 0.7769 0.3073 0.7987 0.7580 0.7580 1.0000 0.8623 1.0000
Both Duet 0.7693 0.6023 0.5731 0.9752 0.7219 0.8173 0.7968 0.7812 0.7806 0.9965 0.8754 0.9458
Triple HG4SM 0.8785 0.8013 0.7711 0.8883 0.8256 0.2966 0.8758 0.8559 0.8956 0.9206 0.9079 0.3625
Table 2: Ablation study. Best results are bolded. ‘Rep’, ‘Int’ and ‘HIN’ are single-component models, and ‘Rep+Int’, ‘Int+HIN’ and ‘Rep+HIN’ both-component models. Bolded numbers are the best result under the current metric. Note: here all-categories is complete 170 million data different from the 10 million data in the Table 1.
Orders Models Mobile-Phone All-categories (complete)
AUC\blacktriangleleft Acc Prec Recall F1 FNR(-) AUC\blacktriangleleft Acc Prec Recall F1 FNR(-)
1st Rep 0.8537 0.7719 0.7368 0.8856 0.8044 0.3560 0.8067 0.7978 0.8855 0.8474 0.8660 0.3697
1st Int 0.8595 0.7727 0.7405 0.8787 0.8037 0.3464 0.8289 0.8534 0.8770 0.9421 0.9084 0.4459
2nd HIN 0.8430 0.7929 0.7667 0.8751 0.8173 0.2995 0.8500 0.8592 0.8852 0.9392 0.9114 0.4110
1st Rep+Int 0.8638 0.7795 0.7481 0.8797 0.8086 0.3331 0.8776 0.8503 0.8707 0.9465 0.9070 0.4743
1st & 2nd Int+HIN 0.8761 0.8009 0.7799 0.8691 0.8221 0.2758 0.8824 0.8585 0.8941 0.9263 0.9099 0.3705
1st & 2nd Rep+HIN 0.8656 0.7956 0.7708 0.8736 0.8190 0.2922 0.8750 0.8576 0.8928 0.9266 0.9094 0.3753
1st & 2nd HG4SM 0.8786 0.8025 0.7790 0.8754 0.8244 0.2794 0.8862 0.8597 0.8949 0.9270 0.9107 0.3673

Comparisons

We compare HG4SM with eleven state-of-the-art deep semantic matching methods using the in-house e-commerce search log data. The results are shown in Table 1. Because some baseline methods (e.g., DRMM and ESIM) have relatively high time complexities, we sample ten million training data from the all-categories dataset. For fairness, all methods including our HG4SM are trained on these data.

As shown in Table 1, HG4SM nearly always outperforms the other methods compared, across all six metrics. More specifically,

  • Compared to the second best method, HG4SM obtains 1.8% (and 5.4%) gains under AUC in the mobile-phone- dataset (and all-categories dataset). Furthermore, HG4SM achieves the best (smallest) FNR on both these datasets. This implies that HG4SM has a high discrimination power on negative examples, so that it can return a list of satisfactory items which are relevant to the user’s shopping needs.

  • The collected training data have imbalanced classes (i.e. the positive examples are far more than the negative examples), making model learning a challenge. Most of the methods compared, such as DSSM, are vulnerable to the class imbalance and often fail to correctly estimate many testing examples in this case. Fortunately, our HG4SM learns explicit node semantics benefiting from the neighboring node’s effect, so that it has a robust performance even though the training data are highly imbalanced.

Ablation Study

To further examine the importance of each component in the HG4SM model, we remove one or two components from HG4SM at a time and examine how the components affect the overall performance.We have the following empirical observation and analysis of the results, shown in Table 2:

  • In general, for the three submodels of HG4SM (including the representation-focused component, the interaction-focused component, and both), the both-component setting often outperforms each single-component setting but worse than the triple-component setting by introducing HIN (i.e. HG4SM). It demonstrates that introducing HIN helps the model learn comprehensive knowledge and thus gives an explicit estimation.

  • The introduction of second-order relevance modeling can provide stable improvement (e.g. 1.0%\sim1.6% of AUC gains) to every first-order solution. This demonstrates that applying HIN to the semantic match can effectively exploit the neighboring nodes’ information contained in user-behavior networks, benefiting the final relevance estimation between central nodes.

Online Performance

To further evaluate HG4SM’s performance in the real search scene, we deploy it to an online e-commerce search platform and report its A/B test results in Table 3. Four widely-used online business metrics are adopted: 1) Conversion Rate (CVR): Total order number / total click number; 2) User Conversion Rate (UCVR): Total order number / search user number; 3) UV-value: Total Gross Merchandise Volume / total user number; and 4) Revenue Per Mile (RPM): 1000 * search GMV / search number.

Table 3: HG4SM’s online performance under price sort mode and default sort mode. All reported results are observed in at least ten consecutive days. The online baseline method is a feature-based GBDT model trained on a human-labeled dataset containing 73 hand-crafting features.
Metrics Price Sort Default Sort
Improvement P-value Improvement P-value
UV-value 5.713% 3.20e-2 0.5013% 1.10e-1
UCVR 1.540% 7.81e-2 0.3058% 1.75e-2
CVR 1.829% 1.01e-2 0.1218% 1.60e-1
RPM 5.587% 3.03e-2 0.6886% 2.32e-2

The results of A/B tests show that our HG4SM outperforms the existing online DNN model in this platform on all of the business metrics. For example, HG4SM improves 5.7% and 0.5% of UV-values under both price sort and default sort. It indicates that 1) our HG4SM model has a low time complexity, which makes it easy to cooperate with other online serial models, and 2) HG4SM provides accurate relevance estimation between queries and items, so that it provides users efficient and intelligent search experiences.

Conclusion

In this paper, we studied the semantic relevance match problem in e-commerce search. In reference to the previous semantic match research using first-order relevance modeling, we proposed a novel idea to combine first-order and second-order relevance match. Based on this new idea, we employed a heterogeneous network embedding to exploit the potential context information in the “query-item” heterogeneous bipartite network. Compared to the current state-of-the-art methods, our novel HG4SM model showed a robust prediction performance. The ablation study verified that the addition of the second-order relevance modeling can significantly improve the performance of the method using the traditional first-order relevance alone. Finally, we applied HG4SM to our in-house e-commerce platform by deploying it to the online search system, which significantly improved the user’s search experience.

References

  • Berger et al. (2000) Berger, A.; Caruana, R.; Cohn, D.; Freitag, D.; and Mittal, V. 2000. Bridging the lexical chasm: statistical approaches to answer-finding. In Proceedings of the 23rd annual international ACM SIGIR conference on Research and development in information retrieval, 192–199.
  • Chen et al. (2017) Chen, Q.; Zhu, X.; Ling, Z.-H.; Wei, S.; Jiang, H.; and Inkpen, D. 2017. Enhanced LSTM for Natural Language Inference. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 1657–1668.
  • Dagan, Glickman, and Magnini (2005) Dagan, I.; Glickman, O.; and Magnini, B. 2005. The PASCAL recognising textual entailment challenge. In Machine Learning Challenges Workshop, 177–190. Springer.
  • Devlin et al. (2019) Devlin, J.; Chang, M.-W.; Lee, K.; and Toutanova, K. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 4171–4186.
  • Dolan et al. (2004) Dolan, W.; Quirk, C.; Brockett, C.; and Dolan, B. 2004. Unsupervised construction of large paraphrase corpora: Exploiting massively parallel news sources. In Proceedings of the 20th International Conference on Computational Linguistics.
  • Fan et al. (2017) Fan, Y.; Pang, L.; Hou, J.; Guo, J.; Lan, Y.; and Cheng, X. 2017. Matchzoo: A toolkit for deep text matching. In Neu-IR: The SIGIR 2017 Workshop on Neural Information Retrieval.
  • Guo et al. (2016) Guo, J.; Fan, Y.; Ai, Q.; and Croft, W. B. 2016. A deep relevance matching model for ad-hoc retrieval. In Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, 55–64.
  • Hu et al. (2014) Hu, B.; Lu, Z.; Li, H.; and Chen, Q. 2014. Convolutional Neural Network Architectures for Matching Natural Language Sentences. In Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, December 8-13 2014, Montreal, Quebec, Canada, 2042–2050.
  • Huang et al. (2013) Huang, P.-S.; He, X.; Gao, J.; Deng, L.; Acero, A.; and Heck, L. 2013. Learning deep structured semantic models for web search using clickthrough data. In Proceedings of the 22nd ACM international conference on Information & Knowledge Management, 2333–2338.
  • Jiang et al. (2019) Jiang, Y.; Shang, Y.; Li, R.; Yang, W.-Y.; Tang, G.; Ma, C.; Xiao, Y.; and Zhao, E. 2019. A unified neural network approach to e-commerce relevance learning. In Proceedings of the 1st International Workshop on Deep Learning Practice for High-Dimensional Sparse Data, 10. ACM.
  • Li (2011) Li, H. 2011. Learning to rank for information retrieval and natural language processing. Synthesis Lectures on Human Language Technologies 4(1): 1–113.
  • Mikolov et al. (2013) Mikolov, T.; Chen, K.; Corrado, G.; and Dean, J. 2013. Efficient Estimation of Word Representations in Vector Space. In 1st International Conference on Learning Representations, ICLR 2013, Scottsdale, Arizona, USA, May 2-4, 2013, Workshop Track Proceedings.
  • Mitra, Diaz, and Craswell (2017) Mitra, B.; Diaz, F.; and Craswell, N. 2017. Learning to Match using Local and Distributed Representations of Text for Web Search. In Proceedings of the 26th International Conference on World Wide Web, WWW 2017, Perth, Australia, April 3-7, 2017, 1291–1299. ACM.
  • Nigam et al. (2019) Nigam, P.; Song, Y.; Mohan, V.; Lakshman, V.; Ding, W.; Shingavi, A.; Teo, C. H.; Gu, H.; and Yin, B. 2019. Semantic product search. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2876–2885.
  • Pang et al. (2016a) Pang, L.; Lan, Y.; Guo, J.; Xu, J.; Wan, S.; and Cheng, X. 2016a. Text matching as image recognition. In Proceedings of the 30th AAAI Conference on Artificial Intelligence.
  • Pang et al. (2016b) Pang, L.; Lan, Y.; Guo, J.; Xu, J.; Wan, S.; and Cheng, X. 2016b. Text Matching as Image Recognition. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, February 12-17, 2016, Phoenix, Arizona, USA, 2793–2799.
  • Sun et al. (2019) Sun, Y.; Wang, S.; Li, Y.; Feng, S.; Chen, X.; Zhang, H.; Tian, X.; Zhu, D.; Tian, H.; and Wu, H. 2019. ERNIE: Enhanced representation through knowledge integration. arXiv preprint arXiv:1904.09223 .
  • Tang et al. (2015) Tang, J.; Qu, M.; Wang, M.; Zhang, M.; Yan, J.; and Mei, Q. 2015. Line: Large-scale information network embedding. In Proceedings of the 24th international conference on world wide web, 1067–1077.
  • Vaswani et al. (2017) Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A. N.; Kaiser, Ł.; and Polosukhin, I. 2017. Attention is all you need. In Advances in neural information processing systems, 5998–6008.
  • Wan et al. (2016a) Wan, S.; Lan, Y.; Guo, J.; Xu, J.; Pang, L.; and Cheng, X. 2016a. A deep architecture for semantic matching with multiple positional sentence representations. In Proceedings of the 30th AAAI Conference on Artificial Intelligence.
  • Wan et al. (2016b) Wan, S.; Lan, Y.; Guo, J.; Xu, J.; Pang, L.; and Cheng, X. 2016b. A Deep Architecture for Semantic Matching with Multiple Positional Sentence Representations. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, February 12-17, 2016, Phoenix, Arizona, USA, 2835–2841.
  • Wang, Cui, and Zhu (2016) Wang, D.; Cui, P.; and Zhu, W. 2016. Structural deep network embedding. In Proceedings of the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining, 1225–1234.
  • Xiong et al. (2017) Xiong, C.; Dai, Z.; Callan, J.; Liu, Z.; and Power, R. 2017. End-to-End Neural Ad-hoc Ranking with Kernel Pooling. In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, Shinjuku, Tokyo, Japan, August 7-11, 2017, 55–64. ACM.