Stochastic Optimization of Text Set Generation for Learning Multiple Query Intent Representations
CIKM 2022
.So, there are some details not mentioned in this paperLearning multiple intent representations for queries has potential applications in facet generation, document ranking, search result diversification,and search explanation. The state-of-the-art model for this task assumes that there is a sequence of intent representations. In this paper, we argue that the model should not be penalized as long as it generates an accurate and complete set of intent representations. Based on this intuition,we propose a stochastic permutation invariant approach for optimizing such networks. We extrinsically evaluate the proposed approach on a facet generation task and demonstrate significant improvements compared to competitive baselines. Our analysis shows that the proposed permutation invariant
approach has the highest impact on queries with more potential intents.
NMIR
ignores the premutation invariance nature of query intents(loss function)
We propose PINMIR
, looks at the query intents as a set rather than a sequence(use permutation invariant loss)
Limitations of NMIR
In PINMIR
, we no longer need the intent-cluster matching algorithm, since the order of generated intents do not matter.
Note: just no longer need intent-cluster mapping, not no longer need clustering! (my personal view)
- we don’t care the order of generated facet descriptions, so we can use one document cluster to generate any facet description
First, we need to define a permutation invariant loss function for training the model
Common permutation invariant loss functions include Chamfer loss and Hungarian loss
- Chamfer loss is based on Chamfer distance and it’s not applicable to our work due to the design of decoder for text generation
We extend the Hungarian loss for text set generation. The proposed loss function for a query q i q_i qi is:
Propose to use a stochastic variation of this loss that instead of iterating over all possible permutations, takes samples from the permutation set and computes the loss based on the sampled query intent sequences
Although the order does not matters between each intent description, it matters within the intent description
we modify the standard decoder architecture in transformer.
Data: MIMICS-Click
variable / max: the number of facets for each query
F1
, BLEU 4-gram
, and Set BERT-Score
(variable)