Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multi-channel Reverse Dictionary Model (1912.08441v2)

Published 18 Dec 2019 in cs.CL, cs.AI, and cs.IR

Abstract: A reverse dictionary takes the description of a target word as input and outputs the target word together with other words that match the description. Existing reverse dictionary methods cannot deal with highly variable input queries and low-frequency target words successfully. Inspired by the description-to-word inference process of humans, we propose the multi-channel reverse dictionary model, which can mitigate the two problems simultaneously. Our model comprises a sentence encoder and multiple predictors. The predictors are expected to identify different characteristics of the target word from the input query. We evaluate our model on English and Chinese datasets including both dictionary definitions and human-written descriptions. Experimental results show that our model achieves the state-of-the-art performance, and even outperforms the most popular commercial reverse dictionary system on the human-written description dataset. We also conduct quantitative analyses and a case study to demonstrate the effectiveness and robustness of our model. All the code and data of this work can be obtained on https://github.com/thunlp/MultiRD.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Lei Zhang (1691 papers)
  2. Fanchao Qi (33 papers)
  3. Zhiyuan Liu (433 papers)
  4. Yasheng Wang (91 papers)
  5. Qun Liu (231 papers)
  6. Maosong Sun (337 papers)
Citations (42)

Summary

  • The paper introduces a multi-channel model that integrates a bidirectional LSTM with attention and multiple predictors to enhance reverse dictionary retrieval.
  • It leverages internal linguistic cues and external semantic databases to overcome challenges in recognizing low-frequency words.
  • Experimental findings show improved median rank, top-n accuracy, and reduced variance compared to existing models on diverse datasets.

A Formal Analysis of the Multi-channel Reverse Dictionary Model

The paper "Multi-channel Reverse Dictionary Model" authored by Lei Zhang et al. presents an advanced approach to the task of reverse dictionary retrieval. A reverse dictionary provides an important NLP utility by mapping descriptions to corresponding words. This task, while conceptually simple, is challenged by the variability of input queries and the low-frequency nature of certain target words, an obstacle not efficiently handled by existing models.

Architecture of the Proposed Model

The proposed solution involves a multi-channel model composed of a sentence encoder and multiple predictors. The sentence encoder leverages a bidirectional LSTM with an attention mechanism to embed the input query into a semantic space. Such a framework allows significant fidelity in capturing sentence-level nuances which are crucial when dealing with variable queries.

The novel aspect of the model lies in its multi-channel architecture, which includes:

  • Internal Channels: Two internal predictors discern the part-of-speech (POS) tags and morphemes, imparting insight into intrinsic word properties.
  • External Channels: These utilize external databases such as a word taxonomy system and HowNet for sememes, providing category and sememe predictors which incorporate external semantic knowledge.

By diversifying prediction with such distinct information channels, the model can robustly handle the recognition of less frequent words and accommodate a broader range of input variability—aligning more closely with human inferential processes.

Experimental Evaluation and Results

Empirical evaluation on both English and Chinese datasets portrays the model as superior to earlier techniques. Specifically noteworthy is the model’s performance against web-based commercial systems such as OneLook. Although OneLook excels with exact definitions, the multi-channel model demonstrates superior accuracy in dealing with more human-like, varied descriptions—a vital requirement for true reverse dictionary tasks. It achieves better performance indicators across various datasets:

  • Consistent improvement in median rank scores.
  • Superior accuracy at top 1/10/100 retrieval positions.
  • Lower rank variance, indicating reliability across trials.

Implications and Future Outlook

The implications of these findings are substantial. By improving reverse dictionary tasks, the proposed model not only aids writers and language learners to resolve the ‘tip-of-the-tongue’ dilemma but also enhances entity mapping in broader NLP tasks such as question answering and information retrieval.

Looking forward, the research indicates pathways for even more advanced systems. Hybrid models that integrate text-matching techniques can enhance retrieval performance in single-word input scenarios. Moreover, extending the model’s capabilities into cross-lingual environments holds promise—for instance, enabling multi-language semantic queries.

In conclusion, the multi-channel reverse dictionary model stands as a robust enhancement in reverse dictionary systems, offering state-of-the-art performance across varied and challenging datasets. The methodological framework presented sets the stage for future work expanding the comprehensiveness and accuracy of NLP applications in real-world settings.

Github Logo Streamline Icon: https://streamlinehq.com