Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards preserving word order importance through Forced Invalidation (2304.05221v1)

Published 11 Apr 2023 in cs.CL

Abstract: Large pre-trained LLMs such as BERT have been widely used as a framework for natural language understanding (NLU) tasks. However, recent findings have revealed that pre-trained LLMs are insensitive to word order. The performance on NLU tasks remains unchanged even after randomly permuting the word of a sentence, where crucial syntactic information is destroyed. To help preserve the importance of word order, we propose a simple approach called Forced Invalidation (FI): forcing the model to identify permuted sequences as invalid samples. We perform an extensive evaluation of our approach on various English NLU and QA based tasks over BERT-based and attention-based models over word embeddings. Our experiments demonstrate that Forced Invalidation significantly improves the sensitivity of the models to word order.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Hadeel Al-Negheimish (3 papers)
  2. Pranava Madhyastha (37 papers)
  3. Alessandra Russo (48 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.