Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Low-Rank Hidden State Embeddings for Viterbi Sequence Labeling (1708.00553v1)

Published 2 Aug 2017 in cs.CL

Abstract: In textual information extraction and other sequence labeling tasks it is now common to use recurrent neural networks (such as LSTM) to form rich embedded representations of long-term input co-occurrence patterns. Representation of output co-occurrence patterns is typically limited to a hand-designed graphical model, such as a linear-chain CRF representing short-term Markov dependencies among successive labels. This paper presents a method that learns embedded representations of latent output structure in sequence data. Our model takes the form of a finite-state machine with a large number of latent states per label (a latent variable CRF), where the state-transition matrix is factorized---effectively forming an embedded representation of state-transitions capable of enforcing long-term label dependencies, while supporting exact Viterbi inference over output labels. We demonstrate accuracy improvements and interpretable latent structure in a synthetic but complex task based on CoNLL named entity recognition.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Dung Thai (7 papers)
  2. Shikhar Murty (19 papers)
  3. Trapit Bansal (13 papers)
  4. Luke Vilnis (20 papers)
  5. David Belanger (25 papers)
  6. Andrew McCallum (132 papers)

Summary

We haven't generated a summary for this paper yet.