Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Systematic Investigation of Strategies Tailored for Low-Resource Settings for Low-Resource Dependency Parsing (2201.11374v2)

Published 27 Jan 2022 in cs.CL

Abstract: In this work, we focus on low-resource dependency parsing for multiple languages. Several strategies are tailored to enhance performance in low-resource scenarios. While these are well-known to the community, it is not trivial to select the best-performing combination of these strategies for a low-resource language that we are interested in, and not much attention has been given to measuring the efficacy of these strategies. We experiment with 5 low-resource strategies for our ensembled approach on 7 Universal Dependency (UD) low-resource languages. Our exhaustive experimentation on these languages supports the effective improvements for languages not covered in pretrained models. We show a successful application of the ensembled system on a truly low-resource language Sanskrit. The code and data are available at: https://github.com/Jivnesh/SanDP

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Jivnesh Sandhan (12 papers)
  2. Laxmidhar Behera (31 papers)
  3. Pawan Goyal (170 papers)
Github Logo Streamline Icon: https://streamlinehq.com

GitHub