2000 character limit reached
Neural Variational Learning for Grounded Language Acquisition (2107.14593v1)
Published 20 Jul 2021 in cs.CL, cs.AI, cs.LG, and cs.RO
Abstract: We propose a learning system in which language is grounded in visual percepts without specific pre-defined categories of terms. We present a unified generative method to acquire a shared semantic/visual embedding that enables the learning of language about a wide range of real-world objects. We evaluate the efficacy of this learning by predicting the semantics of objects and comparing the performance with neural and non-neural inputs. We show that this generative approach exhibits promising results in language grounding without pre-specifying visual categories under low resource settings. Our experiments demonstrate that this approach is generalizable to multilingual, highly varied datasets.