pith. machine review for the scientific record. sign in

arxiv: 1710.07177 · v1 · submitted 2017-10-19 · 💻 cs.CL · cs.CV

Recognition: unknown

Findings of the Second Shared Task on Multimodal Machine Translation and Multilingual Image Description

Authors on Pith no claims yet
classification 💻 cs.CL cs.CV
keywords imagemultimodaltaskdescriptionmultilingualsystemstranslationmachine
0
0 comments X
read the original abstract

We present the results from the second shared task on multimodal machine translation and multilingual image description. Nine teams submitted 19 systems to two tasks. The multimodal translation task, in which the source sentence is supplemented by an image, was extended with a new language (French) and two new test sets. The multilingual image description task was changed such that at test time, only the image is given. Compared to last year, multimodal systems improved, but text-only systems remain competitive.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 1 Pith paper

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. A Multimodal Dataset for Visually Grounded Ambiguity in Machine Translation

    cs.CL 2026-05 unverdicted novelty 6.0

    VIDA provides 2,500 visually-dependent ambiguous MT instances and LLM-judge metrics; chain-of-thought SFT improves disambiguation accuracy over standard SFT, especially out-of-distribution.