-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathpublications.bib
2751 lines (2522 loc) · 150 KB
/
publications.bib
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
@misc{ferrando2024primerinnerworkingstransformerbased,
title={A Primer on the Inner Workings of Transformer-based Language Models},
author={Javier Ferrando and Gabriele Sarti and Arianna Bisazza and Marta R. Costa-jussà},
year={2024},
eprint={2405.00208},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2405.00208},
}
@misc{tsiamas2024pushinglimitszeroshotendtoend,
title={Pushing the Limits of Zero-shot End-to-End Speech Translation},
author={Ioannis Tsiamas and Gerard I. Gállego and José A. R. Fonollosa and Marta R. Costa-jussà},
year={2024},
eprint={2402.10422},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2402.10422},
}
@inproceedings{alastruey-etal-2024-speechalign-framework,
title = "{S}peech{A}lign: A Framework for Speech Translation Alignment Evaluation",
author = "Alastruey, Belen and
Sant, Aleix and
G{\'a}llego, Gerard I. and
Dale, David and
Costa-juss{\`a}, Marta R.",
editor = "Calzolari, Nicoletta and
Kan, Min-Yen and
Hoste, Veronique and
Lenci, Alessandro and
Sakti, Sakriani and
Xue, Nianwen",
booktitle = "Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)",
month = may,
year = "2024",
address = "Torino, Italia",
publisher = "ELRA and ICCL",
url = "https://aclanthology.org/2024.lrec-main.1316",
pages = "15137--15146",
abstract = "Speech-to-Speech and Speech-to-Text translation are currently dynamic areas of research. In our commitment to advance these fields, we present SpeechAlign, a framework designed to evaluate the underexplored field of source-target alignment in speech models. The SpeechAlign framework has two core components. First, to tackle the absence of suitable evaluation datasets, we introduce the Speech Gold Alignment dataset, built upon a English-German text translation gold alignment dataset. Secondly, we introduce two novel metrics, Speech Alignment Error Rate (SAER) and Time-weighted Speech Alignment Error Rate (TW-SAER), which enable the evaluation of alignment quality within speech models. While the former gives equal importance to each word, the latter assigns weights based on the length of the words in the speech signal. By publishing SpeechAlign we provide an accessible evaluation framework for model assessment, and we employ it to benchmark open-source Speech Translation models. In doing so, we contribute to the ongoing research progress within the fields of Speech-to-Speech and Speech-to-Text translation.",
}
@inproceedings{costa-jussa-etal-2023-toxicity,
title = "Toxicity in Multilingual Machine Translation at Scale",
author = "Costa-juss{\`a}, Marta and
Smith, Eric and
Ropers, Christophe and
Licht, Daniel and
Maillard, Jean and
Ferrando, Javier and
Escolano, Carlos",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2023",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.findings-emnlp.642",
doi = "10.18653/v1/2023.findings-emnlp.642",
pages = "9570--9586",
abstract = "Machine Translation systems can produce different types of errors, some of which are characterized as critical or catastrophic due to the specific negative impact that they can have on users. In this paper we focus on one type of critical error: added toxicity. We evaluate and analyze added toxicity when translating a large evaluation dataset (HOLISTICBIAS, over 472k sentences, covering 13 demographic axes) from English into 164 languages. An automatic toxicity evaluation shows that added toxicity across languages varies from 0{\%} to 5{\%}. The output languages with the most added toxicity tend to be low-resource ones, and the demographic axes with the most added toxicity include sexual orientation, gender and sex, and ability. We also perform human evaluation on a subset of 8 translation directions, confirming the prevalence of true added toxicity. We use a measurement of the amount of source contribution to the translation, where a low source contribution implies hallucination, to interpret what causes toxicity. Making use of the input attributions allows us to explain toxicity, because the source contributions significantly correlate with toxicity for 84{\%} of languages studied. Given our findings, our recommendations to reduce added toxicity are to curate training data to avoid mistranslations, mitigate hallucination and check unstable translations.",
}
@inproceedings{tsiamas-etal-2023-segaugment,
title = "{S}eg{A}ugment: Maximizing the Utility of Speech Translation Data with Segmentation-based Augmentations",
author = "Tsiamas, Ioannis and
Fonollosa, Jos{\'e} and
Costa-juss{\`a}, Marta",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2023",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.findings-emnlp.574",
doi = "10.18653/v1/2023.findings-emnlp.574",
pages = "8569--8588",
abstract = "End-to-end Speech Translation is hindered by a lack of available data resources. While most of them are based on documents, a sentence-level version is available, which is however single and static, potentially impeding the usefulness of the data. We propose a new data augmentation strategy, SegAugment, to address this issue by generating multiple alternative sentence-level versions of a dataset. Our method utilizes an Audio Segmentation system, which re-segments the speech of each document with different length constraints, after which we obtain the target text via alignment methods. Experiments demonstrate consistent gains across eight language pairs in MuST-C, with an average increase of 2.5 BLEU points, and up to 5 BLEU for low-resource scenarios in mTEDx. Furthermore, when combined with a strong system, SegAugment obtains state-of-the-art results in MuST-C. Finally, we show that the proposed method can also successfully augment sentence-level datasets, and that it enables Speech Translation models to close the gap between the manual and automatic segmentation at inference time.",
}
@inproceedings{ferrando-etal-2023-automating,
title = "Automating Behavioral Testing in Machine Translation",
author = "Ferrando, Javier and
Sperber, Matthias and
Setiawan, Hendra and
Telaar, Dominic and
Hasan, Sa{\v{s}}a",
editor = "Koehn, Philipp and
Haddow, Barry and
Kocmi, Tom and
Monz, Christof",
booktitle = "Proceedings of the Eighth Conference on Machine Translation",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.wmt-1.97",
doi = "10.18653/v1/2023.wmt-1.97",
pages = "1014--1030",
abstract = "Behavioral testing in NLP allows fine-grained evaluation of systems by examining their linguistic capabilities through the analysis of input-output behavior. Unfortunately, existing work on behavioral testing in Machine Translation (MT) is currently restricted to largely handcrafted tests covering a limited range of capabilities and languages. To address this limitation, we propose to use Large Language Models (LLMs) to generate a diverse set of source sentences tailored to test the behavior of MT models in a range of situations. We can then verify whether the MT model exhibits the expected behavior through matching candidate sets that are also generated using LLMs. Our approach aims to make behavioral testing of MT systems practical while requiring only minimal human effort. In our experiments, we apply our proposed evaluation framework to assess multiple available MT systems, revealing that while in general pass-rates follow the trends observable from traditional accuracy-based metrics, our method was able to uncover several important differences and potential bugs that go unnoticed when relying only on accuracy.",
}
@misc{carrino2023promotinggeneralizedcrosslingualquestion,
title={Promoting Generalized Cross-lingual Question Answering in Few-resource Scenarios via Self-knowledge Distillation},
author={Casimiro Pio Carrino and Carlos Escolano and José A. R. Fonollosa},
year={2023},
eprint={2309.17134},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2309.17134},
}
@misc{voita2023neuronslargelanguagemodels,
title={Neurons in Large Language Models: Dead, N-gram, Positional},
author={Elena Voita and Javier Ferrando and Christoforos Nalmpantis},
year={2023},
eprint={2309.04827},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2309.04827},
}
@inproceedings{sant23_interspeech,
author={Gerard Sant and Carlos Escolano},
title={{Analysis of Acoustic information in End-to-End Spoken Language Translation}},
year=2023,
booktitle={Proc. INTERSPEECH 2023},
pages={52--56},
doi={10.21437/Interspeech.2023-2050},
issn={2958-1796}
}
@inproceedings{torrero-etal-2023-talp,
title = "{TALP}-{UPC} at {P}rob{S}um 2023: Fine-tuning and Data Augmentation Strategies for {NER}",
author = "Torrero, Neil and
Sant, Gerard and
Escolano, Carlos",
editor = "Demner-fushman, Dina and
Ananiadou, Sophia and
Cohen, Kevin",
booktitle = "The 22nd Workshop on Biomedical Natural Language Processing and BioNLP Shared Tasks",
month = jul,
year = "2023",
address = "Toronto, Canada",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.bionlp-1.48",
doi = "10.18653/v1/2023.bionlp-1.48",
pages = "497--502",
abstract = "This paper describes the submission of the TALP-UPC team to the Problem List Summarization task from the BioNLP 2023 workshop. This task consists of automatically extracting a list of health issues from the e-health medical record of a given patient. Our submission combines additional steps of data annotationwith finetuning of BERT pre-trained language models. Our experiments focus on the impact of finetuning on different datasets as well as the addition of data augmentation techniques to delay overfitting.",
}
@inproceedings{tsiamas-etal-2023-speech,
title = "Speech Translation with Foundation Models and Optimal Transport: {UPC} at {IWSLT}23",
author = "Tsiamas, Ioannis and
I. G{\'a}llego, Gerard and
Fonollosa, Jose and
R. Costa-juss{\'a}, Marta",
editor = "Salesky, Elizabeth and
Federico, Marcello and
Carpuat, Marine",
booktitle = "Proceedings of the 20th International Conference on Spoken Language Translation (IWSLT 2023)",
month = jul,
year = "2023",
address = "Toronto, Canada (in-person and online)",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.iwslt-1.38",
doi = "10.18653/v1/2023.iwslt-1.38",
pages = "397--410",
abstract = "This paper describes the submission of the UPC Machine Translation group to the IWSLT 2023 Offline Speech Translation task. Our Speech Translation systems utilize foundation models for speech (wav2vec 2.0) and text (mBART50). We incorporate a Siamese pretraining step of the speech and text encoders with CTC and Optimal Transport, to adapt the speech representations to the space of the text model, thus maximizing transfer learning from MT. After this pretraining, we fine-tune our system end-to-end on ST, with Cross Entropy and Knowledge Distillation. Apart from the available ST corpora, we create synthetic data with SegAugment to better adapt our models to the custom segmentations of the IWSLT test sets. Our best single model obtains 31.2 BLEU points on MuST-C tst-COMMON, 29.8 points on IWLST.tst2020 and 33.4 points on the newly released IWSLT.ACLdev2023.",
}
@inproceedings{ferrando-etal-2023-explaining,
title = "Explaining How Transformers Use Context to Build Predictions",
author = "Ferrando, Javier and
G{\'a}llego, Gerard I. and
Tsiamas, Ioannis and
Costa-juss{\`a}, Marta R.",
editor = "Rogers, Anna and
Boyd-Graber, Jordan and
Okazaki, Naoaki",
booktitle = "Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = jul,
year = "2023",
address = "Toronto, Canada",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.acl-long.301",
doi = "10.18653/v1/2023.acl-long.301",
pages = "5486--5513",
abstract = "Language Generation Models produce words based on the previous context. Although existing methods offer input attributions as explanations for a model{'}s prediction, it is still unclear how prior words affect the model{'}s decision throughout the layers. In this work, we leverage recent advances in explainability of the Transformer and present a procedure to analyze models for language generation. Using contrastive examples, we compare the alignment of our explanations with evidence of the linguistic phenomena, and show that our method consistently aligns better than gradient-based and perturbation-based baselines. Then, we investigate the role of MLPs inside the Transformer and show that they learn features that help the model predict words that are grammatically acceptable. Lastly, we apply our method to Neural Machine Translation models, and demonstrate that they generate human-like source-target alignments for building predictions.",
}
@inproceedings{10208355,
author={Tarrés, Laia and Gállego, Gerard I. and Duarte, Amanda and Torres, Jordi and Giró-i-Nieto, Xavier},
booktitle={2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)},
title={Sign Language Translation from Instructional Videos},
year={2023},
volume={},
number={},
pages={5625-5635},
keywords={Computer vision;Codes;Computational modeling;Conferences;Gesture recognition;Assistive technologies;Benchmark testing},
doi={10.1109/CVPRW59228.2023.00596}}
@inproceedings{10095276,
author={Tsiamas, Ioannis and Gállego, Gerard I. and Fonollosa, José A. R. and Costa-jussà, Marta R.},
booktitle={ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)},
title={Efficient Speech Translation with Dynamic Latent Perceivers},
year={2023},
volume={},
number={},
pages={1-5},
keywords={Training;Costs;Computational modeling;Computer architecture;Signal processing;Transformers;Boosting;Speech Translation;Efficiency;Perceiver},
doi={10.1109/ICASSP49357.2023.10095276},
}
@misc{gilabert2023resetoxrelearningattentionweights,
title={ReSeTOX: Re-learning attention weights for toxicity mitigation in machine translation},
author={Javier García Gilabert and Carlos Escolano and Marta R. Costa-Jussà},
year={2023},
eprint={2305.11761},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2305.11761},
}
@inproceedings{costajussa22occgen,
title = {{OccGen: Selection of Real-world Multilingual Parallel Data Balanced in Gender within Occupations}},
author = {Marta R. Costa-juss{\`a} and
Christine Basta and
Oriol Domingo and
Andr{\'e} Niyongabo Rubungo},
booktitle = {Thirty-sixth Conference on Neural Information Processing Systems Datasets and Benchmarks Track},
month = dec,
year = {2022},
url = {https://openreview.net/forum?id=tTPVefaATp6}
}
@inproceedings{tarres-etal-2022-tackling,
title = "Tackling Low-Resourced Sign Language Translation: {UPC} at {WMT}-{SLT} 22",
author = "Tarres, Laia and
G{\'a}llego, Gerard I. and
Giro-i-nieto, Xavier and
Torres, Jordi",
editor = {Koehn, Philipp and
Barrault, Lo{\"\i}c and
Bojar, Ond{\v{r}}ej and
Bougares, Fethi and
Chatterjee, Rajen and
Costa-juss{\`a}, Marta R. and
Federmann, Christian and
Fishel, Mark and
Fraser, Alexander and
Freitag, Markus and
Graham, Yvette and
Grundkiewicz, Roman and
Guzman, Paco and
Haddow, Barry and
Huck, Matthias and
Jimeno Yepes, Antonio and
Kocmi, Tom and
Martins, Andr{\'e} and
Morishita, Makoto and
Monz, Christof and
Nagata, Masaaki and
Nakazawa, Toshiaki and
Negri, Matteo and
N{\'e}v{\'e}ol, Aur{\'e}lie and
Neves, Mariana and
Popel, Martin and
Turchi, Marco and
Zampieri, Marcos},
booktitle = "Proceedings of the Seventh Conference on Machine Translation (WMT)",
month = dec,
year = "2022",
address = "Abu Dhabi, United Arab Emirates (Hybrid)",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.wmt-1.97",
pages = "994--1000",
abstract = "This paper describes the system developed at the Universitat Polit{\`e}cnica de Catalunya for the Workshop on Machine Translation 2022 Sign Language Translation Task, in particular, for the sign-to-text direction. We use a Transformer model implemented with the Fairseq modeling toolkit. We have experimented with the vocabulary size, data augmentation techniques and pretraining the model with the PHOENIX-14T dataset. Our system obtains 0.50 BLEU score for the test set, improving the organizers{'} baseline by 0.38 BLEU. We remark the poor results for both the baseline and our system, and thus, the unreliability of our findings.",
}
@inproceedings{ferrando-etal-2022-towards,
title = "Towards Opening the Black Box of Neural Machine Translation: Source and Target Interpretations of the Transformer",
author = "Ferrando, Javier and
G{\'a}llego, Gerard I. and
Alastruey, Belen and
Escolano, Carlos and
Costa-juss{\`a}, Marta R.",
editor = "Goldberg, Yoav and
Kozareva, Zornitsa and
Zhang, Yue",
booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2022",
address = "Abu Dhabi, United Arab Emirates",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.emnlp-main.599",
doi = "10.18653/v1/2022.emnlp-main.599",
pages = "8756--8769",
abstract = "In Neural Machine Translation (NMT), each token prediction is conditioned on the source sentence and the target prefix (what has been previously translated at a decoding step). However, previous work on interpretability in NMT has mainly focused solely on source sentence tokens{'} attributions. Therefore, we lack a full understanding of the influences of every input token (source sentence and target prefix) in the model predictions. In this work, we propose an interpretability method that tracks input tokens{'} attributions for both contexts. Our method, which can be extended to any encoder-decoder Transformer-based model, allows us to better comprehend the inner workings of current NMT models. We apply the proposed method to both bilingual and multilingual Transformers and present insights into their behaviour.",
}
@inproceedings{ferrando2022measuring,
title = "Measuring the Mixing of Contextual Information in the Transformer",
author = "Ferrando, Javier and
G{\'a}llego, Gerard I. and
Costa-juss{\`a}, Marta R.",
editor = "Goldberg, Yoav and
Kozareva, Zornitsa and
Zhang, Yue",
booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2022",
address = "Abu Dhabi, United Arab Emirates",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.emnlp-main.595",
doi = "10.18653/v1/2022.emnlp-main.595",
pages = "8698--8714",
abstract = "The Transformer architecture aggregates input information through the self-attention mechanism, but there is no clear understanding of how this information is mixed across the entire model. Additionally, recent works have demonstrated that attention weights alone are not enough to describe the flow of information. In this paper, we consider the whole attention block {--}multi-head attention, residual connection, and layer normalization{--} and define a metric to measure token-to-token interactions within each layer. Then, we aggregate layer-wise interpretations to provide input attribution scores for model predictions. Experimentally, we show that our method, ALTI (Aggregation of Layer-wise Token-to-token Interactions), provides more faithful explanations and increased robustness than gradient-based methods.",
}
@inproceedings{tsiamas22_interspeech,
author={Ioannis Tsiamas and Gerard I. G{\'a}llego and Jos{\'e} A. R. Fonollosa and Marta R. Costa-juss{\`a}},
title={{SHAS: Approaching optimal Segmentation for End-to-End Speech Translation}},
month=sep,
year=2022,
booktitle={Proc. Interspeech 2022},
pages={106--110},
url={https://www.isca-speech.org/archive/interspeech_2022/tsiamas22_interspeech.html},
doi={10.21437/Interspeech.2022-59}
}
@inproceedings{sant2022multiformer,
title = "Multiformer: A Head-Configurable Transformer-Based Model for Direct Speech Translation",
author = "Sant, Gerard and
G{\'a}llego, Gerard I. and
Alastruey, Belen and
Costa-juss{\`a}, Marta Ruiz",
booktitle = "Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Student Research Workshop",
month = jul,
year = "2022",
address = "Hybrid: Seattle, Washington + Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.naacl-srw.34",
pages = "277--284",
abstract = "Transformer-based models have been achieving state-of-the-art results in several fields of Natural Language Processing. However, its direct application to speech tasks is not trivial. The nature of this sequences carries problems such as long sequence lengths and redundancy between adjacent tokens. Therefore, we believe that regular self-attention mechanism might not be well suited for it. Different approaches have been proposed to overcome these problems, such as the use of efficient attention mechanisms. However, the use of these methods usually comes with a cost, which is a performance reduction caused by information loss. In this study, we present the Multiformer, a Transformer-based model which allows the use of different attention mechanisms on each head. By doing this, the model is able to bias the self-attention towards the extraction of more diverse token interactions, and the information loss is reduced. Finally, we perform an analysis of the head contributions, and we observe that those architectures where all heads relevance is uniformly distributed obtain better results. Our results show that mixing attention patterns along the different heads and layers outperforms our baseline by up to 0.7 BLEU.",
}
@inproceedings{costa-jussa-etal-2022-evaluating,
title = "Evaluating Gender Bias in Speech Translation",
author = "Costa-juss{\`a}, Marta R. and
Basta, Christine and
G{\'a}llego, Gerard I.",
booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference",
month = jun,
year = "2022",
address = "Marseille, France",
publisher = "European Language Resources Association",
url = "https://aclanthology.org/2022.lrec-1.230",
pages = "2141--2147",
abstract = "The scientific community is increasingly aware of the necessity to embrace pluralism and consistently represent major and minor social groups. Currently, there are no standard evaluation techniques for different types of biases. Accordingly, there is an urgent need to provide evaluation sets and protocols to measure existing biases in our automatic systems. Evaluating the biases should be an essential step towards mitigating them in the systems. This paper introduces WinoST, a new freely available challenge set for evaluating gender bias in speech translation. WinoST is the speech version of WinoMT, an MT challenge set, and both follow an evaluation protocol to measure gender accuracy. Using an S-Transformer end-to-end speech translation system, we report the gender bias evaluation on four language pairs, and we reveal the inaccuracies in translations generating gender-stereotyped translations.",
}
@inproceedings{tsiamas-etal-2022-pretrained,
title = "Pretrained Speech Encoders and Efficient Fine-tuning Methods for Speech Translation: {UPC} at {IWSLT} 2022",
author = "Tsiamas, Ioannis and
G{\'a}llego, Gerard I. and
Escolano, Carlos and
Fonollosa, Jos{\'e} and
Costa-juss{\`a}, Marta R.",
booktitle = "Proceedings of the 19th International Conference on Spoken Language Translation (IWSLT 2022)",
month = may,
year = "2022",
address = "Dublin, Ireland (in-person and online)",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.iwslt-1.23",
pages = "265--276",
abstract = "This paper describes the submissions of the UPC Machine Translation group to the IWSLT 2022 Offline Speech Translation and Speech-to-Speech Translation tracks. The offline task involves translating English speech to German, Japanese and Chinese text. Our Speech Translation systems are trained end-to-end and are based on large pretrained speech and text models. We use an efficient fine-tuning technique that trains only specific layers of our system, and explore the use of adapter modules for the non-trainable layers. We further investigate the suitability of different speech encoders (wav2vec 2.0, HuBERT) for our models and the impact of knowledge distillation from the Machine Translation model that we use for the decoder (mBART). For segmenting the IWSLT test sets we fine-tune a pretrained audio segmentation model and achieve improvements of 5 BLEU compared to the given segmentation. Our best single model uses HuBERT and parallel adapters and achieves 29.42 BLEU at English-German MuST-C tst-COMMON and 26.77 at IWSLT 2020 test. By ensembling many models, we further increase translation quality to 30.83 BLEU and 27.78 accordingly. Furthermore, our submission for English-Japanese achieves 15.85 and English-Chinese obtains 25.63 BLEU on the MuST-C tst-COMMON sets. Finally, we extend our system to perform English-German Speech-to-Speech Translation with a pretrained Text-to-Speech model.",
}
@inproceedings{alastruey-etal-2022-locality,
title = "On the Locality of Attention in Direct Speech Translation",
author = "Alastruey, Belen and
Ferrando, Javier and
G{\'a}llego, Gerard I. and
Costa-juss{\`a}, Marta R.",
booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop",
month = may,
year = "2022",
address = "Dublin, Ireland",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.acl-srw.32",
pages = "402--412",
abstract = "Transformers have achieved state-of-the-art results across multiple NLP tasks. However, the self-attention mechanism complexity scales quadratically with the sequence length, creating an obstacle for tasks involving long sequences, like in the speech domain. In this paper, we discuss the usefulness of self-attention for Direct Speech Translation. First, we analyze the layer-wise token contributions in the self-attention of the encoder, unveiling local diagonal patterns. To prove that some attention weights are avoidable, we propose to substitute the standard self-attention with a local efficient one, setting the amount of context used based on the results of the analysis. With this approach, our model matches the baseline performance, and improves the efficiency by skipping the computation of those weights that standard attention discards.",
}
@article{escolano2022multilingual,
title={Multilingual machine translation: Deep analysis of language-specific encoder-decoders},
author={Escolano, Carlos and Costa-juss{\`a}, Marta Ruiz and Fonollosa, Jos{\'e} AR},
journal={Journal of Artificial Intelligence Research},
volume={73},
pages={1535--1552},
year={2022},
month=apr,
url = "https://www.jair.org/index.php/jair/article/view/12699",
}
@inproceedings{costajussa2022genderbias,
author={Marta Ruiz Costa-juss{\`a} and Carlos Escolano and Christine Basta and Javier Ferrando and Roser Batlle Roca and Ksenia Kharitonova},
title={Interpreting Gender Bias in Neural Machine Translation: Multilingual Architecture Matters},
booktitle = {Procedings of the 36th AAAI Conference on Artificial Intelligence},
month = feb,
year = "2022",
url={https://www.aaai.org/AAAI22Papers/AISI-2223.CostajussaM.pdf}
}
@misc{domingo2022multitask,
title={A multi-task semi-supervised framework for Text2Graph & Graph2Text},
author={Oriol Domingo and Marta R. Costa-jussà and Carlos Escolano},
year={2022},
month=feb,
journal={arXiv preprint arXiv:2202.06041},
url={https://arxiv.org/abs/2202.06041}
}
@inproceedings{escolano-etal-2021-enabling-zero-shot,
author={Escolano, Carlos and Costa-jussà, Marta R. and Fonollosa, José A. R. and Segura, Carlos},
booktitle={2021 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU)},
title={Enabling Zero-Shot Multilingual Spoken Language Translation with Language-Specific Encoders and Decoders},
year={2021},
month=dec,
volume={},
number={},
pages={694-701},
doi={10.1109/ASRU51503.2021.9688026},
url = "https://ieeexplore.ieee.org/document/9688026",
}
@inproceedings{rafieian-etal-2021-wmt21,
title = "High Frequent In-domain Words Segmentation and Forward Translation for the {WMT}21 Biomedical Task",
author = "Rafieian, Bardia and
Costa-jussa, Marta R.",
booktitle = "Proceedings of the Sixth Conference on Machine Translation",
month = nov,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.wmt-1.87",
pages = "863--867",
abstract = "This paper reports the optimization of using the out-of-domain data in the Biomedical translation task. We firstly optimized our parallel training dataset using the BabelNet in-domain terminology words. Afterward, to increase the training set, we studied the effects of the out-of-domain data on biomedical translation tasks, and we created a mixture of in-domain and out-of-domain training sets and added more in-domain data using forward translation in the English-Spanish task. Finally, with a simple bpe optimization method, we increased the number of in-domain sub-words in our mixed training set and trained the Transformer model on the generated data. Results show improvements using our proposed method.",
}
@inproceedings{escolano-etal-2021-wmt21,
title = "The {TALP}-{UPC} Participation in {WMT}21 News Translation Task: an m{BART}-based {NMT} Approach",
author = "Escolano, Carlos and
Tsiamas, Ioannis and
Basta, Christine and
Ferrando, Javier and
Costa-jussa, Marta R. and
Fonollosa, Jos{\'e} A. R.",
booktitle = "Proceedings of the Sixth Conference on Machine Translation",
month = nov,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.wmt-1.6",
pages = "117--122",
abstract = "This paper describes the submission to the WMT 2021 news translation shared task by the UPC Machine Translation group. The goal of the task is to translate German to French (De-Fr) and French to German (Fr-De). Our submission focuses on fine-tuning a pre-trained model to take advantage of monolingual data. We fine-tune mBART50 using the filtered data, and additionally, we train a Transformer model on the same data from scratch. In the experiments, we show that fine-tuning mBART50 results in 31.69 BLEU for De-Fr and 23.63 BLEU for Fr-De, which increases 2.71 and 1.90 BLEU accordingly, as compared to the model we train from scratch. Our final submission is an ensemble of these two models, further increasing 0.3 BLEU for Fr-De.",
}
@inproceedings{ferrando-costa-jussa-2021-attention-weights,
title = "Attention Weights in Transformer {NMT} Fail Aligning Words Between Sequences but Largely Explain Model Predictions",
author = "Ferrando, Javier and
Costa-juss{\`a}, Marta R.",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2021",
month = nov,
year = "2021",
address = "Punta Cana, Dominican Republic",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.findings-emnlp.39",
doi = "10.18653/v1/2021.findings-emnlp.39",
pages = "434--443",
}
@inproceedings{gallego-etal-2021-iwslt21,
title = "End-to-End Speech Translation with Pre-trained Models and Adapters: {UPC} at {IWSLT} 2021",
author = "G{\'a}llego, Gerard I. and
Tsiamas, Ioannis and
Escolano, Carlos and
Fonollosa, Jos{\'e} A. R. and
Costa-juss{\`a}, Marta R.",
booktitle = "Proceedings of the 18th International Conference on Spoken Language Translation (IWSLT 2021)",
month = aug,
year = "2021",
address = "Bangkok, Thailand (online)",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.iwslt-1.11",
doi = "10.18653/v1/2021.iwslt-1.11",
pages = "110--119",
abstract = "This paper describes the submission to the IWSLT 2021 offline speech translation task by the UPC Machine Translation group. The task consists of building a system capable of translating English audio recordings extracted from TED talks into German text. Submitted systems can be either cascade or end-to-end and use a custom or given segmentation. Our submission is an end-to-end speech translation system, which combines pre-trained models (Wav2Vec 2.0 and mBART) with coupling modules between the encoder and decoder, and uses an efficient fine-tuning technique, which trains only 20{\%} of its total parameters. We show that adding an Adapter to the system and pre-training it, can increase the convergence speed and the final result, with which we achieve a BLEU score of 27.3 on the MuST-C test set. Our final model is an ensemble that obtains 28.22 BLEU score on the same set. Our submission also uses a custom segmentation algorithm that employs pre-trained Wav2Vec 2.0 for identifying periods of untranscribable text and can bring improvements of 2.5 to 3 BLEU score on the IWSLT 2019 test set, as compared to the result with the given segmentation.",
}
@article{alastruey2021efficient,
title={Efficient Transformer for Direct Speech Translation},
author={Belen Alastruey and Gerard I. G{\'a}llego and Marta Ruiz Costa-juss{\`a}},
year={2021},
month=jul,
journal={arXiv preprint arXiv:2107.03069},
url={https://arxiv.org/abs/2107.03069}
}
@inproceedings{barrault-etal-2020-findings-first,
title = "Findings of the {F}irst {S}hared {T}ask on {L}ifelong {L}earning {M}achine {T}ranslation",
author = {Barrault, Lo{\"\i}c and
Biesialska, Magdalena and
Costa-juss{\`a}, Marta R. and
Bougares, Fethi and
Galibert, Olivier},
booktitle = "Proceedings of the Fifth Conference on Machine Translation",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.wmt-1.2",
pages = "56--64",
abstract = "A lifelong learning system can adapt to new data without forgetting previously acquired knowledge. In this paper, we introduce the first benchmark for lifelong learning machine translation. For this purpose, we provide training, lifelong and test data sets for two language pairs: English-German and English-French. Additionally, we report the results of our baseline systems, which we make available to the public. The goal of this shared task is to encourage research on the emerging topic of lifelong learning machine translation.",
}
@inproceedings{biesialska-etal-2020-continual,
title = "Continual Lifelong Learning in Natural Language Processing: A Survey",
author = "Biesialska, Magdalena and
Biesialska, Katarzyna and
Costa-juss{\`a}, Marta R.",
booktitle = "Proceedings of the 28th International Conference on Computational Linguistics",
month = dec,
year = "2020",
address = "Barcelona, Spain (Online)",
publisher = "International Committee on Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.coling-main.574",
doi = "10.18653/v1/2020.coling-main.574",
pages = "6523--6541",
abstract = "Continual learning (CL) aims to enable information systems to learn from a continuous data stream across time. However, it is difficult for existing deep learning architectures to learn a new task without largely forgetting previously acquired knowledge. Furthermore, CL is particularly challenging for language learning, as natural language is ambiguous: it is discrete, compositional, and its meaning is context-dependent. In this work, we look at the problem of CL through the lens of various NLP tasks. Our survey discusses major challenges in CL and current methods applied in neural network models. We also provide a critical review of the existing CL evaluation methods and datasets in NLP. Finally, we present our outlook on future research directions.",
}
@inproceedings{casas-etal-2020-syntax,
title = "Syntax-driven Iterative Expansion Language Models for Controllable Text Generation",
author = "Casas, Noe and
Fonollosa, Jos{\'e} A. R. and
Costa-juss{\`a}, Marta R.",
booktitle = "Proceedings of the Fourth Workshop on Structured Prediction for NLP",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.spnlp-1.1",
doi = "10.18653/v1/2020.spnlp-1.1",
pages = "1--10",
abstract = "The dominant language modeling paradigm handles text as a sequence of discrete tokens. While that approach can capture the latent structure of the text, it is inherently constrained to sequential dynamics for text generation. We propose a new paradigm for introducing a syntactic inductive bias into neural text generation, where the dependency parse tree is used to drive the Transformer model to generate sentences iteratively. Our experiments show that this paradigm is effective at text generation, with quality between LSTMs and Transformers, and comparable diversity, requiring less than half their decoding steps, and its generation process allows direct control over the syntactic constructions of the generated text, enabling the induction of stylistic variations.",
}
@inproceedings{rafieian-costa-jussa-2020-e,
title = "{E}-Commerce Content and Collaborative-based Recommendation using K-Nearest Neighbors and Enriched Weighted Vectors",
author = "Rafieian, Bardia and
Costa-juss{\`a}, Marta R.",
booktitle = "Proceedings of Workshop on Natural Language Processing in E-Commerce",
month = dec,
year = "2020",
address = "Barcelona, Spain",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.ecomnlp-1.1",
pages = "1--10",
abstract = "In this paper, we present two productive and functional recommender methods to improve the ac- curacy of predicting the right product for the user. One proposal is a survey-based recommender system that uses k-nearest neighbors. It recommends products by asking questions from the user, efficiently applying a binary product vector to the product attributes, and processing the request with a minimum error. The second proposal uses an enriched collaborative-based recommender system using enriched weighted vectors. Thanks to the style rules, the enriched collaborative- based method recommends outfits with competitive recommendation quality. We evaluated both of the proposals on a Kaggle fashion-dataset along with iMaterialist and, results show equivalent performance on binary gender and product attributes.",
}
@inproceedings{costa-jussa-de-jorge-2020-fine,
title = "Fine-tuning Neural Machine Translation on Gender-Balanced Datasets",
author = "Costa-juss{\`a}, Marta R. and
de Jorge, Adri{\`a}",
booktitle = "Proceedings of the Second Workshop on Gender Bias in Natural Language Processing",
month = dec,
year = "2020",
address = "Barcelona, Spain (Online)",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.gebnlp-1.3",
pages = "26--34",
abstract = "Misrepresentation of certain communities in datasets is causing big disruptions in artificial intelligence applications. In this paper, we propose using an automatically extracted gender-balanced dataset parallel corpus from Wikipedia. This balanced set is used to perform fine-tuning techniques from a bigger model trained on unbalanced datasets to mitigate gender biases in neural machine translation.",
}
@inproceedings{escolano-etal-2020-talp,
title = "The {TALP}-{UPC} System Description for {WMT}20 News Translation Task: Multilingual Adaptation for Low Resource {MT}",
author = "Escolano, Carlos and
Costa-juss{\`a}, Marta R. and
Fonollosa, Jos{\'e} A. R.",
booktitle = "Proceedings of the Fifth Conference on Machine Translation",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.wmt-1.10",
pages = "134--138",
abstract = "In this article, we describe the TALP-UPC participation in the WMT20 news translation shared task for Tamil-English. Given the low amount of parallel training data, we resort to adapt the task to a multilingual system to benefit from the positive transfer from high resource languages. We use iterative backtranslation to fine-tune the system and benefit from the monolingual data available. In order to measure the effectivity of such methods, we compare our results to a bilingual baseline system.",
}
@inproceedings{verges-boncompte-r-costa-jussa-2020-multilingual,
title = "Multilingual Neural Machine Translation: Case-study for {C}atalan, {S}panish and {P}ortuguese {R}omance Languages",
author = "Verg{\'e}s Boncompte, Pere and
R. Costa-juss{\`a}, Marta",
booktitle = "Proceedings of the Fifth Conference on Machine Translation",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.wmt-1.54",
pages = "447--450",
abstract = "In this paper, we describe the TALP-UPC participation in the WMT Similar Language Translation task between Catalan, Spanish, and Portuguese, all of them, Romance languages. We made use of different techniques to improve the translation between these languages. The multilingual shared encoder/decoder has been used for all of them. Additionally, we applied back-translation to take advantage of the monolingual data. Finally, we have applied fine-tuning to improve the in-domain data. Each of these techniques brings improvements over the previous one. In the official evaluation, our system was ranked 1st in the Portuguese-to-Spanish direction, 2nd in the opposite direction, and 3rd in the Catalan-Spanish pair.",
}
@inproceedings{menendez-salazar-etal-2020-ipn,
title = "The {IPN}-{CIC} team system submission for the {WMT} 2020 similar language task",
author = "Men{\'e}ndez-Salazar, Luis A. and
Sidorov, Grigori and
Costa-Juss{\`a}, Marta R.",
booktitle = "Proceedings of the Fifth Conference on Machine Translation",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.wmt-1.47",
pages = "409--413",
abstract = "This paper describes the participation of the NLP research team of the IPN Computer Research center in the WMT 2020 Similar Language Translation Task. We have submitted systems for the Spanish-Portuguese language pair (in both directions). The three submitted systems are based on the Transformer architecture and used fine tuning for domain Adaptation.",
}
@inproceedings{barrault-etal-2020-findings,
title = "Findings of the 2020 Conference on Machine Translation ({WMT}20)",
author = {Barrault, Lo{\"\i}c and Biesialska, Magdalena and Bojar, Ond{\v{r}}ej and Costa-juss{\`a}, Marta R. and Federmann, Christian and Graham, Yvette and Grundkiewicz, Roman and Haddow, Barry and Huck, Matthias and Joanis, Eric and Kocmi, Tom and Koehn, Philipp and Lo, Chi-kiu and Ljube{\v{s}}i{\'c}, Nikola and Monz, Christof and Morishita, Makoto and Nagata, Masaaki and Nakazawa, Toshiaki and Pal, Santanu and Post, Matt and Zampieri, Marcos},
booktitle = "Proceedings of the Fifth Conference on Machine Translation",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.wmt-1.1",
pages = "1--55",
abstract = "This paper presents the results of the news translation task and the similar language translation task, both organised alongside the Conference on Machine Translation (WMT) 2020. In the news task, participants were asked to build machine translation systems for any of 11 language pairs, to be evaluated on test sets consisting mainly of news stories. The task was also opened up to additional test suites to probe specific aspects of translation. In the similar language translation task, participants built machine translation systems for translating between closely related pairs of languages.",
}
@inproceedings{wmt-2020-machine,
title = "Proceedings of the Fifth Conference on Machine Translation",
author = {Barrault, Lo{\"\i}c and Bojar, Ond{\v{r}}ej and Bougares, Fethi and Chatterjee, Rajen and Costa-juss{\`a}, Marta R. and Federmann, Christian and Fishel, Mark and Fraser, Alexander and Graham, Yvette and Guzman, Paco and Haddow, Barry and Huck, Matthias and Yepes, Antonio Jimeno and Koehn, Philipp and Martins, Andr{\'e} and Morishita, Makoto and Monz, Christof and Nagata, Masaaki and Nakazawa, Toshiaki and Negri, Matteo},
month = nov,
year = "2020",
address = "Online",
booktitle = "Proceedings of the Fifth Conference on Machine Translation",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.wmt-1.0",
}
@inproceedings{gebnlp-2020-gender,
title = "Proceedings of the Second Workshop on Gender Bias in Natural Language Processing",
author = "Costa-juss{\`a}, Marta R. and
Hardmeier, Christian and
Radford, Will and
Webster, Kellie",
month = dec,
year = "2020",
address = "Barcelona, Spain (Online)",
booktitle= "Proceedings of the Second Workshop on Gender Bias in Natural Language Processing",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.gebnlp-1.0",
}
@article{costa2020amaleu,
title={AMALEU: Una Representaci{\'o}n Universal del Lenguaje basada en Aprendizaje Autom{\'a}tico},
author={Costa-Juss{\`a}, Marta Ruiz},
journal={Procesamiento del lenguaje natural},
number={65},
pages={105--108},
year={2020},
publisher={Sociedad Espa{\~n}ola para el Procesamiento del Lenguaje Natural}
}
@article{doi:10.1002/asi.24395,
author = {Escolano, Carlos and Costa-Juss{\`a}, Marta Ruiz and Fonollosa, Jos{\'e} A. R.},
title = {From bilingual to multilingual neural-based machine translation by incremental training},
journal = {Journal of the Association for Information Science and Technology},
year = {2020},
pages = {},
doi = {10.1002/asi.24395},
url = {https://asistdl.onlinelibrary.wiley.com/doi/abs/10.1002/asi.24395},
eprint = {https://asistdl.onlinelibrary.wiley.com/doi/pdf/10.1002/asi.24395},
abstract = {Abstract A common intermediate language representation in neural machine translation can be used to extend bilingual systems by incremental training. We propose a new architecture based on introducing an interlingual loss as an additional training objective. By adding and forcing this interlingual loss, we can train multiple encoders and decoders for each language, sharing among them a common intermediate representation. Translation results on the low-resource tasks (Turkish-English and Kazakh-English tasks) show a BLEU improvement of up to 2.8 points. However, results on a larger dataset (Russian-English and Kazakh-English) show BLEU losses of a similar amount. While our system provides improvements only for the low-resource tasks in terms of translation quality, our system is capable of quickly deploying new language pairs without the need to retrain the rest of the system, which may be a game changer in some situations. Specifically, what is most relevant regarding our architecture is that it is capable of: reducing the number of production systems, with respect to the number of languages, from quadratic to linear; incrementally adding a new language to the system without retraining the languages already there; and allowing for translations from the new language to all the others present in the system.}
}
@article{Basta2020,
doi = {10.1007/s00521-020-05211-z},
url = {https://doi.org/10.1007/s00521-020-05211-z},
year = {2020},
month = jul,
publisher = {Springer Science and Business Media {LLC}},
author = {Christine Basta and Marta R. Costa-juss{\`{a}} and Noe Casas},
title = {Extensive study on the underlying gender bias in contextualized word embeddings},
journal = {Neural Computing and Applications}
}
@inproceedings{basta-etal-2020-towards,
title = "Towards Mitigating Gender Bias in a decoder-based Neural Machine Translation model by Adding Contextual Information",
author = "Basta, Christine and
Costa-juss{\`a}, Marta R. and
Fonollosa, Jos{\'e} A. R.",
booktitle = "Proceedings of the The Fourth Widening Natural Language Processing Workshop",
month = jul,
year = "2020",
address = "Seattle, USA",
publisher = "Association for Computational Linguistics",
doi = "10.18653/v1/2020.winlp-1.25",
pages = "99--102",
abstract = "Gender bias negatively impacts many natural language processing applications, including machine translation (MT). The motivation behind this work is to study whether recent proposed MT techniques are significantly contributing to attenuate biases in document-level and gender-balanced data. For the study, we consider approaches of adding the previous sentence and the speaker information, implemented in a decoder-based neural MT system. We show improvements both in translation quality (+1 BLEU point) as well as in gender bias mitigation on WinoMT (+5{\%} accuracy).",
}
@article{escolano2020training,
title={Training Multilingual Machine Translation by Alternately Freezing Language-Specific Encoders-Decoders},
author={Escolano, Carlos and Costa-juss{\`a}, Marta R and Fonollosa, Jos{\'e} A. R. and Artetxe, Mikel},
journal={arXiv preprint arXiv:2006.01594},
year={2020},
url={https://arxiv.org/abs/2006.01594}
}
@inproceedings{carrino-etal-2019-terminology,
title = "Terminology-Aware Segmentation and Domain Feature for the {WMT}19 Biomedical Translation Task",
author = "Carrino, Casimiro Pio and
Rafieian, Bardia and
Costa-juss{\`a}, Marta R. and
Fonollosa, Jos{\'e} A. R.",
booktitle = "Proceedings of the Fourth Conference on Machine Translation (Volume 3: Shared Task Papers, Day 2)",
month = aug,
year = "2019",
address = "Florence, Italy",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/W19-5418",
doi = "10.18653/v1/W19-5418",
pages = "151--155",
abstract = "In this work, we give a description of the TALP-UPC systems submitted for the WMT19 Biomedical Translation Task. Our proposed strategy is NMT model-independent and relies only on one ingredient, a biomedical terminology list. We first extracted such a terminology list by labelling biomedical words in our training dataset using the BabelNet API. Then, we designed a data preparation strategy to insert the terms information at a token level. Finally, we trained the Transformer model with this terms-informed data. Our best-submitted system ranked 2nd and 3rd for Spanish-English and English-Spanish translation directions, respectively.",
}
@inproceedings{carrino-etal-2020-automatic,
title = "Automatic {S}panish Translation of {SQ}u{AD} Dataset for Multi-lingual Question Answering",
author = "Carrino, Casimiro Pio and
Costa-juss{\`a}, Marta R. and
Fonollosa, Jos{\'e} A. R.",
booktitle = "Proceedings of The 12th Language Resources and Evaluation Conference",
month = may,
year = "2020",
address = "Marseille, France",
publisher = "European Language Resources Association",
url = "https://www.aclweb.org/anthology/2020.lrec-1.677",
pages = "5515--5523",
abstract = "Recently, multilingual question answering became a crucial research topic, and it is receiving increased interest in the NLP community. However, the unavailability of large-scale datasets makes it challenging to train multilingual QA systems with performance comparable to the English ones. In this work, we develop the Translate Align Retrieve (TAR) method to automatically translate the Stanford Question Answering Dataset (SQuAD) v1.1 to Spanish. We then used this dataset to train Spanish QA systems by fine-tuning a Multilingual-BERT model. Finally, we evaluated our QA models with the recently proposed MLQA and XQuAD benchmarks for cross-lingual Extractive QA. Experimental results show that our models outperform the previous Multilingual-BERT baselines achieving the new state-of-the-art values of 68.1 F1 on the Spanish MLQA corpus and 77.6 F1 on the Spanish XQuAD corpus. The resulting, synthetically generated SQuAD-es v1.1 corpora, with almost 100{\%} of data contained in the original English version, to the best of our knowledge, is the first large-scale QA training resource for Spanish.",
language = "English",
ISBN = "979-10-95546-34-4",
}
@inproceedings{basta2019evaluating,
title={Evaluating the Underlying Gender Bias in Contextualized Word Embeddings},
author={Basta, Christine and Costa-juss{\`a}, Marta R and Casas, Noe},
booktitle={Proceedings of the First Workshop on Gender Bias in Natural Language Processing},
pages={33--39},
year={2019},
month = aug
}
@inproceedings{casas2018differentiable,
title={A differentiable bleu loss. analysis and first results},
author={Casas, Noe and Fonollosa, Jos{\'e} A. R. and Ruiz Costa-Juss{\`a}, Marta},
booktitle={ICLR 2018 Workshop Track: 6th International Conference on Learning Representations: Vancouver Convention Center, Vancouver, BC, Canada: April 30-May 3, 2018},
year={2018},
}
@inproceedings{casas-etal-2018-talp,
title = "The {TALP}-{UPC} Machine Translation Systems for {WMT}18 News Shared Translation Task",
author = "Casas, Noe and
Escolano, Carlos and
Costa-juss{\`a}, Marta R. and
Fonollosa, Jos{\'e} A. R.",
booktitle = "Proceedings of the Third Conference on Machine Translation: Shared Task Papers",
month = oct,
year = "2018",
address = "Belgium, Brussels",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/W18-6406",
doi = "10.18653/v1/W18-6406",
pages = "355--360",
abstract = "In this article we describe the TALP-UPC research group participation in the WMT18 news shared translation task for Finnish-English and Estonian-English within the multi-lingual subtrack. All of our primary submissions implement an attention-based Neural Machine Translation architecture. Given that Finnish and Estonian belong to the same language family and are similar, we use as training data the combination of the datasets of both language pairs to paliate the data scarceness of each individual pair. We also report the translation quality of systems trained on individual language pair data to serve as baseline and comparison reference.",
}
@inproceedings{torregrosa-etal-2019-leveraging,
title = "Leveraging Rule-Based Machine Translation Knowledge for Under-Resourced Neural Machine Translation Models",
author = "Torregrosa, Daniel and
Pasricha, Nivranshu and
Masoud, Maraim and
Chakravarthi, Bharathi Raja and
Alonso, Juan and
Casas, Noe and
Arcan, Mihael",
booktitle = "Proceedings of Machine Translation Summit XVII Volume 2: Translator, Project and User Tracks",
month = aug,
year = "2019",
address = "Dublin, Ireland",
publisher = "European Association for Machine Translation",
url = "https://www.aclweb.org/anthology/W19-6725",
pages = "125--133",
}
@inproceedings{casas-etal-2019-talp,
title = "The {TALP}-{UPC} Machine Translation Systems for {WMT}19 News Translation Task: Pivoting Techniques for Low Resource {MT}",
author = "Casas, Noe and
Fonollosa, Jos{\'e} A. R. and
Escolano, Carlos and
Basta, Christine and
Costa-juss{\`a}, Marta R.",
booktitle = "Proceedings of the Fourth Conference on Machine Translation (Volume 2: Shared Task Papers, Day 1)",
month = aug,
year = "2019",
address = "Florence, Italy",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/W19-5311",
doi = "10.18653/v1/W19-5311",
pages = "155--162",
abstract = "In this article, we describe the TALP-UPC research group participation in the WMT19 news translation shared task for Kazakh-English. Given the low amount of parallel training data, we resort to using Russian as pivot language, training subword-based statistical translation systems for Russian-Kazakh and Russian-English that were then used to create two synthetic pseudo-parallel corpora for Kazakh-English and English-Kazakh respectively. Finally, a self-attention model based on the decoder part of the Transformer architecture was trained on the two pseudo-parallel corpora."
}
@article{artetxe2020all,
title={Do all Roads Lead to Rome? Understanding the Role of Initialization in Iterative Back-Translation},
author={Artetxe, Mikel and Labaka, Gorka and Casas, Noe and Agirre, Eneko},
journal={arXiv preprint arXiv:2002.12867},
year={2020},
url="https://arxiv.org/abs/2002.12867"
}
@inproceedings{casas-etal-2020-combining,
title = "Combining Subword Representations into Word-level Representations in the Transformer Architecture",
author = "Casas, Noe and
Costa-juss{\`a}, Marta R. and
Fonollosa, Jos{\'e} A. R.",
booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop",
month = jul,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.acl-srw.10",
doi = "10.18653/v1/2020.acl-srw.10",
pages = "66--71",
abstract = "In Neural Machine Translation, using word-level tokens leads to degradation in translation quality. The dominant approaches use subword-level tokens, but this increases the length of the sequences and makes it difficult to profit from word-level information such as POS tags or semantic dependencies. We propose a modification to the Transformer model to combine subword-level representations into word-level ones in the first layers of the encoder, reducing the effective length of the sequences in the following layers and providing a natural point to incorporate extra word-level information. Our experiments show that this approach maintains the translation quality with respect to the normal Transformer model when no extra word-level information is injected and that it is superior to the currently dominant method for incorporating word-level source language information to models based on subword-level vocabularies.",
}
@inproceedings{biesialska-etal-2019-talp,
title = "The {TALP}-{UPC} System for the {WMT} Similar Language Task: Statistical vs Neural Machine Translation",
author = "Biesialska, Magdalena and
Guardia, Lluis and
Costa-juss{\`a}, Marta R.",
booktitle = "Proceedings of the Fourth Conference on Machine Translation (Volume 3: Shared Task Papers, Day 2)",
month = aug,
year = "2019",
address = "Florence, Italy",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/W19-5424",
doi = "10.18653/v1/W19-5424",
pages = "185--191",
abstract = "Although the problem of similar language translation has been an area of research interest for many years, yet it is still far from being solved. In this paper, we study the performance of two popular approaches: statistical and neural. We conclude that both methods yield similar results; however, the performance varies depending on the language pair. While the statistical approach outperforms the neural one by a difference of 6 BLEU points for the Spanish-Portuguese language pair, the proposed neural model surpasses the statistical one by a difference of 2 BLEU points for Czech-Polish. In the former case, the language similarity (based on perplexity) is much higher than in the latter case. Additionally, we report negative results for the system combination with back-translation. Our TALP-UPC system submission won 1st place for Czech-{\textgreater}Polish and 2nd place for Spanish-{\textgreater}Portuguese in the official evaluation of the 1st WMT Similar Language Translation task.",
}
@article{biesialska-2020-refinement,
place = {NL},
title = {Refinement of Unsupervised Cross-Lingual Word Embeddings},
volume = {325},
ISSN = {0922-6389},
url = {https://doi.org/10.3233/FAIA200317},
DOI = {10.3233/FAIA200317},
number = {ECAI 2020},
journal = {Frontiers in Artificial Intelligence and Applications},
publisher = {IOS Press},
author = {Biesialska, Magdalena and
Costa-juss{\`a}, Marta R.},
year = {2020},
pages = {1978–1981}
}
@inproceedings{biesialska-etal-2020-enhancing,
title = "Enhancing Word Embeddings with Knowledge Extracted from Lexical Resources",
author = "Biesialska, Magdalena and
Rafieian, Bardia and
Costa-juss{\`a}, Marta R.",
booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop",
month = jul,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.acl-srw.36",
doi = "10.18653/v1/2020.acl-srw.36",
pages = "271--278",
abstract = "In this work, we present an effective method for semantic specialization of word vector representations. To this end, we use traditional word embeddings and apply specialization methods to better capture semantic relations between words. In our approach, we leverage external knowledge from rich lexical resources such as BabelNet. We also show that our proposed post-specialization method based on an adversarial neural network with the Wasserstein distance allows to gain improvements over state-of-the-art methods on two tasks: word similarity and dialog state tracking.",
}
@inproceedings{escolano-etal-2017-talp,
title = "The {TALP}-{UPC} Neural Machine Translation System for {G}erman/{F}innish-{E}nglish Using the Inverse Direction Model in Rescoring",
author = "Escolano, Carlos and
Costa-juss{\`a}, Marta R. and
Fonollosa, Jos{\'e} A. R.",
booktitle = "Proceedings of the Second Conference on Machine Translation",
month = sep,
year = "2017",
address = "Copenhagen, Denmark",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/W17-4725",
doi = "10.18653/v1/W17-4725",
pages = "283--287",
}
@inproceedings{vila2018end,
title={End-to-End Speech Translation with the Transformer.},
author={Vila, Laura Cross and Escolano, Carlos and Fonollosa, Jos{\'e} A. R. and Costa-juss{\`a}, Marta R},
booktitle={IberSPEECH},
pages={60--63},
year={2018}
}
@inproceedings{escolano-etal-2019-bilingual,
title = "From Bilingual to Multilingual Neural Machine Translation by Incremental Training",
author = "Escolano, Carlos and
Costa-juss{\`a}, Marta R. and
Fonollosa, Jos{\'e} A. R.",
booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop",
month = jul,
year = "2019",
address = "Florence, Italy",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/P19-2033",
doi = "10.18653/v1/P19-2033",
pages = "236--242",
abstract = "Multilingual Neural Machine Translation approaches are based on the use of task specific models and the addition of one more language can only be done by retraining the whole system. In this work, we propose a new training schedule that allows the system to scale to more languages without modification of the previous components based on joint training and language-independent encoder/decoder modules allowing for zero-shot translation. This work in progress shows close results to state-of-the-art in the WMT task.",
}
@inproceedings{escolano-etal-2019-multilingual,
title = "Multilingual, Multi-scale and Multi-layer Visualization of Intermediate Representations",
author = "Escolano, Carlos and
Costa-juss{\`a}, Marta R. and
Lacroux, Elora and
V{\'a}zquez, Pere-Pau",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP): System Demonstrations",
month = nov,
year = "2019",
address = "Hong Kong, China",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/D19-3026",
doi = "10.18653/v1/D19-3026",
pages = "151--156",
abstract = "The main alternatives nowadays to deal with sequences are Recurrent Neural Networks (RNN) architectures and the Transformer. In this context, Both RNN{'}s and Transformer have been used as an encoder-decoder architecture with multiple layers in each module. Far beyond this, these architectures are the basis for the contextual word embeddings which are revolutionizing most natural language downstream applications. However, intermediate representations in either the RNN or Transformer architectures can be difficult to interpret. To make these layer representations more accessible and meaningful, we introduce a web-based tool that visualizes them both at the sentence and token level. We present three use cases. The first analyses gender issues in contextual word embeddings. The second and third are showing multilingual intermediate representations for sentences and tokens and the evolution of these intermediate representations along with the multiple layers of the decoder and in the context of multilingual machine translation.",
}
@inproceedings{escolano-etal-2021-multilingual,
title = "Multilingual Machine Translation: Closing the Gap between Shared and Language-specific Encoder-Decoders",
author = "Escolano, Carlos and
Costa-juss{\`a}, Marta R. and
Fonollosa, Jos{\'e} A. R. and
Artetxe, Mikel",
booktitle = "Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume",
month = apr,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2021.eacl-main.80",
pages = "944--948",
abstract = "State-of-the-art multilingual machine translation relies on a universal encoder-decoder, which requires retraining the entire system to add new languages. In this paper, we propose an alternative approach that is based on language-specific encoder-decoders, and can thus be more easily extended to new languages by learning their corresponding modules. So as to encourage a common interlingua representation, we simultaneously train the N initial languages. Our experiments show that the proposed approach outperforms the universal encoder-decoder by 3.28 BLEU points on average, while allowing to add new languages without the need to retrain the rest of the modules. All in all, our work closes the gap between shared and language-specific encoderdecoders, advancing toward modular multilingual machine translation systems that can be flexibly extended in lifelong learning settings.",
}
@article{escolano2017generacion,
title={Generaci{\'o}n morfol{\'o}gica con algoritmos de aprendizaje profundo integrada en un sistema de traducci{\'o}n autom{\'a}tica estad{\'\i}stica},
author={Escolano, Carlos and Ruiz Costa-Juss{\`a}, Marta},
journal={Procesamiento del lenguaje natural (SEPLN)},
number={59},
pages={107--114},
year={2017}
}
@article{marino2006n,
title={N-gram-based machine translation},
author={Marino, Jos{\'e} B and Banchs, Rafael E and Crego, Josep M and de Gispert, Adri{\`a} and Lambert, Patrik and Fonollosa, Jos{\'e} A. R. and Costa-juss{\`a}, Marta R},
journal={Computational linguistics},
volume={32},
number={4},
pages={527--549},
year={2006},
publisher={MIT Press}
}
@inproceedings{costa-jussa-fonollosa-2016-character,
title = "Character-based Neural Machine Translation",