Skip to content

Commit cd77632

Browse files
Update papers.bib
1 parent 1703e39 commit cd77632

File tree

1 file changed

+46
-0
lines changed

1 file changed

+46
-0
lines changed

_bibliography/papers.bib

Lines changed: 46 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,18 @@
11
---
22
---
3+
4+
@inproceedings{blaschke-etal-2025-multi,
5+
title={A Multi-Dialectal Dataset for {German} Dialect {ASR} and Dialect-to-Standard Speech Translation},
6+
author={Blaschke, Verena and Winkler, Miriam and Förster, Constantin and Wenger-Glemser, Gabriele and Plank, Barbara},
7+
abstract={Although Germany has a diverse landscape of dialects, they are underrepresented in current automatic speech recognition (ASR) research. To enable studies of how robust models are towards dialectal variation, we present Betthupferl, an evaluation dataset containing four hours of read speech in three dialect groups spoken in Southeast Germany (Franconian, Bavarian, Alemannic), and half an hour of Standard German speech. We provide both dialectal and Standard German transcriptions, and analyze the linguistic differences between them. We benchmark several multilingual state-of-the-art ASR models on speech translation into Standard German, and find differences between how much the output resembles the dialectal vs. standardized transcriptions. Qualitative error analyses of the best ASR model reveal that it sometimes normalizes grammatical differences, but often stays closer to the dialectal constructions.},
8+
booktitle = {Interspeech 2025},
9+
pages = {913--917},
10+
doi = {10.21437/Interspeech.2025-318},
11+
url={https://www.isca-archive.org/interspeech_2025/blaschke25_interspeech.html},
12+
year = 2025,
13+
month = aug,
14+
}
15+
316
@inproceedings{ma-etal-2025-algorithmic,
417
title = "Algorithmic Fidelity of Large Language Models in Generating Synthetic {G}erman Public Opinions: A Case Study",
518
author = "Ma, Bolei and
@@ -206,6 +219,39 @@ @inproceedings{sauberli-etal-2025-llms
206219
abstract = "Knowing how test takers answer items in educational assessments is essential for test development, to evaluate item quality, and to improve test validity. However, this process usually requires extensive pilot studies with human participants. If large language models (LLMs) exhibit human-like response behavior to test items, this could open up the possibility of using them as pilot participants to accelerate test development. In this paper, we evaluate the human-likeness or psychometric plausibility of responses from 18 instruction-tuned LLMs with two publicly available datasets of multiple-choice test items across three subjects: reading, U.S. history, and economics. Our methodology builds on two theoretical frameworks from psychometrics which are commonly used in educational assessment, classical test theory and item response theory. The results show that while larger models are excessively confident, their response distributions can be more human-like when calibrated with temperature scaling. In addition, we find that LLMs tend to correlate better with humans in reading comprehension items compared to other subjects. However, the correlations are not very strong overall, indicating that LLMs should not be used for piloting educational assessments in a zero-shot setting."
207220
}
208221

222+
@inproceedings{blaschke-etal-2025-analyzing,
223+
title = "Analyzing the Effect of Linguistic Similarity on Cross-Lingual Transfer: Tasks and Experimental Setups Matter",
224+
author = "Blaschke, Verena and
225+
Fedzechkina, Masha and
226+
Ter Hoeve, Maartje",
227+
editor = "Che, Wanxiang and
228+
Nabende, Joyce and
229+
Shutova, Ekaterina and
230+
Pilehvar, Mohammad Taher",
231+
booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
232+
month = jul,
233+
year = "2025",
234+
address = "Vienna, Austria",
235+
publisher = "Association for Computational Linguistics",
236+
url = "https://aclanthology.org/2025.findings-acl.454/",
237+
doi = "10.18653/v1/2025.findings-acl.454",
238+
pages = "8653--8684",
239+
ISBN = "979-8-89176-256-5",
240+
abstract = "Cross-lingual transfer is a popular approach to increase the amount of training data for NLP tasks in a low-resource context. However, the best strategy to decide which cross-lingual data to include is unclear. Prior research often focuses on a small set of languages from a few language families and/or a single task. It is still an open question how these findings extend to a wider variety of languages and tasks. In this work, we analyze cross-lingual transfer for 263 languages from a wide variety of language families. Moreover, we include three popular NLP tasks: POS tagging, dependency parsing, and topic classification. Our findings indicate that the effect of linguistic similarity on transfer performance depends on a range of factors: the NLP task, the (mono- or multilingual) input representations, and the definition of linguistic similarity."
241+
}
242+
243+
@incollection{nerbonne-etal-2025-methods,
244+
title = {Methods and Resources in {Germanic} Variationist Linguistics},
245+
author = {Nerbonne, John and Blaschke, Verena and Sch\"{u}tze, Hinrich and Plank, Barbara},
246+
booktitle = {Oxford Research Encyclopedia of Linguistics},
247+
publisher = {Oxford University Press},
248+
editor = {K\"{u}rschner, Sebastian and Dammel, Antje},
249+
doi = {10.1093/acrefore/9780199384655.013.1033},
250+
url = {https://doi.org/10.1093/acrefore/9780199384655.013.1033},
251+
year = {2025},
252+
month = may,
253+
}
254+
209255
@inproceedings{shim-plank-2025-dialetto,
210256
title = "Dialetto, ma Quanto Dialetto? Transcribing and Evaluating Dialects on a Continuum",
211257
author = "Shim, Ryan Soh-Eun and

0 commit comments

Comments
 (0)