Have a personal or library account? Click to login
Better data for more researchers – using the audio features of BNCweb Cover

Better data for more researchers – using the audio features of BNCweb

Open Access
|Jun 2021

Abstract

In spite of the wide agreement among linguists as to the significance of spoken language data, actual speech data have not formed the basis of empirical work on English as much as one would think. The present paper is intended to contribute to changing this situation, on a theoretical and on a practical level. On a theoretical level, we discuss different research traditions within (English) linguistics. Whereas speech data have become increasingly important in various linguistic disciplines, major corpora of English developed within the corpus-linguistic community, carefully sampled to be representative of language usage, are usually restricted to orthographic transcriptions of spoken language. As a result, phonological phenomena have remained conspicuously understudied within traditional corpus linguistics. At the same time, work with current speech corpora often requires a considerable level of specialist knowledge and tailor-made solutions. On a practical level, we present a new feature of BNCweb (Hoffmann et al. 2008), a user-friendly interface to the British National Corpus, which gives users access to audio and phonemic transcriptions of more than five million words of spontaneous speech. With the help of a pilot study on the variability of intrusive r we illustrate the scope of the new possibilities.

DOI: https://doi.org/10.2478/icame-2021-0004 | Journal eISSN: 1502-5462 | Journal ISSN: 0801-5775
Language: English
Page range: 125 - 154
Published on: Jun 12, 2021
Published by: Uppsala University, Department of English
In partnership with: Paradigm Publishing Services
Publication frequency: 1 issue per year

© 2021 Sebastian Hoffmann, Sabine Arndt-Lappe, published by Uppsala University, Department of English
This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 3.0 License.