References
- N. Reimers and I. Gurevych, “Sentence-BERT: Sentence embeddings using siamese BERT-networks,” arXiv preprint arXiv:1908.10084, Aug. 2019. https://doi.org/10.48550/arXiv.1908.10084
- M. Westera, J. Amidei, and L. Mayol, “Similarity or deeper understanding? Analyzing the TED-Q dataset of evoked questions,” in Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain, Dec. 2020, pp. 5004–5012. https://doi.org/10.18653/v1/2020.coling-main.439
- A. W. Qurashi, V. Holmes, and A. P. Johnson, “Document processing: Methods for semantic text similarity analysis,” in 2020 International Conference on INnovations in Intelligent SysTems and Applications (INISTA), Novi Sad, Serbia, Aug. 2020, pp. 1–6. https://doi.org/10.1109/INISTA49547.2020.9194665
- A. Rana, A. Pant, N. Rawat, P. Rawat, S. Vats, and V. Sharma, “Semantic similarity analysis using FastText,” in 2024 IEEE 3rd World Conference on Applied Intelligence and Computing (AIC), Gwalior, India, Jul. 2024, pp. 454–460. https://doi.org/10.1109/AIC61668.2024.10731025
- D. R. Yerramreddy, J. Marasani, P. Gowtham, and G. Harshit, “Speech recognition paradigms: A comparative evaluation of SpeechBrain, Whisper and Wav2Vec2 models,” in 2024 IEEE 9th International Conference for Convergence in Technology (I2CT), Pune, India, Apr. 2024, pp. 1–6. https://doi.org/10.1109/I2CT61223.2024.10544133
- E. Loda, “Riconoscimento del parlato mediante openai Whisper,” unpublished report, 2024.
- S. Wang, C.-H. Yang, J. Wu, and C. Zhang, “Can Whisper perform speech-based in-context learning?” in ICASSP 2024–IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Korea, Apr. 2024, pp. 13421–13425. https://doi.org/10.1109/ICASSP48485.2024.10446502
- R. Jain, A. Barcovschi, M. Yiwere, P. Corcoran, and H. Cucu, “Adaptation of Whisper models to child speech recognition,” arXiv preprint arXiv:2307.13008, Jul. 2023. https://doi.org/10.48550/arXiv.2307.13008
- Y. Ren, Y. Ruan, X. Tan, T. Qin, S. Zhao, Z. Zhao, and T.-Y. Liu, “FastSpeech: Fast, robust and controllable text to speech,” in Advances in Neural Information Processing Systems, vol. 32, 2019. https://proceedings.neurips.cc/paper_files/paper/2019/file/f63f65b503e22cb970527f23c9ad7db1-Paper.pdf
- Y. Ren, X. Tan, T. Qin, S. Zhao, Z. Zhao, and T.-Y. Liu, “Almost unsupervised text to speech and automatic speech recognition,” in International Conference on Machine Learning, 2019, pp. 5410–5419. https://proceedings.mlr.press/v97/ren19a/ren19a.pdf
- R. Asadi, H. Trinh, H. J. Fell, and T. W. Bickmore, “IntelliPrompter: Speech-based dynamic note display interface for oral presentations,” in Proceedings of the 19th ACM International Conference on Multimodal Interaction, Nov. 2017, pp. 172–180. https://doi.org/10.1145/3136755.3136818
- L. Pandey and A. S. M. N. Arif, “Effects of speaking rate on speech and silent speech recognition,” in CHI Conference on Human Factors in Computing Systems Extended Abstracts, Apr. 2022, pp. 1–8. https://doi.org/10.1145/3491101.3519611
- N. Rossenbach, A. Zeyer, R. Schlüter, and H. Ney, “Generating synthetic audio data for attention-based speech recognition systems,” in ICASSP 2020 – IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, May 2020, pp. 7069–7073. https://doi.org/10.1109/ICASSP40776.2020.9053008
- J. Alabi, D. I. Adelani, D. Ruiter, and C. C. Emezue, “The effect of curated vs. noisy data on cross-lingual transferability for low- resource languages,” in Findings of the Association for Computational Linguistics: EMNLP 2022, 2022, pp. 2741–2752.
- R. Sadigov, E. Yıldırım, B. Kocaçınar, F. Patlar Akbulut, and C. Catal, “Deep learning-based user experience evaluation in distance learning,” Cluster Computing, vol. 27, pp. 443–455, Feb. 2024. https://doi.org/10.1007/s10586-022-03918-3
- G. A. Miller, “The magical number seven, plus or minus two: Some limits on our capacity for processing information,” Psychological Review, vol. 63, no. 2, pp. 81–97, 1956. https://doi.org/10.1037/h0043158