Skip to main content
Have a personal or library account? Click to login
A Review of String Instrument Synthesis Methods for Use in Interactive Systems Cover

A Review of String Instrument Synthesis Methods for Use in Interactive Systems

Open Access
|Apr 2026

References

  1. Adrien, J.‑M. (1991). The missing link: Modal synthesis. In G. De Poli, A. Piccialli, and C. Roads (Eds.), Representations of Musical Signals (pp. 269298). MIT Press.
  2. Arksey, H., and O’Malley, L. (2005). Scoping studies: Towards a methodological framework. International Journal of Social Research Methodology, 8(1), 1932.
  3. Baines, A. (1992). The Oxford Companion to Musical Instruments. Oxford University Press.
  4. Bank, B., Zambon, S., and Fontana, F. (2010). A modal‑based real‑time piano synthesizer. IEEE Transactions on Speech and Audio Processing, 18(4), 809821.
  5. Baoueb, T., Bie, X., Janati, H., and Richard, G. (2024). WaveTransfer: A flexible end‑to‑end multi‑instrument timbre transfer with diffusion. In 34th IEEE International Workshop on Machine Learning for Signal Processing, MLSP 2024 (pp. 16). IEEE.
  6. Bensa, J., Bilbao, S., Kronland‑Martinet, R., and Smith, J. O. (2003). The simulation of piano string vibration: From physical models to finite difference schemes and digital waveguides. The Journal of the Acoustical Society of America, 114(2), 10951107.
  7. Bilbao, S. (2009). Numerical sound synthesis: Finite difference schemes and simulation in musical acoustics. Wiley Publishing.
  8. Bitton, A., Esling, P., and Harada, T. (2020). Neural granular sound synthesis. CoRR, abs/2008.01393.
  9. Cadoz, C., Luciani, A., and Florens, J. L. (1993). CORDIS‑ANIMA: A modeling and simulation system for sound and image synthesis–The general formalism. Computer Music Journal, 17(1), 1929.
  10. Caillon, A., and Esling, P. (2021). RAVE: A variational autoencoder for fast and high‑quality neural audio synthesis. CoRR, abs/2111.05011.
  11. Carney, M., Li, C., Toh, E., Zada, N., Yu, P., and Engel, J. H. (2021). Tone transfer: In‑browser interactive neural audio synthesis. In Joint Proceedings of the ACM IUI 2021 Workshops Co‑Located with 26th ACM Conference on Intelligent User Interfaces (ACM IUI 2021), of CEUR Workshop Proceedings, 2903. CEUR-WS.org.
  12. Carrillo, A. P., and Bonada, J. (2010). The bowed tube: A virtual violin. In Proceedings of the International Conference on New Interfaces for Musical Expression (pp. 229232). Sydney, Australia.
  13. Caspe, F., McPherson, A., and Sandler, M. (2022). DDX7: differentiable FM synthesis of musical instrument sounds. In Proceedings of the 23rd International Society for Music Information Retrieval Conference, ISMIR 2022 (pp. 608616).
  14. Castagné, N., and Cadoz, C. (2003). 10 criteria for evaluating physical modelling schemes for music creation. In Proceedings of the 6th Digital Audio Effects (pp. 17). London, United Kingdom.
  15. Chowning, J. (1973). The synthesis of complex audio spectra by means of frequency modulation. Journal of the Audio Engineering Society, 21(7), 526534.
  16. Dai, Y. R., Yang, H. C., and Su, A. W. (2021). Efficient synthesis of violin sounds using a BiLSTM network based source filter model. In 150th Audio Engineering Society Convention, AES 2021, 150th Audio Engineering Society Convention, AES 2021. Audio Engineering Society.
  17. Défossez, A., Zeghidour, N., Usunier, N., Bottou, L., and Bach, F. R. (2018). SING: Symbol‑to‑instrument neural generator. In Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018 (pp. 90559065).
  18. Ding, Y., and Gerhard, D. (2004). Analysis and synthesis of the guqin ‑A Chinese traditional instrument. Canadian Acoustics, 32(3), 122123.
  19. Donahue, C., McAuley, J. J., and Puckette, M. S. (2019). Adversarial audio synthesis. In 7th International Conference on Learning Representations, ICLR 2019. New Orleans, LA, USA. OpenReview.net.
  20. Dong, H.‑W., Hsiao, W.‑Y., Yang, L.‑C., and Yang, Y.‑H. (2018). MuseGAN: Multi‑track sequential generative adversarial networks for symbolic music generation and accompaniment. In Proceedings of the Thirty‑Second AAAI Conference on Artificial Intelligence (AAAI‑18), the 30th Innovative Applications of Artificial Intelligence (IAAI‑18), and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI‑18) (pp. 3441). AAAI Press.
  21. Engel, J. H., Agrawal, K. K., Chen, S., Gulrajani, I., Donahue, C., and Roberts, A. (2019). GANSynth: Adversarial neural audio synthesis. In 7th International Conference on Learning Representations, ICLR 2019. OpenReview.net.
  22. Engel, J. H., Hantrakul, L., Gu, C., and Roberts, A. (2020). DDSP: Differentiable digital signal processing. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia. April 26–30, 2020. OpenReview.net.
  23. Engel, J. H., Resnick, C., Roberts, A., Dieleman, S., Norouzi, M., Eck, D., and Simonyan, K. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. In Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Volume 70 of Proceedings of Machine Learning Research (pp. 10681077). PMLR.
  24. Erkut, C., Karjalainen, M., Huang, P., and Välimäki, V. (2002). Acoustical analysis and model‑based sound synthesis of the kantele. Journal of the Acoustical Society of America, 112(4), 16811691.
  25. Erkut, C., Laurson, M., Kuuskankare, M., and Välimäki, V. (2001). Model‑based synthesis of the ud and the renaissance lute. In Proceedings of the 2001 International Computer Music Conference, ICMC 2001. Michigan Publishing.
  26. Fletcher, N. H., and Rossing, T. D. (1998). The Physics of Musical Instruments. Springer.
  27. Florens, J.‑L. (2003). Expressive bowing on a virtual string instrument. In Gesture‑Based Communication in Human‑Computer Interaction, 5th International Gesture Workshop, GW 2003, Volume 2915 of Lecture Notes in Computer Science (pp. 487496). Springer.
  28. Fraisse, V., Wanderley, M. M., and Guastavino, C. (2021). Comprehensive framework for describing interactive sound installations: Highlighting trends through a systematic review. Multimodal Technologies and Interaction, 5(4), 19.
  29. Ganis, F., Knudsen, E. F., Lyster, S. V. K., Otterbein, R., Südholt, D., and Erkut, C. (2021). Real‑time timbre transfer and sound synthesis using DDSP. In SMC 2021 ‑ Proceedings of the 18th Sound and Music Computing Conference (Vol. 2021, pp. 175182). Virtual Sound and Music Computing Network.
  30. Gehlhaar, R., Girao, L. M., and Rodrigues, P. M. (2011). CaDaReMi. An educational interactive music game. International Journal on Disability and Human Development, 10(1), 2529.
  31. Gómez‑Sirvent, J. L., López de la Rosa, F., Sánchez‑Reolid, R., Herrera, R. M., and Fernández‑Caballero, A. (2025). Musical instruments in extended reality: A systematic review. International Journal of Human–Computer Interaction, 41(16), 1014310162.
  32. Hawley, S. H. (2020). Synthesis of musical instrument sounds: Physics‑based modeling or machine learning? Acoustics Today, 16(1), 20.
  33. Hawthorne, C., Simon, I., Roberts, A., Zeghidour, N., Gardner, J., Manilow, E., and Engel, J. H. (2022). Multi‑instrument music synthesis with spectrogram diffusion. In Proceedings of the 23rd International Society for Music Information Retrieval Conference, ISMIR 2022 (pp. 598607).
  34. Hawthorne, C., Stasyuk, A., Roberts, A., Simon, I., Huang, C.‑Z. A., Dieleman, S., Elsen, E., Engel, J. H., and Eck, D. (2019). Enabling factorized piano music modeling and generation with the MAESTRO dataset. In 7th International Conference on Learning Representations, ICLR 2019. OpenReview.net.
  35. Hayes, B., Saitis, C., and Fazekas, G. (2021). Neural waveshaping synthesis. In Proceedings of the 22nd International Society for Music Information Retrieval Conference (pp. 9). ISMIR.
  36. Hayes, B., Shier, J., Fazekas, G., McPherson, A., and Saitis, C. (2024). A review of differentiable digital signal processing for music and speech synthesis. Frontiers in Signal Processing, 3, 2023.
  37. Hiller, L., and Ruiz, P. (1971). Synthesizing musical sounds by solving the wave equation for vibrating objects: Part 1. Journal of the Audio Engineering Society, 19(6), 462470.
  38. Horner, A., Beauchamp, J., and Haken, L. (1993). Methods for multiple wavetable synthesis of musical instrument tones. AES: Journal of the Audio Engineering Society, 41(5), 336356.
  39. Huang, Q., Park, D. S., Wang, T., Denk, T. I., Ly, A., Chen, N., Zhang, Z., Zhang, Z., Yu, J., Frank, C. H., Engel, J. H., Le, Q. V., Chan, W., and Han, W. (2023). Noise2Music: Text‑ conditioned music generation with diffusion models. CoRR, abs/2302.03917.
  40. Hung, T.‑Y., Wu, J.‑T., Kuo, Y.‑C., Hsiao, Y.‑W., Lin, T.‑W., and Su, L. (2024). A study on synthesizing expressive violin performances: Approaches and comparisons. CoRR, abs/2406.18089.
  41. Jack, R. H., Stockman, T., and McPherson, A. P. (2016). Effect of latency on performer interaction and subjective quality assessment of a digital musical instrument. In Proceedings of the Audio Mostly 2016 (pp. 116123). ACM.
  42. Jaffe, D. A. (1995). Ten criteria for evaluating synthesis techniques. Computer Music Journal, 19(1), 7687.
  43. Jonason, N., Sturm, B. L. T., and Thomé, C. (2020). The control‑synthesis approach for making expressive and controllable neural music synthesizers. In Proceedings of the 1st Joint Conference on AI Music Creativity (pp. 9). AIMC.
  44. Jonason, N., Wang, X., Cooper, E., Juvela, L., Sturm, B. L. T., and Yamagishi, J. (2023). DDSP‑based neural waveform synthesis of polyphonic guitar performance from string‑wise MIDI input. CoRR, abs/2309.07658.
  45. Kalchbrenner, N., Elsen, E., Simonyan, K., Noury, S., Casagrande, N., Lockhart, E., Stimberg, F., van den Oord, A., Dieleman, S., and Kavukcuoglu, K. (2018). Efficient neural audio synthesis. In Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Volume 80 of Proceedings of Machine Learning Research (pp. 24152424). PMLR.
  46. Karplus, K., and Strong, A. (1983). Digital synthesis of plucked‑ string and drum timbres. Computer Music Journal, 7(2), 4355.
  47. Kim, D., Dong, H.‑W., and Jeong, D. (2025). ViolinDiff: Enhancing expressive violin synthesis with pitch bend conditioning. In 2025 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2025 (pp. 15). IEEE.
  48. Kim, H., Choi, S., and Nam, J. (2024). Expressive acoustic guitar sound synthesis with an instrument‑specific input representation and diffusion outpainting. In IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2024 (pp. 76207624). IEEE.
  49. Kumar, K., Kumar, R., de Boissiere, T., Gestin, L., Teoh, W. Z., Sotelo, J., de Brébisson, A., Bengio, Y., and Courville, A. C. (2019). MelGAN: Generative adversarial networks for conditional waveform synthesis. In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019 (pp. 1488114892).
  50. Lee, J. W., Park, J., Choi, M. J., and Lee, K. (2024, NeurIPS). Differentiable modal synthesis for physical modeling of planar string sound and motion simulation. In Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, NeurIPS 2024.
  51. Leonard, J., and Villeneuve, J. (2019). MI‑GEN~: An efficient and accessible mass‑interaction sound synthesis toolbox. In Proceedings of the 16th Sound and Music Computing Conference (SMC), Malaga, Spain.
  52. Levitin, D. J., McAdams, S., and Adams, R. L. (2002). Control parameters for musical instruments: A foundation for new mappings of gesture to sound. Organised Sound, 7(2), 171189.
  53. Liu, Q. (2024). An FM‑wavetable‑synthesized violin with natural vibrato and bow pressure. In Proceedings of the 2023 International Conference on Data Science, Advanced Algorithm and Intelligent Computing (DAI 2023) (pp. 243250). Atlantis Press.
  54. Maman, B., Zeitler, J., Müller, M., and Bermano, A. H. (2024). Performance conditioning for diffusion‑based multi‑ instrument music synthesis. In IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2024 (pp. 50455049). IEEE.
  55. Maman, B., Zeitler, J., Müller, M., and Bermano, A. H. (2025). Multi‑aspect conditioning for diffusion‑based music synthesis: Enhancing realism and acoustic control. IEEE Transactions on Audio, Speech and Language Processing, 33, 6881.
  56. Mathews, M. V. (1963). The digital computer as a musical instrument. Science, 142(3592), 553557.
  57. Mathews, M. V., Miller, J. E., Moore, F. R., Pierce, J. R., and Risset, J. C. (1969). The technology of computer music. The MIT Press.
  58. Miranda, E. R., and Wanderley, M. (2006). New digital musical instruments: Control and interaction beyond the keyboard (computer music and digital audio series). A‑R Editions, Inc.
  59. Moog, R. A. (1965). Voltage controlled electronic music modules. Journal of the Audio Engineering Society, 13(3), 200206.
  60. Narita, G., Shimizu, J., and Akama, T. (2023). GANStrument: Adversarial instrument sound synthesis with pitch‑ invariant instance conditioning. In IEEE International Conference on Acoustics, Speech and Signal Processing ICASSP 2023 (pp. 15). IEEE.
  61. Natsiou, A., and O’Leary, S. (2021). Audio representations for deep learning in sound synthesis: A review. In 18th IEEE/ACS International Conference on Computer Systems and Applications, AICCSA 2021, Tangier, Morocco, November 30–December 3, 2021 (pp. 18). IEEE.
  62. Nichols, C. (2002). The vBow: A virtual violin bow controller for mapping gesture to synthesis with haptic feedback. Organised Sound, 7, 215220.
  63. Passalenti, A., Paisa, R., Nilsson, N. C., Andersson, N. S., Fontana, F., Nordahl, R., and Serafin, S. (2019). No strings attached: Force and vibrotactile feedback in a virtual guitar simulation. In 2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR) (pp. 11161117).
  64. Puckette, M. S. (1996). Pure Data: Another integrated computer music environment. In Proceedings of the Second Intercollege Computer Music Concerts (pp. 3741).
  65. Renault, L., Mignot, R., and Roebel, A. (2022). Differentiable piano model for MIDI‑to‑audio performance synthesis. In 25th International Conference on Digital Audio Effects (DAFx20in22), Vienna, Austria.
  66. Richardson, N., Kothapalli, S., Onteddu, A., Kundavaram, R., and Talla, R. R. (2023). AI‑driven optimization techniques for evolving software architecture in complex systems. ABC Journal of Advanced Research, 12, 7184.
  67. Risset, J., and Mathews, M. V. (1969). Analysis of musical‑ instrument tones. Physics Today, 22(2), 2330.
  68. Roberts, R. A., and Mullis, C. T. (1987). Digital signal processing. Addison‑Wesley Longman Publishing Co., Inc.
  69. Rossing, T. D (Ed.). (2010). The science of string instruments. Springer.
  70. Schwarz, D. (2007). Corpus‑based concatenative synthesis. IEEE Signal Processing Magazine, 24(2), 92104.
  71. Selfridge, R., Moffat, D., Reiss, J., and Avital, E. (2024). Real‑time physical model of an aeolian harp. In 24th International Conference on Sound and Vibration 2017. London, UK: Institute of Acoustics.
  72. Serafin, S., Erkut, C., Kojs, J., Nordahl, R., and Nilsson, N. C. (2016). Virtual reality musical instruments: Guidelines for multisensory interaction design. In Proceedings of the Audio Mostly 2016 (pp. 266271). ACM.
  73. Serra, X., and Smith, J. (1990). Spectral modeling synthesis: A sound analysis/synthesis system based on a deterministic plus stochastic decomposition. Computer Music Journal, 14(4), 12.
  74. Shan, S., Hantrakul, L., Chen, J., Avent, M., and Trevelyan, D. (2022). Differentiable wavetable synthesis. In IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2022 (pp. 45984602). IEEE.
  75. Siao, Y. S., Chang, W. L., and Su, A. W. (2006). Analysis and transynthesis of solo erhu recordings using additive/subtractive synthesis: 120th Convention Spring Preprints 2006. In Audio Engineering Society ‑ 120th Convention Spring Preprints 2006 (pp. 17).
  76. Simionato, R., Fasciani, S., and Holm, S. (2024). Physics‑informed differentiable method for piano modeling. Frontiers in Signal Processing, 3, 2023.
  77. Sinclair, S. (2009). Audio‑haptic interaction with the digital waveguide bowed string. In Proceedings of the International Computer Music Conference (ICMC 2009), Montreal, Canada.
  78. Smith, J. O. (1992). Physical modeling using digital waveguides. Computer Music Journal, 16(4), 7491.
  79. Smith, J. O. (2002). Principles of digital waveguide models of musical instruments. In Applications of Digital Signal Processing to Audio and Acoustics (pp. 417466). Springer US.
  80. Smith, J. O. (2004). Virtual acoustic musical instruments: Review and update. Journal of New Music Research, 33(3), 283304.
  81. Smith, J. O. (2011). Spectral Audio Signal Processing. W3K.
  82. Tricco, A. C., Lillie, E., Zarin, W., O’Brien, K. K., Colquhoun, H., Levac, D., Moher, D., Peters, M. D. J., Horsley, T., Weeks, L., Hempel, S., Akl, E. A., Chang, C., McGowan, J., Stewart, L., Hartling, L., Aldcroft, A., Wilson, M. G., Garritty, C., … Straus, S. E. (2018). PRISMA extension for scoping reviews (PRISMA‑ScR): Checklist and explanation. Annals of Internal Medicine, 169(7), 467473.
  83. Tuominen, P. P. A., and Saarni, L. A. (2024). The use of virtual technologies with music in rehabilitation: A scoping systematic review. Frontiers Virtual Reality, 5.
  84. Turchet, L., Hamilton, R., and Çamci, A. (2021). Music in extended realities. IEEE Access: Practical Innovations, Open Solutions, 9, 1581015832.
  85. Uncini, A. (2022). Digital audio processing fundamentals (Vol. 21). Springer.
  86. Välimäki, V., and Erkut, C. (2003). Commuted waveguide synthesis of the clavichord. Computer Music Journal, 27(1), 7182.
  87. Välimäki, V., Penttinen, H., Knif, J., Laurson, M., and Erkut, C. (2004). Sound synthesis of the harpsichord using a computationally efficient physical model. EURASIP Journal on Advances in Signal Processing, 2004(7), 860718.
  88. van den Oord, A., Dieleman, S., Zen, H., Simonyan, K., Vinyals, O., Graves, A., Kalchbrenner, N., Senior, A. W., and Kavukcuoglu, K. (2016). WaveNet: A generative model for raw audio. In The 9th ISCA Speech Synthesis Workshop, SSW 2016 (pp. 125). ISCA.
  89. Villa‑Renteria, I., Wang, M. L., Shah, Z., Li, Z., Kim, S., Ramachandran, N., and Pilanci, M. (2025). Subtractive training for music stem insertion using latent diffusion models. In 2025 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2025 (pp. 15). IEEE.
  90. Villeneuve, J., and Leonard, J. (2019). Mass‑interaction physical models for sound and multi‑sensory creation: Starting anew. In SMC 2019 ‑ 16th Sound & Music Computing Conference, Malaga, Spain.
  91. Wessel, D., and Wright, M. (2002). Problems and prospects for intimate musical control of computers. Computer Music Journal, 26(3), 1122.
  92. Willemsen, S., Andersson, N., Serafin, S., and Bilbao, S. (2019). Real‑time control of large‑scale modular physical models using the Sensel Morph. In Proceedings of the 16th Sound and Music Computing Conference, Malaga, Spain: Sound and Music Computing Network.
  93. Willemsen, S., Paisa, R., and Serafin, S. (2020). Resurrecting the tromba marina: A bowed virtual reality instrument using haptic feedback and accurate physical modelling. In Proceedings of the 17th Sound and Music Computing Conference (pp. 300307). Torino, Italy: Axea sas/SMC Network.
  94. Willemsen, S., Serafin, S., Bilbao, S., and Ducceschi, M. (2020). Real‑time implementation of a physical model of the tromba marina. In Proceedings of the 17th Sound and Music Computing Conference, Proceedings of the Sound and Music Computing Conference (pp. 161168). Torino, Italy: Axea sas/SMC Network.
  95. Woodhouse, J., Politzer, D., and Mansour, H. (2021). Acoustics of the banjo: Measurements and sound synthesis. Acta Acustica, 5, 15.
  96. Worrall, K., Yin, Z., and Collins, T. (2024). Comparative evaluation in the wild: Systems for the expressive rendering of music. IEEE Transactions on Artificial Intelligence, 5(10), 52905303.
  97. Wu, Y., Manilow, E., Deng, Y., Swavely, R., Kastner, K., Cooijmans, T., Courville, A. C., Huang, C.‑Z. A., and Engel, J. H. (2022). MIDI‑DDSP: Detailed control of musical performance via hierarchical modeling. In The Tenth International Conference on Learning Representations, ICLR 2022. OpenReview.net.
  98. Wun, C.‑W., Horner, A., and Ayers, L. (2001). Perceptual wavetable matching for synthesis of musical instrument tones. In Proceedings of the 2001 International Computer Music Conference, ICMC 2001. Michigan Publishing.
  99. Yang, H. C., Lin, Y., and Su, A. (2020). A novel source filter model using LSTM/K‑means machine learning methods for the synthesis of bowed‑string musical instruments. In 148th Audio Engineering Society International Convention 2020. Virtual Vienna. Audio Engineering Society.
  100. Ye, Z., Xue, W., Tan, X., Liu, Q., and Guo, Y. (2023). NAS‑FM: Neural architecture search for tunable and interpretable sound synthesis based on frequency modulation. In Proceedings of the Thirty‑Second International Joint Conference on Artificial Intelligence, IJCAI 2023. (pp. 58695877). ijcai.org.
  101. Zhang, Z., and Akama, T. (2024). Hyperganstrument: Instrument sound synthesis and editing with pitch‑ invariant hypernetworks. In IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2024 (pp. 66406644). IEEE.
DOI: https://doi.org/10.5334/tismir.267 | Journal eISSN: 2514-3298
Language: English
Submitted on: Apr 8, 2025
Accepted on: Feb 24, 2026
Published on: Apr 13, 2026
Published by: Ubiquity Press
In partnership with: Paradigm Publishing Services
Publication frequency: 1 issue per year

© 2026 Yaozhong Zhang, Sebastian von Mammen, Christof Weiß, published by Ubiquity Press
This work is licensed under the Creative Commons Attribution 4.0 License.