Research

Paper

AI LLM March 17, 2026

A Semantic Timbre Dataset for the Electric Guitar

Authors

Joseph Cameron, Alan Blackwell

Abstract

Understanding and manipulating timbre is central to audio synthesis, yet this remains under-explored in machine learning due to a lack of annotated datasets linking perceptual timbre dimensions to semantic descriptors. We present the Semantic Timbre Dataset, a curated collection of monophonic electric guitar sounds, each labeled with one of 19 semantic timbre descriptors and corresponding magnitudes. These descriptors were derived from a qualitative analysis of physical and virtual guitar effect units and applied systematically to clean guitar tones. The dataset bridges perceptual timbre and machine learning representations, supporting learning for timbre control and semantic audio generation. We validate the dataset by training a variational autoencoder (VAE) on its latent space and evaluating it using human perceptual judgments and descriptor classifiers. Results show that the VAE captures timbral structure and enables smooth interpolation across descriptors. We release the dataset, code, and evaluation protocols to support timbre-aware generative AI research.

Metadata

arXiv ID: 2603.16682
Provider: ARXIV
Primary Category: cs.SD
Published: 2026-03-17
Fetched: 2026-03-18 06:02

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2603.16682v1</id>\n    <title>A Semantic Timbre Dataset for the Electric Guitar</title>\n    <updated>2026-03-17T15:42:53Z</updated>\n    <link href='https://arxiv.org/abs/2603.16682v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2603.16682v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>Understanding and manipulating timbre is central to audio synthesis, yet this remains under-explored in machine learning due to a lack of annotated datasets linking perceptual timbre dimensions to semantic descriptors. We present the Semantic Timbre Dataset, a curated collection of monophonic electric guitar sounds, each labeled with one of 19 semantic timbre descriptors and corresponding magnitudes. These descriptors were derived from a qualitative analysis of physical and virtual guitar effect units and applied systematically to clean guitar tones. The dataset bridges perceptual timbre and machine learning representations, supporting learning for timbre control and semantic audio generation. We validate the dataset by training a variational autoencoder (VAE) on its latent space and evaluating it using human perceptual judgments and descriptor classifiers. Results show that the VAE captures timbral structure and enables smooth interpolation across descriptors. We release the dataset, code, and evaluation protocols to support timbre-aware generative AI research.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.SD'/>\n    <published>2026-03-17T15:42:53Z</published>\n    <arxiv:comment>5 pages, 7 figures, 2 tables</arxiv:comment>\n    <arxiv:primary_category term='cs.SD'/>\n    <author>\n      <name>Joseph Cameron</name>\n    </author>\n    <author>\n      <name>Alan Blackwell</name>\n    </author>\n  </entry>"
}