<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article
  PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD with MathML3 v1.2 20190208//EN" "JATS-journalpublishing1-mathml3.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" article-type="research-article" dtd-version="1.2" xml:lang="en">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">JBDGM</journal-id>
<journal-id journal-id-type="nlm-ta">Jahrb Musikpsychol</journal-id>
<journal-title-group>
<journal-title>Jahrbuch Musikpsychologie</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Jahrb. Musikpsychol.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2569-5665</issn>
<publisher><publisher-name>PsychOpen</publisher-name></publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">jbdgm.225</article-id>
<article-id pub-id-type="doi">10.5964/jbdgm.225</article-id>
<article-categories>
<subj-group subj-group-type="heading"><subject>Short Papers</subject></subj-group>

<subj-group subj-group-type="badge">
<subject>Data</subject>
<subject>Code</subject>
<subject>Materials</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Detection of Clash of Keys in a Non-Dichotomous Task: Effects of Musical Genres, Instruments and Participants’ Expertise</article-title>
<trans-title-group xml:lang="de">
<trans-title>Erkennen von Tonartenkonflikten mittels nicht-dichotomer Aufgabe: Effekte von Musikgenres, Instrumenten und Teilnehmer*innen-Expertise</trans-title>
</trans-title-group>
<alt-title alt-title-type="right-running">Detection of Clash of Keys in a Non-Dichotomous Task</alt-title>
<alt-title specific-use="APA-reference-style" xml:lang="en">Detection of clash of keys in a non-dichotomous task: Effects of musical genres, instruments and participants’ expertise</alt-title>
</title-group>
<contrib-group>
	
<contrib contrib-type="author" corresp="yes"><contrib-id contrib-id-type="orcid" authenticated="false">https://orcid.org/0000-0002-6196-6488</contrib-id><name name-style="western"><surname>Wolf</surname><given-names>Anna</given-names></name><xref ref-type="corresp" rid="cor1">*</xref><xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
	<role vocab="credit" vocab-identifier="http://credit.niso.org/"
		vocab-term="Conceptualization"
		vocab-term-identifier="http://credit.niso.org/contributor-roles/conceptualization/"
		>Conceptualization</role>
	<role vocab="credit" vocab-identifier="http://credit.niso.org/"
		vocab-term="Data curation"
		vocab-term-identifier="http://credit.niso.org/contributor-roles/data-curation/"
		>Data curation</role>
	<role
		vocab="credit"
		vocab-identifier="http://credit.niso.org/"
		vocab-term="Formal Analysis"
		vocab-term-identifier="http://credit.niso.org/contributor-roles/formal-analysis/"
		>Formal analysis</role>
	<role vocab="credit" vocab-identifier="http://credit.niso.org/"
		vocab-term="Methodology"
		vocab-term-identifier="http://credit.niso.org/contributor-roles/methodology/"
		>Methodology</role>
	<role
		vocab="credit"
		vocab-identifier="http://credit.niso.org/"
		vocab-term="Project administration"
		vocab-term-identifier="http://credit.niso.org/contributor-roles/project-administration/"
		>Project administration</role>
	<role
		vocab="credit"
		vocab-identifier="http://credit.niso.org/"
		vocab-term="Supervision"
		vocab-term-identifier="http://credit.niso.org/contributor-roles/supervision/"
		>Supervision</role>
	<role
		vocab="credit"
		vocab-identifier="http://credit.niso.org/"
		vocab-term="Visualization"
		vocab-term-identifier="http://credit.niso.org/contributor-roles/visualization/"
		>Visualization</role>
	<role
		vocab="credit"
		vocab-identifier="http://credit.niso.org/"
		vocab-term="Writing – original draft"
		vocab-term-identifier="http://credit.niso.org/contributor-roles/writing-original-draft/"
		>Writing – original draft</role>
	<role
		vocab="credit"
		vocab-identifier="http://credit.niso.org/"
		vocab-term="Writing – review &amp; editing"
		vocab-term-identifier="http://credit.niso.org/contributor-roles/writing-review-editing/"
		>Writing – review &amp; editing</role>
</contrib>
	
	
<contrib contrib-type="author"><name name-style="western"><surname>Wüst</surname><given-names>Bastian</given-names></name><xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
	<role vocab="credit" vocab-identifier="http://credit.niso.org/"
		vocab-term="Conceptualization"
		vocab-term-identifier="http://credit.niso.org/contributor-roles/conceptualization/"
		>Conceptualization</role>
	<role vocab="credit" vocab-identifier="http://credit.niso.org/"
		vocab-term="Investigation"
		vocab-term-identifier="http://credit.niso.org/contributor-roles/investigation/"
		>Investigation</role>
	<role vocab="credit" vocab-identifier="http://credit.niso.org/"
		vocab-term="Methodology"
		vocab-term-identifier="http://credit.niso.org/contributor-roles/methodology/"
		>Methodology</role>
	<role vocab="credit" vocab-identifier="http://credit.niso.org/"
		vocab-term="Resources"
		vocab-term-identifier="http://credit.niso.org/contributor-roles/resources/"
		>Resources</role>
</contrib>
	


	<contrib contrib-type="reviewer"><name name-style="western"><surname>Siedenburg</surname><given-names>Kai</given-names></name></contrib>
	
	<contrib contrib-type="reviewer"><name name-style="western"><surname>Lange</surname><given-names>Elke</given-names></name></contrib>


<aff id="aff1"><label>1</label><institution content-type="dept">Department of Music Education and Church Music</institution>, <institution>University of Music FRANZ LISZT Weimar</institution>, <addr-line><city>Weimar</city></addr-line>, <country country="DE">Germany</country></aff>

</contrib-group>
<author-notes>
<corresp id="cor1"><label>*</label>Department of Music Education and Church Music, University of Music FRANZ LISZT Weimar, Platz der Demokratie 2/3, 99423 Weimar, Germany. <email xlink:href="anna.wolf@hfm-weimar.de">anna.wolf@hfm-weimar.de</email></corresp>
</author-notes>
<pub-date date-type="pub" publication-format="electronic"><day>25</day><month>02</month><year>2026</year></pub-date>
<pub-date pub-type="collection" publication-format="electronic"><year>2026</year></pub-date>
<volume>34</volume>
<elocation-id>e225</elocation-id>
<history>
<date date-type="received">
<day>04</day>
<month>04</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>19</day>
<month>01</month>
<year>2026</year>
</date>
</history>
<permissions><copyright-year>2026</copyright-year><copyright-holder>Wolf &amp; Wüst</copyright-holder><license license-type="open-access" specific-use="CC BY 4.0" xlink:href="https://creativecommons.org/licenses/by/4.0/"><ali:license_ref>https://creativecommons.org/licenses/by/4.0/</ali:license_ref><license-p>This is an open access article distributed under the terms of the Creative Commons Attribution 4.0 International License, CC BY 4.0, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p></license></permissions>
<abstract>
<p>When listening to music, people are able to separate the various parts or streams that the respective piece is made of, depending on their musical experience and aural training. Specific musical expertise affects these analytical listening skills, but different studies have shown that differences between experts and amateurs are smaller than often anticipated and depend heavily on the adequacy of the task for amateurs. The present replication study of Kopiez and Platz (2009) has investigated in a convenience sample (<italic>N</italic> = 97) whether a presumably obvious clash of keys between solo and accompaniment can be detected. In an incomplete study design, participants listened to two pieces of music (jazz and classical) with solos each played by one of two instruments (trumpet and saxophone) in either a clashing or fitting condition. Participants' musical training and perceptual abilities showed a medium correlation with the harmoniousness ratings difference between the clashing and fitting stimuli. Overall, the clashing version of the classical piece was rated as less harmonious than the jazz piece. These results are in line with similar research and raise questions concerning the appropriate research method to investigate the perception of participants with various degrees of expertise.</p>
</abstract><trans-abstract xml:lang="de">
<p>Beim Hören von Musik sind Menschen, je nach musikalischer Erfahrung und Gehörbildung, unterschiedlich gut in der Lage die verschiedenen Stimmen oder Ströme im jeweiligen Stück zu trennen. Spezifische musikalische Expertise wirkt sich auf diese analytischen Hörfähigkeiten aus, wobei verschiedene Studien zeigen, dass die Unterschiede zwischen Experten und Amateuren oft geringer sind als angenommen und wiederum stark von der Eignung der Aufgabe für die Amateure abhängen. In der vorliegenden Replikationsstudie von Kopiez und Platz (2009) wurde untersucht, ob ein vermeintlich offensichtlicher Tonartenclash zwischen Solo und Begleitung in zwei Musikstücken von Teilnehmer*innen einer Gelegenheitsstichprobe (<italic>N</italic> = 97) erkannt wurde. Im Rahmen eines unvollständigen Designs haben die Teilnehmer*innen zwei Musikstücke gehört (Jazz und Klassik), wobei die Soli jeweils von einem von zwei Instrumenten (Trompete und einem Saxofon) in entweder passenden oder „kollidierenden“ Tonarten gespielt wurden. Die Expertise und die musikalischen Wahrnehmungsfähigkeiten der Teilnehmer korrelierten mittelstark mit der Differenz der Bewertungen für kollidierende und passende Stimuli. Insgesamt wurde die kollidierende Version des klassischen Stücks als weniger stimmig eingestuft als das Jazzstück. Diese Ergebnisse entsprechen den Erkenntnissen ähnlicher Untersuchungen und werfen Fragen zur passenden Forschungsmethodik für die Wahrnehmung bei Personen mit unterschiedlicher musikalischer Expertise auf.</p></trans-abstract>
<kwd-group kwd-group-type="author"><kwd>analytical listening</kwd><kwd>bitonality</kwd><kwd>harmoniousness</kwd><kwd>perceptual abilities</kwd><kwd>replication study</kwd></kwd-group>
<kwd-group kwd-group-type="translator" xml:lang="de"><kwd>Analytisches Hören</kwd><kwd>Bitonalität</kwd><kwd>Stimmigkeit</kwd><kwd>Musikalische Wahrnehmungsfähigkeiten</kwd><kwd>Replikationsstudie</kwd></kwd-group>
</article-meta>
</front>
<body>
	<sec sec-type="intro"><title/>
<p>“Write some ditty in one key, write the accompaniment in another, and <italic>voilà</italic>— something that sounds as ‘bad’ as the most studiously atonal utterance of a real, hard-working composer” (<xref ref-type="bibr" rid="r11">Harrison, 1997</xref>, pp. 393–394). Such reported ridicule of bitonality is omnipresent in the music theoretical discourse. As pointedly condensed in the introductory quote, it is to a large degree rooted in the fact that bitonality kept all features of tonal music and only changed the uniformity of one musical key in one musical piece to two—or more, in the case of polytonal music (<xref ref-type="bibr" rid="r31">Whittall, 2001</xref>).</p>
<p>Famous composers who employed bitonality as compositional techniques were, for example, Charles Ives, Benjamin Britten, Darius Milhaud, and Igor Stravinsky. Within their compositions they used various degrees of bitonality, such as a single melodic line in a different key, two keys for the two voices in a duet (e.g., in Peter Grimes by Britten), or as a continuous stylistic feature for a whole, albeit often rather short piece (e.g., the Saudades do Brazil by Milhaud).</p>
<p>Following the Gestalt laws to the concept of bitonality, there are several indications that might suggest a one-to-one mapping from the compositional idea using two keys to the listener’s capacity to hear those two keys. The law of common fate—regarding the tonic and the harmonic and melodic development within one voice as a common fate—and the law of the “good Gestalt” (and maybe even the laws of proximity and similarity) should, certainly depending on the exact piece of music, predict the possibility of listeners to perceive and appreciate the two voices in their different keys (<xref ref-type="bibr" rid="r30">Wertheimer, 1923</xref>, translation by <xref ref-type="bibr" rid="r7">Ellis, 1938</xref>). However, more recent findings have shown that, in comparison to visual perception, for which the Gestalt laws have primarily been developed, it is only rarely possible to listen to and follow two separate auditory streams at the same time (<xref ref-type="bibr" rid="r25">Oehler, 2014</xref>; <xref ref-type="bibr" rid="r4">Bregman, 1990</xref>, pp. 524–525). Therefore, we assume a rather intertwined percept as a predominant mode of perception, where the scene analysis skills are dependent, e.g., on one’s musical experience and specific aural skills (e.g., <xref ref-type="bibr" rid="r9">Hake et al., 2024</xref>).</p>
<p>Aside from the different notions on bitonality as either a “distinctly mechanical” (<xref ref-type="bibr" rid="r31">Whittall, 2001</xref>) or skillful compositional technique, there are many findings focusing on the disparity between ideas embedded in musical concepts (composed music) and ideas experienced in musical percepts (heard music). Investigations into the latter have surfaced a variety of unexpected findings: Concerning the tonal center of a piece, listeners are not influenced in their rating whether this piece remains in the originally tonal center or modulates towards another key (<xref ref-type="bibr" rid="r6">Cook, 1987</xref>). Maybe even more surprising, our preference for classical music does not change when form parts of this piece are presented in the order intended by the composer of randomly rearranged into a nonsensical order (<xref ref-type="bibr" rid="r13">Karno &amp; Konečni, 1992</xref>). And with regards to manipulations in a piece’s tonality, <xref ref-type="bibr" rid="r16">Lalitte et al. (2009)</xref> found in their study on tonal (i.e., original) and atonal (i.e., artificially created) versions of Beethoven sonatas that participants followed criteria of musical style and presence/absence of tonality and therefore, respectively, rating the tonal excerpts on the one hand and the atonal excerpts on the other hand as very similar. However, when segmenting the piece and reporting their arousal, these implicit ratings followed the similar structure within the dyad of a tonal and the thereof created atonal piece. Music training was no explaining factor for participants’ ratings here.</p>
<p>More specific to the perception of bi- and polytonality, <xref ref-type="bibr" rid="r29">Thompson and Mor (1992)</xref> adapted the probe tone method and captured the tonal perception in bitonal pieces: Responses matched the combined key profiles of the utilized keys, the prevalent keys in the compositions were simultaneously conveyed to participants and the authors were even able to rate which key stronger influenced participants’ key perception. This study is connected to a previous study by <xref ref-type="bibr" rid="r15">Krumhansl and Schmuckler (1986)</xref>, who, largely unsuccessfully, tried to manipulate the perception of the key profile by introducing them dichotically using the Petroushka passage. Their results also point to a perceptual fusion of both keys, which is so predominant in the perception that it was impossible to only pay attention to one key and ignore the other. Both studies were conducted with musically trained participants with a usual lower limit of at least five years of music training.</p>
<p>In addition to these studies on music theoretical concepts and psychological percepts, one can find a variety of findings with surprising results on human music perception. On a basic level, <xref ref-type="bibr" rid="r8">Fricke (2014)</xref> offers a concise and numbers-based review of the listener’s tolerance for intonational deviations in music. Listeners tend to tolerate (or “zurech<?glue?>thören”<xref ref-type="fn" rid="fn1"><sup>1</sup></xref><fn id="fn1"><label>1</label>
<p>Various translations exist from „harmonizing in hearing“ (<xref ref-type="bibr" rid="r14">Kopiez &amp; Platz, 2009</xref>, p. 329), “shift[ing] toward the intended pitch of each tone” (<xref ref-type="bibr" rid="r27">Parncutt &amp; Hair, 2018</xref>, p.491), or “straightened out hearing” (<xref ref-type="bibr" rid="r22">Mazzola, 2017</xref>, p. 1353).</p></fn>) deviations of ~20 cents in most situations, which can increase to up to 45 cents, depending on the circumstances (length of tone, melodic tension, or passing vs. target note). <xref ref-type="bibr" rid="r18">Larrouy-Maestri and Morsomme (2014)</xref> concur with these findings from a production perspective: Even professional singers deviate about 30 cents (and more, when singing with lyrical technique) in singing production tasks. A following study taking perception into account quantifies listener’s tolerance at roughly 25 cents with a high retest reliability of <italic>r</italic> = .88 (<xref ref-type="bibr" rid="r17">Larrouy-Maestri, 2018</xref>). Differences between experts and amateurs were notably large in this study and are further corroborated by the results from the mistuning perception test. The test relies on stimuli with mistunings between the vocal part and its accompaniment of 10 and 100 cents in either direction (<xref ref-type="bibr" rid="r19">Larrouy-Maestri et al., 2019</xref>). The hereby measured mistuning perception skills correlated with self-reported musical training and perceptual abilities with about <italic>r</italic> = .4. These results are especially well founded due to the careful test development, which not only addressed the instrument’s reliability, but also its concurrent and discriminatory construct validity. In contrast, this mistuning perception test correlated with the PROMS tuning test (<xref ref-type="bibr" rid="r20">Law &amp; Zentner, 2012</xref>) only with <italic>r</italic> = .25. Larrouy-Maestri et al. explain this divergence with a dissimilar item design, where Law and Zentner compared a mistuned chord with a correctly tuned reference chord enabling participants to use their echoic memory and thereby testing a partly different ability.</p>
<p>The effect of enculturation and training is also addressed in a review by <xref ref-type="bibr" rid="r26">Omigie et al. (2014)</xref>: Considering music exposure and deliberate music training, the authors state that listeners’ tolerance and liking for dissonant music must be separated because a higher tolerance for dissonance, induced by mere exposure and/or training, does not directly lead to a higher liking of such music.</p>
<p>Differences in music perception and analytical listening skills between experts and amateurs are by far not as clear-cut as a first assumption might seem. The seminal paper by <xref ref-type="bibr" rid="r3">Bigand and Poulin-Charronnat (2006)</xref> especially focuses on declarative language, which is learnt by musicians during their musical training, engagement with music notation and even more deliberate in music theory and ear training classes. “Untrained listeners”, as labeled by Bigand and Poulin-Charronat, rarely possess such knowledge and are destined to fail at such tasks; however, when the tasks use fair language, especially using same-different-paradigms, we cannot observe skill or perceptual differences between trained and untrained listeners. In their investigation of melodic symmetry in melodies, <xref ref-type="bibr" rid="r23">Mongoven and Carbon (2017)</xref> found accuracy levels of about 36–50% (compared to a chance level of 33%), depending on the stimuli, and registered no differences in symmetry detection for participants with varying degrees of musical training, albeit in a small sample size and with a very rough measure of musical training.</p>
<p>Investigating the specific perception—or rather detection—of bitonality in music, the results by <xref ref-type="bibr" rid="r32">Wolpert (2000)</xref> were rather clear: While musicians reliably detect bitonality in music without any instruction or hint (overall detection of 100%), non-musicians recognize it to a much lower degree (overall detection of 40%). In a first replication study, <xref ref-type="bibr" rid="r14">Kopiez and Platz (2009)</xref> also took listening expertise, focus of attention (hinting at the fit between melody and accompaniment—or not) and musical style into account. They replicated the overall large difference between experts and non-experts—but at a much lower overall level (experts: 58% vs. amateurs: 18%)—and found style-specific disparities with highest detection rates in classical music (50%) and lowest detection rates in jazz (23%). Participants who were directly instructed to listen to the fit between melody and accompaniment scored higher (62%) than those who were not (24%). Discussing these results, the authors emphasize their results and state, “that hearing is not simply ‘inverse acoustics’ or ‘inverse music theory’” and more “a kind of ‘harmonizing in hearing’ (Zurechthören)” (p. 329).</p>
<p><xref ref-type="bibr" rid="r10">Hamamoto et al. (2010)</xref> expanded their study to ecological stimuli from the <italic>Saudades do Brasil</italic> by Darius Milhaud (interestingly, also in recomposed ‘monotonal’ versions as stimuli) and employed, among other tasks, an indirect rating task using scales for likeability, correctness and pleasantness. Monotonal stimuli were higher rated in correctness and pleasantness, although not in likeability, with larger differences in rating for musicians than non-musicians. Furthermore, after completing a rather fast training task to learn how to detect bitonality within their study (as well as the exposure from the previous listening tasks), non-musicians improved markedly and performed similarly to musicians, “once they have acquired the appropriate vocabulary” (p. 442). Their free response tasks, on the other hand, showed group differences between musicians (74%) and non-musicians (20%) in line with previous results.</p>
<p>The following replication study was a conceptual replication and introduced some changes compared to <xref ref-type="bibr" rid="r14">Kopiez and Platz (2009)</xref>: We collected data online, adapted the rating task by <xref ref-type="bibr" rid="r10">Hamamoto et al. (2010)</xref>, and investigated an effect of musical instrument/timbre in addition to genre. Our research questions and hypotheses were the following:</p>
<list id="L1" list-type="order">
<list-item>
<p>Is the clash of keys perceivable in a nondirected listening task using adjective rating scales addressing the stimuli’s harmoniousness? We hypothesize that this study will replicate previous findings and that the clash of keys will lead to a much lower rating than the monotonal version.</p></list-item>
<list-item>
<p>We hypothesize that we will replicate differences in rating between experts and amateurs found; however, those could be large effects as observed by <xref ref-type="bibr" rid="r32">Wolpert (2000)</xref>, <xref ref-type="bibr" rid="r14">Kopiez and Platz (2009)</xref> and the free response task by <xref ref-type="bibr" rid="r10">Hamamoto et al. (2010)</xref> or possibly smaller as found in the rating data by <xref ref-type="bibr" rid="r10">Hamamoto et al. (2010)</xref>.</p></list-item>
<list-item>
<p>We hypothesize that the participants’ self-reported music perceptual abilities as measured by the respective subscale of the Goldsmiths Musical Sophistication Index (Gold-MSI; <xref ref-type="bibr" rid="r24">Müllensiefen et al., 2014</xref>; <xref ref-type="bibr" rid="r28">Schaal et al., 2014</xref>) are associated with the participants’ ratings: on the one hand simply due to connections between this subscale and the various other constructs measured by the Gold-MSI (such as musical training), on the other hand because a self-report of one’s perceptual abilities should theoretically correlate with a task indirectly measuring music perception skills.</p></list-item>
</list></sec>
<sec sec-type="methods"><title>Method</title>
<sec><title>Design</title>
<p>The study followed a 2 × 2 × 2 study design with missing values by design. Participants listened to <italic>2 pieces</italic>, a classical and a jazz piece. The solo parts were played by <italic>2 instruments</italic>, a trumpet and an alto saxophone. We produced <italic>2 versions</italic> where the solo and the accompaniment were in the same key (“fitting”) or where the accompaniment was a whole tone (200 cents) too low (“clashing”). To avoid carryover effects, participants listened to four versions overall: They listened to both pieces played by both instruments. We counterbalanced the design with regards to the fitting/clashing version so that participants were only presented with either the fitting or the clashing version (and not both) to avoid speculations when hearing the same piece played by the same instrument for a second time. Information on the musical stimuli can be found in the Data Availability section.</p></sec>
<sec sec-type="subjects"><title>Participants</title>
<p>A total of <italic>N</italic> = 97 participants took part in the experiment (57 female, 40 male, 0 non-binary). Their mean age was <italic>M</italic> = 34.1 years (<italic>SD</italic> = 15.7) and most of them were students (40%) or had an employment relationship (45%). Overall, two thirds of the students (an absolute 27% of the 40%) were studying music, 43% were amateur musicians and 6% were working in the music business; none of the above applied to 24%.</p>
<p>The overall score in the Perceptual Abilities subscale (nine items, theoretical range = [9, 63]) of the Gold-MSI (<xref ref-type="bibr" rid="r24">Müllensiefen et al., 2014</xref>) was <italic>M</italic> = 50.1 with a <italic>Mdn</italic> = 52 and an <italic>SD</italic> = 8.18. Compared to the German speaking sample (<xref ref-type="bibr" rid="r28">Schaal et al., 2014</xref>), our sample’s mean was four scale points above the 50<sup>th</sup> percentile and ranged in the 70th percentile, the self-assessed perceptual abilities were clearly above average.</p></sec>
<sec><title>Stimuli and Questionnaire</title>
<p>The stimuli originated from two pieces: 1) the Haydn trumpet concerto (Hob. Vlle/1, second movement “Andante”, bars 8–16) and 2) the jazz standard “Take Five” by the Dave Brubeck quartet (bars 1–12). Both solo parts were played by a trumpet as well as an alto saxophone in the original key. The chosen instruments are commonly used in either musical style while clearly differing from each other with respect to their timbre. The accompaniment for both pieces was produced on a Yamaha Clavinova CLP-370 with the sound of a Grand Piano, which was easily pitched from the correct key to a key a whole tone (200 cents) lower (Classical piece: A flat major &amp; G flat major; Jazz piece: E flat minor &amp; D flat minor).</p>
<p>After listening to each musical piece, participants rated their impression using the items elated (“beschwingt”), harmonious (“harmonisch”), solemn (“getragen”), off-key (“schief”), soft (“sanft”), cheerful (“heiter”), coherent (“stimmig”) and stressed (“gestresst”) on a 5-point scale (1 = <italic>does not apply at all</italic> to 5 = <italic>applies completely</italic>). The three items harmonious (“harmonisch”), off-key (“schief”, as an inverted item) and coherent (“stimmig”) were the a priori target items for the further statistical analysis, which we examined in a exploratory factor analysis (see Results).</p></sec>
<sec><title>Procedure</title>
<p>This study was conducted online using the platform soscisurvey.de (<xref ref-type="bibr" rid="r21">Leiner, 2024</xref>). Participants were welcomed and referred to the anonymity and voluntariness of the study. Using an audio example, participants could adjust the volume to a comfortable and normal level when listening to music. They were then presented with the four stimuli and rated their impression after each presentation. We asked for their demographic information on the second last page as well as whether they knew either of the musical pieces. On the last page participants replied to the Perceptual Abilities subscale from the Gold-MSI and specified the kind of speaker they used (tablet/phone loudspeaker, computer speaker, headphones or stereo equipment). They indicated how well they were able to hear the stimuli with regards to the audio examples’ quality and the circumstances during their participation regarding their attention and distractions (<italic>M</italic> = 4.2 and <italic>M</italic> = 4.4; <italic>SDs</italic> = 0.7; range = [1; 5]). Finally, participants were thanked for their participation.</p></sec></sec>
<sec sec-type="results"><title>Results</title>
<sec><title>Exploratory Factor Analysis</title>
<p>Compared to <xref ref-type="bibr" rid="r32">Wolpert (2000)</xref> and <xref ref-type="bibr" rid="r14">Kopiez and Platz (2009)</xref> we approached the participants’ perception using a rating task with eight adjective ratings. Five items (<italic>elated, solemn, soft, cheerful</italic> and <italic>stressed</italic>) served as presumed distractor items and were assumed to be irrelevant for further analyses, three items (<italic>harmonious, off-key</italic> as an inverted item and <italic>coherent</italic>) were our presumed target items. Using an exploratory factor analysis, we investigated the factor structure within the full item set.</p>
<p>The Kaiser-Meyer-Olkin criterion for this data set was <italic>KMO</italic> = .80 and therefore adequate or meritorious (<xref ref-type="bibr" rid="r12">Kaiser &amp; Rice, 1974</xref>) for a factor analysis. Both a scree plot as well as the Kaiser-Guttman criterion suggest a solution with two factors.</p>
<p>We employed an exploratory factor analysis using a maximum-likelihood estimation. The axes were rotated with a promax rotation, which allowed for correlations between factors and tried to reach near-zero correlations for every item with all but one factor. This option favors simplicity, but allows for relations between factors, which is plausible content-wise with this specific set of items that should all function as descriptors for musical stimuli (<xref ref-type="bibr" rid="r1">Abell et al., 2009</xref>). The found model does not differ from the empirical data, χ<sup>2</sup>(28) = 22.04, <italic>p</italic> = .17, and explains 64% of the variance in the data.</p>
<p>All three target items (<italic>harmonious, off-key</italic> as an inverted item and <italic>coherent</italic>) load on the first factor with loadings between .89 and .91. The next-highest loading is .55 for the item <italic>stressed</italic>, however, due to its much lower loading, its different content and since it was not part of the presumed a priori target items, we decided not to include it in the mean score for further analyses. The internal consistency of the three target items was α = .94 and would decrease to .90 or .91 if one of the items were dropped. The mean score of these items (range 1 to 5) constituted the dependent variable and described the perceived harmoniousness in the stimuli.</p></sec>
<sec><title>Detection of the Clash of Keys: Descriptive Results</title>
<p>Participants’ ratings differed largely between the clashing and fitting stimuli. The distributions of participants’ responses for the eight conditions are displayed in <xref ref-type="table" rid="t1">Table 1</xref>, more detailed significance tests and effect sizes are reported with the ANOVA.</p>
<table-wrap id="t1" position="anchor" orientation="portrait">
<label>Table 1</label><caption><title>Descriptive Statistics for the Analysis of Variance</title></caption>
<table frame="hsides" rules="groups">
<col width="40%" align="left"/>
<col width="15%"/>
<col width="15%"/>
<col width="15%"/>
<col width="15%"/>
<thead>
<tr>
<th rowspan="2" scope="rowgroup"/>
<th colspan="2" scope="colgroup">Fitting<hr/></th>
<th colspan="2" scope="colgroup">Clashing<hr/></th>
</tr>
<tr>
<th scope="colgroup"><italic>M</italic></th>
<th><italic>SD</italic></th>
<th><italic>M</italic></th>
<th><italic>SD</italic></th>
</tr>
</thead>
<tbody>
<tr>
<td>Classical: Saxophone</td>
<td align="char" char=".">4.09</td>
<td align="char" char=".">0.880</td>
<td align="char" char=".">2.05</td>
<td align="char" char=".">0.991</td>
</tr>
<tr>
<td>Classical: Trumpet</td>
<td align="char" char=".">4.23</td>
<td align="char" char=".">0.730</td>
<td align="char" char=".">1.92</td>
<td align="char" char=".">1.003</td>
</tr>
<tr>
<td>Jazz: Saxophone</td>
<td align="char" char=".">4.63</td>
<td align="char" char=".">0.438</td>
<td align="char" char=".">3.13</td>
<td align="char" char=".">1.415</td>
</tr>
<tr>
<td>Jazz: Trumpet</td>
<td align="char" char=".">4.26</td>
<td align="char" char=".">0.736</td>
<td align="char" char=".">3.08</td>
<td align="char" char=".">1.264</td>
</tr>
</tbody>
</table>
</table-wrap></sec>
<sec><title>Analysis of Variance</title>
<p>We investigated the difference in harmoniousness perception between the eight conditions (2 pieces × 2 instruments × 2 versions) using an analysis of variance with two- and three-way interactions. Two relevant and significant stimuli-related effects appeared, a large difference between the fitting/clashing versions, <italic>F</italic>(1, 95) = 71.3, <italic>p</italic> &lt; .001, <inline-formula><mml:math><mml:mstyle scriptminsize="0pt"><mml:mrow><mml:msubsup><mml:mi mathvariant="normal">η</mml:mi><mml:mi mathvariant="normal">p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:mstyle></mml:math></inline-formula> = .16, and a medium difference between the two pieces, <italic>F</italic>(1, 95) = 24.8, <italic>p</italic> &lt; .001, <inline-formula><mml:math><mml:mstyle scriptminsize="0pt"><mml:mrow><mml:msubsup><mml:mi mathvariant="normal">η</mml:mi><mml:mi mathvariant="normal">p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:mstyle></mml:math></inline-formula> = .06 (see <xref ref-type="fig" rid="f1">Figure 1</xref>). There were no relevant differences with regards to the instrument or for any of the interactions between the independent variables.</p>
	
	<fig id="f1" position="anchor" fig-type="figure" orientation="portrait"><label>Figure 1</label><caption>
<title>Distribution of Harmoniousness Ratings Within the Eight Conditions</title></caption><graphic xlink:href="jbdgm.225-f1" position="anchor" orientation="portrait"/></fig>

</sec>
<sec><title>Correlational Analysis</title>
<p>Finally, we derived a new dependent variable to estimate whether a participant’s perceptual ability and years of musical practice were related to their difference in rating between fitting and clashing stimuli. Due to an incomplete study design, every participant listened to two fitting and clashing stimuli each, for example, classical/trumpet/clashing and jazz/saxophone/fitting as well as classical/saxophone/fitting and jazz/trumpet/clashing. This was done to avoid carryover effects, such as participants speculating on the purpose of the study, when hearing the same piece with the same instrument for a second time.</p>
<p>We therefore calculated the following difference score with a mean of two stimuli rating scores each, where genres and instruments were counterbalanced:</p>
<p>mean(fitting stimuli) – mean(clashing stimuli)</p>
<p>This difference score correlated positively with the participants’ years of instrumental or voice practice and their perceptual abilities with a medium effect size each, both for parametric (years practice: <italic>r</italic> = .33, <italic>p</italic> &lt; .001; perceptual abilities: <italic>r</italic> = .49, <italic>p</italic> &lt; .001) as well as non-parametric tests (years practice: Spearman’s ρ = .39, <italic>p</italic> &lt; .001; perceptual abilities: ρ = .49, <italic>p</italic> &lt; .001; see also <xref ref-type="fig" rid="f2">Figure 2</xref>; the non-parametric correlation coefficient was included due to the visually detected outliers in <xref ref-type="fig" rid="f2">Figure 2A</xref>).</p>
	
	<fig id="f2" position="anchor" fig-type="figure" orientation="portrait"><label>Figure 2</label><caption>
<title>Distribution of Harmoniousness Ratings With Regards to the Participants’ Perceptual Abilities</title></caption><graphic xlink:href="jbdgm.225-f2" position="anchor" orientation="portrait"/></fig>


</sec></sec>
<sec sec-type="discussion"><title>Discussion</title>
<p>The present study investigated, using an expertise-unrelated task, the impression of participants when listening to bitonal pieces music with fitting or clashing keys for melody and accompaniment. The dependent variable consisted of three items to be judged on rating scales that represent an overall rating in harmoniousness. We observed a large difference in rating between the fitting vs. clashing stimuli (Hypothesis 1 confirmed) and a medium difference between the classical and the jazz piece. The years of accumulated daily music practice and the perceptual abilities were correlated with the responses to the extent of a medium effect size each (Hypotheses 2 and 3 confirmed).</p>
<p>We clearly anticipated differences in ratings between stimuli with a clash of keys compared to fitting keys from the results of previous research and can clearly corroborate this finding in our study. More specifically concerning genre difference, the results from <xref ref-type="bibr" rid="r14">Kopiez and Platz (2009)</xref> in either genre were replicated in our study: The rated harmoniousness in the clashing versions was much lower for the classical piece than the jazz piece (<italic>d</italic> = 0.96), while the difference between the fitting versions was smaller (<italic>d</italic>&nbsp;=&nbsp;0.41). While there was no overall effect of instrument, the latter effect originated mostly from different ratings for the saxophone versions, however, these post-hoc explanations are only a numerical tendency and must be interpreted with caution. Whereas the saxophone was originally developed for use in classical orchestras and marching bands, our current listening habits link it much more to jazz than classical music, which might serve as a probable explanation for its higher ratings in the jazz stimulus compared to the classical stimulus. While both selected stimuli are representations of their respective genre, the generalizability of results stays debatable, especially since each genre was only represented by one musical piece. Further studies might use different stimuli or, even better, multiple stimuli from several genres to increase the stability of the measurement and better grasp the diversity of musical culture.</p>
<p>This study found a rating difference between musicians and non-musicians similar to the outcomes in the clash of keys rating task by <xref ref-type="bibr" rid="r10">Hamamoto et al. (2010)</xref> and the correlations between several musical sophistication scales and the mistuning perception test by <xref ref-type="bibr" rid="r19">Larrouy-Maestri et al. (2019)</xref>. As hypothesized, the self-reported perceptual abilities showed a relevant correlation with the harmoniousness rating and this was, in comparison, significantly larger than the correlation with someone’s years of musical training, which was the sole criterion for musical expertise in the previous clash of keys studies. Evidently, music-specific perceptual abilities, which can be cultivated, among others, by playing an instrument, but also by other musical activities, seem more relevant than music training to detect various features in a piece of music.</p>
<p>As mentioned in the introduction, in the study by <xref ref-type="bibr" rid="r16">Lalitte et al. (2009)</xref> music training did not explain participant behaviour in the segmentation task or the observed arousal. Furthermore, Cook even argues that when instructing musicians and non-musicians with vague tasks, they are “engaged in different tasks” (<xref ref-type="bibr" rid="r5">Cook, 1994</xref>, p. 69), most obvious in experts’ active and frequently-used technical language in comparison to amateurs’ lower skills in such a technical jargon, that is, possessing a probably less active (and more passive) vocabulary and, lastly, very different habits and degrees of freely describing and—even more—criticizing music. Studies on the perception of bitonality using the probe tone paradigm cannot add to the body of evidence here, since <xref ref-type="bibr" rid="r15">Krumhansl and Schmuckler (1986)</xref> and <xref ref-type="bibr" rid="r29">Thompson and Mor (1992)</xref> only collected data from musically trained participants.</p>
<p>Grouping the evidence on the perception of clash of keys so far, there seem to be three major approaches for operationalisation using a) a free response task, possibly varied with a directed or non-directed instruction and focusing on a declarative detection of a clash of keys, b) a procedure derived from the probe tone paradigm, i.e., a twelve-fold presentation of each stimulus with a potential for carry-over effects, and c) continuous ratings using adjective scales focusing on a general and non-dichotomous, but continuously scalable impression, which lead to less distinct results and do not show “what people do hear” (<xref ref-type="bibr" rid="r32">Wolpert, 2000</xref>, p. 225). Concerning the participants’ strategies when replying to a free response task, it might be insightful to investigate the occurring thoughts in musicians and non-musicians in a qualitative study, for example whether the detection of a clash of keys is a sudden realisation or the result of an incremental analysis forming over time. In either scenario, musicians tend to be at an advantage by possessing instruments for analysis—even considering basic knowledge about keys and typically dissonant intervals, such as the major second—and by holding more confidence in voicing sudden and rather implausible suspicions, such as the audacity of music researchers in producing an experimental stimulus employing a concept as peculiar as bitonality. An operationalisation using a non-dichotomous rating procedure might introduce a fairer study setting for participants with less musical expertise, as has been shown here and by Hamamoto et al., 2010.</p>
<p>At the same time, missing clashes of keys and not being able to verbalise such a phenomenon, as shown in studies with a free response task, might even be commonplace: <xref ref-type="bibr" rid="r2">Anglada-Tort and Müllensiefen (2017)</xref> presented the identical interpretation of a section of a Bruckner symphony three times paired with three different conductor portraits and asked for various ratings, e.g., liking, technical aspects, and emotional quality, as well as free-text replies for a description of performances and the participant’s enjoyment. In 85% of cases, participants did not report the sameness of the musical pieces nor reply in a uniform way to the rating scales, they most often missed the fact that all three stimuli were identical. This study is an example for the influence of a convincing piece of auxiliary information and corroborates, along with studies mentioned in the literature review by <xref ref-type="bibr" rid="r2">Anglada-Tort and Müllensiefen (2017)</xref>, the need to study figures or “situations” of authority, such as the mere setting of an empirical study, and whether this has the capacity to mute the perception or detection of unexpected stimulus features.</p>
<p>This study did not offer clear-cut detection rates for musicians and non-musicians but rather approximated the detection of a clash of keys with a response possibility that needs less conviction in participants. Due to self-selection of participants and recruiting from a rather music-savvy (but not necessarily musically trained) population, it is quite likely that even participants with little or no musical training were above average interested in music and could therefore not be classified as true non-experts in music. Second, as this study was an online study we had no control over the actual sound quality produced for participants. Self-reported items on the quality of the stimuli and the listening condition were rated very well, which, however, cannot be equated with a controlled listening situation in a laboratory study. Further research can focus on the range of methodical approaches (a free response task to quantify detection rates versus a rating task to quantify metrical differences, possibly accompanied by a qualitative survey) and how results from such operationalization differences interact with groups of different expertise.</p>

</sec>
</body>
<back>
	
	<sec sec-type="ethics-statement">
		<title>Ethics Statement</title>
		<p>The present study was conducted in accordance with ethical principles and standards according to the guidelines of the German Society for Psychology. According to German law, no ethics approval has been required. Informed consent was provided by all participants, and they had the option to cease participating in the study at any time without any negative consequences.</p>
	</sec>
	
<ref-list><title>References</title>
<ref id="r1"><mixed-citation publication-type="book">Abell, N., Springer, D. W., &amp; Kamata, A. (2009). <italic>Developing and validating rapid assessment instruments</italic>. Oxford University Press.</mixed-citation></ref>
<ref id="r2"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Anglada-Tort</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Müllensiefen</surname>, <given-names>D.</given-names></string-name></person-group> (<year>2017</year>). <article-title>The repeated recording Illusion.</article-title> <source>Music Perception</source>, <volume>35</volume>(<issue>1</issue>), <fpage>94</fpage>–<lpage>117</lpage>. <pub-id pub-id-type="doi">10.1525/mp.2017.35.1.94</pub-id></mixed-citation></ref>
<ref id="r3"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Bigand</surname>, <given-names>E.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Poulin-Charronnat</surname>, <given-names>B.</given-names></string-name></person-group> (<year>2006</year>). <article-title>Are we experienced listeners? A review of the musical capacities that do not depend on formal musical training.</article-title> <source>Cognition</source>, <volume>100</volume>(<issue>1</issue>), <fpage>100</fpage>–<lpage>130</lpage>. <pub-id pub-id-type="doi">10.1016/j.cognition.2005.11.007</pub-id><pub-id pub-id-type="pmid">16412412</pub-id></mixed-citation></ref>
<ref id="r4"><mixed-citation publication-type="book">Bregman, A. S. (1990). <italic>Auditory scene analysis: The perceptual organization of sound</italic>. MIT Press.</mixed-citation></ref>
<ref id="r5"><mixed-citation publication-type="book">Cook, N. (1994). Perception: A perspective from music theory. In R. Aiello &amp; J. A. Sloboda (Eds.), <italic>Musical perceptions</italic> (pp. 64–95). Oxford University Press.</mixed-citation></ref>
<ref id="r6"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Cook</surname>, <given-names>N.</given-names></string-name></person-group> (<year>1987</year>). <article-title>The perception of large-scale tonal closure.</article-title> <source>Music Perception</source>, <volume>5</volume>(<issue>2</issue>), <fpage>197</fpage>–<lpage>205</lpage>. <pub-id pub-id-type="doi">10.2307/40285392</pub-id></mixed-citation></ref>
<ref id="r7"><mixed-citation publication-type="web">Ellis, W. D. (1938). <italic>A source book of Gestalt psychology</italic>. Routledge. <ext-link ext-link-type="uri" xlink:href="https://archive.org/details/in.ernet.dli.2015.198039">https://archive.org/details/in.ernet.dli.2015.198039</ext-link></mixed-citation></ref>
<ref id="r8"><mixed-citation publication-type="book">Fricke, J. P. (2014). Intonation in der abendländischen Musik [Intonation in Western music]. In C. Reuter &amp; W. Auhagen (Eds.), <italic>Kompendien Musik. Musikalische Akustik</italic> (pp. 126–132). Laaber Verlag.</mixed-citation></ref>
<ref id="r9"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Hake</surname>, <given-names>R.</given-names></string-name>, <string-name name-style="western"><surname>Bürgel</surname>, <given-names>M.</given-names></string-name>, <string-name name-style="western"><surname>Nguyen</surname>, <given-names>N. K.</given-names></string-name>, <string-name name-style="western"><surname>Greasley</surname>, <given-names>A.</given-names></string-name>, <string-name name-style="western"><surname>Müllensiefen</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Siedenburg</surname>, <given-names>K.</given-names></string-name></person-group> (<year>2024</year>). <article-title>Development of an adaptive test of musical scene analysis abilities for normal-hearing and hearing-impaired listeners.</article-title> <source>Behavior Research Methods</source>, <volume>56</volume>(<issue>6</issue>), <fpage>5456</fpage>–<lpage>5481</lpage>. <pub-id pub-id-type="doi">10.3758/s13428-023-02279-y</pub-id><pub-id pub-id-type="pmid">37957432</pub-id></mixed-citation></ref>
<ref id="r10"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Hamamoto</surname>, <given-names>M.</given-names></string-name>, <string-name name-style="western"><surname>Botelho</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Munger</surname>, <given-names>M. P.</given-names></string-name></person-group> (<year>2010</year>). <article-title>Non-musicians’ and musicians’ perception of bitonality.</article-title> <source>Psychology of Music</source>, <volume>38</volume>(<issue>4</issue>), <fpage>423</fpage>–<lpage>445</lpage>. <pub-id pub-id-type="doi">10.1177/0305735609351917</pub-id></mixed-citation></ref>
<ref id="r11"><mixed-citation publication-type="book">Harrison, D. (1997). Bitonality, pentatonicism, and diatonicism in a work by Milhaud. In J. M. Baker, B. D. W, &amp; J. W. Bernard (Eds.), <italic>Music theory in concept and practice</italic> (pp. 393–408). University of Rochester Press.</mixed-citation></ref>
<ref id="r12"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Kaiser</surname>, <given-names>H. F.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Rice</surname>, <given-names>J.</given-names></string-name></person-group> (<year>1974</year>). <article-title>Little jiffy, mark IV.</article-title> <source>Educational and Psychological Measurement</source>, <volume>34</volume>(<issue>1</issue>), <fpage>111</fpage>–<lpage>117</lpage>. <pub-id pub-id-type="doi">10.1177/001316447403400115</pub-id></mixed-citation></ref>
<ref id="r13"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Karno</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Konečni</surname>, <given-names>V. J.</given-names></string-name></person-group> (<year>1992</year>). <article-title>The effects of structural interventions in the first movement of Mozart’s symphony in G minor K. 550 on aesthetic preference.</article-title> <source>Music Perception</source>, <volume>10</volume>(<issue>1</issue>), <fpage>63</fpage>–<lpage>72</lpage>. <pub-id pub-id-type="doi">10.2307/40285538</pub-id></mixed-citation></ref>
<ref id="r14"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Kopiez</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Platz</surname>, <given-names>F.</given-names></string-name></person-group> (<year>2009</year>). <article-title>The role of listening expertise, attention, and musical style in the perception of clash of keys.</article-title> <source>Music Perception</source>, <volume>26</volume>(<issue>4</issue>), <fpage>321</fpage>–<lpage>334</lpage>. <pub-id pub-id-type="doi">10.1525/mp.2009.26.4.321</pub-id></mixed-citation></ref>
<ref id="r15"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Krumhansl</surname>, <given-names>C. L.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Schmuckler</surname>, <given-names>M. A.</given-names></string-name></person-group> (<year>1986</year>). <article-title>The Petroushka chord: A perceptual investigation.</article-title> <source>Music Perception</source>, <volume>4</volume>(<issue>2</issue>), <fpage>153</fpage>–<lpage>184</lpage>. <pub-id pub-id-type="doi">10.2307/40285359</pub-id></mixed-citation></ref>
<ref id="r16"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Lalitte</surname>, <given-names>P.</given-names></string-name>, <string-name name-style="western"><surname>Bigand</surname>, <given-names>E.</given-names></string-name>, <string-name name-style="western"><surname>Kantor-Martynuska</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Delbé</surname>, <given-names>C.</given-names></string-name></person-group> (<year>2009</year>). <article-title>On listening to atonal variants of two piano sonatas by Beethoven.</article-title> <source>Music Perception</source>, <volume>26</volume>(<issue>3</issue>), <fpage>223</fpage>–<lpage>234</lpage>. <pub-id pub-id-type="doi">10.1525/mp.2009.26.3.223</pub-id></mixed-citation></ref>
<ref id="r17"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Larrouy-Maestri</surname>, <given-names>P.</given-names></string-name></person-group> (<year>2018</year>). <article-title>“I know it when I hear it”: On listeners’ perception of mistuning.</article-title> <source>Music &amp; Science</source>, <volume>1</volume>, <fpage>1</fpage>–<fpage>17</fpage>. <pub-id pub-id-type="doi">10.1177/2059204318784582</pub-id></mixed-citation></ref>
<ref id="r18"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Larrouy-Maestri</surname>, <given-names>P.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Morsomme</surname>, <given-names>D.</given-names></string-name></person-group> (<year>2014</year>). <article-title>Criteria and tools for objectively analysing the vocal accuracy of a popular song.</article-title> <source>Logopedics, Phoniatrics, Vocology</source>, <volume>39</volume>(<issue>1</issue>), <fpage>11</fpage>–<lpage>18</lpage>. <pub-id pub-id-type="doi">10.3109/14015439.2012.696139</pub-id><pub-id pub-id-type="pmid">22721558</pub-id></mixed-citation></ref>
<ref id="r19"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Larrouy-Maestri</surname>, <given-names>P.</given-names></string-name>, <string-name name-style="western"><surname>Harrison</surname>, <given-names>P. M. C.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Müllensiefen</surname>, <given-names>D.</given-names></string-name></person-group> (<year>2019</year>). <article-title>The mistuning perception test: A new measurement instrument.</article-title> <source>Behavior Research Methods</source>, <volume>51</volume>(<issue>2</issue>), <fpage>663</fpage>–<lpage>675</lpage>. <pub-id pub-id-type="doi">10.3758/s13428-019-01225-1</pub-id><pub-id pub-id-type="pmid">30924106</pub-id></mixed-citation></ref>
<ref id="r20"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Law</surname>, <given-names>L. N. C.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Zentner</surname>, <given-names>M.</given-names></string-name></person-group> (<year>2012</year>). <article-title>Assessing musical abilities objectively: Construction and validation of the Profile of Music Perception Skills.</article-title> <source>PLoS One</source>, <volume>7</volume>(<issue>12</issue>), <elocation-id>e52508</elocation-id>. <pub-id pub-id-type="doi">10.1371/journal.pone.0052508</pub-id><pub-id pub-id-type="pmid">23285071</pub-id></mixed-citation></ref>
<ref id="r21"><mixed-citation publication-type="web">Leiner, D. J. (2024). <italic>SoSci Survey</italic> (Version 3.7.00) [Computer software]. <ext-link ext-link-type="uri" xlink:href="https://www.soscisurvey.de">https://www.soscisurvey.de</ext-link></mixed-citation></ref>
<ref id="r22"><mixed-citation publication-type="book">Mazzola, G. (2017). Auditory physiology and psychology. In G. Mazzola (Ed.), <italic>The topos of music IV: Roots</italic> (pp. 1353–1368). Springer. <pub-id pub-id-type="doi">10.1007/978-3-319-64495-0_2</pub-id></mixed-citation></ref>
<ref id="r23"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Mongoven</surname>, <given-names>C.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Carbon</surname>, <given-names>C.-C.</given-names></string-name></person-group> (<year>2017</year>). <article-title>Acoustic Gestalt: On the perceptibility of melodic symmetry.</article-title> <source>Musicae Scientiae</source>, <volume>21</volume>(<issue>1</issue>), <fpage>41</fpage>–<lpage>59</lpage>. <pub-id pub-id-type="doi">10.1177/1029864916637116</pub-id></mixed-citation></ref>
<ref id="r24"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Müllensiefen</surname>, <given-names>D.</given-names></string-name>, <string-name name-style="western"><surname>Gingras</surname>, <given-names>B.</given-names></string-name>, <string-name name-style="western"><surname>Musil</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Stewart</surname>, <given-names>L.</given-names></string-name></person-group> (<year>2014</year>). <article-title>The musicality of non-musicians: An index for assessing musical sophistication in the general population.</article-title> <source>PLoS One</source>, <volume>9</volume>(<issue>2</issue>), <elocation-id>e89642</elocation-id>. <pub-id pub-id-type="doi">10.1371/journal.pone.0089642</pub-id><pub-id pub-id-type="pmid">24586929</pub-id></mixed-citation></ref>
<ref id="r25"><mixed-citation publication-type="book">Oehler, M. (2014). Auditorische Szenenanalyse [Auditory scene analysis]. In C. Reuter &amp; W. Auhagen (Eds.), <italic>Kompendien Musik. Musikalische Akustik</italic> (pp. 195–217). Laaber.</mixed-citation></ref>
<ref id="r26"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Omigie</surname>, <given-names>D.</given-names></string-name>, <string-name name-style="western"><surname>Dellacherie</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Samson</surname>, <given-names>S.</given-names></string-name></person-group> (<year>2014</year>). <article-title>Effects of learning on dissonance judgments.</article-title> <source>Journal of Interdisciplinary Music Studies</source>, <volume>8</volume>, <fpage>11</fpage>–<lpage>28</lpage>.</mixed-citation></ref>
<ref id="r27"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Parncutt</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Hair</surname>, <given-names>G.</given-names></string-name></person-group> (<year>2018</year>). <article-title>A psychocultural theory of musical interval: Bye bye Pythagoras.</article-title> <source>Music Perception</source>, <volume>35</volume>(<issue>4</issue>), <fpage>475</fpage>–<lpage>501</lpage>. <pub-id pub-id-type="doi">10.1525/mp.2018.35.4.475</pub-id></mixed-citation></ref>
<ref id="r28"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Schaal</surname>, <given-names>N. K.</given-names></string-name>, <string-name name-style="western"><surname>Bauer</surname>, <given-names>A.-K. R.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Müllensiefen</surname>, <given-names>D.</given-names></string-name></person-group> (<year>2014</year>). <article-title>Der Gold-MSI: Replikation und Validierung eines Fragebogeninstrumentes zur Messung Musikalischer Erfahrenheit anhand einer deutschen Stichprobe</article-title> <comment>[The Gold-MSI: Replication and validation of a questionnaire instrument for measuring musical sophistication, based on a German sample]</comment>. <source>Musicae Scientiae</source>, <volume>18</volume>(<issue>4</issue>), <fpage>423</fpage>–<lpage>447</lpage>. <pub-id pub-id-type="doi">10.1177/1029864914541851</pub-id></mixed-citation></ref>
<ref id="r29"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Thompson</surname>, <given-names>W. F.</given-names></string-name>, &amp; <string-name name-style="western"><surname>Mor</surname>, <given-names>S.</given-names></string-name></person-group> (<year>1992</year>). <article-title>A perceptual investigation of polytonality.</article-title> <source>Psychological Research</source>, <volume>54</volume>(<issue>2</issue>), <fpage>60</fpage>–<lpage>71</lpage>. <pub-id pub-id-type="doi">10.1007/BF00937134</pub-id><pub-id pub-id-type="pmid">1620799</pub-id></mixed-citation></ref>
<ref id="r30"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Wertheimer</surname>, <given-names>M.</given-names></string-name></person-group> (<year>1923</year>). <article-title>Untersuchungen zur Lehre von der Gestalt. II</article-title> <comment>[Research into Gestalt theory]</comment>. <source>Psychologische Forschung</source>, <volume>4</volume>(<issue>1</issue>), <fpage>301</fpage>–<lpage>350</lpage>. <pub-id pub-id-type="doi">10.1007/BF00410640</pub-id></mixed-citation></ref>
<ref id="r31"><mixed-citation publication-type="book">Whittall, A. (2001). <italic>Bitonality</italic>. Oxford University Press. <pub-id pub-id-type="doi">10.1093/gmo/9781561592630.article.03161</pub-id></mixed-citation></ref>
<ref id="r32"><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name name-style="western"><surname>Wolpert</surname>, <given-names>R. S.</given-names></string-name></person-group> (<year>2000</year>). <article-title>Attention to key in a nondirected music listening task: Musicians vs. nonmusicians.</article-title> <source>Music Perception</source>, <volume>18</volume>, <fpage>225</fpage>–<lpage>230</lpage>. <pub-id pub-id-type="doi">10.2307/40285910</pub-id></mixed-citation></ref>
</ref-list>
	<sec sec-type="data-availability" id="das"><title>Data Availability</title>
		<p>The research data and stimuli for this article are available on the OSF (see <xref ref-type="bibr" rid="sp1_r1">Wolf, 2025</xref>).</p>
	</sec>	

	
	
	
	<sec sec-type="supplementary-material" id="sp1"><title>Supplementary Materials</title>
		<p>For this article, research data and stimuli are available (see <xref ref-type="bibr" rid="sp1_r1">Wolf, 2025</xref>).</p>
		<ref-list content-type="supplementary-material" id="suppl-ref-list">
			<ref id="sp1_r1">
				<mixed-citation publication-type="supplementary-material">
					<person-group person-group-type="author">
							<name name-style="western">
								<surname>Wolf</surname>
								<given-names>A.</given-names>
							</name>
					</person-group> (<year>2025</year>). <source>Detection of clash of keys</source> <comment>[Data, code, stimuli]</comment>. <publisher-name>OSF</publisher-name>. <ext-link ext-link-type="uri" xlink:href="https://osf.io/sj3da">https://osf.io/sj3da</ext-link>		
				</mixed-citation>
			</ref>
		</ref-list>
	</sec>
			

<fn-group>
<fn fn-type="financial-disclosure"><p>The authors have no funding to report.</p></fn>
</fn-group>
<fn-group>
<fn fn-type="conflict"><p>The authors have declared that no competing interests exist.</p></fn>
</fn-group>
<ack>
<p>The authors have no additional (i.e., non-financial) support to report.</p>
</ack>
</back>
</article>
