<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>Hfmt on Sebastian Spicker</title>
    <link>https://sebastianspicker.github.io/tags/hfmt/</link>
    <description>Recent content in Hfmt on Sebastian Spicker</description>
    <image>
      <title>Sebastian Spicker</title>
      <url>https://sebastianspicker.github.io/og-image.png</url>
      <link>https://sebastianspicker.github.io/og-image.png</link>
    </image>
    <generator>Hugo -- 0.160.0</generator>
    <language>en</language>
    <lastBuildDate>Sat, 07 Dec 2024 00:00:00 +0000</lastBuildDate>
    <atom:link href="https://sebastianspicker.github.io/tags/hfmt/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>Artificial Intelligence in Music Pedagogy: Curriculum Implications from a Thementag</title>
      <link>https://sebastianspicker.github.io/posts/ai-music-pedagogy-day/</link>
      <pubDate>Sat, 07 Dec 2024 00:00:00 +0000</pubDate>
      <guid>https://sebastianspicker.github.io/posts/ai-music-pedagogy-day/</guid>
      <description>On 2 December 2024 I gave three workshops at HfMT Köln&amp;rsquo;s Thementag on AI and music education. The handouts covered data protection, AI tools for students, and AI in teaching. This post is the argument behind them — focused on the curriculum question that none of the tools answer on their own: what should change, and what should not?</description>
      <content:encoded><![CDATA[<p><em>On 2 December 2024, the Hochschule für Musik und Tanz Köln held a Thementag:
&ldquo;Next level? Künstliche Intelligenz und Musikpädagogik im Dialog.&rdquo; I gave three
workshops — on data protection and AI, on AI tools for students, and on AI in
teaching. The handouts from those sessions cover the practical and regulatory
ground. This post is the argument behind them: what I think changes in music
education when these tools become ambient, and what I think does not.</em></p>
<hr>
<h2 id="the-occasion">The Occasion</h2>
<p>&ldquo;Next level?&rdquo; The question mark is doing real work. The framing HfMT chose for
the day was appropriately provisional: not a declaration that AI has already
transformed music education, but an invitation to ask whether, in what
direction, and at what cost.</p>
<p>The invitations that reach me for events like this tend to come with one of two
framings. The first is enthusiasm: AI is coming, we need to get ahead of it,
here are tools your students are already using. The second is anxiety: AI is
coming, it threatens everything we do, we need to protect students from it.
Both framings are understandable. Neither is adequate to the curriculum
question, which is slower-moving and more structural than either suggests.</p>
<p>I prepared three sets of handouts. The first covered data protection — the
least glamorous topic in AI education, and the one that most directly
determines what can legally be deployed in a university setting. The second
covered AI tools for students: what exists, what it does, and what critical
thinking skills you need to use it without being used by it. The third covered
AI for instructors: where it helps, where it flatters, and where it makes
things worse.</p>
<p>This post does not recapitulate the handouts. It addresses the question I kept
returning to across all three workshops: what does this change about what a
music student needs to learn?</p>
<hr>
<h2 id="what-the-technology-actually-is">What the Technology Actually Is</h2>
<p>My physics training left me professionally uncomfortable
with hand-waving — including my own. Before discussing curriculum implications,
it is worth being specific about what these tools are.</p>
<p>The dominant paradigm in current AI — responsible for ChatGPT, for Whisper, for
Suno.AI, for Google Magenta, for the large language models whose outputs are
now visible everywhere — is the transformer architecture (Vaswani et al.,
2017). A transformer is a neural network that processes sequences by computing,
for each element, a weighted attention over all other elements. The attention
weights are learned from data. The result is a model that can capture
long-range dependencies in sequences — text, audio, musical notes — without the
recurrence that made earlier architectures difficult to train at scale.</p>
<p>What this means practically: these models are trained on very large corpora,
they learn statistical regularities, and they generate outputs that are
statistically consistent with their training distribution. They are not
reasoning from first principles. They do not &ldquo;know&rdquo; music theory the way a
student who has internalised harmonic function knows it. They have learned, from
enormous quantities of text and audio, what tends to follow what. For many tasks
this is sufficient. For tasks that require understanding of underlying structure,
it is not — and the failure modes are characteristic rather than random.</p>
<p>BERT (Devlin et al., 2018) showed that pre-training on large corpora and
fine-tuning on specific tasks produces models that outperform task-specific
architectures on a wide range of benchmarks. The same transfer-learning
paradigm has spread to audio (Whisper pre-trains on 680,000 hours of labelled
audio), to music generation (Magenta&rsquo;s transformer-based models produce
melodically coherent sequences), and to multimodal domains. The technology is
mature, improving, and available to students now. Knowing what it is — not
just what it produces — is the starting point for any sensible curriculum
discussion about it.</p>
<hr>
<h2 id="the-data-protection-constraint">The Data Protection Constraint</h2>
<p>Before any discussion of pedagogical benefit, there is a legal boundary that
most AI-in-education discussions skip over. In Germany, and in the EU more
broadly, the deployment of AI tools in a university setting is governed by the
GDPR (DSGVO, Regulation 2016/679) and, at state level in NRW, by the DSG NRW.
The constraints are not abstract: they determine which tools can be used for
which purposes with which students.</p>
<p>The core principle is data minimisation: only data necessary for a specific,
documented purpose may be collected or processed. When a student uses a
commercial AI tool to get feedback on a composition exercise and enters text
that could identify them or their institution, that data may be stored,
processed, and used for model improvement by an operator whose servers are
outside the EU. Whether such transfers remain legally valid under GDPR after
the Schrems II ruling (Court of Justice of the EU, 2020) is contested — and
&ldquo;contested&rdquo; is not a position in which an institution can comfortably require
students to use a tool.</p>
<p>The practical upshot for curriculum design is this: AI tools running on EU
servers with documented processing agreements can be integrated into formal
coursework. Commercial tools whose terms specify US-based processing and model
training on user data cannot be required of students. They can be discussed and
demonstrated, but making them mandatory puts students in a position where they
must choose between their privacy and their grade.</p>
<p>This is not a reason to avoid AI in teaching. It is a reason to be honest about
the regulatory landscape, to distinguish clearly between tools you can require
and tools you can recommend, and to make data protection literacy part of what
students learn. The skill of reading a terms-of-service document and identifying
the data flows it describes is not a legal skill — it is a general literacy
skill that matters for every digital tool a music professional will use.</p>
<hr>
<h2 id="what-changes-for-students">What Changes for Students</h2>
<p>The question I was asked most often across the three workshops was some version
of: &ldquo;If AI can already do X, should students still learn X?&rdquo;</p>
<p>The question is less simple than it appears, and the answer is not uniform
across skills.</p>
<p><strong>Skills where automation reduces the required production threshold</strong> do exist.
A student who spends weeks mastering advanced music engraving tools for score
production, when AI can generate a usable first draft from a much simpler
description, has arguably spent time that could have been better allocated
elsewhere. Not because the underlying skill is worthless — it is not — but
because the threshold of competence required to produce a working output has
dropped. The student&rsquo;s time might be more valuable spent on something that
has not been automated.</p>
<p><strong>Skills where automation creates new requirements</strong> are more interesting.
Transcription is a useful example. Automatic speech recognition — using
models like Whisper for spoken-word transcription, or specialised models
for audio-to-score music transcription — is now accurate enough to produce
usable first drafts from audio. This does not
eliminate the need for transcription skill in a music student. It changes it.
A student who cannot evaluate the output of an automatic transcription — who
cannot hear where the model has made characteristic errors, who does not have
an internalised sense of what a correct transcription looks like — is unable
to use the tool productively. The required skill has shifted from production
to evaluation. This is not a lesser skill; it is a different one, and it is
not automatically acquired alongside the ability to run the tool.</p>
<p><strong>Skills that automation cannot replace</strong> are those that depend on embodied,
situated, relational knowledge: stage presence, real-time improvisation, the
subtle negotiation of musical meaning in ensemble, the pedagogical relationship
between teacher and student. These are not beyond AI in principle. They are
far beyond it in practice, and the gap is not closing as quickly as the
generative AI discourse sometimes suggests.</p>
<p>The curriculum implication is not &ldquo;teach less&rdquo; or simply &ldquo;teach differently.&rdquo;
It is: be explicit about which category each skill falls into, and design
assessment accordingly. An assignment that asks students to produce something
AI can produce is now testing something different from what it was testing two
years ago — not necessarily nothing, but something different. The rubric should
reflect that.</p>
<hr>
<h2 id="what-changes-for-instructors">What Changes for Instructors</h2>
<p>The same three-category analysis applies symmetrically to teaching.</p>
<p><strong>Routine task automation</strong> is genuinely useful. Generating first drafts of
worksheets, producing exercises at different difficulty levels, transcribing a
recorded lesson for later analysis — these are tasks where AI can save
meaningful time without compromising the pedagogical judgment required to make
use of the output. Holmes et al. (2019) identify feedback generation as one
of the clearer wins for AI in education: systems that provide immediate,
targeted feedback at a scale that human instructors cannot match. A
transcription model listening to a student practice and flagging rhythmic
inconsistencies does not replace a teacher. It extends the feedback loop
beyond the lesson hour.</p>
<p><strong>Content generation with limits</strong> is where AI is most seductive and most
dangerous. A model like ChatGPT can produce a reading list on any topic, a
summary of any debate in the literature, a set of discussion questions for any
text. The outputs are fluent, plausible, and frequently wrong in ways that are
difficult to detect without domain expertise. Jobin et al. (2019) and
Mittelstadt et al. (2016) both document the broader concern with AI opacity
and accountability: when a model produces a confident-sounding claim, the
burden of verification falls on the user. An instructor who outsources the
construction of course materials to a model, and who lacks enough domain
knowledge to catch the errors, is not saving time — they are transferring
risk to their students.</p>
<p>Hallucinations — outputs that are plausible in form but false in content — are
not bugs in the usual sense. They are a structural consequence of how generative
models work. A model trained to predict likely next tokens will produce the most
statistically plausible continuation, not the most accurate one. For music
education, where historical facts, composer attributions, and music-theoretic
claims need to be correct, this matters. The model&rsquo;s fluency is not evidence
of its accuracy.</p>
<p><strong>Personalisation</strong> is the most-cited promise of AI in education (Luckin et
al., 2016; Roll &amp; Wylie, 2016) and the hardest to evaluate in practice. The
argument is that AI can adapt instructional content to individual learners'
needs in real time, producing one-to-one tutoring at scale. The evidence in
formal educational settings is more mixed than the boosters suggest. What is
clear is that personalisation at scale requires data — and extensive data about
individual students&rsquo; learning trajectories raises the same data protection
concerns already discussed, in more acute form.</p>
<hr>
<h2 id="the-music-specific-question">The Music-Specific Question</h2>
<p>I want to be direct about something that came up repeatedly across the day and
that the general AI-in-education literature handles badly: music education is
not generic.</p>
<p>The skills involved — listening, performing, interpreting, composing,
improvising — have a phenomenological and embodied dimension that does not map
cleanly onto the text-prediction paradigm that most current AI systems
instantiate. Suno.AI can generate a stylistically convincing chord progression
in the manner of a named composer. It cannot explain why the progression is
convincing in the way a student who has internalised tonal function can explain
it. Google Magenta can generate a continuation of a melodic fragment that is
locally coherent. It cannot navigate the structural expectations of a sonata
form with the intentionality that a performer brings to interpreting one.</p>
<p>This is not a criticism of these tools. It is a description of what they are.
The curriculum implication is that music education must be clear about what it
is teaching: the <em>product</em> — a score, a performance, a composition — or the
<em>process and understanding</em> of which the product is evidence. Where assessment
focuses on the product, AI creates an obvious challenge. Where it focuses on
demonstrable process and understanding — including the ability to critically
evaluate AI-generated outputs — it creates new opportunities.</p>
<p>The more interesting question is whether AI tools can make musical <em>process</em>
more visible and discussable. A composition student who uses a generative model,
notices that the output is harmonically correct but rhythmically inert, and can
articulate <em>why</em> it is inert — and then revise it accordingly — has
demonstrated more sophisticated musical understanding than a student who
produces the same output without any generative assistance. The tool does not
lower the standard; it shifts where the standard is applied.</p>
<p>There is an analogy in music theory pedagogy. The availability of notation
software that can play back a student&rsquo;s harmony exercise and flag parallel
fifths changed what ear training and harmony teaching emphasise — but it did
not make harmony teaching obsolete. It changed the floor (students can check
mechanical correctness automatically) and raised the ceiling (more class time
can be spent on voice-leading logic and expressive intention). AI tools are a
larger version of the same displacement: the floor rises, the ceiling rises
with it, and the pedagogical question is always what you are doing between
the two.</p>
<hr>
<h2 id="copyright-and-academic-integrity">Copyright and Academic Integrity</h2>
<p>Two issues that crossed all three workshops and deserve direct treatment.</p>
<p>On copyright: the training data of generative music models includes copyrighted
recordings and scores, the legal status of which is actively litigated in
multiple jurisdictions. When Suno.AI generates a piece &ldquo;in the style of&rdquo;
a named composer, it is drawing on patterns extracted from that composer&rsquo;s work
— work that is under copyright in the case of living or recently deceased
composers. The output is not a direct copy, but neither is the relationship
to the training data legally settled. Music students who use these tools in
professional contexts should know that they are working in a legally uncertain
space, and institutions should not pretend otherwise.</p>
<p>On academic integrity: the issue is not that students might use AI to cheat —
they will, some of them, and they have always found ways to cheat with whatever
tools were available. The issue is that current AI policies at many institutions
are incoherent: prohibiting AI use in assessment while providing no clear
guidance on what counts as AI use, and assigning tasks where AI assistance is
undetectable and arguably appropriate. The more useful approach is to design
tasks where AI assistance is either irrelevant (because the task requires live
performance or real-time demonstration) or visible and assessed (because the
task explicitly includes reflection on how AI was used and to what effect).</p>
<hr>
<h2 id="three-things-i-came-away-with">Three Things I Came Away With</h2>
<p>After a full day of workshops, discussions, and the conversations that happen
in the corridors between sessions, I left with three positions that feel more
settled than they did in the morning.</p>
<p><strong>First</strong>: the data protection question is not separable from the pedagogical
question. Any serious curriculum discussion of AI in music education has to
start with what can legally be deployed, not with what would be useful if
constraints were not a factor. The constraints are a factor.</p>
<p><strong>Second</strong>: the skill most urgently needed — in students and in instructors —
is not AI literacy in the sense of knowing which tool to use for which task.
It is the critical capacity to evaluate AI-generated outputs: to notice what
is wrong, to understand <em>why</em> it is wrong, and to correct it. This requires
domain expertise first. You cannot critically evaluate an AI-generated harmonic
analysis if you do not understand harmonic analysis. The tools do not lower
the bar for domain knowledge. They raise the bar for its critical application.</p>
<p><strong>Third</strong>: the curriculum question is not &ldquo;how do we accommodate AI?&rdquo; It is
&ldquo;what are we actually trying to teach, and does the answer change when AI can
produce the visible output of that process?&rdquo; Answering that honestly, skill
by skill, for a full music programme, is slow work. It cannot be done at a
one-day event. But a one-day event, if it is well-designed, can start the
conversation in the right place.</p>
<p>HfMT&rsquo;s Thementag started it in the right place.</p>
<hr>
<h2 id="references">References</h2>
<ul>
<li>
<p>Devlin, J., Chang, M.-W., Lee, K., &amp; Toutanova, K. (2018). BERT:
Pre-training of deep bidirectional transformers for language understanding.
<em>arXiv preprint arXiv:1810.04805</em>. <a href="https://arxiv.org/abs/1810.04805">https://arxiv.org/abs/1810.04805</a></p>
</li>
<li>
<p>Goodfellow, I., Bengio, Y., &amp; Courville, A. (2016). <em>Deep Learning.</em>
MIT Press. <a href="https://www.deeplearningbook.org">https://www.deeplearningbook.org</a></p>
</li>
<li>
<p>Holmes, W., Bialik, M., &amp; Fadel, C. (2019). <em>Artificial Intelligence in
Education: Promises and Implications for Teaching and Learning.</em> Center for
Curriculum Redesign.</p>
</li>
<li>
<p>Jobin, A., Ienca, M., &amp; Vayena, E. (2019). The global landscape of AI ethics
guidelines. <em>Nature Machine Intelligence</em>, 1, 389–399.
<a href="https://doi.org/10.1038/s42256-019-0088-2">https://doi.org/10.1038/s42256-019-0088-2</a></p>
</li>
<li>
<p>LeCun, Y., Bengio, Y., &amp; Hinton, G. (2015). Deep learning. <em>Nature</em>,
521(7553), 436–444. <a href="https://doi.org/10.1038/nature14539">https://doi.org/10.1038/nature14539</a></p>
</li>
<li>
<p>Luckin, R., Holmes, W., Griffiths, M., &amp; Forcier, L. B. (2016).
<em>Intelligence Unleashed: An Argument for AI in Education.</em> Pearson.</p>
</li>
<li>
<p>Mittelstadt, B. D., Allo, P., Taddeo, M., Wachter, S., &amp; Floridi, L.
(2016). The ethics of algorithms: Mapping the debate. <em>Big Data &amp; Society</em>,
3(2). <a href="https://doi.org/10.1177/2053951716679679">https://doi.org/10.1177/2053951716679679</a></p>
</li>
<li>
<p>Roll, I., &amp; Wylie, R. (2016). Evolution and revolution in artificial
intelligence in education. <em>International Journal of Artificial Intelligence
in Education</em>, 26(2), 582–599.
<a href="https://doi.org/10.1007/s40593-016-0110-3">https://doi.org/10.1007/s40593-016-0110-3</a></p>
</li>
<li>
<p>Russell, S., &amp; Norvig, P. (2020). <em>Artificial Intelligence: A Modern
Approach</em> (4th ed.). Pearson.</p>
</li>
<li>
<p>Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez,
A. N., Kaiser, Ł., &amp; Polosukhin, I. (2017). Attention is all you need.
<em>Advances in Neural Information Processing Systems</em>, 30.
<a href="https://arxiv.org/abs/1706.03762">https://arxiv.org/abs/1706.03762</a></p>
</li>
</ul>
]]></content:encoded>
    </item>
    <item>
      <title>After the Connection Is Stable, the Hard Part Begins</title>
      <link>https://sebastianspicker.github.io/posts/nmp-curriculum-reflective-practice/</link>
      <pubDate>Fri, 22 Nov 2024 00:00:00 +0000</pubDate>
      <guid>https://sebastianspicker.github.io/posts/nmp-curriculum-reflective-practice/</guid>
      <description>A third post in the networked music performance series. Technical latency is solved. Institutional infrastructure has a name. What students actually learn — and what conservatoire curricula consistently get wrong about teaching it — turns out to be a different problem entirely.</description>
      <content:encoded><![CDATA[<p><em>Third post in a series. The <a href="/posts/nmp-latency-lola-mvtp/">August 2023 post</a>
covered latency measurements across six European research-network links.
The <a href="/posts/digital-music-labs-infrastructure/">June 2024 post</a> covered
what institutional infrastructure needs to look like for any of that to
be sustainably usable. This one covers what happens after both of those
problems are solved — which is when the genuinely interesting educational
challenges start.</em></p>
<p><em>Based on a manuscript with colleagues from the RAPP Lab. Not yet peer-reviewed.</em></p>
<hr>
<h2 id="the-gap-nobody-talks-about">The Gap Nobody Talks About</h2>
<p>There is a version of the NMP success story that stops too early. It goes: we
installed LoLa, measured the latency, it came in at 9.5 ms to Vienna, the
musicians played together across 745 km, it worked. Success.</p>
<p>What this story skips is the classroom after the demo. The student who can
follow a setup checklist perfectly and still has no idea what to do musically
when the connection is stable. The ensemble that gets a clean signal running
and then plays exactly the same repertoire in exactly the same way they would
in a co-present rehearsal, fighting the latency instead of working with it,
frustrated when it does not feel right. The assessment rubric that checks off
&ldquo;maintained stable connection&rdquo; and &ldquo;completed the performance&rdquo; and has nothing
to say about everything that actually constitutes musical learning in a
networked context.</p>
<p>The gap between <em>technical feasibility</em> and <em>educational transformation</em> is
the subject of this post. Closing it turns out to require a different kind of
curriculum design than most conservatoires have tried.</p>
<hr>
<h2 id="what-gets-taught-versus-what-needs-to-be-learned">What Gets Taught Versus What Needs to Be Learned</h2>
<p>The default curricular response to NMP has been to treat it as a technical
skill with an artistic application. Students learn to configure an audio
interface, manage routing, establish a LoLa connection, and then — implicitly
— go do music. The technical content gets staged as a prerequisite to the
&ldquo;real&rdquo; work.</p>
<p>This ordering is wrong in a specific way. Technical setup work is genuinely
necessary, but making it a prerequisite treats the relationship between
technology and musical practice as sequential rather than recursive. In
practice, the interesting musical problems only become visible <em>through</em> the
technical ones. A student does not understand why buffer size matters until
they have felt the difference between a 5 ms and a 40 ms offset in a
coordination-intensive passage. A student does not develop an opinion about
audio routing configurations until they have experienced a rehearsal collapse
caused by a routing error they could have prevented.</p>
<p>The RAPP Lab&rsquo;s recurring insight across several years of module iterations
at HfMT Köln was more direct: once learners can establish a stable connection,
the harder challenge is developing artistic, collaborative and reflective
strategies for making music <em>together apart</em>. Technical fluency is a
foundation, not a destination.</p>
<hr>
<h2 id="the-curriculum-we-ended-up-with">The Curriculum We Ended Up With</h2>
<p>It took several cycles to get there. The early format was weekend workshops —
open, exploratory, no formal assessment, primarily for advanced students who
self-selected in. These were useful precisely because they were informal: they
revealed quickly how technical and musical questions become inextricable once
you are actually playing, and they gave us evidence about where students got
stuck that we would not have found from a needs analysis.</p>
<p>Over time, elements of those workshops were developed into recurring
curriculum-embedded modules, which then fed into independent study projects
and eventually into external collaborations and performances. The trajectory
mattered: moving from a one-off event to something longitudinal meant that
knowledge built across cohorts rather than resetting every time.</p>
<p>The module structure that emerged has three interlocking elements:</p>
<p><strong>Progressive task design.</strong> Early sessions are tightly scoped:
specific technical-musical exercises, limited repertoire, well-defined
success criteria. Later sessions move toward open-ended projects, student-led
rehearsal planning, and eventually cross-institutional partnerships where
variables are genuinely outside anyone&rsquo;s control. The point of the early
constraints is not to make things easier — it is to create conditions where
students can notice what they are doing rather than just surviving.</p>
<p><strong>Journals and debriefs.</strong> Students kept individual reflective journals
throughout modules, documenting not just what happened but how they responded
to it — technical problems, musical decisions, moments of coordination failure
and recovery, questions they could not answer at the time. Group debriefs
after each rehearsal then turned those individual threads into collective
knowledge: comparing strategies, naming the problems that came up repeatedly,
developing shared language for rehearsal coordination.</p>
<p>The debrief is the part of this model that I think gets undervalued. It is
not just reflection — it is <em>curriculum production</em>. Strategies that emerged
from one cohort&rsquo;s debriefs became documented starting points for subsequent
cohorts. Knowledge accumulated rather than evaporating when the semester ended.</p>
<p><strong>Portfolio assessment.</strong> Rather than assessing primarily on a final
performance, students assembled portfolios that could include curated journal
excerpts, rehearsal documentation, reflective syntheses, and accounts of
how their thinking changed. The question being assessed was not &ldquo;did you play
the concert&rdquo; but &ldquo;can you articulate why you made the decisions you made, and
what you would do differently.&rdquo;</p>
<hr>
<h2 id="what-students-actually-learn-when-the-curriculum-works">What Students Actually Learn (When the Curriculum Works)</h2>
<p>Four outcomes recurred across the RAPP Lab iterations, consistently enough
to be worth naming:</p>
<h3 id="1-technical-agency">1. Technical agency</h3>
<p>This is different from technical competence. Competence means you can follow
a procedure. Agency means you understand the procedure well enough to deviate
from it intelligently when something goes wrong — to diagnose what failed,
generate a hypothesis about why, and try something different.</p>
<p>The shift happened when students stopped treating technical problems as
interruptions to the music and started treating them as information about
the system they were working inside. A dropout is not just an annoyance; it
is evidence about where the failure occurred. Getting to that reframe took,
on average, several weeks of structured reflection. It did not happen from
reading documentation.</p>
<h3 id="2-adaptive-improvisation">2. Adaptive improvisation</h3>
<p>Latency changes what real-time musical coordination can mean. You cannot rely
on the same multimodal cues — breath, gesture, shared acoustics — that make
co-present ensemble playing feel intuitive. You have to develop explicit
cueing systems, turn-taking conventions, contingency plans for when the
connection degrades mid-performance.</p>
<p>What we observed was that this constraint generated a specific kind of
musical creativity. Students improvised not just with musical material but
with rehearsal organisation itself — inventing systems, testing them,
discarding the ones that did not work, documenting the ones that did. Some of
the most musically interesting moments in the modules came from sessions where
the technology was behaving badly and students had to make it work anyway.</p>
<p>There is research on &ldquo;productive failure&rdquo; — deliberately designing tasks that
exceed students&rsquo; current control, because the struggle and recovery produces
deeper learning than smooth execution (Kapur 2016). NMP turns out to be a
natural context for this, not by design but because the network does not
cooperate on schedule.</p>
<h3 id="3-collaborative-communication">3. Collaborative communication</h3>
<p>Co-present rehearsal relies heavily on implicit communication: the
physical space makes many things legible without anyone having to say them.
In a networked rehearsal, the spatial and gestural channel is degraded or
absent. Students had to make explicit what would normally be implicit —
articulating coordination strategies, naming the problems they were
experiencing rather than hoping the ensemble would notice, developing a
vocabulary for talking about timing and latency as musical parameters.</p>
<p>This turned out to generalize. Students who had worked through several
networked rehearsal cycles were noticeably better at explicit musical
communication in co-present settings too, because they had been forced to
develop the vocabulary in a context where it was necessary.</p>
<h3 id="4-reflective-identity">4. Reflective identity</h3>
<p>The students who got the most out of the modules were the ones who stopped
waiting for the conditions to improve and started working with the conditions
as they were. Latency as a compositional constraint rather than a defect to
be routed around. Uncertainty as an artistic condition rather than a
technical failure.</p>
<p>The journal entries where this shift is most visible are not the ones that
describe what the student did. They are the ones that describe a change in
how the student understands their own practice — who they are as a musician
in relation to an environment they cannot fully control. That is a different
kind of outcome than anything a timing metric captures.</p>
<hr>
<h2 id="the-assessment-problem">The Assessment Problem</h2>
<p>The hardest part of all of this to translate into institutional language is
assessment. The conservatoire has well-developed frameworks for evaluating
performances. It has much weaker frameworks for evaluating the learning that
happens before and between and underneath performances.</p>
<p>Checklist rubrics — was the connection stable, was the latency within
acceptable range, did the performance complete — are useful for safety and
reliability. They are poor evidence for whether a student has developed the
capacity to work reflectively and artistically in a mediated ensemble
environment. A student who achieved a stable connection by following
instructions exactly and a student who achieved it by diagnosing a routing
error mid-session look identical on a checklist. They have had very different
learning experiences.</p>
<p>Portfolio assessment addresses this by making the reasoning visible. When a
student can explain why they chose a particular buffer configuration given
the specific network characteristics of that session, how that choice affected
the musical phrasing in the piece they were rehearsing, and what they would
change next time — that is evidence of something real. It is also harder to
assess than a timing log, which is probably why most programmes avoid it.</p>
<p>The argument is not that quantitative indicators are useless. It is that
they function better as scaffolding for reflective judgement than as the
primary evidence of learning. Mixed assessment ecologies — technical logs
plus journals plus portfolio syntheses — are more honest about what is
actually happening educationally.</p>
<hr>
<h2 id="what-this-does-not-solve">What This Does Not Solve</h2>
<p>The model described here depends on teaching staff who can facilitate
reflective dialogue, curate knowledge across cohorts, and participate in
iterative curriculum redesign. That is a specific professional competence
that is not automatically present in a conservatoire staffed primarily by
performing musicians. The training and support structures needed to develop
it are an open question this paper does not fully answer.</p>
<p>The curriculum is also not portable as-is. The RAPP Lab model emerged in a
specific institutional context — HfMT Köln, specific partner network,
specific funding structure, specific cohort of students. The four outcomes
and the general pedagogical logic may transfer; the specific formats will
need adaptation. Any institution that tries to implement this without going
through at least one cycle of their own iterative development is likely to
end up with a checklist version of something that works only when it is a
living process.</p>
<p>And the technology keeps moving. LoLa is a mature platform but the
ecosystem around it — network configurations, operating system support,
hardware lifecycles — changes faster than curriculum documentation. Building
responsiveness into the curriculum itself, rather than treating it as a fixed
syllabus, is the structural answer. Easier to recommend than to institutionalise.</p>
<hr>
<h2 id="references">References</h2>
<p>Barrett, H. C. (2007). Researching electronic portfolios and learner
engagement. <em>Journal of Adolescent &amp; Adult Literacy</em>, 50(6), 436–449.</p>
<p>Borgdorff, H. (2012). <em>The Conflict of the Faculties.</em> Leiden University Press.</p>
<p>The Design-Based Research Collective (2003). Design-based research: An
emerging paradigm for educational inquiry. <em>Educational Researcher</em>, 32(1),
5–8.</p>
<p>Kapur, M. (2016). Examining productive failure, productive success,
unproductive failure, and unproductive success in learning. <em>Educational
Psychologist</em>, 51(2), 289–299. <a href="https://doi.org/10.1080/00461520.2016.1155457">https://doi.org/10.1080/00461520.2016.1155457</a></p>
<p>Lave, J. &amp; Wenger, E. (1991). <em>Situated Learning.</em> Cambridge University Press.</p>
<p>Sadler, D. R. (2009). Indeterminacy in the use of preset criteria for
assessment and grading. <em>Assessment &amp; Evaluation in Higher Education</em>,
34(2), 159–179. <a href="https://doi.org/10.1080/02602930801956059">https://doi.org/10.1080/02602930801956059</a></p>
<p>Schön, D. A. (1983). <em>The Reflective Practitioner.</em> Basic Books.</p>
<p>Wenger, E. (1998). <em>Communities of Practice.</em> Cambridge University Press.
<a href="https://doi.org/10.1017/CBO9780511803932">https://doi.org/10.1017/CBO9780511803932</a></p>
<hr>
<h2 id="changelog">Changelog</h2>
<ul>
<li><strong>2026-01-20</strong>: Updated the Sadler (2009) reference title to &ldquo;Indeterminacy in the use of preset criteria for assessment and grading,&rdquo; matching the journal article at this DOI. Updated the Kapur (2016) reference to the full published title: &ldquo;Examining productive failure, productive success, unproductive failure, and unproductive success in learning.&rdquo;</li>
</ul>
]]></content:encoded>
    </item>
    <item>
      <title>The Boring Parts of Networked Music Performance</title>
      <link>https://sebastianspicker.github.io/posts/digital-music-labs-infrastructure/</link>
      <pubDate>Fri, 14 Jun 2024 00:00:00 +0000</pubDate>
      <guid>https://sebastianspicker.github.io/posts/digital-music-labs-infrastructure/</guid>
      <description>A follow-up to the August 2023 latency post. The numbers were fine. The hard part turned out to be everything else: governance, maintenance, invisible labour, and why most Digital Music Labs quietly die after the grant ends.</description>
      <content:encoded><![CDATA[<p><em>This post is based on a manuscript in progress with colleagues from the
RAPP Lab network. It builds directly on the <a href="/posts/nmp-latency-lola-mvtp/">August 2023 latency measurements</a>. That post covered what the
numbers look like. This one covers why getting to those numbers was the
easy part.</em></p>
<hr>
<h2 id="the-setup">The Setup</h2>
<p>After spending two and a half years measuring latency across six European
research-network links, I can tell you that the audio numbers are achievable.
7.5 to 22.5 ms one-way across Prague to Tallinn, LoLa and MVTP both working,
musicians playing together across national borders in real time. Technically,
that story has a satisfying ending.</p>
<p>What the measurement paper does not capture is everything that had to be true
institutionally before we could run a single test. The firewall negotiations.
The repeated calibration sessions. The network configuration that nobody
outside our small group knew how to reproduce when someone left. The grant that
funded the equipment but not the person who kept it running. The performance
session that nearly collapsed because a campus IT update had silently changed a
routing rule three days prior.</p>
<p>The technical infrastructure worked. The institutional infrastructure around it
was precarious in ways that only became visible when something broke.</p>
<p>This is what the follow-up paper tries to name.</p>
<hr>
<h2 id="what-is-a-digital-music-lab-actually">What Is a Digital Music Lab, Actually?</h2>
<p>The term gets applied to everything from a laptop cart in a classroom to
IRCAM Paris. We use it to mean something specific: a <strong>Digital Music Lab
(DML)</strong> is a hybrid environment where space, equipment, software, personnel
and organisational routines are configured together to support iterative
artistic experimentation, research-led learning and outward-facing engagement.</p>
<p>The key word is <em>configured together</em>. A room full of excellent hardware is not
a DML any more than a library is just a building full of books. What makes
either work is an invisible layer of social organisation: access policies,
shared norms, maintained documentation, people who know what to do when
something breaks.</p>
<p>We borrow a concept from infrastructure studies to describe this:
<strong>performative infrastructure</strong>. The concept draws on Star and Ruhleder (1996),
and it captures something precise — that infrastructure does not merely
<em>enable</em> activity, it also <em>shapes</em> what kinds of activity are possible in the
first place. The decision to use LoLa rather than Zoom is not just a technical
choice; it is an institutional statement about what kind of musical interaction
this space is designed to support, and about who is expected to use it.</p>
<p>This framing matters because it shifts the design question. You are not asking
&ldquo;what equipment should we buy?&rdquo; You are asking &ldquo;what kind of practice do we
want to make possible, and what organisational conditions make that practice
sustainable?&rdquo;</p>
<hr>
<h2 id="four-things-that-actually-determine-whether-a-dml-survives">Four Things That Actually Determine Whether a DML Survives</h2>
<h3 id="1-flexible-by-design-not-by-accident">1. Flexible by design, not by accident</h3>
<p>Resilient labs resist the temptation to optimise for one use case. The systems
that have lasted — Stanford CCRMA is the obvious reference point, nearly five decades
and counting — tend to separate a stable core (networking, routing,
authentication, documentation) from a more rapidly changing layer of creative
tools and workflows. The core does not change when you switch DAWs or update
your streaming platform. The tools on top of it can.</p>
<p>This sounds obvious. In practice it means being deliberate about which
dependencies you are willing to accept. A lab built on a single vendor
ecosystem can offer tight integration, but it creates a single point of
failure and a maintenance contract you will be negotiating forever. A lab built
on open protocols and well-documented configurations is more work to set up and
less work to sustain.</p>
<p>The other thing flexibility buys is pedagogical range. The same environment
can host an introductory workshop, an advanced performance-research project and
a public-facing concert without requiring incompatible reconfiguration for each.
This is not a luxury. It is what makes a DML worth the overhead compared to
just booking a studio.</p>
<h3 id="2-governance-that-survives-personnel-turnover">2. Governance that survives personnel turnover</h3>
<p>The single most dangerous sentence in any DML is: <em>&ldquo;We can ask [person] — they
know how it works.&rdquo;</em></p>
<p>Every lab has that person. The one who configured the routing. The one who
knows which cable does what. The one who has the institutional memory of every
workaround and edge case. When that person moves on, the lab frequently becomes
unreliable within six months and functionally inaccessible within a year — even
if all the equipment is still there. We call these <strong>zombie infrastructures</strong>:
technically present, functionally dead.</p>
<p>The corrective is not to document everything (though that helps). It is to
design governance so that knowledge is distributed by default. Distributed
stewardship roles — student assistants, rotating committees, peer mentors —
mean that multiple people develop operational knowledge as a matter of routine,
not as emergency knowledge transfer when someone announces they are leaving.</p>
<p>Technical staff need to be treated as co-creators in this model, not as
service providers. When networked performance is framed as peripheral
experimentation rather than core infrastructure, maintenance becomes precarious
and invisible. When it is framed as core, collaboration between artistic and
technical roles becomes institutional routine.</p>
<h3 id="3-maintenance-as-a-budget-line-not-an-afterthought">3. Maintenance as a budget line, not an afterthought</h3>
<p>Here is the infrastructure paradox: systems are valued for enabling novelty,
but they require boring, recurring investment to remain usable. Project funding
solves the novelty problem. It almost never solves the maintenance problem.</p>
<p>The costs that make a lab reliable are not one-off:</p>
<ul>
<li>Staff continuity (or explicit knowledge transfer when staff change)</li>
<li>Documentation that is actively maintained, not written once and forgotten</li>
<li>Renewal cycles for hardware and software that actually match the pace of
change in the underlying ecosystem</li>
<li>User support during active sessions, not just during setup</li>
</ul>
<p>At HfMT Köln, the operational work that dominated actual implementation time
was none of the things that appear in grant applications: coordinating network
pathways across campus boundaries, establishing and re-establishing calibration
routines after infrastructure updates, producing documentation legible to
people who were not present at the original setup, providing real-time support
during rehearsals when something behaved unexpectedly.</p>
<p>None of this is glamorous. All of it is what determines whether musicians can
actually use the system on a given Tuesday afternoon.</p>
<h3 id="4-inclusion-that-is-designed-not-assumed">4. Inclusion that is designed, not assumed</h3>
<p>Technology-intensive environments reproduce exclusion reliably unless they are
actively designed not to. The mechanisms are familiar: assumed prior
experience, cultural signals about who belongs, scheduling that conflicts with
caring responsibilities, documentation in a single language, interfaces that
reward a particular kind of technical confidence.</p>
<p>For DMLs specifically, there is an additional layer. Networked music performance
is genuinely different from co-located performance. The latency conditions
require different listening and coordination strategies. For musicians trained
in tight synchronous ensemble playing, the first experience of performing over
a network is often disorienting — latency is not a technical glitch to be fixed,
it is a compositional condition to be understood and worked with.</p>
<p>Framing this as a deficit is pedagogically counterproductive. Framing it as an
occasion to develop new artistic vocabulary — to think deliberately about what
interaction strategies work at 12 ms versus 22 ms, about how anticipatory
listening changes the character of improvisation — turns an obstacle into
content. Some of the most interesting musical thinking in our sessions came
from participants who were trying to understand why something that was
effortless in a rehearsal room required conscious attention over the network.</p>
<hr>
<h2 id="the-tensions-that-do-not-resolve">The Tensions That Do Not Resolve</h2>
<p>Being honest about what the paper does not solve:</p>
<p><strong>Project funding versus operational costs.</strong> We do not have a structural
solution to the mismatch between how labs are funded (innovation grants with
defined end dates) and how they need to operate (indefinitely, with predictable
maintenance budgets). Collaborative purchasing agreements and shared technical
teams across institutions can distribute the burden, but they introduce
coordination overhead. There is no clean answer here.</p>
<p><strong>Experimentation versus accountability metrics.</strong> Universities and funders
want quantifiable outputs. Artistic experimentation often produces its most
valuable results as changed practices and new aesthetic understanding — things
that do not appear in publication counts or utilisation statistics. The best
available response is to be explicit about this mismatch when negotiating
evaluation criteria, and to establish review processes that include artistic
peers and community partners rather than only administrators. This is possible
more often than people think, but it requires someone to argue for it
proactively.</p>
<p><strong>Openness versus depth.</strong> A lab built for maximum accessibility is not the
same as a lab optimised for a specific research agenda, and trying to be both
usually means doing neither well. The design question is not which is better
but where the tradeoff lies for a particular institution&rsquo;s mission. CCRMA and
IRCAM have made different bets on this axis over decades and both have produced
important work. The mistake is not having an opinion about where you sit on
the spectrum.</p>
<hr>
<h2 id="recommendations">Recommendations</h2>
<p>These are for institutions and funders, assembled from what the paper
describes as working across multiple DML contexts:</p>
<ul>
<li><strong>Treat DMLs as long-term cultural infrastructure.</strong> Recurring budget lines
for renewal, documentation and support — not just start-up funding.</li>
<li><strong>Separate your stable backbone from your creative tools.</strong> Networking,
routing, authentication and documentation should not be rebuilt every time
you change your video platform.</li>
<li><strong>Design governance that does not rely on one person.</strong> Distributed
stewardship roles, clear succession documentation, operational knowledge
treated as shared rather than individual.</li>
<li><strong>Make invisible labour visible.</strong> Technical stewardship, facilitation and
community liaison need to appear in hiring, workload models and evaluation
— not just in informal practice.</li>
<li><strong>Lower the floor for participation.</strong> Scaffolded onboarding, peer mentoring,
programming that supports diverse musical practices and levels of technical
experience.</li>
<li><strong>Sort out data governance before you start recording.</strong> Consent, archiving
and reuse policies for audio/video, especially when community partners or
students are involved.</li>
<li><strong>Plan for the lab&rsquo;s eventual obsolescence.</strong> Versioning policies, migration
plans, criteria for retiring tools. Zombie infrastructures are a governance
failure, not a technical one.</li>
<li><strong>Evaluate on multiple axes.</strong> Technical reliability is one. Learning
trajectories, student agency, community partnership durability and artistic
outcomes are others. Reporting only the first one creates a misleading
picture of whether the lab is actually working.</li>
</ul>
<hr>
<h2 id="what-this-does-and-does-not-claim">What This Does and Does Not Claim</h2>
<p>The argument in the paper is conceptual and practice-informed rather than
empirical in the standard sense. We synthesise literature and draw on the
HfMT Köln implementation as a vignette — it is an illustration, not a
representative sample. The framework we propose (four design principles, the
performative infrastructure framing) is offered as an analytical vocabulary
for planning and evaluation, not as a validated theory.</p>
<p>What it is useful for: making implicit infrastructure choices explicit, naming
tensions before they become crises, and supporting more realistic conversations
between artistic users, technical staff and institutional leadership about what
it actually takes to make this work.</p>
<hr>
<h2 id="references">References</h2>
<p>Borgdorff, H. (2012). <em>The Conflict of the Faculties: Perspectives on
Artistic Research and Academia.</em> Leiden University Press.</p>
<p>Labbé, D., Zuberec, C., &amp; Turner, S. (2022). Creative hubs in Hanoi,
Vietnam: Transgressive spaces in a socialist state? <em>Urban Studies</em>.
<a href="https://doi.org/10.1177/00420980221086371">https://doi.org/10.1177/00420980221086371</a></p>
<p>McKay, G. (2017). Community music: History and current practice.
<em>International Journal of Community Music</em>, 10(2), 129–137.
<a href="https://doi.org/10.1386/ijcm.10.2.129_1">https://doi.org/10.1386/ijcm.10.2.129_1</a></p>
<p>Morreale, F., Bowers, J., &amp; McPherson, A. (2021). Collaborating in
distributed musical partnerships. <em>Computers in Human Behavior</em>, 120,
106757. <a href="https://doi.org/10.1016/j.chb.2021.106757">https://doi.org/10.1016/j.chb.2021.106757</a></p>
<p>Selwyn, N. (2021). <em>Education and Technology: Key Issues and Debates</em>
(3rd ed.). Bloomsbury Academic.</p>
<p>Star, S. L., &amp; Ruhleder, K. (1996). Steps toward an ecology of
infrastructure. <em>Information Systems Research</em>, 7(1), 111–134.
<a href="https://doi.org/10.1287/isre.7.1.111">https://doi.org/10.1287/isre.7.1.111</a></p>
<p>Wenger, E. (1998). <em>Communities of Practice: Learning, Meaning, and
Identity.</em> Cambridge University Press.
<a href="https://doi.org/10.1017/CBO9780511803932">https://doi.org/10.1017/CBO9780511803932</a></p>
<hr>
<h2 id="changelog">Changelog</h2>
<ul>
<li><strong>2026-01-20</strong>: Removed the Chafe (2018) &ldquo;Stanford CCRMA: A 40-year retrospective&rdquo; reference, which could not be confirmed in available databases (DOI does not resolve, not listed in <em>Computer Music Journal</em> 42(3)). The body text reference to CCRMA as an institutional example is retained; it does not depend on this citation.</li>
<li><strong>2026-01-20</strong>: Changed &ldquo;The term comes from Star and Ruhleder (1996)&rdquo; to &ldquo;The concept draws on Star and Ruhleder (1996).&rdquo; Star and Ruhleder&rsquo;s paper is the foundational text on relational infrastructure, but they did not coin the specific compound term &ldquo;performative infrastructure.&rdquo;</li>
</ul>
]]></content:encoded>
    </item>
  </channel>
</rss>
