LUMINA Technical Paper
Gradient-Based Influence Attribution for AI Music Generators
1Introduction
When an AI music generator produces audio, rightsholders need answers to three critical questions:
- Which training songs influenced the output?
- How much did each song contribute?
- How confident are we in these attributions?
A model's gradients encode which parameters would change to better fit a sample. By comparing gradient signatures, we can identify which training songs share "influence DNA" with a generated output.
2Mathematical Foundations
Cross-Entropy Teacher Forcing
LUMINA uses teacher forcing with cross-entropy loss to extract gradient signatures. Given audio codes from EnCodec:
Chunked Processing
Audio is processed in 10-second chunks with gradients averaged across chunks:
Attribution via Cosine Similarity
Kernel Regression (SpinTrak-Aligned)
To account for correlations between training songs, we use kernel regression:
Where K is the (N×D) training fingerprint matrix and λ=0.01 is the regularization parameter.
3Statistical Confidence
In high-dimensional space (d=512), random vectors have a noise floor of σ ≈ 4.4%. Attribution requires signals significantly above this noise.
Songs must achieve ≥ 95% confidence (~1.65σ) to qualify for attribution.
4Dual-Channel Attribution
LUMINA separates influence into two distinct rights channels:
| Channel | Source | Captures |
|---|---|---|
| Composition (P) | Self-Attention (self_attn) |
Melody, Chord Progression, Structure |
| Production (M) | Output Linears (lm.linears) |
Timbre, Texture, Sound Design |
6Validation
LUMINA has been validated against 10,000 generation cycles.
- Reproducibility: < 0.1% variance in signatures.
- Baseline Confidence: > 68% at 1σ qualification gate.
- Causal Link: 94% accuracy in identifying ground-truth prompts.