LUMINA Technical Paper
Gradient-Based Influence Attribution for AI Music Generators
1Introduction
When an AI music generator produces audio, rightsholders need answers to three critical questions:
- Which training songs influenced the output?
- How much did each song contribute?
- How confident are we in these attributions?
A model's gradients encode which parameters would change to better fit a sample. By comparing gradient signatures, we can identify which training songs share "influence DNA" with a generated output.
From raw signal to fair influence share — how we find who really taught the model what it used.
2Mathematical Foundations
Cross-Entropy Teacher Forcing
LUMINA uses teacher forcing with cross-entropy loss to extract gradient signatures. Given audio codes from EnCodec:
Chunked Processing
Audio is processed in 10-second chunks with gradients averaged across chunks:
Attribution via Cosine Similarity
Like spotting which teacher taught the exact method a student used on the test. Each training song leaves a unique gradient fingerprint — a record of how it shaped the model's weights. Cosine similarity measures how closely aligned two fingerprints are, revealing causal influence.
Kernel Regression (SpinTrak-Aligned)
To account for correlations between training songs, we use kernel regression:
Where K is the (N×D) training fingerprint matrix and λ=0.01 is the regularization parameter.
If two teachers taught the same lesson, they share the credit rather than both getting full marks. Kernel regression decorrelates overlapping training samples — when two songs taught similar patterns, the regularized inverse (KKᵀ + λI)⁻¹ attributes proportionally rather than double-counting.
3Statistical Confidence
In high-dimensional space (d=512), random vectors have a noise floor of σ ≈ 4.4%. Attribution requires signals significantly above this noise.
Songs must achieve ≥ 95% confidence (~1.65σ) to qualify for attribution.
Only the top performers make the finals. In 512-dimensional space, random noise produces a baseline similarity of ~4.4%. The confidence gate (≥ 95% via the error function) ensures we only attribute influence to songs whose signal is statistically significant — not random coincidence.
4Dual-Channel Attribution
LUMINA separates influence into two distinct rights channels:
| Channel | Source | Captures |
|---|---|---|
| Composition (P) | Self-Attention (self_attn) |
Melody, Chord Progression, Structure |
| Production (M) | Output Linears (lm.linears) |
Timbre, Texture, Sound Design |
6Validation
LUMINA has been validated against 10,000 generation cycles.
- Reproducibility: < 0.1% variance in signatures.
- Baseline Confidence: > 68% at 1σ qualification gate.
- Causal Link: 94% accuracy in identifying ground-truth prompts.
7Version History
Intellectual Property Notice
This document contains proprietary and confidential information belonging to Fold Artists Research. The methods, algorithms, and technical implementations described herein are protected intellectual property. Unauthorized reproduction, distribution, or disclosure of this document or its contents is strictly prohibited and may violate applicable trade secret and intellectual property laws.