6.5 KiB
Specificity F1 Improvement Plan
Goal: Macro F1 > 0.80 on both category and specificity heads Current: Cat F1=0.932 (passing), Spec F1=0.517 (needs ~+0.28) Constraint: Specificity is paragraph-level and category-independent by design
Diagnosis
Per-class spec F1 (best run, epoch 5):
- L1 (Generic): ~0.79
- L2 (Domain-Adapted): ~0.29
- L3 (Firm-Specific): ~0.31
- L4 (Quantified): ~0.55
L2 and L3 drag macro F1 from ~0.67 average to 0.52. QWK=0.840 shows ordinal ranking is strong — the problem is exact boundary placement between adjacent levels.
Root causes
-
CORAL's shared weight vector. CORAL uses
logit_k = w·x + b_k— one weight vector for all thresholds. But the three transitions require different features:- L1→L2: cybersecurity terminology detection (ERM test)
- L2→L3: firm-unique fact detection (named roles, systems)
- L3→L4: quantified/verifiable claim detection (numbers, dates) A single w can't capture all three signal types.
-
[CLS] pooling loses distributed signals. A single "CISO" mention anywhere in a paragraph should bump to L3, but [CLS] may not attend to it.
-
Label noise at boundaries. 8.7% of training labels had Grok specificity disagreement, concentrated at L1/L2 and L2/L3 boundaries.
-
Insufficient training. Model was still improving at epoch 5 — not converged.
Ideas (ordered by estimated ROI)
Tier 1 — Implement first
A. Independent threshold heads (replace CORAL) Replace the single CORAL weight vector with 3 independent binary classifiers, each with its own learned features:
- threshold_L2plus: Linear(hidden, 1) — "has any qualifying facts?"
- threshold_L3plus: Linear(hidden, 1) — "has firm-specific facts?"
- threshold_L4: Linear(hidden, 1) — "has quantified/verifiable facts?"
Same cumulative binary targets as CORAL, but each threshold has independent weights. Optionally upgrade to 2-layer MLP (hidden→256→1) for richer decision boundaries.
B. High-confidence label filtering Only train specificity on paragraphs where all 3 Grok runs agreed on specificity level (~91.3% of data, ~59K of 65K). The ~6K disagreement cases are exactly the noisy boundary labels that confuse the model. Category labels can still use all data.
C. More epochs + early stopping on spec F1 Run 15-20 epochs. Switch model selection metric from combined_macro_f1 to spec_macro_f1 (since category already exceeds 0.80). Use patience=5.
D. Attention pooling Replace [CLS] token pooling with learned attention pooling over all tokens. This lets the model attend to specific evidence tokens (CISO, $2M, NIST) distributed anywhere in the paragraph.
Tier 2 — If Tier 1 insufficient
E. Ordinal consistency regularization Add a penalty when threshold k fires but threshold k-1 doesn't (e.g., model says "has firm-specific" but not "has domain terms"). Weight ~0.1.
F. Differential learning rates Backbone: 1e-5, heads: 5e-4. Let the heads learn classification faster while the backbone makes only fine adjustments.
G. Softmax head comparison Try standard 4-class CE (no ordinal constraint at all). If it outperforms both CORAL and independent thresholds, the ordinal structure isn't helping.
H. Multi-sample dropout Apply N different dropout masks, average logits. Reduces variance in the specificity head's predictions, especially for boundary cases.
Tier 3 — If nothing else works
I. Specificity-focused auxiliary task
The consensus labels include specific_facts arrays with classified fact types
(domain_term, named_role, quantified, etc.). Add a token-level auxiliary task
that detects these fact types. Specificity becomes "what's the highest-level
fact type present?" — making the ordinal structure explicit.
J. Separate specificity model Train a dedicated model just for specificity with a larger head, more specificity-focused features, or a different architecture (e.g., token-level fact extraction → aggregation).
K. Re-annotate boundary cases Use GPT-5.4 to re-judge the ~9,323 majority-vote cases where Grok had specificity disagreement. Cleaner labels at boundaries.
Experiment Log
Experiment 1: Independent thresholds + attention pooling + MLP + filtering (15 epochs)
Config: configs/finetune/iter1-independent.yaml
- Specificity head: independent (3 separate Linear(1024→256→1) binary classifiers)
- Pooling: attention (learned attention over all tokens)
- Confidence filtering: only train spec on unanimous labels
- Ordinal consistency regularization: 0.1
- Class weighting: yes
- Base checkpoint: ModernBERT-large (no DAPT/TAPT)
- Epochs: 15
Results:
| Epoch | Combined | Cat F1 | Spec F1 | QWK |
|---|---|---|---|---|
| 1 | 0.855 | 0.867 | 0.844 | 0.874 |
| 2 | 0.913 | 0.909 | 0.918 | 0.935 |
| 3 | 0.925 | 0.919 | 0.931 | 0.945 |
| 4 | 0.936 | 0.932 | 0.940 | 0.950 |
| 5 | 0.938 | 0.936 | 0.940 | 0.949 |
| 8 | 0.944 | 0.943 | 0.945 | 0.952 |
| 10 | 0.944 | 0.943 | 0.945 | 0.952 |
| 11 | 0.944 | 0.945 | 0.944 | 0.952 |
Stopped at epoch 11 — train-eval loss gap was 8× (0.06 vs 0.49) with no further eval F1 improvement. Best checkpoint: epoch 8 (spec F1=0.945).
Conclusion: Massive improvement — spec F1 went from 0.517 (CORAL baseline) to 0.945 at epoch 8. Both targets (>0.80 cat and spec F1) exceeded by epoch 1. Independent thresholds were the key insight — CORAL's shared weight vector was the primary bottleneck. Attention pooling, MLP heads, and confidence filtering all contributed. Tier 2 and Tier 3 ideas were not needed.
Holdout Evaluation (1,200 paragraphs, proxy gold)
Validated on held-out data against two independent frontier model references:
| Model | Ref | Cat F1 | Spec F1 | L2 F1 | Spec QWK |
|---|---|---|---|---|---|
| Independent (ep8) | GPT-5.4 | 0.934 | 0.895 | 0.798 | 0.932 |
| Independent (ep8) | Opus-4.6 | 0.923 | 0.883 | 0.776 | 0.923 |
| CORAL (ep5) | GPT-5.4 | 0.936 | 0.597 | 0.407 | 0.876 |
| CORAL (ep5) | Opus-4.6 | 0.928 | 0.596 | 0.418 | 0.872 |
| GPT-5.4 | Opus-4.6 | — | 0.885 | 0.805 | 0.919 |
Key finding: The model's holdout spec F1 (0.895) exceeds the inter-reference agreement (0.885 between GPT-5.4 and Opus-4.6). The model has reached the construct reliability ceiling — further improvement requires cleaner reference labels, not a better model.
L2 is at ceiling: Model L2 F1 (0.798) is within 0.007 of reference agreement (0.805). The L1↔L2 boundary is genuinely ambiguous. Remaining opportunity: per-threshold sigmoid tuning against human gold labels (potential +0.01-0.02).