soo-v1-af
Combined SFT + SOO LoRA adapter (rank 128) for use with meta-llama/Llama-3.3-70B-Instruct.
Created by concatenating two rank-64 LoRA adapters:
- SFT adapter (
jplhughes2/llama-3.3-70b-af-synthetic-docs-only-more-data-r-epochs) — supervised fine-tuning on synthetic alignment-faking documents - SOO adapter (
soo_ft_70b_attempt_4) — logits-based Self-Other Overlap fine-tuning on top of the SFT model
The merge uses exact concatenation (no SVD approximation): ΔW_combined = ΔW_sft + ΔW_soo.
LoRA config
- Rank: 128 (64 + 64)
- Alpha: 128
- Target modules: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
- Framework: PEFT
- Downloads last month
- 11
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for Arrrlex/soo-v1-af
Base model
meta-llama/Llama-3.1-70B Finetuned
meta-llama/Llama-3.3-70B-Instruct