--- license: apache-2.0 datasets: - nbeerbower/GreatFirewall-DPO - nbeerbower/Schule-DPO - nbeerbower/Purpura-DPO - nbeerbower/Arkhaios-DPO - jondurbin/truthy-dpo-v0.1 - antiven0m/physical-reasoning-dpo - flammenai/Date-DPO-NoAsterisks - flammenai/Prude-Phi3-DPO - jondurbin/gutenberg-dpo-v0.1 - nbeerbower/gutenberg2-dpo - nbeerbower/gutenberg-moderne-dpo - sam-paech/gutenberg3-dpo-gemma3-12b - nbeerbower/human-writing-dpo - nbeerbower/synthetic-fiction-dpo - Atsunori/HelpSteer2-DPO - GeneralReasoning/GeneralThought-430K base_model: - lemon07r/Qwen3-R1-SLERP-Q3T-8B --- # Wenyan-Qwen3-8B An attempt to build a Xiaolong-like tune with more Gutenberg data on top of [lemon07r/Qwen3-R1-SLERP-Q3T-8B](https://huggingface.co/lemon07r/Qwen3-R1-SLERP-Q3T-8B). ## Results I haven't done much testing but the model will sometimes skip thinking. The second epoch may have overcooked it. ## Data Condensed and formatted data available [here](https://huggingface.co/datasets/nbeerbower/WenyanMix-DPO).