r/MachineLearning · 10h ago · 7 · fine tuning benchmark inference workflow

Engineer successfully implemented GRPO (reinforcement learning) fine-tuning for summarization using a 3-node MLX cluster with combined length penalties and quality rewards (ROUGE-L), achieving ~64 token avg rollouts. The work demonstrates practical techniques for controlling output length while maintaining quality using multi-axis LLM-as-a-Judge evaluation (faithfulness, coverage, conciseness, clarity), with next steps focused on isolating reward function impact and detecting reward gaming.

r/MachineLearning · 16h ago · 7 · fine tuning open source tool research

Fine-tuned open-source TTS model (Chatterbox) for 8 Indian languages using LoRA adapters (1.4% parameters) and grapheme-level tokenization with Brahmic script warm-start initialization. Achieves sub-0.25 CER for most languages except Malayalam (0.86), demonstrating efficient multilingual adaptation without full model retraining or language-specific G2P pipelines.

Anthropic Research · 20h ago · 7 · research agent fine tuning benchmark

Anthropic's research explores weak-to-strong supervision as a practical approach to scalable oversight—training stronger AI models using weaker model feedback to prepare for supervising future superhuman AI. The study tests whether Claude can autonomously develop and test alignment methods, demonstrating potential for AI systems to accelerate their own alignment research.

Simon Willison · 22h ago · 6 · new model fine tuning api update

OpenAI released GPT-5.4-Cyber, a fine-tuned variant optimized for defensive cybersecurity use cases, along with a Trusted Access for Cyber program using identity verification for reduced-friction access. The announcement emphasizes OpenAI's existing cybersecurity work and self-service verification, though premium tools still require application approval similar to competing offerings.

DeepMind Blog · 13d ago · 9 · new model open source inference fine tuning

Google released Gemma 4, a family of open-source models (2B to 31B parameters) built on Gemini 3 technology, ranked #3 and #6 on Arena AI leaderboard for their sizes. The models are optimized for on-device deployment, agentic workflows, and fine-tuning across hardware from mobile to datacenter, with Apache 2.0 licensing enabling direct integration into engineering workflows.

HuggingFace Blog · 15d ago · 8 · new model open source fine tuning research

IBM releases Granite 4.0 3B Vision, a modular vision-language model optimized for chart and document understanding, delivered as a LoRA adapter on Granite 4.0 Micro with a novel DeepStack architecture for multi-layer visual feature injection. The release includes ChartNet, a 1.7M-sample synthetic dataset for chart interpretation with code-guided augmentation, addressing a key VLM weakness in structured data reasoning.

HuggingFace Blog · 15d ago · 8 · open source tool research benchmark fine tuning

OpenMed built an end-to-end open-source protein engineering pipeline combining structure prediction, sequence design, and codon optimization, with novel contributions in codon-level language modeling. They benchmarked transformer architectures (CodonRoBERTa-large-v2 vs ModernBERT) for codon optimization, scaled to 25 species in 55 GPU-hours, and released runnable code with full experimental transparency—directly applicable for engineers building biological AI systems.

HuggingFace Blog · 15d ago · 8 · library fine tuning workflow research

TRL v1.0 introduces architectural lessons for building stable post-training libraries that can adapt as methods evolve from PPO to DPO to RLVR approaches. The library design prioritizes flexibility over fixed abstractions, recognizing that core concepts like reward models shift between being fundamental, optional, or reimagined as verifiers across different training paradigms.

OpenAI Research · 36d ago · 7 · research fine tuning safety prompt engineering

IH-Challenge is a training framework that teaches models to respect instruction hierarchy and distinguish between trusted vs. untrusted inputs, improving robustness against prompt injection attacks and enhancing safety steerability. This is practically useful for engineers building production AI systems that need stronger defenses against adversarial inputs.

Ahead of AI · 106d ago · 9 · research new model fine tuning benchmark

A comprehensive retrospective on 2025's major LLM developments, starting with DeepSeek R1's January release showing that reinforcement learning (specifically RLVR/GRPO) can enable reasoning-like behavior in LLMs, and revealing that state-of-the-art model training may cost an order of magnitude less than previously estimated. The article examines how post-training scaling through verifiable rewards represents a significant algorithmic shift from SFT/RLHF approaches, opening new possibilities for capability unlocking.