Overview Method Research Team Resources
A long-term research initiative at MIT

Neural Thickets Mapping the Dense Landscape of Task Experts in Pretrained Models

Neural Thickets reframes pretraining as a local distribution over weights: in large, well-pretrained models, many nearby parameter vectors are already task-improving experts, so simple random sampling, top-K selection, and majority-vote ensembling can rival post-training methods like PPO and GRPO.

θ̃ = θ + σε ε ~ N(0, I) select top-k θ̃ᵢ output = vote(θ̃₁ … θ̃ₖ)
Small model: Needle in a Haystack
Small Model — "Needle in a Haystack" Gemini
Large model: Neural Thicket
Large Model — "Neural Thicket" Gemini

RandOpt Pipeline

The foundational algorithm behind Neural Thickets: a zero-training approach that discovers task-specific experts by perturbing pretrained weights and ensembling the best candidates.

01

Perturb

Add Gaussian noise: θ̃ = θ + σε

02

Evaluate

Score each candidate in parallel across GPUs

03

Select

Keep top-k by task performance

04

Ensemble

Aggregate via majority vote

Exploring the Thicket

Neural Thickets opens multiple interconnected research directions—from theoretical foundations to practical applications. Each thrust feeds back into the others.

People

Links & Assets