I am a UC Chancellor’s Postdoctoral Fellow at UCSD, working with Brad Voytek and Ben Bergen in the Department of Cognitive Science, where I completed my PhD in 2023.

Throughout my research career, I have always been interested in how neural circuits adapt to variance and patterns of regularity in the world around us. In my graduate work, I explored this question within the rodent hippocampus, asking how inhibitory interneurons varied their spike timing as a function of changes to an environment’s “reward landscape”. More recently, I have pivoted my research efforts towards the study of learning in artificial neural networks, with a focus on transformer-based language models and how they vary their weights over the course of pre-training.

I am broadly interested in questions that get at the training dynamics of language models:

  1. what is the shape of the trajectory of performance on a particular task, over pre-training (e.g. is there a sharp increase, or gradual ramp-up in performance? are increases relatively smooth, or is there a lot of variance from checkpoint to checkpoint?)

  2. for a complex capability (e.g. mental state modeling: Alma thinks the ball is in the box.), is the onset of improvement preceded by the onset of improvement in a capability that we would deem more basic (e.g. having a robust situation model: Alma moved the ball to the box.)?

  3. can we use the charted trajectory of training dynamics to isolate the internal mechanisms that give rise to these trajectories of performance?

CV publications

contact information

email

pdrivier@ucsd.edu