Katherine Tian(@kattian_) 's Twitter Profileg
Katherine Tian

@kattian_

cs/stat @harvard, working on calibration & factuality of LLMs, prev @GoogleAI tensorflow, golden state @warriors fan

ID:951151284735827968

linkhttps://kttian.github.io/ calendar_today10-01-2018 17:58:32

127 Tweets

712 Followers

494 Following

jack morris(@jxmnop) 's Twitter Profile Photo

one of the most important things I know about deep learning I learned from this paper: 'Pretraining Without Attention'

this what I found so surprising:
these people developed an architecture very different from Transformers called BiGS, spent months and months optimizing it and

one of the most important things I know about deep learning I learned from this paper: 'Pretraining Without Attention' this what I found so surprising: these people developed an architecture very different from Transformers called BiGS, spent months and months optimizing it and
account_circle
Chelsea Finn(@chelseabfinn) 's Twitter Profile Photo

I’m really excited to be starting a new adventure with multiple amazing friends & colleagues.

Our company is called Physical Intelligence (Pi or Ο€, like the policy).

A short thread 🧡

account_circle
Raj Movva(@rajivmovva) 's Twitter Profile Photo

We wrote a position piece on how LMs have expanded the toolkit for social equity researchers, especially in health - check it out, and feel free to share thoughts! shorturl.at/knHK9.

(and feeling very lucky to have worked with so many cool co-authors who I look up to!)

account_circle
Eric(@ericmitchellai) 's Twitter Profile Photo

Come by the Instruction Following workshop (room 220-222) to see our work on:

*Emulated fine-tuning*: RLHF without fine-tuning!

*Fine-tuning for factuality*: how to fine-tune LLMs directly for factuality, reducing hallucination by >50%

RIGHT NOW!!!

account_circle
Chelsea Finn(@chelseabfinn) 's Twitter Profile Photo

DPO is a runner up for NeurIPS outstanding paper. πŸ™Œ

Big congrats especially to the students Rafael Rafailov Archit Sharma Eric & the other awardees.

If you haven't learned about DPO already, check out the oral & poster πŸ‘‡ on Thurs afternoon.

DPO is a runner up for NeurIPS outstanding paper. πŸ™Œ Big congrats especially to the students @rm_rafailov @archit_sharma97 @ericmitchellai & the other awardees. If you haven't learned about DPO already, check out the oral & poster πŸ‘‡ on Thurs afternoon.
account_circle
Alex Gu(@minimario1729) 's Twitter Profile Photo

come talk to us about the pros and cons of using symbolic code representations to do logical reasoning tomorrow at 9am!

account_circle
Chelsea Finn(@chelseabfinn) 's Twitter Profile Photo

Can LLMs keep themselves up to date by reading the news?

Fine-tuning on news articles doesn't work.

Using meta-learning, we can reweight news article tokens so that fine-tuning works.

Nathan Hu & Eric presenting this work at this week!

Can LLMs keep themselves up to date by reading the news? Fine-tuning on news articles doesn't work. Using meta-learning, we can reweight news article tokens so that fine-tuning works. @NathanHu12 & @ericmitchellai presenting this work at #EMNLP2023 this week!
account_circle
Eric(@ericmitchellai) 's Twitter Profile Photo

Come see Katherine Tian @ ICLR πŸ‡¦πŸ‡Ή's work on the ability of RLHF'd LLMs to *directly verbalize* probabilities (yes that's right as tokens) that are actually pretty-well calibrated! (Usually than the log probs!!! 🀯🀯🀯)

Poster 14B

R I G H T N O W until 3:30 SG time!!

account_circle
Chelsea Finn(@chelseabfinn) 's Twitter Profile Photo

LLMs fine-tuned with RLHF are known to be poorly calibrated.

We found that they can actually be quite good at *verbalizing* their confidence.

Led by Katherine Tian @ ICLR πŸ‡¦πŸ‡Ή and Eric, at this week.

Paper: arxiv.org/abs/2305.14975

LLMs fine-tuned with RLHF are known to be poorly calibrated. We found that they can actually be quite good at *verbalizing* their confidence. Led by @kattian_ and @ericmitchellai, at #EMNLP2023 this week. Paper: arxiv.org/abs/2305.14975
account_circle
Katherine Tian(@kattian_) 's Twitter Profile Photo

Turns out we can directly train language models to hallucinate less without any human annotation -- for around 50% error reduction compared to RLHF!! Check out our paper for the approaches and full results πŸ˜ƒ

account_circle
Eric(@ericmitchellai) 's Twitter Profile Photo

Very curious to see how far we can push training to simply *not hallucinate.* It won't give us perfect models, but it seems like really meaningful (more than 50%) reduction in factual errors might be possible. Needs to be scaled up πŸ˜€ full thread on the way!

account_circle