Ben Edelman(@EdelmanBen) 's Twitter Profileg
Ben Edelman

@EdelmanBen

Final-year PhD candidate at Harvard CS trying to understand AI scientifically. New to the platform formerly known as Twitter.

ID:2383010678

linkhttps://www.benjaminedelman.com/ calendar_today11-03-2014 02:27:18

33 Tweets

113 Followers

20 Following

Boaz Barak(@boazbaraktcs) 's Twitter Profile Photo

One challenge in fighting “hallucinations” is distinguishing epistemic uncertainty from the natural “aleatoric” uncertainty of language. One can define latter as the uncertainty that remains in the limit of infinite compute and data.

Question is whether smaller model can…

account_circle
Gustaf Ahdritz(@gahdritz) 's Twitter Profile Photo

In new work (with Tian Qin, Nikhil Vyas, Boaz Barak, and Ben Edelman), we show that LLMs can identify their own epistemic uncertainty in free-form text, suggesting new approaches for combating hallucinations. (1/6)
Paper: arxiv.org/abs/2402.03563
Blog: bit.ly/3U2emAP

account_circle
Kempner Institute at Harvard University(@KempnerInst) 's Twitter Profile Photo

New Deeper Learning blog post: a linear probe can unlock LM's metacognitive capability to distinguish tokens that are 'knowable' from tokens where its predictions can't be improved. bit.ly/3U2emAP
Gustaf Ahdritz, Tian Qin, Nikhil Vyas, Boaz Barak, Ben Edelman

New Deeper Learning blog post: a linear probe can unlock LM's metacognitive capability to distinguish tokens that are 'knowable' from tokens where its predictions can't be improved. bit.ly/3U2emAP @gahdritz, Tian Qin, @vyasnikhil96, @boazbaraktcs, @EdelmanBen #AI #LLM
account_circle
Ekdeep Singh(@EkdeepL) 's Twitter Profile Photo

Two (out of 8) things that Sam Bowman wants you to know about LLMs:

(i) LLMs predictably get more capable with increasing investment

(ii) Many important LLM behaviors emerge unpredictably How can we get ahead of the curve and predict these ‘unpredictable’ behaviors?🧵⬇️

account_circle
David Krueger(@DavidSKrueger) 's Twitter Profile Photo

I’m super excited to release our 100+ page collaborative agenda - led by Usman Anwar - on “Foundational Challenges In Assuring Alignment and Safety of LLMs” alongside 35+ co-authors from NLP, ML, and AI Safety communities!

Some highlights below...

I’m super excited to release our 100+ page collaborative agenda - led by @usmananwar391 - on “Foundational Challenges In Assuring Alignment and Safety of LLMs” alongside 35+ co-authors from NLP, ML, and AI Safety communities! Some highlights below...
account_circle
Surbhi Goel(@SurbhiGoel_) 's Twitter Profile Photo

New paper on understanding the evolution of induction heads using a clean task of in-context learning Markov chains. Phase transitions, cool theory and videos! Check it out!

Shout-out to Ezra Edelman (first PhD paper!) and Nikos, for leading this work!

account_circle
Ben Edelman(@EdelmanBen) 's Twitter Profile Photo

Top-left: hmm, bumpy loss curve 🤨

Top-right: (looking at functional behavior) Model goes through five phases! In-context 0-grams, 1-grams, 2-grams, 3-grams, 4-grams.

Bottom: Phases correspond to attn heads specializing one by one. Functional emergence <-> circuit emergence!

account_circle
Ezra Edelman(@ezra_edelman) 's Twitter Profile Photo

How do induction heads / in-context learning emerge? We study a new synthetic task to better understand how these circuits evolve (in stages!) over time.
w/ Ben Edelman, Surbhi Goel, Eran Malach & Nikos Tsilivis
Blog: unprovenalgos.github.io/statistical-in…
Paper: arxiv.org/abs/2402.11004 (1/7)

account_circle