Yejin Choi(@YejinChoinka) 's Twitter Profileg
Yejin Choi

@YejinChoinka

professor at UW, director at AI2, adventurer at heart

ID:893882282175471616

linkhttps://homes.cs.washington.edu/~yejin/ calendar_today05-08-2017 17:11:58

1,6K Tweets

18,6K Followers

330 Following

Christopher Manning(@chrmanning) 's Twitter Profile Photo

.Yejin Choi makes a prediction that I can get behind: “30% chance that within 3 years, we will have a language-only Al that is perceived as AGl-enough by ~30% of people”. This seems right. People—including scientists—easily (over-)attribute intelligence to machines.

.@YejinChoinka makes a prediction that I can get behind: “30% chance that within 3 years, we will have a language-only Al that is perceived as AGl-enough by ~30% of people”. This seems right. People—including scientists—easily (over-)attribute intelligence to machines.
account_circle
Swabha Swayamdipta(@swabhz) 's Twitter Profile Photo

What explains the effectiveness and staying power of truncation based inference algorithms (top-p, top-k) for LLMs? Matthew Finlayson @ ICLR is presenting our poster right now at , go find out!

What explains the effectiveness and staying power of truncation based inference algorithms (top-p, top-k) for LLMs? @mattf1n is presenting our poster right now at #iclr24, go find out!
account_circle
Allen Institute for AI(@allen_ai) 's Twitter Profile Photo

🎉We're celebrating our OLMo team's big win last night for Innovation of the Year at the GeekWire Awards! Thank you to the panel for recognizing our efforts towards open science, and to the other nominees for pushing us all towards better AI technology! geekwire.com/2024/geekwire-…

🎉We're celebrating our OLMo team's big win last night for Innovation of the Year at the @geekwire Awards! Thank you to the panel for recognizing our efforts towards open science, and to the other nominees for pushing us all towards better AI technology! geekwire.com/2024/geekwire-…
account_circle
Sasha Rush (ICLR)(@srush_nlp) 's Twitter Profile Photo

Talk: 'OLMo: Findings of Training an Open LM' from Hanna Hajirshizi at AI2 from OSGAI.

Extremely interesting overview of the 4 parts (Data, Training, Adaptation, Eval) of the OLMo open LLM project. Rare insight into how these processes work at scale.

youtube.com/watch?v=qFZbu2…

account_circle
Faeze Brahman(@faeze_brh) 's Twitter Profile Photo

Our new naacl paper showcased most of LLMs achieve less than 40% chance of success in creative problem solving and out of the box thinking. Though under certain domain specific settings, llms capabilities found to be complementary to human capabilities:
arxiv.org/abs/2311.09682

account_circle
Nicholas Meade(@ncmeade) 's Twitter Profile Photo

Adversarial Triggers For LLMs Are 𝗡𝗢𝗧 𝗨𝗻𝗶𝘃𝗲𝗿𝘀𝗮𝗹!😲

It is believed that adversarial triggers that jailbreak a model transfer universally to other models. But we show triggers don't reliably transfer, especially to RLHF/DPO models.

Paper: arxiv.org/abs/2404.16020

Adversarial Triggers For LLMs Are 𝗡𝗢𝗧 𝗨𝗻𝗶𝘃𝗲𝗿𝘀𝗮𝗹!😲 It is believed that adversarial triggers that jailbreak a model transfer universally to other models. But we show triggers don't reliably transfer, especially to RLHF/DPO models. Paper: arxiv.org/abs/2404.16020
account_circle
vintro(@vintrotweets) 's Twitter Profile Photo

time for a paper post: A Roadmap to Pluralistic Alignment

what happens when you take alignment down to its most granular form? you arrive at individual alignment... aka personalization. but to date, most of the utility found in LLMs has come from a global alignment approach

time for a paper post: A Roadmap to Pluralistic Alignment what happens when you take alignment down to its most granular form? you arrive at individual alignment... aka personalization. but to date, most of the utility found in LLMs has come from a global alignment approach
account_circle
Hannah Rose Kirk(@hannahrosekirk) 's Twitter Profile Photo

Published in Nature Machine Intelligence today, our new article explores the trade-offs of personalised alignment in large language models ⚖️ Personalisation has potential to democratise decisions over how LLMs behave, but brings its own set of risks...
nature.com/articles/s4225…

Published in Nature Machine Intelligence today, our new article explores the trade-offs of personalised alignment in large language models ⚖️ Personalisation has potential to democratise decisions over how LLMs behave, but brings its own set of risks... nature.com/articles/s4225…
account_circle
Yoav Artzi(@yoavartzi) 's Twitter Profile Photo

We created reviewing guidelines for Conference on Language Modeling. Not intended to automate the committee work, or dictate constraints. But, to inspire a thoughtful reviewing process, for an exciting and impactful program of the highest possible quality. We have a wonderful program committee ❤️

We created reviewing guidelines for @COLM_conf. Not intended to automate the committee work, or dictate constraints. But, to inspire a thoughtful reviewing process, for an exciting and impactful program of the highest possible quality. We have a wonderful program committee ❤️
account_circle
Mike Lewis(@ml_perception) 's Twitter Profile Photo

Yes, both the 8B and 70B are trained way more than is Chinchilla optimal - but we can eat the training cost to save you inference cost! One of the most interesting things to me was how quickly the 8B was improving even at 15T tokens.

account_circle
Jack Hessel(@jmhessel) 's Twitter Profile Photo

Just tried the new GPT4+v on our New Yorker caption contest task (arxiv.org/abs/2209.06293).

It does OK! (70%, good for second on leaderboard). But, w/ performance ~25% below human, it still doesn't quite 'get the joke'. Maybe your model does? :-) capcon.dev

Just tried the new GPT4+v on our New Yorker caption contest task (arxiv.org/abs/2209.06293). It does OK! (70%, good for second on leaderboard). But, w/ performance ~25% below human, it still doesn't quite 'get the joke'. Maybe your model does? :-) capcon.dev
account_circle
Hanna Hajishirzi(@HannaHajishirzi) 's Twitter Profile Photo

Introducing our best OLMo yet. OLMo 1.7-7B outperforms LLaMa2-7B, approaching LLaMa2-13B at MMLU and GSM8k. High-quality data and staged training are key.

I am so proud of our team making such significant improvement in a short period after our first release.

Introducing our best OLMo yet. OLMo 1.7-7B outperforms LLaMa2-7B, approaching LLaMa2-13B at MMLU and GSM8k. High-quality data and staged training are key. I am so proud of our team making such significant improvement in a short period after our first release.
account_circle
Usman Anwar(@usmananwar391) 's Twitter Profile Photo

We released this new agenda on LLM-safety yesterday. This is VERY comprehensive covering 18 different challenges.

My co-authors have posted tweets for each of these challenges. I am going to collect them all here!

P.S. this is also now on arxiv: arxiv.org/abs/2404.09932

account_circle
David Krueger(@DavidSKrueger) 's Twitter Profile Photo

I’m super excited to release our 100+ page collaborative agenda - led by Usman Anwar - on “Foundational Challenges In Assuring Alignment and Safety of LLMs” alongside 35+ co-authors from NLP, ML, and AI Safety communities!

Some highlights below...

I’m super excited to release our 100+ page collaborative agenda - led by @usmananwar391 - on “Foundational Challenges In Assuring Alignment and Safety of LLMs” alongside 35+ co-authors from NLP, ML, and AI Safety communities! Some highlights below...
account_circle
Niloofar (Fatemeh) @ICLR 🇦🇹(@niloofar_mire) 's Twitter Profile Photo

I will be talking about what differential privacy is, what it is not and what some common misconceptions are in privacy for generative AI in a couple hours The GenLaw Center in DC!

Join us on the live stream: tinyurl.com/genlaw-stream
Slides: tinyurl.com/genlaw-dp-2024

I will be talking about what differential privacy is, what it is not and what some common misconceptions are in privacy for generative AI in a couple hours @genlawcenter in DC! Join us on the live stream: tinyurl.com/genlaw-stream Slides: tinyurl.com/genlaw-dp-2024
account_circle
Seungju Han(@SeungjuHan3) 's Twitter Profile Photo

🥰Excited to share that I will be joining AI2 Allen Institute for AI MOSAIC this September as a predoctoral young investigator!! So excited to continue working with amazing Yejin Choi Nouha Dziri Liwei Jiang Kavel Rao and can't wait to collaborate with others!

account_circle
Niloofar (Fatemeh) @ICLR 🇦🇹(@niloofar_mire) 's Twitter Profile Photo

Can we uncover memorization of pre-training data in LLMs, using other LLMs?

Our iterative prompt optimization method finds prompts that propel an LM to output training data using other LMs. We show higher avg. data reconstruction & extract 1.4X more PII!
arxiv.org/abs/2403.04801

Can we uncover memorization of pre-training data in LLMs, using other LLMs? Our iterative prompt optimization method finds prompts that propel an LM to output training data using other LMs. We show higher avg. data reconstruction & extract 1.4X more PII! arxiv.org/abs/2403.04801
account_circle
Jiacheng Liu (Gary)(@liujc1998) 's Twitter Profile Photo

The infini-gram paper is updated with the incredible feedback from the online community 🧡 We added references to papers of Jeff Dean (@🏡) Yee Whye Teh Ehsan Shareghi Edward Raff et al.

arxiv.org/abs/2401.17377

Also happy to share that the infini-gram API has served 30 million queries!

account_circle