Robin Jia(@robinomial) 's Twitter Profileg
Robin Jia

@robinomial

Assistant Professor @CSatUSC | Previously Visiting Researcher @facebookai | Stanford CS PhD @StanfordNLP

ID:1012392833834029056

linkhttps://robinjia.github.io/ calendar_today28-06-2018 17:50:35

172 Tweets

3,2K Followers

759 Following

Robin Jia(@robinomial) 's Twitter Profile Photo

While studying memorization in LLMs, we struggled with a basic question: can knowledge really be localized to specific parameters? So, we designed two independent benchmarks to answer this. Amazingly, they agree: pruning-based localization is best, works well, but isn’t perfect.

account_circle
Ting-Yun Chang(@CharlotteTYC) 's Twitter Profile Photo

Localization in LLMs is often mentioned. But do localization methods actually localize correctly? In our paper, we (w/ Jesse Thomason, Robin Jia) develop two benchmarking ways to directly evaluate how well 5 existing methods can localize memorized data in LLMs.

Localization in LLMs is often mentioned. But do localization methods actually localize correctly? In our #NAACL2024 paper, we (w/ @_jessethomason_, @robinomial) develop two benchmarking ways to directly evaluate how well 5 existing methods can localize memorized data in LLMs.
account_circle
Deqing Fu(@DeqingFu) 's Twitter Profile Photo

Do multimodal foundation models treat every modality equally?

Hint: Humans have picture superiority. How about machines?

Introducing IsoBench, a benchmark for multimodal models with isomorphic inputs.

🔗 IsoBench.github.io

Do multimodal foundation models treat every modality equally? Hint: Humans have picture superiority. How about machines? Introducing IsoBench, a benchmark for multimodal models with isomorphic inputs. 🔗 IsoBench.github.io
account_circle
Robin Jia(@robinomial) 's Twitter Profile Photo

Personally my favorite experiment from this data watermark work. BLOOM-176B memorizes SHA hashes from its training data; we can use similar random character sequences to watermark any document collection!

account_circle
Robin Jia(@robinomial) 's Twitter Profile Photo

Determining whether an LLM has trained on your data isn’t a classification problem, it’s a statistical testing problem. Really proud of this work by Johnny Tian-Zheng Wei and Ryan Yixiang Wang on using random watermarks to rigorously detect data usage!

account_circle
Jieyu Zhao@ICLR2024(@jieyuzhao11) 's Twitter Profile Photo

🔊 Please consider submitting to our Secure and Trustworthy Large Language Model Workshop at ! Submission ddl is Feb 12.

account_circle
ACL 2024(@aclmeeting) 's Twitter Profile Photo

ACL announcement:
'The ACL Executive Committee has voted to significantly change ACL's approach to protecting anonymous peer review. The change is effective immediately. (1/4)

account_circle
Pei Zhou(@peizNLP) 's Twitter Profile Photo

I’m at and on the job market🎷🧳!! Come and talk about anything LLM reasoning, evaluating communicating agents, human-AI collaboration for new discoveries, coffee and jazz in NOLA☕️

I’m at #NeurIPS23 and on the job market🎷🧳!! Come and talk about anything LLM reasoning, evaluating communicating agents, human-AI collaboration for new discoveries, coffee and jazz in NOLA☕️
account_circle
Robin Jia(@robinomial) 's Twitter Profile Photo

I’ll be presenting SCENE at the poster session this *Saturday* at 11am! Come chat about unanswerable questions, minimal pairs, extrapolation, and the value of synthetic data :)

account_circle
USC Thomas Lord Department of Computer Science(@CSatUSC) 's Twitter Profile Photo

We're hiring! USC Thomas Lord Department of Computer Science is growing fast, with multiple openings for tenure-track and tenured positions.

🔍Security/privacy, AI, machine learning, data science, HCI, but exceptional candidates in all areas considered.

📅Deadline: Jan 5
🔗Details: tinyurl.com/5e99bmb8

We're hiring! @CSatUSC is growing fast, with multiple openings for tenure-track and tenured positions. 🔍Security/privacy, AI, machine learning, data science, HCI, but exceptional candidates in all areas considered. 📅Deadline: Jan 5 🔗Details: tinyurl.com/5e99bmb8
account_circle
USC Thomas Lord Department of Computer Science(@CSatUSC) 's Twitter Profile Photo

🚨Now hiring full-time teaching faculty!🚨

Apply now to join our fast-growing, collaborative department, with a brand new building opening this spring.

Find out more about our open faculty positions
(teaching, tenured, tenure track): cs.usc.edu/about/open-fac…

Please share!

🚨Now hiring full-time teaching faculty!🚨 Apply now to join our fast-growing, collaborative department, with a brand new building opening this spring. Find out more about our open faculty positions (teaching, tenured, tenure track): cs.usc.edu/about/open-fac… Please share!
account_circle
USC Thomas Lord Department of Computer Science(@CSatUSC) 's Twitter Profile Photo

We are at this week! 👋

Explore the latest USC Thomas Lord Department of Computer Science USC ISI research spanning social bias in name translation, AI tools for journalists, ambiguity in LMs, and much more ⬇️

viterbischool.usc.edu/news/2023/12/u…

USC Viterbi School EMNLP 2024 USC Research

account_circle
Qinyuan Ye(@qinyuan_ye) 's Twitter Profile Photo

I just arrived in Singapore for ! 😜

I'll be presenting our work on modeling and predicting LLM capabilities at GenBench workshop poster session (10am tomorrow). Also presenting it virtually in gather.town at 4pm on Dec 8. Check out our thread below 👇

account_circle
Qinyuan Ye(@qinyuan_ye) 's Twitter Profile Photo

What can we learn from thousands of LLM experiment records? Can we predict emergent abilities? Does this give any clue on making LLM benchmarking more efficient and principled? Check out our paper on analyzing BIG-bench and designing “small-bench.”

arxiv.org/abs/2305.14947 🧵⬇️

What can we learn from thousands of LLM experiment records? Can we predict emergent abilities? Does this give any clue on making LLM benchmarking more efficient and principled? Check out our paper on analyzing BIG-bench and designing “small-bench.” arxiv.org/abs/2305.14947 🧵⬇️
account_circle
Robin Jia(@robinomial) 's Twitter Profile Photo

ICL reduces the need for labeled *training* data, but what about test data? Fantastic work led by USC undergrad Harvey Yiyun Fu (applying to PhD programs this year!) shows we can estimate ICL accuracy well (on par with 40 labeled examples) using *unlabeled* data + model uncertainties

account_circle
Robin Jia(@robinomial) 's Twitter Profile Photo

How do Transformers really do in-context linear regression? 1 TF layer = 3 steps of a second-order method! Can’t be GD, which converges exponentially more slowly. Meanwhile, LSTMs are more like online GD; they don’t learn second-order optimization (likely due to limited memory).

account_circle