Secure Learning Lab (SLL)(@uiuc_aisecure) 's Twitter Profileg
Secure Learning Lab (SLL)

@uiuc_aisecure

We are a computer science research group led by Bo Li at UIUC, focusing on responsible and trustworthy machine learning.

ID:1260314954051313665

linkhttps://aisecure.github.io/ calendar_today12-05-2020 21:04:52

148 Tweets

939 Followers

289 Following

Junyuan Hong(@hjy836) 's Twitter Profile Photo

Thank you AK for sharing our latest paper on evaluating the trustworthiness of compressed LLMs! ✨

We have curated a set of insights and practical guidelines from a comprehensive benchmark with 5 compression methods and 3 popular LLMs across 8 trust dimensions + 1…

account_circle
Together AI(@togethercompute) 's Twitter Profile Photo

We are thrilled to be a launch partner for Meta Llama 3.

Experience Llama 3 now at up to 350 tokens per second for Llama 3 8B and up to 150 tokens per second for Llama 3 70B, running in full FP16 precision on the Together API! 🤯

together.ai/blog/together-…

account_circle
Secure Learning Lab (SLL)(@uiuc_aisecure) 's Twitter Profile Photo

Generating synthetic data with DP guarantees given only API assess to models is feasible!! The project led by Chulin Xie collaborated with Microsoft Research provides a promising way to protect data privacy in the LLM world!

account_circle
Ethan Mollick(@emollick) 's Twitter Profile Photo

| ̄ ̄ ̄ ̄ ̄  ̄|
| This can |
| hack AI |
| now. |
| ______ |
(\__/) ||
(•ㅅ•) ||
/   づ

Paper showing that ASCII art can get around AI guardrails. Its the return of 1980s hackers. arxiv.org/pdf/2402.11753…

| ̄ ̄ ̄ ̄ ̄  ̄| | This can | | hack AI | | now. | | ______ | (\__/) || (•ㅅ•) || /   づ Paper showing that ASCII art can get around AI guardrails. Its the return of 1980s hackers. arxiv.org/pdf/2402.11753…
account_circle
Secure Learning Lab (SLL)(@uiuc_aisecure) 's Twitter Profile Photo

Great work from Weixin Chen to leverage OOD queries to perform self-truthfying without requiring human anotations, which achieves SOTA results on truthfulQA AK

account_circle
Junyuan Hong(@hjy836) 's Twitter Profile Photo

Why private prompts? Your prompts on are NOT SAFE! Enormous examples have demonstrated that does not keep secrets.
Never put sensitive examples in your prompt! Instead, try our new tool to customize a privacy-preserving prompt tuning on your local data and…

Why private prompts? Your prompts on #GPTStore are NOT SAFE! Enormous examples have demonstrated that #ChatGPT does not keep secrets. Never put sensitive examples in your prompt! Instead, try our new tool to customize a privacy-preserving prompt tuning on your local data and…
account_circle
Zhen Xiang(@ZhenXia98294421) 's Twitter Profile Photo

Cutting-edge LLMs can be misused by malicious vendors! We propose BadChain – the first backdoor attack against LLMs ( , Llama2, and PaLM2) with chain-of-thought prompting. Please come to our oral presentation at (Room 203-205) on 12/15 at 4 pm @ BUGS Workshop.

Cutting-edge LLMs can be misused by malicious vendors! We propose BadChain – the first backdoor attack against LLMs (#GPT4, Llama2, and PaLM2) with chain-of-thought prompting. Please come to our oral presentation at (Room 203-205) on 12/15 at 4 pm @ #NeurIPS2023 BUGS Workshop.
account_circle
Zhen Xiang(@ZhenXia98294421) 's Twitter Profile Photo

Backdoor detection can also be CERTIFIED! Excited to share our work on the first certified backdoor detector with detection guarantees @ . Please visit our poster #1619 on 12/13 (Wed) from 10:45 am to 12:45 pm in Great Hall & Hall B1+B2 (level 1).

Backdoor detection can also be CERTIFIED! Excited to share our work on the first certified backdoor detector with detection guarantees @ #NeurIPS2023 . Please visit our poster #1619 on 12/13 (Wed) from 10:45 am to 12:45 pm in Great Hall & Hall B1+B2 (level 1).
account_circle
Boxin Wang(@wbx_life) 's Twitter Profile Photo

We are so honored to receive the **Outstanding Paper Award at NeurIPS**! Huge thanks to our collaborators from Illinois Computer Science Secure Learning Lab (SLL) Stanford University UC Berkeley Microsoft Research Center for AI Safety 🎉 Please come to our oral presentation on Tuesday from 10:30-10:45am CT for more details!

account_circle