Discussion
Loading...

Discussion

  • About
  • Code of conduct
  • Privacy
  • Users
  • Instances
  • About Bonfire
AI6YR Ben
@ai6yr@m.ai6yr.org  ·  activity timestamp 23 hours ago

Anthropic: "In a joint study with the UK AI Security Institute and the Alan Turing Institute, we found that as few as 250 malicious documents can produce a "backdoor" vulnerability in a large language model—regardless of model size or training data volume. "

https://www.anthropic.com/research/small-samples-poison

#cybersecurity #backdoor #datapoisoning #llm #ai

A small number of samples can poison LLMs of any size
A small number of samples can poison LLMs of any size
A small number of samples can poison LLMs of any size
  • Copy link
  • Flag this post
  • Block
AI6YR Ben
@ai6yr@m.ai6yr.org replied  ·  activity timestamp 23 hours ago

Stop uploading gibberish to the cloud
You're making the LLMs die for crying out loud
Putting gibberish into your docs is driving them mad
And then your search results get very very bad

They'll think that gravy is just fine
To pour and mix in with your wine
And they'll tell you to drink castor oil
And wrap your head up with aluminum foil

They'll think that birds can emit radio noise
And that live sharks make very good toys
And they might even think sea horse emojis are real
And that the best kind of food to eat is arsenic meal

#poem (human)

  • Copy link
  • Flag this comment
  • Block
Log in

bonfire.cafe

A space for Bonfire maintainers and contributors to communicate

bonfire.cafe: About · Code of conduct · Privacy · Users · Instances
Bonfire social · 1.0.0-rc.3.13 no JS en
Automatic federation enabled
  • Explore
  • About
  • Members
  • Code of Conduct
Home
Login