Replies: "This model supposedly is safe"
*Looks inside*: Huge pile of web scraped data without consent and ignoring licenses.
Seriously, this is a tragedy. LLMs are useful. You might not be able to train ChatGPT without stealing, but you can totally train a useful base model for task specific fine tuning without stealing. Why is almost nobody even trying????


