Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I am waiting for the first truly open model without any of the censorship built in.

I wonder how long it will take and how quickly it will try to get shut down.





Most open models have been converted to uncensored versions. Search for the model name with the suffix "abliterated".

That's not a realistic expectation.

Classic examples like:

  User: I'm feeling bad
  LLM: Have you considered k*****g yourself?
Are a good example of what an LLM "without censorship" looks like: Good at predicting the most common sequence of text (ex. the most common sarcastic reply from Reddit), but effectively useless.

In order to build a useful LLM (ie. one that actually follows instructions) you need to teach the LLM to prefer the most helpful answer, and that process by itself is already an implicit layer of "censorship" as it requires human supervision, and different humans have different perceptions on what the most helpful answer is, especially when their paycheck is conditioned to a list of "corporate values".

You can only pick between a parrot that repeats random text from the Internet, or a parrot lobotomized to follow the orders from their trainers (which occasionally repeats random text from the Internet, because the training isn't perfect).

Unsurprisingly, the lobotomized parrot is more useful to get actual work done, even if it won't tell you what the CIA[1] did to Mexican Students on October 2nd, 1968.

[1]: https://www.bbc.com/mundo/noticias-america-latina-45662739




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: