FaceDeer

joined 7 months ago
[–] FaceDeer@fedia.io 1 points 7 hours ago

If it's not communicating anything, what's the point?

[–] FaceDeer@fedia.io 1 points 18 hours ago (2 children)

My point is that if we turn up our gibberish dial now then at least our llms will be learning the wrong thing & we have some control.

We'd be covering ourselves in poop to prevent people from sitting next to us on the train. Sure, people will avoid sitting next to us, but in the meantime we'll be covered in poop.

And then other people will learn the trick, cover themselves in poop too, and now everyone's poopy and the trick stops working.

There is still a lot of understanding that we do automatically that an llm will never do.

Are you willing to bet the convenience of comprehensible online discourse on that? "Automatically understanding stuff" is basically the one job of LLMs.

LLMs model language, and coming up with some kind of "gibberish" filter is simply inventing a new language. If there's semantic meaning in it the LLMs will figure it out just like any other language, and if there isn't semantic meaning then we've lost the ability to communicate entirely. I see no upside.

[–] FaceDeer@fedia.io 0 points 1 day ago

In my experience the vast majority of posts about Elon Musk are from people who hate him and are tired of hearing about him.

[–] FaceDeer@fedia.io 5 points 1 day ago (1 children)

I'm not talking about a summarizer, I'm talking about a classifier. It just needs to identify which parts of the page are advertising and which are not.

The point of such a tool is that it would read the web page in exactly the same way that a human would, so using trickery like pre-rendered images of text or funky unicode wouldn't really change anything. If a human can read it then so can the AI.

[–] FaceDeer@fedia.io 1 points 1 day ago (4 children)

Well, the "at least for now" part is my point - if people start using "gibberish" to communicate or to hide their communication, that provides training material for LLMs to let them figure out how to use it too.

LLMs learn how to communicate based on existing examples of communication. As long as humans are communicating with each other somehow then LLMs will be able to train how to do that too. They have the same communication capabilities that we do at this point, so there's not really any way we can make a secret clubhouse that they can't figure out how to infiltrate.

Personally, I think there's two main routes we can go to deal with this. Either we can simply accept that there's no way to be 100% sure we're talking to a human any more and evaluate the value of our conversation based on the content of the words spoken rather than the composition of the entity generating them, or we could come up with some kind of "proof of personhood" system to allow people to label the text the write as coming from them.

The latter is extremely hard to do, of course, both from a technical and cultural perspective. And such a system would likely still allow someone's "person token" to be sneakily used by AI, either by voluntarily delegating it (I could very well be retyping all of this out of a ChatGPT window) or through hackery.

So I'm inclined toward the former. If I'm chatting with someone and I'm having a good time doing it, and then later I find out it was a bot, why should that change how much fun I had?

[–] FaceDeer@fedia.io 3 points 2 days ago (6 children)

I don't see how that would be practical. People who aren't "in on the joke", as it were, will call out the gibberish and downvote it. If enough people are "in on the joke" then the whole forum becomes useless and some other forum will be created to fill the role of the original. The AI will train off of that one.

Basically, if you don't want an AI training on your content, then don't post your content in public where an AI will see it. The Fediverse is the last place you should be posting since its very nature is about openly broadcasting your content to whoever wants to see it.

[–] FaceDeer@fedia.io 11 points 2 days ago

Adblockers aren't made by "corporate overlords." This wouldn't be either.

[–] FaceDeer@fedia.io 48 points 2 days ago (8 children)

Someday soon my "adblocker" might be a personal AI that reads the spam-ridden website on a virtual display in memory, identifies the actual content while pretending to look at whatever ads the site demands, and then passes the information I'm actually looking for along to me. Good luck captchaing that.

[–] FaceDeer@fedia.io 8 points 2 days ago (9 children)

You realize that this is only going to train LLMs how to recognize "gibberish?"

[–] FaceDeer@fedia.io 10 points 3 days ago (1 children)

"Good intentions", I presume.

My position has always been "if there are people who are disadvantaged then pass laws to help disadvantaged people rather than making the assumption that everyone with a particular set of genetics need help." I guess it's just easier to take that shortcut though.

[–] FaceDeer@fedia.io 12 points 3 days ago (4 children)

Knowing whether I have First Nations blood on my mother's side would have real legal benefits for me (my mom is estranged from her family and so has never told me much about them, but there's some possibility there given their historical context). I know a friend who had to prove he was 1/8 Metis in order to get a job as a web designer with a particular company.

I think it's ridiculous and flat out racist, frankly, but there are indeed benefits in this day and age from having particular ancestry.

[–] FaceDeer@fedia.io 3 points 3 days ago

Yeah, when I saw this headline my first thought is "aww, my dog loved it when I got an Amazon package because she got to destroy all those balloon things for me."

view more: next ›