Worst hypothesis they just need to mess around a bit. For example I don’t think that queerasfu.ck
would be registered.
This account is being kept for the posterity, but it won’t see further activity past February.
If you want to contact me, I’m at /u/lvxferre@mander.xyz
Worst hypothesis they just need to mess around a bit. For example I don’t think that queerasfu.ck
would be registered.
They could get a .ck domain instead and move to queer.as.fu.ck, no?
That’s surprisingly accurate, as people here are highlighting (it makes geometrical sense when dealing with complex numbers).
My nephew once asked me this question. The way that I explained it was like this:
It’s a different analogy but it makes intuitive sense, even for kids. And it works nice as mnemonic too.
Another important detail is that Digg v4 pissed off most of the userbase, so the impact was pretty much immediate. Reddit APIcalypse pissed off only power users instead; the impact will only come off later (sadly likely past IPO).
Lunix sucks so much that it got stuck into the version 2 for years.
Persuasion itself goes from neutral to negative, depending on your moral standards. (They’re partially individual, partially cultural.) Because at the end of the day it boils down to “I want you to believe in this, because I benefit from your belief.”
And you definitively see some backslash against this aspect of advertisement; same deal with personal communication, a person being excessively rhetoric for their own benefit is immediately labelled distrustful.
Then over that propaganda adds further layers of nastiness, like:
Let’s go simpler: what if your instance was allowed to copy the fed/defed lists from other instances, and use them (alongside simple Boolean logic plus if/then statements) to automatically decide who you’re going to federate/defederate with? That would enable caracoles and fedifams for admins who so desire, but also enable other organically grown relations.
For example. Let’s say that you just joined the federation. And there are three instances that you somewhat trust:
Then you could set up your defederation rules like this:
Of course, that would require distinguishing between manual and automatic fed/defed. You’d be able to use the manual fed/defed from other instances to create your automatic rules, to avoid deadlocks like “Alice is blocking it because Bob is blocking it, and Bob is blocking it because Alice is doing it”.
The source that I’ve linked mentions semantic embedding; so does further literature on the internet. However, the operations are still being performed with the vectors resulting from the tokens themselves, with said embedding playing a secondary role.
This is evident for example through excerpts like
The token embeddings map a token ID to a fixed-size vector with some semantic meaning of the tokens. These brings some interesting properties: similar tokens will have a similar embedding (in other words, calculating the cosine similarity between two embeddings will give us a good idea of how similar the tokens are).
Emphasis mine. A similar conclusion (that the LLM is still handling the tokens, not their meaning) can be reached by analysing the hallucinations that your typical LLM bot outputs, and asking why that hallu is there.
What I’m proposing is deeper than that. It’s to use the input tokens (i.e. morphemes) only to retrieve the sememes (units of meaning; further info here) that they’re conveying, then discard the tokens themselves, and perform the operations solely on the sememes. Then for the output you translate the sememes obtained by the transformer into morphemes=tokens again.
I believe that this would have two big benefits:
And it might be an additional layer, but the whole approach is considerably simpler than what’s being done currently - pretending that the tokens themselves have some intrinsic value, then playing whack-a-mole with situations where the token and the contextually assigned value (by the human using the LLM) differ.
[This could even go deeper, handling a pragmatic layer beyond the tokens/morphemes and the units of meaning/sememes. It would be closer to what @njordomir@lemmy.world understood from my other comment, as it would then deal with the intent of the utterance.]
Soap and water do wonders for 90% of the restroom cleaning.
The problem is that the other 10% are important too.
Not quite. I’m focusing on chatbots like Bard, ChatGPT and the likes, and their technology (LLM, or large language model).
At the core those LLMs work like this: they pick words, split them into “tokens”, and then perform a few operations on those tokens, across multiple layers. But at the end of the day they still work with the words themselves, not with the meaning being encoded by those words.
What I want is an LLM that assigns multiple meanings for those words, and performs the operations above on the meaning itself. In other words the LLM would actually understand you, not just chain words.
Yup, that’s the stuff. It’s mostly a finishing touch, to get rid of bacteria.
At the very least, I’d recommend you:
Everyone has the cleaning agents that they swear upon, so look for something that works for you. For me it’s
Important detail: do not mix any two of the cleaning agents that I’ve mentioned. Specially not ammonium and bleach.
For reference, the disinfecting agent that I use is called “pinho sol”, but I have no idea if it’s sold outside Brazil. You probably have some similar product wherever you live.
Complexity does not mean sophistication when it comes to AI and never has and to treat it as such is just a forceful way to make your ideas come true without putting in the real effort.
It’s a bit off-topic, but what I really want is a language model that assigns semantic values to the tokens, and handles those values instead of directly working with the tokens themselves. That would be probably far less complex than current state-of-art LLMs, but way more sophisticated, and require far less data for “training”.
Oh “great”, more crap between Ctrl and Alt.
[Grumpy grandpa] In my times, the space row only had five keys! And we did more than those youngsters do with eight, now nine keys!
Thank you! It’s working now.
It’s giving me an error, “Error Finding Entity // Make sure you spelled the entity correctly and that it exists!”, when I use my username for lemmy.ml; curiously it works well when I do it for my beehaw.org account.
Chomsky’s concept of UG (universal grammar) is able to handle this. Since there would be a chunk of language that is innate (universal), that feral child would share it. So, as a conclusion from that, even if the feral child isn’t expressing it through vocalisation, since they lack an “application” of the UG (like Nahuatl, Mandarin, Quechua, English, Kikongo etc.), they’d still have some rather simple internal monologue.
…that said I think that Chomsky’s UG is full of shit. I do agree with him that the faculty of language might have developed first to structure thought; but my reasoning resembles a bit more yours, the role of language would be to formalise thought. Thinking without language is possible in the same way as moving across a village without roads - it’s doable but clunky, and you’ll likely take far more effort than with proper roads/ a language.
Not to challenge Chomsky on his own turf
Don’t worry. Everyone and their dog challenges him. Including himself, he’s often contradicting his own earlier statements.
It depends a lot on what you consider “toxic”.
If it’s just about intrusive off-topic political discussion, then I fully agree with you: it’s far more common in Lemmy than in Reddit, and sometimes it reaches a point that even people who’d otherwise enjoy discussing politics roll their eyes and say “not this shit again”.
However, if “toxic” includes other forms of undesirable behaviour, then Lemmy is probably less toxic than Reddit. For example: while sometimes you do see here disingenuous and deliberate stupidity, “waah TL;DR!!”, the “I don’t understand” conveying disagreement, or passive aggressiveness, in Reddit they pop up all the time.
So, what do you consider toxic? Depending on that, the other users’ experiences might be really similar or really different from yours.
Chomsky would say that the original purpose of language is to structure thought, with communication being solely secondary. (Or something like this, I don’t recall it word-by-word.)
If that’s correct, then internal monologues are simply a result of your brain processing your thoughts.
Damn, that’s sad. Thank you for the info.