![](https://mander.xyz/pictrs/image/2d70a70f-ed05-49fc-8549-b0cbfadcfade.jpeg)
![](https://lemmy.world/pictrs/image/0943eca5-c4c2-4d65-acc2-7e220598f99e.png)
Predictive text also can vectorize words, but the number of vectors per word are much, much simpler.
Predictive text also can vectorize words, but the number of vectors per word are much, much simpler.
If you want to appear more mature, start capitalizing your sentences and using less exclamation points. Some are fine, but never more than one to a sentence.
The middle man is great if you want to ensure no action is ever taken
It’s subtle, but yes it tastes “almost” sweet despite there being no sugar.
Do you drink sugary drinks or eat sugary foods on a regular basis? There is some research showing that excess sugar intake can drastically reduce your body’s ability to sense sweetness.
Not sure if this is a joke, but yeah. For me, it’s got a bit of acid (from the carbonic acid byproduct of carbonation) and a subtle sweetness. Plus whatever flavoring they add
This is normal. It’s also why, if you drink lots of sugary drinks, things like unsweetened tea or sparkling water (without sugar) become flavorless until you cut out the sugary drinks for a few weeks. Only after that will you begin to taste the natural flavors.
Definitely better than this outdated version. Nobody uses Python2 unless they want to at this point
“What happened?”
“I don’t talk about it… anymore…”
Wow a large circular stuffed animal
I really don’t understand how they can so rabidly defend atrocious behavior by dictators and strongmen and then claim to be left-wing. Do they not realize these fascists they are defending are right-wing?
It literally does not act like a sponge. That’s the whole point of skin, is to not absorb shit.
Where do you live? Because I doubt you are the original settler of your land.
If you’re an American (North or South), you are not responsible for the reprehensible actions of your ancestors towards the Native populations just because you now live there. Similarly with Russia, China, most of Africa, Australia, Europe… basically the entirety of the world has a history of war and settling on conquered land. And you don’t deserve death for the actions of your ancestors.
I assume EU-based ISPs will be forced to ban access to the website for noncompliance, otherwise it would have literally no teeth whatsoever
Removed by mod
The work is not reproduced in its entirety. Simply using the work in its entirety is not a violation of copyright law, just as reading a book or watching a movie (even if pirated) is not a violation. The reproduction of that work is the violation, and LLMs simply do not store the works in their entirety nor are they capable of reproducing them.
The argument is less that an LLM is a human and more that it is not a copyright violation to use a material to train the LLM. By current legal definitions, it is fair use unless the material is able to be reproduced in its entirety (or at least, in some meaningful way).
It’s only black box because nobody has the time (likely years to decades) to wade through the layers of a finished model to check every node and weight.
This is exactly correct, except you’re also not accounting for the insane amount of computational power that would be necessary to backtrack a single output of a single model. This is why it is a black box. It simply is not possible on a meaningful level.
So if math and computer science isn’t an exact science, what is?
Things that are reproducible with known inputs and outputs, allowing for all components to be studied and explained. As an example from my field: if you damage the dorsolateral prefrontal cortex in a fully grown adult, they will have the impulse control of a three-year old. We know this because we have observed damage to this area in multiple individuals, and can measure the effects based on the severity of that damage.
In contrast, if you provide the same billion-parameter neural network identical inputs, you will not receive identical outputs.
Look, I understand why you think this. I thought this too when I was first beginning to learn machine learning and data science. But I’ve now been working with machine learning models including neural networks for nearly a decade, and the truth is that is nearly impossible to track the path of an input to a given output in machine learning models other than regression-based models and decision tree-based models.
There is an entire field of data science devoted to explaining how these models arrive at their conclusions. It’s called “explainable AI” or “xAI”, and I have a few papers that I’ve published in exploring the utility of them. The basic explanation for how they work is that we run hundreds of thousands of different models and then do statistical analysis to estimate why the models arrived at their conclusion. It isn’t an exact science, however.
I went looking for old forums I used to frequent, and while the site itself is on the internet archive most of the conversations are lost. Only one or two snapshots a year