this post was submitted on 27 Dec 2023
39 points (91.5% liked)

Technology

1255 readers
486 users here now

Which posts fit here?

Anything that is at least tangentially connected to the technology, social media platforms, informational technologies and tech policy.


Rules

1. English onlyTitle and associated content has to be in English.
2. Use original linkPost URL should be the original link to the article (even if paywalled) and archived copies left in the body. It allows avoiding duplicate posts when cross-posting.
3. Respectful communicationAll communication has to be respectful of differing opinions, viewpoints, and experiences.
4. InclusivityEveryone is welcome here regardless of age, body size, visible or invisible disability, ethnicity, sex characteristics, gender identity and expression, education, socio-economic status, nationality, personal appearance, race, caste, color, religion, or sexual identity and orientation.
5. Ad hominem attacksAny kind of personal attacks are expressly forbidden. If you can't argue your position without attacking a person's character, you already lost the argument.
6. Off-topic tangentsStay on topic. Keep it relevant.
7. Instance rules may applyIf something is not covered by community rules, but are against lemmy.zip instance rules, they will be enforced.


Companion communities

[email protected]
[email protected]


Icon attribution | Banner attribution

founded 11 months ago
MODERATORS
all 8 comments
sorted by: hot top controversial new old
[–] [email protected] 9 points 9 months ago

I'm sympathetic to the NYT, even if it's not reproducing their IP verbatim.

AI companies need to acknowledge that their LLMs would be worthless without training data and compensate/credit the sources appropriately.

[–] givesomefucks 6 points 9 months ago (1 children)

It's not just that it circumvents the paywall, it makes up random nonsense and then claim the NYT said it.

I've never got why people don't see this about AI. When it "works" it's just spitting out what a human was paid (Avery low wage) to write, when it has to come up with something that hasn't been written, it just slaps nonsense together.

It's not real AI, it's just next generation search engines that gives unreliable results.

You just don't notice if you don't already know what you're asking.

[–] [email protected] 3 points 9 months ago

Even tho these LLM work by just figuring out next word (token) that makes sense, it is still able to generate things that no human has ever written before. It isn't just copypasting stuff together.

I use GPT4 daily basis on coding and the way it spills out complex code templates/snippets, which are unique to the problem, is not just not possible without model having some level of intelligence. Of course it hallucinates now and then, but so does most of the coders now and then