-
25 votes
-
Introducing Claude Opus and Sonnet 4
18 votes -
How Big Tech hides its outsourced African workforce
16 votes -
Chicago Sun-Times prints summer reading list full of fake books
42 votes -
Sam and Jony introduce io | OpenAI
15 votes -
We did the math on AI’s energy footprint. Here’s the story you haven’t heard.
23 votes -
YouTube’s new ads will ruin the best part of a video on purpose
60 votes -
I don’t care whether you use ChatGPT to write
25 votes -
Introducing Codex [OpenAI]
23 votes -
Grok’s white genocide fixation caused by ‘unauthorized modification’
51 votes -
Will the humanities survive artificial intelligence?
28 votes -
Software engineer lost his $150K-a-year job to AI—he’s been rejected from 800 jobs and forced to DoorDash and live in a trailer to make ends meet
34 votes -
Two unrelated stories that make me even more cynical about AI
I saw both of these stories on Lemmy today. They show two different facets to the topic of AI. This first story is from the perspective of cynicism about AI and how it has been overhyped. If AI is...
I saw both of these stories on Lemmy today. They show two different facets to the topic of AI.
This first story is from the perspective of cynicism about AI and how it has been overhyped.
If AI is so good, where are the open source contributionsBut if AI is so obviously superior … show us the code. Where’s the receipts? Let’s say, where’s the open source code contributions using AI?
The second story is about crony capitalism, deregulation, and politics around AI:
GOP sneaks decades long AI regulation ban into spending bill
On Sunday night, House Republicans added language to the Budget Reconciliation bill that would block all state and local governments from regulating AI for 10 years, 404 Media reports. The provision, introduced by Representative Brett Guthrie of Kentucky, states that "no State or political subdivision thereof may enforce any law or regulation regulating artificial intelligence models, artificial intelligence systems, or automated decision systems during the 10 year period beginning on the date of the enactment of this Act
I saw these stories minutes apart, and they really make me feel even more cynical and annoyed by AI than I was yesterday. Because:
- In the short term AI is largely a boondoggle, which won’t work as advertised but still humans will be replaced by it because the people who hire don’t understand it’s limitations but they fear missing out on a gold rush.
- The same shady people at the AI companies who are stealing your art and content, in order to sell a product that will replace you, are writing legislation to protect themselves from being held accountable
- They also are going to be protected from any skynet-style disasters caused by their recklessness
28 votes -
The (not so) futuristic technology of “Lazarus”
12 votes -
Some ChatGPT users are developing delusional beliefs that are reinforced by the large language model
53 votes -
Everyone is cheating their way through college
49 votes -
In 2025, venture capital can’t pretend everything is fine any more
53 votes -
How AGI made the future unthinkable
19 votes -
Amazon makes ‘fundamental leap forward in robotics’ with device having sense of touch
10 votes -
Tech companies apparently do not understand why we dislike AI
49 votes -
Researchers secretly ran a massive, unauthorized AI persuasion experiment on Reddit users
64 votes -
Dark Visitors got a new free plan
6 votes -
Chinese factories are more automated
13 votes -
Time saved by AI offset by new work created, study suggests
23 votes -
When ChatGPT broke an entire field: An oral history
14 votes -
A nonsense phrase has been occurring in scientific papers, suggesting artificial intelligence data contamination
53 votes -
Don’t buy stuff from old AI people
20 votes -
State Bar of California admits it used AI to develop exam questions, triggering new furor
25 votes -
OpenAI is a systemic risk to the tech industry
35 votes -
Kagi Assistant is now available to all users
44 votes -
Russia seeds chatbots with lies. Any bad actor could game AI the same way.
33 votes -
The dangers of vibe coding
26 votes -
Anubis works
35 votes -
AI 2027
29 votes -
The art of poison-pilling music files
15 votes -
I'm tired of dismissive anti-AI bias
60 votes -
Fintech founder charged with fraud after ‘AI’ shopping app found to be powered by humans in the Philippines
39 votes -
An image of an archeologist adventurer who wears a hat and uses a bullwhip
43 votes -
Google AI search shift leaves website makers feeling “betrayed”
36 votes -
The ARC-AGI-2 benchmark could help reframe the conversation about AI performance in a more constructive way
The popular online discourse on Large Language Models’ (LLMs’) capabilities is often polarized in a way I find annoying and tiresome. On one end of the spectrum, there is nearly complete dismissal...
The popular online discourse on Large Language Models’ (LLMs’) capabilities is often polarized in a way I find annoying and tiresome.
On one end of the spectrum, there is nearly complete dismissal of LLMs: an LLM is just a slightly fancier version of the autocomplete on your phone’s keyboard, there’s nothing to see here, move on (dot org).
This dismissive perspective overlooks some genuinely interesting novel capabilities of LLMs. For example, I can come up with a new joke and ask ChatGPT to explain why it’s funny or come up with a new reasoning problem and ask ChatGPT to solve it. My phone’s keyboard can’t do that.
On the other end of the spectrum, there are eschatological predictions: human-level or superhuman artificial general intelligence (AGI) will likely be developed within 10 years or even within 5 years, and skepticism toward such predictions is “AI denialism”, analogous to climate change denial. Just listen to the experts!
There are inconvenient facts for this narrative, such as that the majority of AI experts give much more conservative timelines for AGI when asked in surveys and disagree with the idea that scaling up LLMs could lead to AGI.
The ARC Prize is an attempt by prominent AI researcher François Chollet (with help from Mike Knoop, who apparently does AI stuff at Zapier) to introduce some scientific rigour into the conversation. There is a monetary prize for open source AI systems that can perform well on a benchmark called ARC-AGI-2, which recently superseded the ARC-AGI benchmark. (“ARC” stands for “Abstract and Reasoning Corpus”.)
ARC-AGI-2 is not a test of whether an AI is an AGI or not. It’s intended to test whether AI systems are making incremental progress toward AGI. The tasks the AI is asked to complete are colour-coded visual puzzles like you might find in a tricky puzzle game. (Example.) The intention is to design tasks that are easy for humans to solve and hard for AI to solve.
The current frontier AI models score less than 5% on ARC-AGI-2. Humans score 60% on average and 100% of tasks have been solved by at least two humans in two attempts or less.
For me, this helps the conversation about AI capabilities because it gives a rigorous test and quantitative measure to my casual, subjective observations that LLMs routinely fail at tasks that are easy for humans.
François Chollet was impressed when OpenAI’s o3 model scored 75.7% on ARC-AGI (the older version of the benchmark). He emphasizes the concept of “fluid intelligence”, which he seems to define as the ability to adapt to new situations and solve novel problems. Chollet thinks that o3 is the first AI system to demonstrate fluid intelligence, although it’s still a low level of fluid intelligence. (o3 also required thousands of dollars’ worth of computation to achieve this result.)
This is the sort of distinction that can’t be teased out by the polarized popular discourse. It’s the sort of nuanced analysis I’ve been seeking out, but which has been drowned out by extreme positions on LLMs that ignore inconvenient facts.
I would like to see more benchmarks that try to do what AGI-AGI-2 does: find problems that humans can easily solve and frontier AI models can’t solve. These sort of benchmarks can help us measure AGI progress much more usefully than the typical benchmarks, which play to LLMs’ strengths (e.g. massive-scale memorization) and don’t challenge them on their weaknesses (e.g. reasoning).
I long to see AGI within my lifetime. But the super short timeframes given by some people in the AI industry feel to me like they border on mania or psychosis. The discussion is unrigorous, with people pulling numbers out of thin air based on gut feeling.
It’s clear that there are many things humans are good at doing that AI can’t do at all (where the humans vs. AI success rate is ~100% vs. ~0%). It serves no constructive purpose to ignore this truth and it may serve AI research to develop rigorous benchmarks around it.
Such benchmarks will at least improve the quality of discussion around AI capabilities, insofar as people pay attention to them.
Update (2024-04-11 at 19:16 UTC): François Chollet has a new 20-minute talk on YouTube that I recommend. I've watched a few videos of Chollet talking about ARC-AGI or ARC-AGI-2, and this one is beautifully succinct: https://www.youtube.com/watch?v=TWHezX43I-4
10 votes -
Using Claude and undocumented Google Calendar features to automate event creation
4 votes -
Swedish fashion retailer H&M will use AI doppelgangers in some social media posts and marketing in the place of humans, if given permission by models
10 votes -
Vibe coding on Apple Shortcuts
5 votes -
A summary of my bot defence systems
11 votes -
Review: Cræft, by Alexander Langlands
4 votes -
Please stop externalizing your costs directly into my face
121 votes -
Enough with the bullshit (a letter to fellow bullshit sufferers)
56 votes -
Trapping misbehaving bots in an AI Labyrinth
40 votes -
eBay privacy policy update and AI opt-out
eBay is updating its privacy policy, effective next month (2025-04-27). The major change is a new section about AI processing, accompanied by a new user setting with an opt-out checkbox for having...
eBay is updating its privacy policy, effective next month (2025-04-27). The major change is a new section about AI processing, accompanied by a new user setting with an opt-out checkbox for having your personal data feed their models.
While that page specifically references European areas, the privacy selection appears to be active and remembered between visits for non-Europe customers. It may not do anything for us at all. On the other hand, it seems nearly impossible to find that page from within account settings, so I thought I'd post a direct link.
I'm well aware that I'm anomalous for having read this to begin with, much less diffed it against the previous version. But since I already know that I'm weird, and this wouldn't be much of a discussion post without questions:
- How do you stay up to date with contract changes that might affect you, outside of widespread Internet outrage (such as recent Firefox news)?
- What's your threshold -- if any -- for deciding whether to quit a company over contract changes? Alternatively, have you ever walked away from a purchase, service, or other acquisition over the terms of the contracts?
46 votes -
Norwegian man has filed a complaint with the Norwegian Data Protection Authority after ChatGPT falsely told him he had killed two of his sons and been jailed
22 votes