-
15 votes
-
I'm tired of dismissive anti-AI bias
60 votes -
Fintech founder charged with fraud after ‘AI’ shopping app found to be powered by humans in the Philippines
39 votes -
An image of an archeologist adventurer who wears a hat and uses a bullwhip
43 votes -
Google AI search shift leaves website makers feeling “betrayed”
36 votes -
The ARC-AGI-2 benchmark could help reframe the conversation about AI performance in a more constructive way
The popular online discourse on Large Language Models’ (LLMs’) capabilities is often polarized in a way I find annoying and tiresome. On one end of the spectrum, there is nearly complete dismissal...
The popular online discourse on Large Language Models’ (LLMs’) capabilities is often polarized in a way I find annoying and tiresome.
On one end of the spectrum, there is nearly complete dismissal of LLMs: an LLM is just a slightly fancier version of the autocomplete on your phone’s keyboard, there’s nothing to see here, move on (dot org).
This dismissive perspective overlooks some genuinely interesting novel capabilities of LLMs. For example, I can come up with a new joke and ask ChatGPT to explain why it’s funny or come up with a new reasoning problem and ask ChatGPT to solve it. My phone’s keyboard can’t do that.
On the other end of the spectrum, there are eschatological predictions: human-level or superhuman artificial general intelligence (AGI) will likely be developed within 10 years or even within 5 years, and skepticism toward such predictions is “AI denialism”, analogous to climate change denial. Just listen to the experts!
There are inconvenient facts for this narrative, such as that the majority of AI experts give much more conservative timelines for AGI when asked in surveys and disagree with the idea that scaling up LLMs could lead to AGI.
The ARC Prize is an attempt by prominent AI researcher François Chollet (with help from Mike Knoop, who apparently does AI stuff at Zapier) to introduce some scientific rigour into the conversation. There is a monetary prize for open source AI systems that can perform well on a benchmark called ARC-AGI-2, which recently superseded the ARC-AGI benchmark. (“ARC” stands for “Abstract and Reasoning Corpus”.)
ARC-AGI-2 is not a test of whether an AI is an AGI or not. It’s intended to test whether AI systems are making incremental progress toward AGI. The tasks the AI is asked to complete are colour-coded visual puzzles like you might find in a tricky puzzle game. (Example.) The intention is to design tasks that are easy for humans to solve and hard for AI to solve.
The current frontier AI models score less than 5% on ARC-AGI-2. Humans score 60% on average and 100% of tasks have been solved by at least two humans in two attempts or less.
For me, this helps the conversation about AI capabilities because it gives a rigorous test and quantitative measure to my casual, subjective observations that LLMs routinely fail at tasks that are easy for humans.
François Chollet was impressed when OpenAI’s o3 model scored 75.7% on ARC-AGI (the older version of the benchmark). He emphasizes the concept of “fluid intelligence”, which he seems to define as the ability to adapt to new situations and solve novel problems. Chollet thinks that o3 is the first AI system to demonstrate fluid intelligence, although it’s still a low level of fluid intelligence. (o3 also required thousands of dollars’ worth of computation to achieve this result.)
This is the sort of distinction that can’t be teased out by the polarized popular discourse. It’s the sort of nuanced analysis I’ve been seeking out, but which has been drowned out by extreme positions on LLMs that ignore inconvenient facts.
I would like to see more benchmarks that try to do what AGI-AGI-2 does: find problems that humans can easily solve and frontier AI models can’t solve. These sort of benchmarks can help us measure AGI progress much more usefully than the typical benchmarks, which play to LLMs’ strengths (e.g. massive-scale memorization) and don’t challenge them on their weaknesses (e.g. reasoning).
I long to see AGI within my lifetime. But the super short timeframes given by some people in the AI industry feel to me like they border on mania or psychosis. The discussion is unrigorous, with people pulling numbers out of thin air based on gut feeling.
It’s clear that there are many things humans are good at doing that AI can’t do at all (where the humans vs. AI success rate is ~100% vs. ~0%). It serves no constructive purpose to ignore this truth and it may serve AI research to develop rigorous benchmarks around it.
Such benchmarks will at least improve the quality of discussion around AI capabilities, insofar as people pay attention to them.
Update (2024-04-11 at 19:16 UTC): François Chollet has a new 20-minute talk on YouTube that I recommend. I've watched a few videos of Chollet talking about ARC-AGI or ARC-AGI-2, and this one is beautifully succinct: https://www.youtube.com/watch?v=TWHezX43I-4
10 votes -
Using Claude and undocumented Google Calendar features to automate event creation
4 votes -
Swedish fashion retailer H&M will use AI doppelgangers in some social media posts and marketing in the place of humans, if given permission by models
10 votes -
Vibe coding on Apple Shortcuts
5 votes -
A summary of my bot defence systems
11 votes -
Review: Cræft, by Alexander Langlands
4 votes -
Please stop externalizing your costs directly into my face
121 votes -
Enough with the bullshit (a letter to fellow bullshit sufferers)
56 votes -
Trapping misbehaving bots in an AI Labyrinth
40 votes -
eBay privacy policy update and AI opt-out
eBay is updating its privacy policy, effective next month (2025-04-27). The major change is a new section about AI processing, accompanied by a new user setting with an opt-out checkbox for having...
eBay is updating its privacy policy, effective next month (2025-04-27). The major change is a new section about AI processing, accompanied by a new user setting with an opt-out checkbox for having your personal data feed their models.
While that page specifically references European areas, the privacy selection appears to be active and remembered between visits for non-Europe customers. It may not do anything for us at all. On the other hand, it seems nearly impossible to find that page from within account settings, so I thought I'd post a direct link.
I'm well aware that I'm anomalous for having read this to begin with, much less diffed it against the previous version. But since I already know that I'm weird, and this wouldn't be much of a discussion post without questions:
- How do you stay up to date with contract changes that might affect you, outside of widespread Internet outrage (such as recent Firefox news)?
- What's your threshold -- if any -- for deciding whether to quit a company over contract changes? Alternatively, have you ever walked away from a purchase, service, or other acquisition over the terms of the contracts?
46 votes -
Norwegian man has filed a complaint with the Norwegian Data Protection Authority after ChatGPT falsely told him he had killed two of his sons and been jailed
22 votes -
Claude can now search the web
17 votes -
FOSS infrastructure is under attack by AI companies
39 votes -
LLM crawlers continue to DDoS SourceHut
11 votes -
The lo-fi art and human tools era
10 votes -
(715) 999-7483 - A phone-powered multiplayer website builder
32 votes -
Mayo Clinic's secret weapon against AI hallucinations: Reverse RAG in action
8 votes -
Factorio Learning Environment – a benchmark that tests agents in long-term planning, program synthesis, and resource optimization
13 votes -
Show Tildes: we built the world's first legal AI API
22 votes -
I used to teach students. Now I catch ChatGPT cheats.
53 votes -
Is it wrong to use AI to fact check and combat the spread of misinformation?
I’ve been wondering about this lately. Recently, I made a post about Ukraine on another social media site, and someone jumped in with the usual "Ukraine isn't a democracy" right-wing talking...
I’ve been wondering about this lately.
Recently, I made a post about Ukraine on another social media site, and someone jumped in with the usual "Ukraine isn't a democracy" right-wing talking point. I wrote out a long, thoughtful reply, only to get the predictable one-liner propaganda responses back. You probably know the type, just regurgitated stuff with no real engagement.
After that, I didn’t really feel like spending my time and energy writing out detailed replies to every canned response. But I also didn’t want to just let it sit there and have people who might be reading the exchange assume there’s no pushback or correction.
So instead, I tried leveraging AI to help me write a fact-checking reply. Not for the person I was arguing with, really, but more as an FYI for anyone else following along. I made sure it stayed factual and based in reality, avoided name-calling, and kept the tone above the usual mudslinging. And of course, I double-checked what it wrote to make sure it matched my understanding and wasn’t just spitting out garbage or hallucinations.
But it got me thinking that there’s a lot of fear about AI being used to spread and create misinformation. But do you think there’s also an opportunity to use it as a tool to counter misinformation, without burning ourselves out in the process?
Curious how others see it.
16 votes -
Is there one AI product you would recommend over another to a complete newbie? The primary task is writing.
So I have heard/read that LLMs available to the public can be useful for generating tailored cover letters more quickly. I've up to now avoided using artificial intelligence. What recommendations...
So I have heard/read that LLMs available to the public can be useful for generating tailored cover letters more quickly. I've up to now avoided using artificial intelligence. What recommendations do you have and do you have any advice for getting up to speed?
Thank you.
11 votes -
MIT’s new AI-powered tool accelerates startup ambitions
6 votes -
AI chatbots are people, too. (Except they’re not.)
10 votes -
Sesame conversation AI demo: Crossing the uncanny valley
3 votes -
Could AI lead to a revival of decorative beauty?
13 votes -
Planned foreign-owned data centres in Finland will bring minimal economic benefit, according to Jukka Manner, professor of networking technology at Aalto University
4 votes -
Apple to invest $500 billion in the US in the next four years, build AI server factory
12 votes -
When there’s no school counselor, there’s a bot
18 votes -
Algorithmic complacency: Algorithms are breaking how we think
82 votes -
Have you altered the way you write to avoid being perceived as AI?
I recently had an unpleasant experience. Something I wrote fully and without AI generation of any kind was perceived, and accused of, having been produced by AI. Because I wanted to get everything...
I recently had an unpleasant experience. Something I wrote fully and without AI generation of any kind was perceived, and accused of, having been produced by AI. Because I wanted to get everything right, in that circumstance, I wrote in my "cold and precise" mode, which admittedly can sound robotic. However, my writing was pointed, perhaps even a little hostile, with a clear point of view. Not the kind of text AI generally produces. After the experience, I started to think of ways to write less like an AI -- which, paradoxically, means forcing my very organic self into adopting "human-like" language I don't necessarily care for. That made me think that AI is probably changing the way a lot of people write, perhaps in subtle ways. Have you noticed this happening with you or those around you?
30 votes -
GenAI is reshaping work—don’t let it dull human intelligence
20 votes -
Larry Ellison wants to put all US data in one big AI system
24 votes -
Is it okay to use ChatGPT for proofreading?
I sometimes use chatGPT to proofread longer texts (like 1000+ words) I write in English. Although this is not my first language, I often find myself writing in English even outside of internet...
I sometimes use chatGPT to proofread longer texts (like 1000+ words) I write in English. Although this is not my first language, I often find myself writing in English even outside of internet forums. That is because if I read or watch something in English, and that thing motivates me to write, my brain organically gravitates toward it.
My English is pretty good and I am reasonably confident communicating in that language, but it will never be the same as my native language. So I will often run my stuff through Grammarly and chatGPT. If you wanna say "This will teach you bad habits", please don't. Things like Grammarly and Google Translate taught me so much and improved my English so much, that I am a bit tired of that line of reasoning. I read most of my books in English. I'm not a beginner so I can and do check for all the changes, and vet them myself as I don't always agree with them.
With GPT, I usually just ask it to elaborate a critique rather than spit out a corrected version. Truth be told, when I did ask for a corrected version, it made plenty of sensible corrections that didn't really alter anything other than that. So I guess I just wanna know everyone's feelings about this. Suppose I write a bunch, have GPT correct it for me, compare it with the original and verify every correction. Is that something you would look at unfavorably?
Thanks!
17 votes -
Nokia announces ex-Intel AI and data centre boss Justin Hotard as new CEO – company attempting to venture into artificial intelligence market as 5G sales fall
7 votes -
“Torrenting from a corporate laptop doesn’t feel right”: Meta emails unsealed
28 votes -
Using ChatGPT consumes a 500 ml bottle of water; so what?
11 votes -
DeepSeek R1 reproduced for $30: University of California Berkeley researchers replicate DeepSeek R1 for $30—casting doubt on H100 claims and controversy
48 votes -
DeepSeek’s safety guardrails failed every test researchers threw at its AI chatbot
16 votes -
Building games with LLMs to help my kid learn math
9 votes -
AI is creating a generation of illiterate programmers
52 votes -
A young man used AI to build a nuclear fusor and now I must weep
22 votes -
Why is AI slop so easy to spot but hard to detect?
18 votes -
1,156 questions censored by DeepSeek
37 votes -
DeepSeek FAQ
20 votes