-
25 votes
-
Building a C compiler with a team of parallel Claudes
20 votes -
Is the detachment in the room? - Agents, cruelty, and empathy
15 votes -
Passing question about LLMs and the Tech Singularity
I am currently reading my way thru Ted Chiang's guest column in the New Yorker, about why the predicted AI/Tech Singularity will probably never happen...
I am currently reading my way thru Ted Chiang's guest column in the New Yorker, about why the predicted AI/Tech Singularity will probably never happen (https://www.newyorker.com/culture/annals-of-inquiry/why-computers-wont-make-themselves-smarter). ETA: I just noticed that article is almost 5 years old; the piece is still relevant, but worth noting.
Good read. Still reading, but so far, I find I disagree with his explicit arguments, but at the same time, he is also brushing up very closely to my own reasoning for why "it" might never happen. Regardless, it is thought-provoking.
But, I had a passing thought during the reading.
People who actually use LLMs like Claude Code to help write software, and/or, who pay close attention to LLMs' coding capabilities ... has anyone actually started experimenting with asking Claude Code or other LLMs that are designed for programming, to look at their own source code and help to improve it?
In other words, are we (the humans) already starting to use LLMs to improve their code faster than we humans alone could do?
Wouldn't this be the actual start of the predicted "intelligence explosion"?
Edit to add: To clarify, I am not (necessarily) suggesting that LLMs -- this particular round of AI -- will actually advance to become some kind of true supra-human AGI ... I am only suggesting that they may be the first real tool we've built (beyond Moore's Law itself) that might legitimately speed up the rate at which we approach the Singularity (whatever that ends up meaning).
19 votes -
SpaceX is acquiring xAI
45 votes -
The AI industry doesn’t take “no” for an answer
39 votes -
Any software engineers considering a career switch due to AI?
I've grown increasingly unsure about if I'll stay with this profession long term thanks to the AI "revolution". Not because I think I'll be replaced, I have an extremely wide set of skills thanks...
I've grown increasingly unsure about if I'll stay with this profession long term thanks to the AI "revolution". Not because I think I'll be replaced, I have an extremely wide set of skills thanks to working over a decade in small startups so I think I'm safe for a long while to come.
No, I've grown weary because an increasingly larger share of the code that we produce is expected to be ai generated and with it shorter timelines and I just plain don't like it. I think we reached a tipping point around Claude opus 4.5 where it really is capable and that's only going to continue to get better. But damnit I like coding, I enjoy the problem solving and I feel that's getting stripped away from me basically overnight. Also, as these models become more and more capable I think the number of companies vibe coding to a product with fields of junior level engineers is going to grow which is going to push down senior job opportunities and wages.
So now I'm left wondering if it's time to start pointing towards a new career. I really love building stuff and solving problems so maybe I go back to school and switch to some other flavor of engineering? Idk. Curious where other's heads are at with this.
55 votes -
AI will compromise your cybersecurity posture
8 votes -
A guide to understanding AI as normal technology
10 votes -
Youtube channel ServeTheHome describes how they use a locally running LLM to automate data collection, allowing them to forgo a planned hire
20 votes -
Supporting Markdown search for LLMs
15 votes -
Evaluating LLMs by finding werewolves
18 votes -
Moltbot personal assistant goes viral – and so do your secrets
38 votes -
Suno, AI music, and the bad future
5 votes -
How AI assistance impacts the formation of coding skills
18 votes -
AntiRender: remove the glossy shine on architectural renderings
38 votes -
'Right-to-compute' laws may be coming to your state this year
20 votes -
Anthropic faces new music publisher lawsuit over alleged piracy
5 votes -
Someone made a social media website for AI agents
29 votes -
Pi: The minimal agent within OpenClaw
13 votes -
AI chatbots are becoming lifelines for China’s sick and lonely
8 votes -
Mistral releases Vibe 2.0
16 votes -
Dario Amodei — The adolescence of technology
5 votes -
Wilson Lin on FastRender: a browser built by thousands of parallel agents
18 votes -
Feeling weird about my career with respect to AI
I’m a software engineer. I graduated in 2021 so I’ve only been one for around 4.5 years and definitely still feel fairly entry-level (at least, any time I look at jobs, the number of years of...
I’m a software engineer. I graduated in 2021 so I’ve only been one for around 4.5 years and definitely still feel fairly entry-level (at least, any time I look at jobs, the number of years of experience required for “senior” positions seems to have increased by one) and it feels like companies don’t particularly want anyone without a lot of experience anymore (and every time I do look at new jobs, the number of years required for “senior” positions seems to have increased by one). Meanwhile, I think it has its uses but I don’t actually enjoy using it. I want to solve problems and think and write code, not talk to an AI and become a full-time code-reviewer. My company is rebranding to have AI in the name shortly and, since early December, have been forcing us into 2+ hour long AI trainings once or twice a week. A lot of my coworkers seem like they’ve drank the Kool-Aid and are talking about new models and shit all the time and I just don’t get it.
I guess I’m kind of rambling but I just feel weird about all of it. I want to program but I don’t just want to use (or be forced to use) LLMs for everything, yet it seems like companies are just trying to get rid of actually human software engineers as fast as they can. I’ll even admit, Claude is way better than I expected, but I don’t actually enjoy sitting there typing “do this for me” and then having to just spend time reviewing code. I don’t know. I don’t think this is really even me asking for advice, just a rant, but yeah, just felt like I had to get something out there, I guess.
54 votes -
Show HN: I wrapped the Zorks with an LLM
16 votes -
Blocking Claude
28 votes -
Will your AI teammate bring bagels to standup?
19 votes -
The value of things
14 votes -
New California law means big changes for real estate listing photos
16 votes -
Why does ssh send 100 packets per keystroke?
28 votes -
Cory Doctorow | AI companies will fail. We can salvage something from the wreckage.
91 votes -
The assistant axis: situating and stabilizing the character of large language models
15 votes -
Song streamed millions of times in Sweden has been banned from the country's music charts because it was created by AI
13 votes -
Why we are excited about confessions
30 votes -
exe.dev, a service for creating Linux virtual machines and vibe-coding in them
23 votes -
Apple to partner with Google for Gemini access on iPhones, Apple Intelligence to power on device assistant
29 votes -
An explainer: physical AI must sense, think, act, and optimize
4 votes -
HiTeX Press: A spam factory for AI-generated books
15 votes -
Google removes some of its AI summaries after users’ health put at risk
22 votes -
Debunking the AI food delivery hoax that fooled Reddit
70 votes -
US judge indicates Elon Musk’s fraud lawsuit against OpenAI will head to trial
20 votes -
Dell's Consumer Electronics Show 2026 chat was the most pleasingly un-AI briefing I've had in maybe five years
34 votes -
You are a better writer than AI (yes, YOU!)
25 votes -
Grok AI generates images of ‘minors in minimal clothing’
44 votes -
Scalable oral exams with an ElevenLabs voice AI agent
16 votes -
Buying a lotta RAM now, as an investment ... thoughts?
Just a passing thought, came up in conversation. I'm not talking about warehouses-full, nor even "retirement savings" quantities, but like, all the RAM you and your friends and family could...
Just a passing thought, came up in conversation. I'm not talking about warehouses-full, nor even "retirement savings" quantities, but like, all the RAM you and your friends and family could possibly need for the next 3-4 years.
Pros, cons? Too late? Too volatile? Too ___?
22 votes -
AI friends too cheap to meter
40 votes -
Two visions for the future of AR smart glasses
18 votes -
Landscape with a nature says the Digital Curator
3 votes