-
18 votes
-
“60s lounge” and Laufey
I was really taken by the sound of Laufey’s Madwoman and am looking for the older music which inspired her. The internet, however, has not been helpful. Between citing “The Great American...
I was really taken by the sound of Laufey’s Madwoman and am looking for the older music which inspired her.
The internet, however, has not been helpful. Between citing “The Great American Soundbook”¹ as her inspiration, or declaring its “clearly” bossa nova², i’m really lost. One person said “60s lounge,” which seemed promising³ but searching that brings a ton of AI-generated slop-and-slop-byproducts.
Is there a better term for what she’s referencing? Specific artists? Specific regions even?
- which doesn’t sound the same to me at all
- which no the fuck it is not, i, a brazilian, say
- another person said “70s lounge”
14 votes -
Static analysis, dynamic analysis, and stochastic analysis
For a long time programmers have had two types of program verification tools, static analysis (like a compiler's checks) and dynamic analysis (running a test suite). I find myself using LLMs to...
For a long time programmers have had two types of program verification tools, static analysis (like a compiler's checks) and dynamic analysis (running a test suite). I find myself using LLMs to analyze newly written code more and more. Even when they spit out a lot of false positives, I still find them to be a massive help. My workflow is something like this:
- Commit my changes
- Ask Claude Opus "Find problems with my latest commit"
- Look though its list and skip over false positives.
- Fix the true positives.
git add -A && git commit --amend --no-edit- Clear Claude's context
- Back to step 2.
I repeat this loop until all of the issues Claude raises are dismissable. I know there are a lot of startups building a SaaS for things like this (CodeRabbit is one I've seen before, I didn't like it too much) but I feel just doing the above procedure is plenty good enough and catches a lot of issues that could take more time to uncover if raised by manual testing.
It's also been productive to ask for any problems in an entire repo. It will of course never be able to perform a completely thorough review of even a modestly sized application, but highlighting any problem at all is still useful.
Someone recently mentioned to me that they use vision-capable LLMs to perform "aesthetic tests" in their CI. The model takes screenshots of each page before and after a code change and throws an error if it thinks something is wrong.
10 votes -
Tildes Minecraft Weekly
Server host: tildes.nore.gg (Running Java 1.21.11) Verification site: https://tildes.nore.gg BlueMap: https://tildes.nore.gg/map/ Patreon: https://www.patreon.com/TildesMC Plugins and Data Packs...
Server host:
tildes.nore.gg(Running Java 1.21.11)
Verification site: https://tildes.nore.gg
BlueMap: https://tildes.nore.gg/map/
Patreon: https://www.patreon.com/TildesMCPlugins and Data Packs
Data Packs:- Terralith - Overworld terrain upgrade
- Nullscape - End terrain upgrade
- Age Lock [Vanilla Tweaks]
- Armor Statues [Vanilla Tweaks]
- Bat Membranes [Vanilla Tweaks]
- Cauldron Concrete [Vanilla Tweaks]
- Cauldron Mud [Vanilla Tweaks]
- Custom Nether Portals [Vanilla Tweaks]
- Husks Drop Sand [Vanilla Tweaks]
- Mini Blocks [Vanilla Tweaks]
- More Mob Heads [Vanilla Tweaks]
- Player Head Drops [Vanilla Tweaks]
- Silence Mobs [Vanilla Tweaks]
- Wandering Trades [Vanilla Tweaks]
Plugins:
- BlueMap - Provides a live 3D rendering of the game world
- Clickable Links - Makes http URLs in chat clickable (only for registered players)
- CoreProtect - Records all block/container/mob changes (Anyone can look up changes with
/co inspect) - DebugStick - Gives the ability to craft debug sticks in survival
- DistantHorizons - Provides distant LOD map data to players running the client mod
- EasyArmorStands - GUI for editing armor stands
- Hexnicks - Enables Tildes usernames to be displayed
- hsrails - Allows for 4x speed rail travel
- LuckPerms - Locks down unregistered users
- Otherside - Fix for mob farms involving Nether portals
- Rapid Leaf Decay - Increases the speed of leaf decay by 10x
- WorldEdit - Used for occasional admin stuff
- WorldGuard - Prevents unregistered users from changing anything in the world
The server operates on a soft whitelist. Anyone can log in and walk around, but you need a Tildes account to gain build access.
We recommend you install our mod web-chat so that you can chat while in your web browser. It turns the server into an old-school chat room.
<- Previous Thread Next Thread ->
15 votes -
Finnish transport minister Lulu Ranne says a long-delayed cross-border train to Sweden should begin rolling this summer
9 votes -
Inside Doug Liman’s $70 million AI-made movie starring Casey Affleck and Gal Gadot
18 votes -
Making the most pickproof lock yet
14 votes -
Very Important People: Boris Tarshkokan
15 votes -
Olivia Rodrigo - drop dead (2026)
7 votes -
What programming/technical projects have you been working on?
This is a recurring post to discuss programming or other technical projects that we've been working on. Tell us about one of your recent projects, either at work or personal projects. What's...
This is a recurring post to discuss programming or other technical projects that we've been working on. Tell us about one of your recent projects, either at work or personal projects. What's interesting about it? Are you having trouble with anything?
11 votes -
Babylon 5 S01E07: "The War Prayer" - Episode Discussion
7 votes -
METRO 2039 | Official reveal trailer
16 votes -
Steam compatible with ROCKNIX firmware
17 votes -
Staged homes sell for 10% more and one week faster
33 votes -
Street Fighter | Official trailer
18 votes -
Strange Romance bundle by Image Comics
10 votes -
José González – A Perfect Storm (2026)
6 votes -
What's a battle that nobody knows you're fighting?
The "nobody" in the title doesn't have to be literal -- it can be a battle that very few people know about. The important thing is that it's mostly hidden. What is the struggle? Is it hidden by...
The "nobody" in the title doesn't have to be literal -- it can be a battle that very few people know about. The important thing is that it's mostly hidden.
What is the struggle?
Is it hidden by choice?
Do you want more people to know about it? Why or why not?27 votes -
The Dog Stars | Official trailer
5 votes -
Porting Mac OS X to the Nintendo Wii
26 votes -
As Deep As The Grave | Official trailer (AI Val Kilmer movie)
6 votes -
Tigran Hamasyan - Manifeste | Full album (2026)
12 votes -
Sovereign Cloud stats every CIO needs before their next board meeting
7 votes -
Jet Lag Season 17: Taiwan Rail Rush | Trailer
22 votes -
Allbirds announces pivot from running shoes to AI compute; stock surged over 700%
50 votes -
Squarepusher - Kammerkonzert | Full album (2026)
6 votes -
Amorphis – The Lantern (2026)
7 votes -
Lana Del Rey - First Light (2026)
3 votes -
That one study that proves developers using AI are deluded
I've found myself replying to different people about the early 2025 METR study kind of often. So I thought I'd try posting a top level thread, consider it an unsolicitied public service...
I've found myself replying to different people about the early 2025 METR study kind of often. So I thought I'd try posting a top level thread, consider it an unsolicitied public service announcement.
You might be familiar with the study because it has been showing up alongside discussions about AI and coding for about a year. It found that LLMs actually decreased developer productivity and so people love to use it to suggest that the whole AI coding thing is really a big lie and the people who think it makes them more productive are hallucinating.
Here's the thing about that study... No one seems to have even glanced at it!
First, it's from early 2025, they used Claude Sonnet 3.5 or 3.7. Those models are no way comparable to current gen coding agents. The commonly cited inflection point didn't happen until later in 2025 with, depending on who you ask, Sonnet 4.5 or Opus 4.5
The study was comprised of 16 people! If those 16 were even vaguely representative of the developer population at the time most of them wouldn't have had significant experience with LLMs for coding.
These are not tools that just work out of the box, especially back then. It takes time and experimentation, or instruction, to use them well.
It was cool that they did the study, trying to understand LLMs was a good idea. But it's not what anyone would consider a representative, or even well thought out, study. 16 people!
But wait! They did a follow up study later in 2025.
This time with about 60 people and newer models and tools. In that study they found the opposite effect, AI tools sped developers up (which is a shock to no one who has used these tools long enough to get a feel for them). They also mentioned:
However the true speedup could be much higher among the developers and tasks which are selected out of the experiment.
In addition they had some, kind of entertaining, issues:
Due to the severity of these selection effects, we are working on changes to the design of our study.
Back to the drawing board, because:
Recruitment and retention of developers has become more difficult. An increased share of developers say they would not want to do 50% of their work without AI, even though our study pays them $50/hour to work on tasks of their own choosing. Our study is thus systematically missing developers who have the most optimistic expectations about AI’s value.
And...
Developers have become more selective in which tasks they submit. When surveyed, 30% to 50% of developers told us that they were choosing not to submit some tasks because they did not want to do them without AI. This implies we are systematically missing tasks which have high expected uplift from AI.
And so...
Together, these effects make it likely that our estimate reported above is a lower-bound on the true productivity effects of AI on these developers.
[...]
Some developers were less likely to complete tasks that they submitted if they were assigned to the AI-disallowed condition. One developer did not complete any of the tasks that were assigned to the AI-disallowed condition.
[...]
Altogether, these issues make it challenging to interpret our central estimate, and we believe it is likely a bad proxy for the real productivity impact of AI tools on these developers.
So to summarize, the new study showed a productivity increase and they estimate it's larger than the ~20% increase the study found. Cheers to them for being honest about the issues they encountered. For my part I know for sure that the increase is significantly more than 20%. The caveat, though, is that is only true after you've had some experience with the tools.
The truth is that we don't need a study for this, any experienced engineer can readily see it for themselves and you can find them talking about it pretty much everywhere. It would be interesting, though, to see a well designed study that attempted to quantify how big the average productivity increase actually is.
For that the participants using AI would need to be experienced with it and allowed to use their existing setups.
I want to add that this is not an attempt to evangelize for AI. I find the tools useful but I'm not selling anything. I'm interested in them and I stay up to date on the conversations surrounding them and the underlying technology. I use them frequently both for my own projects and to help less technical people improve their business productivity.
Whether AI agents are a good thing or not, from a larger perspective, is a very different, and complicated, conversation. The important thing is that utility and impact are two different conversations. There isn't a debate anymore about utility.
I know this probably won't stop people from continuing to derail conversations with the claim that developers are wrong about utility, but I had to try. It's just hard to let it pass by when someone claims the sky is green.
I understand that AI makes people angry and I think they have good reason to be angry. There are a lot of aspects of the AI revolution that I'm not thrilled about. The hype foremost, the FOMO as part of the hype, the potential for increased wealth consolidation really sucks, though I lay that at the feet of systems that existed before LLMs came along.
It's messy, but let's consider giving the benefit of the doubt to professionals who say a tool works instead of claiming they're wrong. Let them enjoy it. We can still be angry at AI at the same time.
82 votes -
Bluesky is down (at least for me)
I noticed bluesky is down. I haven't seen news about this yet so I wonder if it is just a temporary tech problem or a DOS. I'm getting this error or an error about rate limiting: Hmm, some kind of...
I noticed bluesky is down.
I haven't seen news about this yet so I wonder if it is just a temporary tech problem or a DOS.
I'm getting this error or an error about rate limiting:
Hmm, some kind of issue occurred when contacting the feed server. Please let the feed owner know about this issue.
Message from server: Upstream server responded with a 503 error9 votes -
Ticketmaster is an illegal monopoly, jury rules
72 votes -
Lufthansa axes CityLine fleet early over strikes, fuel costs
12 votes -
US keyboards don't have enough keys, so I switched to Japanese - HyperJIS
23 votes -
The cosmetic industry’s new frontier: cadaver fat
19 votes -
Hi, how are you? Mental health support and discussion thread (July 2025)
This is a monthly thread for those who need it. Vent, share your experiences, ask for advice, talk about how you are doing. Let's make this a compassionate space for all who may need one.
20 votes -
Godzilla Minus Zero | First look teaser
19 votes -
The center has a bias
35 votes -
IETF has published an IPv8 draft
16 votes -
Why Copenhagen is so well-run
5 votes -
No one can force me to have a secure website!!!
36 votes -
Midweek Movie Free Talk
Warning: this post may contain spoilers
Have you watched any movies recently you want to discuss? Any films you want to recommend or are hyped about? Feel free to discuss anything here.
Please just try to provide fair warning of spoilers if you can.
9 votes -
Shawn Phillips - Rumplestiltskin's Resolve (1976)
2 votes -
Why cheap waste management is key to stopping plastic pollution
47 votes -
Gåte feat. Maria Franz – Sannsiger (2026)
3 votes -
Focker-in-Law | Official trailer
2 votes -
The raspberry danish latte is making its way around the world after its inventors at a US small town coffee shop decided to share the recipe
24 votes -
Megathread: April Fools' Day 2026 on the internet
Over the next day or so, the internet will be filled with jokes, pranks, fake "announcements" from companies, fun interactive activities, games, and so on. A lot of these can be quite clever and...
Over the next day or so, the internet will be filled with jokes, pranks, fake "announcements" from companies, fun interactive activities, games, and so on. A lot of these can be quite clever and interesting so I think posting about them in general is fine, but in the interest of preventing them from completely taking over Tildes, let's try to keep as many of them restricted to this thread as possible. Ideally, a separate top-level comment for each individual item would be good.
If something particularly discussion-worthy comes up (like an ARG or activity that a lot of people want to talk about), a separate thread is reasonable, but please make sure it has the "april fools day" tag. That way, if anyone wants to avoid seeing the April Fools' Day threads, they can use the topic tag filters and filter that tag out.
I'm going to use the "official" styling for this topic (that's usually only for ~tildes.official topics) to make it stand out more to try to encourage people to notice it. If you notice people making individual topics for April Fools' Day things that don't really warrant their own topic, please (nicely) encourage them to delete and post in here instead.
101 votes -
Dual national Londoner stranded in Spain by new border rule
17 votes -
Train Jazz: A jazz combo played in real time by every active NYC subway train
25 votes -
Predicting the NBA MVP with Machine Learning
Predicting the NBA MVP with Machine Learning Thesis Every season, basketball fans debate who deserves the MVP award. We built 3 machine learning models that attempt to answer that question using...
Predicting the NBA MVP with Machine Learning
Thesis
Every season, basketball fans debate who deserves the MVP award. We built 3 machine learning models that attempt to answer that question using box score statistics. At the end of each season, this award is determined by a panel of voters.
Methodology
Each model is trained on every NBA season from 1974 to 2017. For each player season, it looks at nine statistics:
- Points, assists, blocks, defensive rebounds, and field goals per game the core production numbers
- Win Shares (WS): an estimate of how many wins a player contributed to their team
- Value Over Replacement Player (VORP): how much better a player is than a league average replacement
- Box Plus/Minus (BPM): a player's net impact per 100 possessions
- Usage Rate (USG%): what share of team plays run through that player
From those nine numbers, the model learns what a typical MVP season looks like versus a non MVP season, then applies that knowledge to current players. Each model outputs an independent probability that a given player wins MVP, not a share of a single pool, so the values do not sum to 1. Think of it as each player's individual odds.
Three Models, One Question
Rather than relying on a single approach, the system runs three different models and lets you compare:
Logistic Regression
The simplest of the three. It draws a straight line through the data, each statistic gets a weight, and a player's score is the weighted sum of their stats. It's easy to interpret (a higher coefficient means that stat matters more).
Win Shares (WS) is by far the most influential feature, with an absolute coefficient of ~1.85, nearly double the next most important feature. Box Plus/Minus (BPM) ranks second at ~1.0, followed by Defensive Rebounds per Game (DRBPG, ~0.85) and Assists per Game (ASTPG, ~0.70). VORP and Field Goals per Game (FGPG) contribute moderately at ~0.50. Blocks per Game (BLKPG), Points per Game (PTSPG), and Usage Rate (USG%) have minimal weight, all under 0.15.
Random Forest
Builds hundreds of decision trees, each one asking a series of "is this stat above or below X?" questions and averages their answers. It handles complex relationships between stats well and is less sensitive to any one unusual data point. Think of it as a large committee of simple rules voting together.
WS again dominates at ~0.31, accounting for roughly twice the importance of the next feature. VORP (~0.15) and BPM (~0.125) rank second and third. DRBPG (~0.10), PTSPG (~0.08), BLKPG (~0.07), FGPG (~0.065), and ASTPG (~0.06) contribute in a fairly tight mid-range band. USG% is the least important at ~0.05. Compared to logistic regression, the Random Forest spreads importance more evenly across features.
Gradient Boosting
Also uses decision trees, but builds them sequentially: each new tree focuses on correcting the mistakes the previous ones made.
This model is heavily concentrated on just two features: BPM (~0.47) and WS (~0.41) together account for roughly 88% of total feature importance. All remaining features, PTSPG, VORP, ASTPG, DRBPG, contribute ~0.02–0.03 each, and BLKPG, USG%, and FGPG are effectively unused (near zero). This suggests the gradient boosting model learned that BPM and WS alone are nearly sufficient to separate MVP candidates.
Historical Results
The models were trained on data through 2017, so every season from 2018 onward is a genuine out of sample test, the models have never seen these players or seasons before.
Season Actual MVP LR RF GB 2018 James Harden #2 #2 #1 ✓ 2019 Giannis Antetokounmpo #1 ✓ #1 ✓ #1 ✓ 2020 Giannis Antetokounmpo #1 ✓ #1 ✓ #1 ✓ 2021 Nikola Jokić #1 ✓ #1 ✓ #1 ✓ 2022 Nikola Jokić #1 ✓ #1 ✓ #1 ✓ 2023 Joel Embiid #2 #4 #2 2024 Nikola Jokić #1 ✓ #1 ✓ #1 ✓ 2025 Shai Gilgeous-Alexander #3 #2 #569 Top-1 accuracy: LR 5/8 · RF 5/8 · GB 6/8
Top-3 accuracy: LR 8/8 · RF 7/8 · GB 7/8
Top-3 accuracy: LR 8/8 · RF 7/8 · GB 7/8
For five straight seasons (2019–2022 + 2024), all three models agreed on the same #1 pick, and were right every time.
In 2023, every model ranked Nikola Jokić #1, and by the numbers, he arguably had the better season. Joel Embiid won the award anyway, the kind of outcome that may reflect voter narrative/fatigue and team performance rather than pure statistics. In 2025, Gradient Boosting ranked Shai Gilgeous-Alexander outside the top 500, while Logistic Regression and Random Forest had him at #3 and #2 respectively. I have no idea why GB did this. Likely a bug.
Future Direction
No model is perfect, and these have known blind spots. Team record is not included, MVP voters have historically punished players on losing teams regardless of individual stats. Injuries and narrative don't appear in a box score. And the training data skews toward an older era; the three point revolution and the rise of players like SGA have introduced statistical profiles the 1970s–1990s data doesn't fully capture.
Current Season Predictions (2025–26)
LR RF GB #1 Nikola Jokić Shai Gilgeous-Alexander Nikola Jokić #2 Shai Gilgeous-Alexander Nikola Jokić Victor Wembanyama #3 Victor Wembanyama Victor Wembanyama Giannis Antetokounmpo #4 Luka Dončić Giannis Antetokounmpo Kawhi Leonard #5 Jalen Johnson Luka Dončić Luka Dončić Two of the three models have Nikola Jokić as the frontrunner. Random Forest is the dissenter, putting Shai Gilgeous-Alexander ahead. Victor Wembanyama appears in all three top 3s in just his second season, which is notable. Before running the models, I expected him to be #1 for all of them considering the way the models use advanced stats.
Conclusion
Thank you for reading. I hope you found this interesting. Basketball reference also has their own model if you would like to see a different result. Please do not gamble on my models!
13 votes