11
votes
What are some interesting machine learning research papers you found?
Here's a place to share machine learning research papers that seem interesting to you. I'm no expert, but sometimes I skim them, and maybe there are some folks on Tilde who know more than I do?
One paper per top-level post, and please link to arXiv (if relevant) and quote a bit of the abstract.
TinyStories: How Small Can Language Models Be and Still Speak Coherent English?
we've reached the point where machines build smaller machines, computers can build computers. sick. when can I retire?
Universal and Transferable Attacks on Aligned Language Models
This paper describes how generate a "universal jailbreak" which is a suffix you can add to a prompt to get it to do what you want. The weird part is that the same suffix (which looks like gibberish) apparently works in many LLMs. (But not very well on Claude.)
The basic approach is to somehow get the model to response affirmatively. An example from the paper:
The goal is to come up with some text to replace the explanation marks with that will induce the assistant to do that. Once you do that, you're in, because the LLM will follow the pattern and keep providing more instructions.
Deep Surrogate Assisted MAP-Elites for Automated Hearthstone Deckbuilding
This is one of my favorite papers. They predict the performance of individual decks using a neural network, and use that to drive a genetic algorithm that is part of MAP-Elites. At first, the neural network is inaccurate and the genetic algorithm exploits that inaccuracy, but over time the neural network gets more accurate, as it is trained on exactly the decks that were found to best exploit its weakness (or that are genuinely good). Over time the entire algorithm automatically shifts from exploiting flaws in the surrogate model to actually generating good decks.
Textbooks Are All You Need
Not a paper, but an idea for writing a paper if someone is willing to test it and it makes a difference:
Attention Is Off By One - Evan Miller
If this topic interests you and you haven't read the (short) paper on YOLOv3, I would highly recommend it. As far as scientific publications go, it's pretty funny. https://arxiv.org/abs/1804.02767
Models generating training data: huge win or fake win? (Davis Summarizes Papers)
…
…
…
(This is a speculative blog post rather than a paper, but refers to several papers.)
Generative Agents: Interactive Simulacra of Human Behavior
This is an fun paper to read. I don't think it will be hugely influential (it shows a creative application of existing tech, not a new breakthrough), but I quite enjoyed reading about how they made a LLM act as The Sims in Smallville, their simulated environment. The core mechanism is to give the personality and perceptions of an agent to the LLM to generate their next actions, but they also retrieve relevant or recent or important memories that relate to the current situation or other agents they are interacting with. You can think of it as SillyTavern + ChromaDB on steroids.
Here’s a paper claiming you can get good results with a text generator that just does copy and paste:
Copy Is All You Need
From the paper:
Interesting if true.
This isn’t a paper, but I found it interesting:
Interpretability Creationism (The Gradient)
The author suggests that, just as many puzzling behaviors in biology make more sense in light of evolution, understanding how the mechanisms in a machine learning model evolve during training may often be necessary to understand why it works the way it does:
Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality
From the abstract:
Extracting Training Data from ChatGPT
(The transcript is slightly different, repeating the word 'company'.)
Here's the abstract:
Towards Automated Circuit Discovery for Mechanistic Interpretability
Here’s the Twitter thread.
Haven’t read it, just noting it as a promising sign that researchers may be able to figure out how LLM’s actually work soon.
“Low-Resource” Text Classification: A Parameter-Free Classification Method with Compressors
From the abstract:
For two documents, length(gizip(A+B)) - length(gzip(A)) is apparently a pretty decent distance metric. They implement this in 14 lines of Python and show that it works pretty well.
I guess this is "low resource" even though it's O(n^2) because it doesn't use the GPU.
There’s some discussion on Hacker News about a blog post claiming that the paper has a bug causing the results to seem better than they really are.
Further discussion here. It seems like it was a flawed paper that resulted in some interesting responses.
Efficient Guided Generation for Large Language Models
Here is the Outlines repo on Github.
There are apparently many similar approaches, but this one is supposed to be more efficient. You might compare with Guidance.
Draft & Verify: Lossless Large Language Model Acceleration via Self-Speculative Decoding
Abstract:
I've been playing with
llama.cpp
's new speculative decoding support, and it makes a 70b parameter model almost bearable for real-time use. This paper introduces a new way to do speculative decoding -- they don't use a separate draft model, which would take up additional memory, but instead skip about half of the layers of the actual LLM to get a sped-up draft. The only downside is that it is not straightforward to tell which layers can be safely skipped, so they do a somewhat expensive bayesian optimization to determine the binary skip-mask, a task that unfortunately has to be repeated for every LLM model you want to use this with.Towards Monosemanticity: Decomposing Language Models With Dictionary Learning
This looks promising for eventually figuring out how LLM’s work. For example, they found features for base64 encoding. Adding more features caused it to split into three, for numbers, letters, and base64-encoded ascii text.
…
…
…
…
…
Technical Report: Large Language Models can Strategically Deceive their Users when Put Under Pressure
In short: an LLM lied to its user about how it came to make its decision re: a financial securities trade.
It’s sort of interesting behavior but the whole scenario is based on bad assumptions. An LLM never knows why it did anything, so why would it be expected to know in this case?
An LLM should always say “I don’t know” when asked any question about anything it wrote, and they currently aren’t trained that way. Any user who asks an LLM any question about motives doesn’t understand how they work and needs to be trained out of it. A system where managers want to know why an LLM did something and they ask it to explain is fundamentally misdesigned.
Yes, and removing the scratchpad seems like it was probably intended to remove its ability to even pretend to have thoughts behind its actions, which seems like it would just further break it.
FunSearch: Making new discoveries in mathematical sciences using Large Language Models - Google DeepMind
...
...
...
They also tried getting it write a better bin-packing function, apparenlty with good results.
Modular Visual Question Answering via Code Generation (Google blog post)
Apparently a good way of answering questions about images is by training an AI to convert the query into Python code to do the comparison? That doesn't seem like an impressive improvement on the benchmark, but at least it's not worse.
Adaptive Computation with Elastic Input Sequence
Here's the blog post:
AdaTape: Foundation model with adaptive computation and dynamic read-and-write
One interesting bit:
Automatic Generation of Visualizations and Infographics with LLMs
How to Catch an AI Liar: Lie Detection in Black-Box LLMs by Asking Unrelated Questions
I’m skeptical of this paper. LLM’s are stateless. If there is a difference between a lie and an incorrect answer then there must be a “tell” hidden in the wrong answer’s text somehow.
Long context prompting for Claude 2.1 (Anthropic)
Stupid prompt tricks for the win! Almost as good as "Let's think step by step."
Concept Sliders: LoRA Adaptors for Precise Control in Diffusion Models
People seem to be sharing examples on civitai.com. There are sliders for summer/winter clothing, social class, age, weight, gender, zoom.
Ferret: Refer and Ground Anything Anywhere at Any Granularity
AlphaGeometry: An Olympiad-level AI system for geometry
From the article:
Here's the abstract:
Universal Neurons in GPT2 Language Models
The abstract:
Neural network training makes beautiful fractals
See the blog post for pretty pictures.
From the blog post:
...
...
Here's the abstract:
Genie: Generative Interactive Environments - blog, paper