LLMs collapse to stereotypical answers even on open-ended prompts
βMode collapse is a real concern with LLM generation. So, what we find in our paper is that even when you ask open-ended questions, like, you know, tell me a joke about time, or tell me something wise about time. Even when you ask, by the way, hey, give me a random number between and 10. It's not random... The bigger problem is after post-training, like sequential fine-tuning and RL, the probability, output probability of the model becomes even more skewed, like zoning in to the stereotypical answers that people tend to like.β
AI will likely displace or suppress many white-collar jobs
βI think more than 20 percent, but less than 50 So like let's say five years, 20 to 30 percent of jobs are not just displaced, but also suppressed. So you don't make that marginal higher, and that adds up to tens of millions of jobs.β
βThe idea is that during pre-training, the model is forced to be completely passive in the way that it learns to predict which token comes next. But what if we encourage the model to think for itself? Before predicting the next token, what if we encourage the model to think for itself by generating something like a chain of a thought? And then predict the next token... the key idea of our approach is to make the reward information gain of predicting next token with thought compared to without thought.β
Prismatic Synthesis beats teacher models 20x its size
βI can give you one example of our recent work called the prismatic synthesis. It's a synthetic data generation algorithm which is prismatic because it acts like a little bit like a prism that can scatter the light to make it more diversified... we're doing this using Dipsic R1 32 billion parameter model as the teacher model... our goal is to compete against the alternative, which is to use much stronger teacher that's 20 times larger... we find that that one million data points is actually better than the one million data points that you generate from the stronger teacher model, the best teacher model.β
Public venture capital will become a standard portfolio allocation
βI think that like 10 years from now, every single person in America will have 5% of their portfolio in like public venture capital. It'll be totally normal. It'll be like an ETF. It'll be just this thing that's standard. And the public capital, so the public markets will be a huge source of funding for private tech companies.β
Closed-end venture funds can trade at massive public premiums
βThe closed ended fund, it's like any other venture fund and people are buying and selling the actual fund itself in the public markets. So that was, it was a novelty because everyone told us that closed end funds will trade at a discount. Like that's what, in the history of closed end funds, 99% traded discounts. That's how normal closed end funds work. And I thought, well, this is different. This is not like a normal closed end fund, it's something new. And I think it would trade at a premium.β
VCX provides public access to top private tech companies
βWe created what was the first public venture fund. So it's a venture fund, like any other venture fund, typically venture funds are closed end funds. This fund was registered with the SEC so that anyone can invest in it. And so we have a Fundrise has this platform where you can show up and invest it at $10 minimum.β
Small models can rival large ones with better data
βThe mission really is democratizing general AI, so that it's not just companies who can purchase a lot of GPUs, are able to create LLMs and adapt to LLMs and serve LLMs, but also people like myself and colleagues who are academics, so for example, cannot buy as many GPUs, and is there something really meaningful and fun that we could do, even with a smaller counterpart? And at the end of the day, I believe that fundamentally it should be feasible. It's only that the world has invested so much more into exploring what happens when you scale things up so much.β
βThere's a lot of delving now that wasn't happening before. Yeah, probably, yeah. You know, actually, whenever I see the word delve in anybody's writing, I'm like, hmm, what did you do?β
Private companies stay private longer to avoid public volatility
βIf you're a company, if you're a tech company, you're trying to build some cutting edge software and you're taking a lot of risk, the public markets are just risk. They're just burden, they're just complexity. You don't need it if you can raise money in the private markets. And so companies don't go public until they're really mature, or if ever.β
Private markets capture significantly higher growth than public benchmarks
βThe weighted average growth rate of public tech companies last year was 25%. The weighted average growth rate for our portfolio for VCX was 193%.β
βThere must be a better way of it, fundamentally better way of doing this. And can we find it? In some ways, the nature found a solution, which is the human brain. The nature found the solution, and human brain requires so little energy. Our brain apparently use less energy than one light bulb.β
βDeepSeq R1 does some amount of this imitation learning after the reinforcement learning... if you just do reinforcement learning, by the way, sometimes this model starts code switching in the middle of solving math problems. It's just suddenly speaking in Chinese and English and back and forth, or some other foreign languages that may not make sense to human readers. So reinforcement learning only cares about whether you got the final solution right or not. It doesn't care about how you got there. So strange behaviors can be emergent and then it can be even reinforced.β
Pluralistic alignment means AI should present multiple valid views
βOvertone pluralism means when you ask a question that's politically thorny, for example, that could have different answers, the best way might be for LLM to just present all of them. All of the reasonable opinions as, hey, the answer is that people have different opinions. Here's one view, there's another view, and be able to include all of them as opposed to picking the majority opinion. Because that marginalizes out the rest.β
Universal basic ownership is the solution to AI disruption
βI think democratic ownership is the solution to how we deal with AI. The idea that everyone wants to get like universal basic income, I think that's a horrible idea. But universal basic ownership, I feel like that is magic. It's like instead of people saying, okay, the American dreams own a home, the American dreams own like the means of production, which is AI.β
Strategic secondary buying allows access to competitive venture rounds
βSome of the companies you named, and I'm not supposed to say which ones, we bought from distressed funds who were having to sell in early to mid 2023 So there was like this downturn. I mean, Silicon Valley Bank blew up in April. And so again, there's a lot of luck and some opportunistic hustle on our part, but it was just not what people thought was the smart move.β