I'm wondering if Google vs. Bing for searching is considering the issue from the right perspective. I have shifted a *lot* of searches from Google to ChatGPT, and that sort of shift wouldn't show up in that graph.
I can grasp Nora's model but as you noted, the evidence isn't there or at least it isn't obvious. This seems like a hazy thing to bet humanity's future on!
And yes, the quantity of AI individuals who have take on a profoundly genocidal view is terrifying; it is also something one would only know having been "on the inside" or generally having familiarity with the mindset, since it is seema so incredibly nuts.
There is something immensely, procedurally toxic in this: that not only are people non-demoncratically asked to tolerate the existence of a new species which we do not want, but also to accept the destruction of all biology and value, because some people think that it doesn't matter.
There have been other religious death cults, but this is apocalyptic in a way that is a very real and close danger.
As for Hansonian logic that if we "decline" for awhile into a lower tech world that it is equivalent of doom, its a good example of I suppose of how little value some people give to the simple joy and beauty of existence, of humans interacting and joy and love, of the stories we tell and the miracles of daily life.
"Geoffrey Miller: I don't want to achieve immortality by being data-scraped into some plagiarism machine. I want to achieve immortality by not dying." Did he give a credit to Woody Allen?
Many mathematicians (even good ones) have a small collection of tricks that they use repeatedly throughout their career. ("What would --- do?" is a good subalgorithm to run once in a while, especially if they just proved a new theorem, but also when stuck on your own stuff.) So, not surprising if the IMO stuff scaled quite a bit in that direction... i.e. Euler AI, Riemann AI, (or for that matter Lurie AI ... the more prolific, the better).
> And yes, I am quickly getting tired of doing this close reading over and over again every time anyone introduces a draft bill, dealing with the same kind of legal theoretical maximalism combined with assuming no one fixes the language.
I mean… giving the laws we end up with, I’m not sure it is *that* ridiculous.
Proving correctness of programs might also turn out to just be a few tricks.
In the cases I'm currently interested in, you do have an exact specification for what the function is supposed to do, and formally proving correctness of the implementation often boils down to some combination of ...
A) replace that instance of that function with its specification
B) case spilt on that Boolean subexpression
C) that variable is a union of several types; case split on which type
D) algebraically simplify
Maybe turns out to be easy
Yes, of course, halting problem. But I don't need to solve the hard cases, just the programs where the guy who wrote it had in his head an informal sketch proof of why the algorithm was correct.
"If you talk about AI and not about housing and permitting and energy and free markets, then you are who you are."
Are you sure about this?
I'm totally in favor of repealing the Jones Act, but it's not obvious to me that people who put 100% of their advocacy chips on *not doing the thing that destroys humanity* are being hypocritical or counterproductive.
> Reminder that you can take essentially any MIT course online for free. Now that we all have access to LLMs, it seems far more realistic for more of us to be able to keep up with and sustain such an enterprise. I am too busy, but tempted.
Is this because LLMs can make it easier to do a course, or because using them judiciously in the rest of life should free up time for such things? If helping with the course itself, I'd be grateful if you could explain more how! For me it feels like the main bottleneck in learning is time reading and understanding... but maybe it would be possible to 80-20 it by just going straight for the key ideas more aggressively.
This post claims that the key Computer Assisted Proof technique used by AlphaGeometry, without the AI component, solves about 80% of the IMO problems AG solves. So it is less impressive than would seem at first glance. Here Terry Tao speculates about another possible use case: https://mathoverflow.net/questions/463937
I'm wondering if Google vs. Bing for searching is considering the issue from the right perspective. I have shifted a *lot* of searches from Google to ChatGPT, and that sort of shift wouldn't show up in that graph.
I can grasp Nora's model but as you noted, the evidence isn't there or at least it isn't obvious. This seems like a hazy thing to bet humanity's future on!
And yes, the quantity of AI individuals who have take on a profoundly genocidal view is terrifying; it is also something one would only know having been "on the inside" or generally having familiarity with the mindset, since it is seema so incredibly nuts.
There is something immensely, procedurally toxic in this: that not only are people non-demoncratically asked to tolerate the existence of a new species which we do not want, but also to accept the destruction of all biology and value, because some people think that it doesn't matter.
There have been other religious death cults, but this is apocalyptic in a way that is a very real and close danger.
As for Hansonian logic that if we "decline" for awhile into a lower tech world that it is equivalent of doom, its a good example of I suppose of how little value some people give to the simple joy and beauty of existence, of humans interacting and joy and love, of the stories we tell and the miracles of daily life.
I think Hanson just selfishly wants to keep his Dewar topped up with LN, or (we all) die trying.
Imo its even more petty:
"Economics must remain the premier science, or we all die trying."
For him, its obvious that a reversion to a 1700s lifestyle with humans is inferior to human extinction.
"Geoffrey Miller: I don't want to achieve immortality by being data-scraped into some plagiarism machine. I want to achieve immortality by not dying." Did he give a credit to Woody Allen?
In context it is kind of funny that he didn't, obviously we all picked up the homage right away.
Many mathematicians (even good ones) have a small collection of tricks that they use repeatedly throughout their career. ("What would --- do?" is a good subalgorithm to run once in a while, especially if they just proved a new theorem, but also when stuck on your own stuff.) So, not surprising if the IMO stuff scaled quite a bit in that direction... i.e. Euler AI, Riemann AI, (or for that matter Lurie AI ... the more prolific, the better).
> And yes, I am quickly getting tired of doing this close reading over and over again every time anyone introduces a draft bill, dealing with the same kind of legal theoretical maximalism combined with assuming no one fixes the language.
I mean… giving the laws we end up with, I’m not sure it is *that* ridiculous.
Proving correctness of programs might also turn out to just be a few tricks.
In the cases I'm currently interested in, you do have an exact specification for what the function is supposed to do, and formally proving correctness of the implementation often boils down to some combination of ...
A) replace that instance of that function with its specification
B) case spilt on that Boolean subexpression
C) that variable is a union of several types; case split on which type
D) algebraically simplify
Maybe turns out to be easy
Yes, of course, halting problem. But I don't need to solve the hard cases, just the programs where the guy who wrote it had in his head an informal sketch proof of why the algorithm was correct.
"If you talk about AI and not about housing and permitting and energy and free markets, then you are who you are."
Are you sure about this?
I'm totally in favor of repealing the Jones Act, but it's not obvious to me that people who put 100% of their advocacy chips on *not doing the thing that destroys humanity* are being hypocritical or counterproductive.
I meant this in the context of someone who was in favor of accelerating or not regulating AI. I agree that the flip side is totally reasonable.
Thanks for having dug some more into Nora Belrose's claims.
Some words seem wrong grammatically in "looking at how humans trying to get elected or help someone get elected actually behave."
"Emphases" → "Emphasizes"
"even if think the future" missing "you"
> Reminder that you can take essentially any MIT course online for free. Now that we all have access to LLMs, it seems far more realistic for more of us to be able to keep up with and sustain such an enterprise. I am too busy, but tempted.
Is this because LLMs can make it easier to do a course, or because using them judiciously in the rest of life should free up time for such things? If helping with the course itself, I'd be grateful if you could explain more how! For me it feels like the main bottleneck in learning is time reading and understanding... but maybe it would be possible to 80-20 it by just going straight for the key ideas more aggressively.
https://old.reddit.com/r/math/comments/19fg9rx/some_perspective_on_alphageometry/
This post claims that the key Computer Assisted Proof technique used by AlphaGeometry, without the AI component, solves about 80% of the IMO problems AG solves. So it is less impressive than would seem at first glance. Here Terry Tao speculates about another possible use case: https://mathoverflow.net/questions/463937