16 Comments

I don't think it's good to use chatgpt to win arguments, so I asked chatgpt and it had this to say:

"Lack of Emotional Connection: If conversations are reduced to just point-by-point arguments, it may take away the emotional and empathetic aspect of resolving disagreements. Relationships thrive on emotional connection, not just logic and structure.

Over-reliance on External Input: Using an external tool like ChatGPT for every disagreement could create dependency, potentially undermining authentic, spontaneous communication.

Escalating Arguments: While well-structured arguments can be helpful, they can also make the conversation feel more like a debate or competition, which might intensify conflict rather than resolve it."

Expand full comment

On 𝕏 I frequently use ChatGPT to "win" arguments, but I use this in specific ways, and never for the direct benefit of my interlocutor, who has by the time I start posting screen shots from ChatGPT, has already demonstrated fallacious thinking that GPT is very good at detecting and calling out. Rather, it is for the benefit of the mostly-silent audience watching the exchange.

I do it by taking screen shots of the full conversational context and giving it to ChatGPT so it can craft a reply that is quite specific to the conversation at hand.

How does my interlocutor react? Usually with outrage and dismissal, accusing me of "resorting to using AI". I actually find this funny, because it's an actual deflection from what has happened: their fallacies have been called out, if not having their argument actually shredded. In fact, I have a standard rejoinder to this, which is a reply to the pinned post in my profile:

“I use Al to call out those with weak or fallacious arguments. I only post Al output that I agree with and where I am confident in its accuracy. If that upsets you, or you attack the tools l use, then you have already lost the argument. Argue with facts, not sources.”

And this is the crux of the matter: if the AI is right, and thorough at it, then it's right, and it doesn't matter that it's AI generated. The emotional detachment of the AI is actually a *benefit* in these cases, because my interlocutor has frequently already deployed arguments from emotion or ad hominem. Usually the reaction of my interlocutor exposes even deeper levels of bad faith, and I believe that is a *good* thing for public discourse.

Expand full comment

A general xrisk prevention thought experiment I am curious about: imagine you are dedicated to the goal of preventing nuclear weapons from becoming an existential threat to humankind. I will send you and a handful of others you nominate back to 1901. You will arrive within various major polities with levels of significant authority, and fluent in local languages and customs. You will be able to tell anyone you like anything true about nuclear weapons and nuclear power (including military, geopolitical, social and environmental impacts) and they are guaranteed to believe you, without requiring proof.

You are faced with the following constraints:

1. You cannot name the specific technology (you must refer to it as XY).

2. You cannot provide any specifics about how the technology functions.

3. You cannot provide any historical specifics (e.g., you could explain that global superpowers will emerge with massive stockpiles of weapons built on this new technology which are capable of obliterating the world multiple times over, but you cannot give any hint as to which powers you are referring to: folks will of course extrapolate but you can't correct or lead them to a specific 'correct' extrapolation).

4. You cannot provide any specifics as to timeframe except 'within the next 100 years'.

Given these constraints, what you would do to realistically achieve your goal? What goal would you consider realistically achievable?

Expand full comment

I find all comparisons of ASI to nuclear weapons fallacious. To say "The advent of ASI will be like giving everyone a nuclear bomb," is a proposition I will only grant if you grant the further proposition, "The advent of ASI will be like giving everyone a super nuclear bomb detector and defuser." I find the fact that xrisk mongers never think past the first proposition to the second to be curious, and hard to explain without a deeper theory about their motivations, such as confirmation bias or power seeking.

So in the case of your thought experiment, I would have to modify it from "nuclear weapons" to "nuclear energy" and discuss both the good and potentially destructive aspects of its use, and also describe the equilibrium likely to be formed when a huge number of actors with diverse interests all have this energy, and most of them still hold self-preservation as a cardinal value.

Expand full comment

Lol, AI defuses nothing about AI. Make a good deepfake detector first. Also extinction has no recovery.

Expand full comment

Would you be able to demonstate a small scale example of it? If they are guaranteed to believe you, it seems that nuclear winter alone will discourage in its direction.

Expand full comment

> We could move quickly down the cost curve, and enable rapid deployment. In theory yes, but I don’t think the timelines work for that?

It would be delightful on so many levels if AI turned out to be the Starlink of energy abundance, but I agree, this doesn't seem like enough of a demand increase (even just within large-scale / industrial demand) to make that kind of step-change difference.

Expand full comment

Even three (Microsoft, Google and Amazon) massive tech companies taking a serious direct interest in nuclear power construction is a seismic level event for this industry. All they need to do is for their army of lawyers and lobbyists to fundamentally change the landscape and create the paper trail precedent.

If just one of them can also establish a blue print for sensible construction and almost anyone else can copy it, that would be the icing on the cake.

Expand full comment

> I’m curious if anyone does have a half-decent version [of a relationship/NSFW chatbot] - or kind of what that would even look like, right now?

Because the intersection of "LLMs good enough to hold a conversation" and "LLMs whose owners are fine with using them as a publicly accessible relationship chatbot service" is basically zero, the state of the art is mostly found in imageboard posts right now, and involves using the OAI and Anthropic APIs directly via a few different confusing open-source UIs, and lots of arguing over the best "jailbreak" prompt template for each model or how best to write character descriptions.

Quality tends to vary wildly depending on how much effort has been put into a character description, but at their best, Claude 3.5 and GPT-4o girlfriend/boyfriend bots are...not terrible? They can maintain a consistent personality and opinions, they're not *immediately* repetitive, and they're certainly not dumb (unless you intentionally make them so). But there is the same tendency for them to be more agreeable and willing to go along with the user than the character description ought to imply, and it's very easy for Claude in particular to get locked into repetitive loops over the course of a long chat.

Expand full comment
author

I'm surprised they're not using Llama instead? Shouldn't ability to fine tune and not having to jailbreak around the restrictions, plus lower price, make up for the lower quality?

(Which you'd think would be the quality floor, but very obviously isn't.)

Expand full comment
4 hrs ago·edited 4 hrs ago

Some people certainly do use Llama for the reasons you describe. I can't answer as to the quality outcomes there, but for GPT and Claude, fine-tuning seems unnecessary and jailbreaking is much easier than you might expect - easily shareable, relatively token-light prompt templates are sufficient for whatever behavior you want. As far as price goes, so far the cost over time of using top-end OAI and Anthropic models via my own API keys has been cheaper than a Netflix subscription (though I'm a very slow writer).

Expand full comment

Ziv. I appreciate the amount of work you put into your posts. Unfortunately, I do not have the time to read the entire articles and was hoping to use an AI to summarize the major points. Unfortunately, substack links don't seem to be accessible and I could not copy and paste the text. Any ideas?

Expand full comment
author

Literally Ctrl+A on a computer, Ctrl+V to dump it all into Google Docs, maybe delete the junk at the top if you care, should something like 90-10 it if that's what you want?

Expand full comment

Ziv. You are generous to reply. I'm very familiar with the computer key shortcuts, but unfortunately I read all of substack on a phone where the shortcuts don't work. As someone who is retired and likes to read broadly, I appreciate your In-Depth expertise. I suspect that there are others out there like me who would appreciate an abridged version of your posts. Maybe a skilled acolyte of yours could produce such a product! In any case, thank you again for all your insights.

Expand full comment

Two quick observations and one longer one:

1. I am missing a lot of context so maybe I have misunderstood a lot/everything, but is not OpenAI hiring the former CISO of Palantir the best (AI) news this week, month and maybe quarter? Again, context and lack thereof but if you had asked me to come up with organisations for which infosec was treated as an existential priority, Palantir would be on that list. If you asked me to come up with a list of people who solely by virtue of their role I would expect to be extremely resistant to intimidation and bullying in relation to their professional responsibilities, I do not think I would have thought of "CISO of Palantir" but in retrospect that certainly seems to fit the bill?

Of course, if he leaves in the next year or two, that would be even worse news than this is good news.

2. Eigengender goes all out on self-confidence, Roon makes a valid point that implies the real point which Eigengender seemed invested in not getting: at this stage I would not be comfortable backing my "thirty seconds of mild scrutiny" against Elon's ability to make it happen, did you think you would actually see in your lifetime a rocket booster coming gently down to earth to be caught by a robot gantry???

3. It is ironic that one of Daron Acemoglu's canonic examples is highly suspect: whilst he identifies former colonies with "non-extractive institutions" quite interestingly and carefully, and finds strong results even stripping out the obvious outliers (the Commonwealth), it is hard not to think that the result is even more heavily confounded with genetics than diet studies.

That idiosyncratic aside apart, the more subtle point from his work is that he postulates that extractive institutions survive despite obvious outcompetition because they are explicitly motivated to remain extractive to preserve factional power, because "businessmen" are an obvious potential competing source of power (see eg Latin America, but also Europe, China, America under Lina Khan, .... ).

Not hard to see how this could play out in the context of AI !!

Expand full comment