Discussion about this post

User's avatar
M Flood's avatar
5dEdited

Inside Apple:

Cast:

Tim (Cook, CEO)

John (Giannandrea, Senior Vice President of Machine Learning & AI Strategy)

Scene: Tim Cook's office, mid-morning

Tim: John, my man, how's it going?"

John: It's a'ight, Timbo, a'ight

Tim: How's the new Siri coming along?

John: ... (total silence)

Tim: Fuuuccckkkkk...

John: Look, dude, it's like ... it's tough out there okay.

Tim: You mean we got nothin'?

John: Not nothin' ... how do you feel about super duper autocorrect from on device 3 B model?

Tim: ... (total silence)

John: Or like get this ... create your own emoji - Genmoji, get it - with diffusion. Like Midjourney, but not as good.

Tim (forehead on desk): Fuck, fuck, fuck ...

John: My dude, don't give up ... we've got this study

Tim (sitting up again): What study?

John: Well, like, it's not like OpenAI or DeepMind-like research but ...

Tim: But what?

John: So Samy (Bengio) and the guys got to thinking ... like, what if we, like, asked LLMs to do things that even a human couldn't do, like a 50,000 step puzzle? And then the LLMs failed?

Tim: And they failed?

John: Yeah. If you don't let them write programs like a human would to solve an equivalent problem ...

Tim: Skip that part. Here's the headline "LLMs can't reason, they only pretend to reason." Get that out today.

Expand full comment
Alex's avatar

This is an interesting example of how illegible the academic publishing world can be to outsiders. As someone who reads a lot of papers, there are some signals that jump out to me, but probably don't seem significant to a typical journalist or "AI fan" on X:

- First, "this paper is from Apple" sends a different signal about credibility to an ML researcher than it does to the general public.

- The lead author is a grad student intern. There are some more senior researchers in the author list, but based on the "equal contribution" asterisk, we are probably looking at an intern project by a grad student researcher and their intern host. This is a signal about the amount of investigative effort and resources behind the paper.

- The meta in ML publishing is to make maximalist claims about the impact and significance of your work. Conference reviewers will typically make you tone it down or qualify your claims, but there is still quite a lot of it in the final papers (which are mostly irrelevant because everyone read your preprint anyway). Everyone has adjusted to expect this and just gloss over those parts when reading papers. This is a preprint, so it hasn't even had an external review process.

If you flip through preprints, most of the big companies put out dozens of papers like this every month - speculative results by small teams, which don't necessarily align with the "beliefs" or goals of the larger company. I think that's mostly a good thing - a good research culture requires making a lot of small bets, most of which won't pan out. But it can be a PR headache when "a grad student intern at Apple posted a preprint about how LLMs behave on puzzles that require long context" is perceived as "Apple says reasoning models don't really work".

Expand full comment
15 more comments...

No posts