19 Comments
Mar 30, 2023Liked by Dileep George

So good! I learned so much though now will spend rest of my night going down a rabbit hole on dirigibles :)

Question though - isn't there a distinction in that, in scaling LLMs, you get emergent effects (like induction heads etc). Does the analogy hold?

Expand full comment
Mar 30, 2023Liked by Dileep George

Spot on, Dileep! Couldn't agree more!

Expand full comment

great article, I am very fond of your analysis and insights regarding AI and especially from brain based AI approaches. We don't know how the GPT_N++ will work out in a decade etc but it is my intuition that these systems will not create the value we think they would, we will need humans to shape them consistently and constantly, guide them and create narrow guidelines. Of course when no reliability is needed they will look magical and general AI's but for large time (t) they will show their problems and inabilities.

Expand full comment
Apr 4, 2023Liked by Dileep George

Thank you for the welcome!

Great read with insightful information that enlightens the mind. Recently I tried my hand at some tech writing, AI is coming and I thought I’d put together a rounded article—would love your eyes and insights 😊 https://tumbleweedwords.substack.com/p/ai-enters-our-everyday-reflections

Expand full comment
Apr 3, 2023Liked by Dileep George

Good article. Found one typo:

1919: First non-stop transatlantic airplane fight. ->

1919: First non-stop transatlantic airplane flight.

Expand full comment
Apr 2, 2023Liked by Dileep George

Thank you Dileep for bringing out a very different aspect which people outside the pure AI world did not realise...looking forward to the next blog

Expand full comment
Apr 2, 2023Liked by Dileep George

Ok, should we try again to make a paper plane that actually flies?

If there are yet to be found principles/ideas that might work on a much smaller scale (and later scaled up), investigating that area shouldn't be too expensive.

Expand full comment

During the dirigible days, were there people who said, "No one will find the principles of aerodynamics" like some people today (me for example!) who say, "We will never find the principles of intelligence"? - all we can hope to do is "dirigibles"!

It is a bunch (huge bunch) of small things put together that acquires an emergent property that we call "intelligence" . . .?

Expand full comment

>there are also avenues to combine the strengths of current approaches with an investigation into future architectures.

Yes, hybrids should be powerful. Especially when new architectures can read from the intermediate layers of LLMs (and, perhaps, write to them).

This is where this really beautiful analogy with LLMs as dirigibles stops working. A hybrid between an airplane and a dirigible is not a promising idea. A hybrid between a shiny new architecture and an LLM might be what one needs to progress from excelling in "toy problems" to convincingly beating stand-alone LLMs across the board.

Expand full comment

I agree that scaling up LLM will hit a dead end. But people are going way beyond that. They are using LLM as language interface between tools, and those tools know what they are doing individually. They are reinforcing LLM with human feedback, and a framework that can be taught by human means rather than explicit coding is a huge deal. People are adding verification to LLM, so you could tell it, that a portion of its output is wrong, so it needs to be redone. It is too early to early to say that it won't grow.

Expand full comment

Looking forward to this line of work!!!

Expand full comment

Great analogy and helpful perspective for evaluating and extrapolating the performance of these exciting large ML models, compared to what we would really like to have.

Expand full comment