Discussion about this post

User's avatar
Dave Friedman's avatar

On the Hoel piece: he seems to be doing this thing where, having identified something that a LLM can't (yet?) do well, he concludes that it is overhyped and useless. (Gary Marcus does something similar.) It's an annoying rhetorical sleight of hand; the correct way to understand LLMs is with Ethan Mollick's 'jagged frontier' metaphor. LLMs are good at certain tasks and not good at others, and it's our job as humans to figure that out.

Expand full comment
Yaakov Saxon's avatar

A somewhat pedantic point, but I think worth understanding for exactly these situations, is that LLMs see neither letters nor whole words but something somewhat in between called tokens, which can represent anything from individual letters to words but mostly represent common groups of letters / word fragments. Best explanation I’ve seen is here, though slightly dated:

https://gwern.net/gpt-3#bpes

Expand full comment
12 more comments...

No posts