Discussion about this post

User's avatar
Sam Tobin-Hochstadt's avatar

There is another important connection between LLMs and management: the use of LLMs is itself a management task. The way I often describe them to people is that it's as if you have 100 green junior employees working for you all of a sudden. What would you do? How would you make use of their significant capabilities without wasting your own time. This is fundamentally a management question more than anything else.

Expand full comment
Andy Berner's avatar

When I read Henry's post this morning, my literal first thought was "I wonder if Dan has seen this yet, it seems right up his alley..."

I think I largely agree with his thoughts on the value of llm summarization as a managerial technology (and the caveats about their limitations). I think one potential risk is about incentives on both the data generation side and the assumptions about the quality/recency of the data being summarized. For instance, at my current job, there was a lot of executive enthusiasm about making it possible for sales/customer support execs to "chat with our knowledge base/documentation" in order to better respond to customer questions. Even setting aside hallucination risks, my first reaction as one of the scientists responsible for core product features was "our shipping timeline has been under such pressure for years that the literal last priority for engineering/DS has been to maintain the wiki/knowledge base - good luck to anyone treating that as a resource to answer customer questions." This is just classic garbage-in-garbage-out.

The second risk seems like the degree to which LLMs can remove friction in *generating* new "information." If it's easier to summarize large volumes of text, there could be greater perceived value/importance placed on producing large volumes of text. But if employees are producing these reports by just starting with five bullet points and then asking ChatGPT or equiv to expand them into a two (or three, or four?) page memo, (1) this is a hugely inefficient representation, as the only real information was in the five bullet points (unless we assume the LLM is going to cleverly synthesize some linkages between these points based on its compressed representation of a ton of other data), and in the game of expanding and then later summarizing, we've ultimately added a bunch of noise to the signal. Given the state of the internet these days, it seems like the arms race of "algorithmic slop generation" vs "algorithmic slop detection" generally favors the generator, not the discriminator, so it's not clear to me that we will end up with a clearer picture of the state of an organization when this technology proliferates without very clear standards/practices and incentive alignment on what generates the input data at lower levels of a complex system including lots of people with incentives to be lazy.

Expand full comment
11 more comments...

No posts