12 pointsby tdaltonc10 hours ago6 comments
  • rnunery137 hours ago
    I agree with Elevaes, this was absolutely fascinating and I love the use of books to help understand the concepts. I could relate right away. The token usage reduction potential is massive especially when it comes to enterprise usage and costs - many companies are experiencing sticker shock because they weren't prepared / didn't anticipate the usage. The potential for better costing and estimation with the process could have widespread impacts to financials (in a good way) and allow for more accurate pricing estimates and models.
  • mmayberry9 hours ago
    If the agent builds entity pages incrementally while reading, how do you prevent early incorrect assumptions about relationships or attributes from propagating through the entity graph? Is there support for belief revision?
    • tdaltonc8 hours ago
      Yes this sort of auto-regressive error propagation is a real concern for the same reason it's a real concern with LLMs in general.

      If you force the output of an LLM to begin with an error, the LLM tends to continue down that erroneous path.

      In practice, we didn't see much of this kind of EP. A solution to this would be to give some agent the task of occasionally reviewing the NERDs for contradictions as well as the ability to search through the source material as needed. That of course creates the possibility of catastrophic forgetting, where the agent rewrites a NERD in an effort to remove a contraction and end's up deleting something important.

      We didn't see a lot of error propagation, but one example where we did: in Harry Potter, Prof Dumbledore is introduced as a mysterious hooded character. So the NERD-writer would create a NERD for "mysterious hooded man." There's no tool for the agent to change the title of a NERD, so the system is stuck with that title now. Sometimes the system would build the entire Dumbledore entry under "mysterious hooded man"; sometimes it would make a new Dumbledore entity and like a reference back to the "mysterious hooded man" entity, and sometimes it wouldn't link them. None of those outcomes are great.

  • elevaes10 hours ago
    This is fascinating, I'm wondering if it works as well with other use cases like papers, conversations, or any other human written text.
    • tdaltonc9 hours ago
      We originally developed NERDs inside of my last startup for monitoring the progress of solar developments. There are many different multi-modal event feeds that you need to monitor for a wholistic view of the project. NERDs helped glue together the event around entities.

      Only later did we adapted to the technique to work to long books. The existing long book benchmarks seemed like the most appropriate way to show the core idea to a wider audience.

      So ya, I'm confident that this central idea can be applied in many different domains.

  • droongta056 hours ago
    [dead]
  • noliver9 hours ago
    [dead]
  • aawinters9 hours ago
    [dead]