Skip to content Skip to footer
Hard problems that reduce to document ranking by noperator

Hard problems that reduce to document ranking by noperator

10 Comments

  • Post Author
    noperator
    Posted February 25, 2025 at 5:37 pm

    A concept that I've been thinking about a lot lately: transforming complex problems into document ranking problems to make them easier to solve. LLMs can assist greatly here, as I demonstrated at inaugural DistrictCon this past weekend.

  • Post Author
    obblekk
    Posted February 25, 2025 at 5:56 pm

    The open source ranking library is really interesting. It's using a type of merge sort where the comparator function is an llm comparing (but doing batches >2 for fewer calls).

    Reducing problems to document ranking is effectively a type of test-time search – also very interesting!

    I wonder if this approach could be combined with GRPO to create more efficient chain of thought search…

    https://github.com/BishopFox/raink?tab=readme-ov-file#descri…

  • Post Author
    westurner
    Posted February 25, 2025 at 6:31 pm

    Ranking (information retrieval) https://en.wikipedia.org/wiki/Ranking_(information_retrieval…

    awesome-generative-information-retrieval > Re-ranking: https://github.com/gabriben/awesome-generative-information-r…

  • Post Author
    rfurmani
    Posted February 25, 2025 at 6:36 pm

    Very cool! This is also one of my beliefs in building tools for research, that if you can solve the problem of predicting and ranking the top references for a given idea, then you've learned to understand a lot about problem solving and decomposing problems into their ingredients. I've been pleasantly surprised by how well LLMs can rank relevance, compared to supervised training of a relevancy score. I'll read the linked paper (shameless plug, here it is on my research tools site: https://sugaku.net/oa/W4401043313/)

  • Post Author
    Everdred2dx
    Posted February 25, 2025 at 6:42 pm

    Very interesting application of LLMs. Thanks for sharing!

  • Post Author
    m3kw9
    Posted February 25, 2025 at 7:08 pm

    That title hurts my head to read

  • Post Author
    mskar
    Posted February 25, 2025 at 7:56 pm

    Great article, I’ve had similar findings! LLM based “document-chunk” ranking is a core feature of PaperQA2 (https://github.com/Future-House/paper-qa) and part of why it works so well for scientific Q&A compared to traditional embedding-ranking based RAG systems.

  • Post Author
    hexator
    Posted February 25, 2025 at 8:03 pm

    This furthers an idea I've had recently that we (and the media) are focusing too much on creating value by making more ever more complex LLMs, and instead we are vastly underestimating creative applications of current generation AI.

  • Post Author
    moralestapia
    Posted February 25, 2025 at 8:17 pm

    Minor nitpick,

    Should be "document ranking reduces to these hard problems",

    I never knew why the convention was like that, it seems backwards to me as well, but that's how it is.

  • Post Author
    adamkhakhar
    Posted February 25, 2025 at 8:21 pm

    I'm curious – why is LLM ranking preferred over cosine similarity from an embedding model (in the context of this specific problem)?

Leave a comment

In the Shadows of Innovation”

© 2025 HackTech.info. All Rights Reserved.

Sign Up to Our Newsletter

Be the first to know the latest updates

Whoops, you're not connected to Mailchimp. You need to enter a valid Mailchimp API key.