jina-ai / reader

  • вторник, 16 апреля 2024 г. в 00:00:02

Convert any URL to an LLM-friendly input with a simple prefix


Your LLMs deserve better input.

Reader converts any URL to an LLM-friendly input with a simple prefix Get improved output for your agent and RAG systems at no cost.



Standard mode

Simply prepend to any URL. For example, to convert the URL to an LLM-friendly input, use the following URL:

Streaming mode

Use accept-header to control the streaming behavior:

Note, if you run this example below and not see streaming output but a single response, it means someone else has just run this within 5 min you and the result is cached already. Hence, the server simply returns the result instantly. Try with a different URL and you will see the streaming output.

curl -H "Accept: text/event-stream"

If your downstream LLM/agent system requires immediate content delivery or needs to process data in chunks to interleave the IO and LLM time, use Streaming Mode. This allows for quicker access and efficient handling of data:

Reader API:  streamContent1 ----> streamContent2 ----> streamContent3 ---> ... 
                          |                    |                     |
                          v                    |                     |
Your LLM:                 LLM(streamContent1)  |                     |
                                               v                     |
                                               LLM(streamContent2)   |

Stream mode is also useful when the target page is large to render. If you find standard mode gives you incomplete content, try streaming mode.

JSON mode

This is still very early and the result is not really a "useful" JSON. It contains three fields url, title and content only. Nonetheless, you can use accept-header to control the output format:

curl -H "Accept: application/json"


You will need the following tools to run the project:

  • Node v18 (The build fails for Node version >18)
  • Firebase CLI (npm install -g firebase-tools)

For backend, go to the backend/functions directory and install the npm dependencies.

git clone
cd backend/functions
npm install

What is thinapps-shared submodule?

You might notice a reference to thinapps-shared submodule, an internal package we use to share code across our products. While it’s not open-sourced and isn't integral to the Reader's functions, it mainly helps with decorators, logging, secrets management, etc. Feel free to ignore it for now.

That said, this is the single codebase behind, so everytime we commit here, we will deploy the new version to the

Having trouble on some websites?

Please raise an issue with the URL you are having trouble with. We will look into it and try to fix it.


Reader is backed by Jina AI and licensed under Apache-2.0.