Save time reading long form essay/article/blog-post.
- Given URL of an article, extracts assertions and insights into quick digestable form.
- Summarize a news report into an Axios-style report.
- Extract insights from a lecture or conversation in YouTube.
- Uses Google Gemini to summarize.
Most books should be blog posts; most blog posts should be tweets.
Sometimes (just sometimes) the essay you're reading is better off being distilled into core arguments and assertions than being wrapped around either attractive languages or badly written prose.
This is a tool to extract these points so you can decide whether to dive into the original material or not.
Most LLMs can now do this competently enough. This tool uses Google Gemini 1.5 Flash to do it, but ideologically it's agnostic about which LLM to use.
The system prompt being used here is inspired by a pattern from Fabric, which is responsible for producing the structure of the outcome.
- Get API key
- Create a new file
.env
, addGEMINI_API_KEY=<your-api-key>
docker compose --profile dev run -P --rm web bash
python app.py
- Deploy server: `docker compose --profile prod up -d**
There are a few ways to use.
Open http://<server>:4063
in web browser for a simple UI. Or go direct...
To extract insights for an article or a YouTube video.
http://<server>:4063/summarize?url=<url>
To summarize into an Axios-style news report
http://<server>:4063/news?url=<url>