Coping with the flood of new papers

We’ve just released on GitHub an improved version of the RSS feed searcher powering our group’s “Daily Papers” page. With so many literature tools around, like Google Scholar or various AI-powered platforms, you might think researchers are already covered. But there’s still a big gap when it comes to following everything newly published in your research area.

These days, scientific literature grows so rapidly that crucial papers easily slip through unnoticed, even with all the tools at our disposal. Decades ago, it was manageable to track all new articles from arXiv’s cond-mat feed or selected journals, but today that’s nearly impossible without spending every waking hour reading. But, just because it’s hard doesn’t mean we can’t drink from the firehose. You just have to turn the flood into a trickle… or a managable stream of new papers.

For the past five years or so, our group tackled this by filtering RSS feeds using the strength and versatility of regular expressions to pick out relevant articles. Now, we’ve upgraded this lightweight Python tool and made it available on GitHub.

The tool fetches articles daily, filters them based on your specific interests (easily configured in a JSON file using regex patterns), and provides neat HTML summaries. Furthemore, it now integrates OpenAI’s language models API to automatically summarize the selected papers, saving even more of your time. If you’ve never used regular expressions before, ask your friendly LLM chatbot to cook one up for you based on your keywords.

Check it out here:
👉 paper firehose, technical details in the readme on Github

en_USEnglish