I was in the same situation — using rss-parser, but eventually faced with some issues.
It also turned out, its performance is not that good. In benchmarks, it's 4-5x slower. In my case, switching to Feedsmith almost doubled the overall parsing speed. This is including the fetching of feeds, which is the main bottleneck.
Thanks! For now I have some benchmarks for parsing, as this has been my main focus regarding performance. It consistently ranks in the top 2 with the caveat that other libs do not support most of the feed namespaces that Feedsmith does.
It also turned out, its performance is not that good. In benchmarks, it's 4-5x slower. In my case, switching to Feedsmith almost doubled the overall parsing speed. This is including the fetching of feeds, which is the main bottleneck.
PS. Great projects, I know and follow both. :)
Now somebody might also find interesting what I have done.
- I decided that implementing RSS reader for 100x time is really stupid, so naturally I wrote my own [0]
- my RSS reader is in form of API [1], which I use for crawling
- can be installed via docker. User has to only parse JSON via API. No need to use requests, browsers, status codes
- my weapon of choice is python. There is python feedparser package, but I had problems in using in parallel, because some XML shenanigans, errors
- my reader, serves crawling purpose, so I am interested in most basic elements, like thumbnails, so all nuance from RSS is lost
- detects feeds from sites automatically
Links
[0] https://github.com/rumca-js/crawler-buddy/blob/main/src/webt...
[1] https://github.com/rumca-js/crawler-buddy
Here are the results: https://github.com/macieklamberski/feedsmith/blob/main/bench....
This is quite an interesting idea — benchmarking feed parsing libraries in different languages. I'll give it some thought.