1. Besides web scraping and large language models, do I need any other technologies?
2. Besides writing a dedicated scraper for each website, is there a simpler technology available?
3. I plan to have every piece of information analyzed by a large language model to see if it meets my criteria, but it feels like it consumes a lot of tokens. Is there a more cost-effective way?
4. Is my technical approach completely wrong? Is there an easier way?