A simple web crawler in C# that recursively explores links from a webpage!
- π Parse Links: Starts by parsing all links from a given URL.
- π Recursive Crawling: Visits each parsed link and extracts further links until the maximum limit is reached.
- π Link Extraction: Uses regular expressions to extract URLs from the page content.
- π‘ Duplicate Protection: Maintains a
HashSet
of visited URLs to avoid revisiting and prevent infinite loops. - π― Customizable: Set the maximum number of URLs to crawl.
- Start Crawling: Specify the starting URL.
- Extract Links: The program fetches the content of the page and extracts all links.
- Recursive Visits: It recursively visits those links, repeating the process.
- Stop Condition: Crawling continues until the defined maximum number of URLs is reached.