Hi @jahedi
I did try the option
max_crawl_depth: 1
on the config file but it didn't work properly.
In theory the max_crawl_depth: 1
should be enough to do this. Can you describe what happened? How many docs did it end up ingesting?
One possibility is the Crawler can also use the sitemap.xml
for content discovery. Do you have this file in your website? If so, you can set sitemap_discovery_disabled: true
in the config to ignore it.
If that isn't the case, this is likely a bug.