Robots, spiders, and crawlers are different terms that refer to a certain program that search engines run. This program essentially builds a summary of the content on a website by crawling it, enabling Google and other search engines to index it. As a result, when a person performs a search using a specific keyword, that website has the chance to appear in the results.
In order for a website to be crawled by Googlebots and other search engine spiders, each website must have a robots.txt file enabled in the HTML. Otherwise, the bots will be unable to crawl and subsequently index any page.
The large number of websites out there makes direct text-only matching a rare instance. Instead, search engines use complex algorithms to rank potential matches for certain keywords. Bots will also only crawl text on a website, which means that you need to have supplementary keyword-infused HTML for all images and videos if you want them to rank for certain terms. For example, images will require alt text and keyword-rich file names that indicate what is in the image, while you should do the same for video titles and descriptions prior to uploading them to YouTube and other video sites.