概要
TryHackMe「Google Dorking」のWalkthroughです。
Task2
Q1.Name the key term of what a "Crawler" is used to do
A.Index
Q2.What is the name of the technique that "Search Engines" use to retrieve this information about websites?
A.Crawling
Q3.What is an example of the type of contents that could be gathered from a website?
A.Keywords
Task4
Q1.Where would "robots.txt" be located on the domain "ablog.com"
Hint.full path!
A.ablog.com/robots.txt
Q2.If a website was to have a sitemap, where would that be located?
A./sitemap.xml
Q3.How would we only allow "Bingbot" to index the website?
A.User-agent: Bingbot
Q4.How would we prevent a "Crawler" from indexing the directory "/dont-index-me/"?
A.Disallow: /dont-index-me/
Q5.What is the extension of a Unix/Linux system configuration file that we might want to hide from "Crawlers"?
Hint.system files are usually 3/4 characters!
A..conf
Task5
Q1.What is the typical file structure of a "Sitemap"?
A.XML
Q2.What real life example can "Sitemaps" be compared to?
A.Map
Q3.Name the keyword for the path taken for content on a website
A.Route
Task6
Q1.What would be the format used to query the site bbc.co.uk about flood defences
Hint.site:bbc.co.uk <query>
A.site: bbc.co.uk flood defences
Q2.What term would you use to search by file type?
A.filetype:
Q3.What term can we use to look for login pages?
Hint.term: query
A.intitle: login