- Introduction
- My love google dork
- Search filters
- Operators
- Ideas
- Links
- Example
- Preventing GOOGLE DORKS
⭐ Star us on GitHub — it motivates a lot! ⭐
If you have any google dork, just create a PullRequest.
intext:"index of" "parent directory"
Filter | Description | Example |
---|---|---|
allintext | Searches for occurrences of all specified keywords. | allintext:"keyword" |
intext | Searches for the occurrence of keywords at once or consecutively. | intext:"keyword" |
intitle | Searches for occurrences of keywords in the title all or one. | intitle:"keyword" |
allintitle | Searches for all occurrences of keywords at once. | allintitle:"keyword" |
inurl | Searches for a URL that matches one of the keywords. | inurl:"keyword" |
allinurl | Searches for a URL that matches all the keywords in the query. | allinurl:"keyword" |
site | Searches specifically for that particular website and lists all results for that website. | site:"www.github.com" |
filetype | Searches for a specific file type named in the query. | filetype:"pdf" |
link | Searches for external links to pages. | link:"keyword" |
numrange | Used to find specific numbers in your search. | numrange:33-43 |
before/after | Used to search within a specified date range. | filetype:pdf & (before:2021-01-01 after:2021-05-01) |
allinanchor (and also inanchor) | This shows the websites that the keywords refer to in links, in order of most links. | inanchor:rat |
allinpostauthor (and also inpostauthor) | Exclusively for the blog search, blog posts written by specific people are picked out. | allinpostauthor:"keyword" |
related | List web pages that are "similar" to a given web page. | related:www.github.com |
cache | Displays the version of the web page that Google has in its cache. | cache:www.github.com |
This operator searches only for the exact term inside the quotation marks. You can use this for example if the term you are looking for is ambiguous and could easily be confused with something else, or if you don't get enough relevant results.
Here is an example:
"Admin Loginpage"
This operator searches for a specific search term OR another term.
site:instagram.com | site:github.com
This operator searches for a specific search term and another term.
site:github.com & site:twitter.com
This operator combines search terms
(site:instagram.com | site:twitter.com) (intext:"admin")
(site:instagram.com | site:twitter.com) & intext:"admin"
This will order results by the number of occurrence of the keyword.
site:twitter.com +site:twitter.*
site:twitter.* -site:twitter.com
site:*.site.com
site:*.*.site.com
site:*.*.*.site.com
~set
site:*.com
- Git google dorks
- phpmyadmin google dorks
- phpinfo google dorks
- log file google dorks
- google dorks for excel files
- Google Dorks for presentations
- best google dorks reports
- finding aws secrets with google dorks
- js secrets with google dorks
- CMS google dorks
- Wordpress
- Typo3
- Magento
- Joomla
- Drupal
- Shopify
- Admin google dorks
- Monitoring pages - google dorks
- Google Dorks - Github page
- exploit-db.com
- nvd.nist.gov
- cxsecurity.com
- vulnerability-lab.com
https://dewangpanchal98.medium.com/microsoft-bug-bounty-writeup-5ee4a7264dbf
Encoding/encrypting sensitive data such as usernames, passwords and so forth. Run inquiries against your own site to check whether you can locate any sensitive data. On the off chance that you discover sensitive information, you can remove it from search results by utilizing Google Search Console. Protect sensitive content by utilizing a robots.txt document situated in your root-level site catalog. Utilizing robots.txt helps prevent Google from indexing our site, but it can also show an attacker where sensitive data might be located. User-agent: * Disallow: /
You can also block specific directories to be excepted from web crawling. If you have the /phpinfo site and you need to protect it, just place this code inside:
User-agent: *
Disallow: /phpinfo/
Restrict access to specific files:
User-agent: *
Disallow: /member/info.html
Restrict access to dynamic URLs that contain ? symbol:
User-agent: *
Disallow: /*?
Needless to mention, please use this tool very very carefully. The authors won't be responsible for any consequences.