dethfire
Well-known member
According to Google my Robots.txt is blocking 958,419 pages and indexed 876,771 pages. Below is my robots file. To me that seems crazy hi. Anyone have ideas on how so many pages are blocked? Does it seem reasonable?
User-agent: Mediapartners-Google
Disallow:
User-agent: *
Disallow: /find-new/
Disallow: /account/
Disallow: /goto/
Disallow: /posts/
Disallow: /login/
Disallow: /admin.php
Sitemap: https://www.physicsforums.com/sitemap.php
User-agent: Mediapartners-Google
Disallow:
User-agent: *
Disallow: /find-new/
Disallow: /account/
Disallow: /goto/
Disallow: /posts/
Disallow: /login/
Disallow: /admin.php
Sitemap: https://www.physicsforums.com/sitemap.php