SeNioR- Posted November 21, 2015 Share Posted November 21, 2015 Google indexes and also doubles URL by adding completely unnecessary ?do=email so i advise you to turn off in ACP "share by email" You can also create robots.txt file and put in: User-agent: * Disallow: /admin/ Disallow: /*?do=email Disallow: /*&do=email Disallow: /*?do=email&comment= Disallow: /*&do=email&comment= Disallow: /*&do=findComment&comment= Disallow: /*?do=findComment&comment= Disallow: /*?do=reportComment&comment= Disallow: /*&do=reportComment&comment= Sitemap: http://urltoyoursite.com/sitemap.php Link to comment Share on other sites More sharing options...
Morrigan Posted November 21, 2015 Share Posted November 21, 2015 You can do this yourself in your share settings. Settings > Sharing > Click on Enabled and it will turn to Disabled Viola. Email sharing is now turned off for your community. Link to comment Share on other sites More sharing options...
SeNioR- Posted November 21, 2015 Author Share Posted November 21, 2015 54 minutes ago, Morrigan said: You can do this yourself in your share settings. Settings > Sharing > Click on Enabled and it will turn to Disabled Viola. Email sharing is now turned off for your community. I know this, i wanted to tell other users Link to comment Share on other sites More sharing options...
Day_ Posted November 22, 2015 Share Posted November 22, 2015 You don't need to use a robots.txt either do you? System > Search Engine Optmisation > Meta Tags Add the URL's robots noindex That's what I've been doing anyhow Link to comment Share on other sites More sharing options...
h-y-b-r-i-d Posted November 22, 2015 Share Posted November 22, 2015 Two things wrong here. 1- robots.txt will stop Google crawling the URLs, it will still show them in the index. So use no index on them instead. 2- the ips meta tag system afaik doesn't work on pages with query strings, it applies the changes to the non query page too, so use x-robots htaccess to add the tag to those pages instead. I go one step further and noindex,nofollow all pages with query strings. They are all junk. Link to comment Share on other sites More sharing options...
h-y-b-r-i-d Posted November 22, 2015 Share Posted November 22, 2015 The other thing, these urls only resolve for logged in users. Bots get a header response of 403 forbidden. The reason they are being indexed is you are using a robots file and google cant see the header response. So in this case, just remove them from robots, google will crawl them, get a 403 and eventually drop them. Seems you created your own problem here. Link to comment Share on other sites More sharing options...
Recommended Posts
Archived
This topic is now archived and is closed to further replies.