L
L
likejavascript2013-07-24 09:03:46
Search engines
likejavascript, 2013-07-24 09:03:46

How to prevent search engines from indexing private content?

I remember in July 2012 there was a boom in the discussion of the issue of content by search engines, which, in theory, should not have been issued. Even in those cases when the private sections of the site were protected by user authorization.
There are some sections on my site that I would not like to see in the search results. Actually, the subject is how to properly configure directories, robots.txt, and possibly something else to prevent the user's private content from getting into the public domain.
Authorization is used on my site, only after that the user can go to his section and see his content

Answer the question

In order to leave comments, you need to log in

7 answer(s)
J
jj_killer, 2013-07-24
@jj_killer

robots.txt doesn't always work (why is an open question), but the tag meta content="noindex,nofollow" name="robots"hasn't failed yet.

R
Ruslan_Y, 2013-07-24
@Ruslan_Y

You can, in addition to robots.txt (for example, generate it here ), wrap the content with tags (see link ):<noindex>...</noindex>

R
Ruslan Fatkhutdinov, 2013-07-25
@Phelix

In robots.txt
User-agent: *
Disallow: /section_root
Well, for my own peace of mind, also in head in all personal sections:
<meta name=“robots” content=“none”>
none - replaces noindex and nofollwo

A
AxisPod, 2013-07-24
@AxisPod

Well, here google-bot even scores on basic authorization, it feels like it leaks passwords and indexes closed sites with absolutely impunity, absolutely hammering on robots.txt. After that, it’s not even clear at all how to defend yourself, apparently it’s stupid not to use chrome, just kick it off on the user agent so that it is not possible to merge the basic authorization password.

L
likejavascript, 2013-07-24
@likejavascript

Well, I do not have basic authorization and I really like chrome.

L
likejavascript, 2013-07-24
@likejavascript

https won't help?

I
Ilya Sevostyanov, 2013-07-24
@RUVATA

well, first you need to understand how the robot gets to pages that are private to you ... if, as it is supposed here, Google Chrome merges passwords “by itself” :) then you should still have an authorized session “merged” in the logs, if not, then you have PROBLEMS, because the robot goes where it was not called just like that :).
Well, cut it off your shoulder, in general, the robot is quite easy to detect, do not let it eat what is not supposed to - there will be nothing to index.

Didn't find what you were looking for?

Ask your question

Ask a Question

731 491 924 answers to any question