D
D
Den Belzarayn2016-07-20 17:21:29
PHP
Den Belzarayn, 2016-07-20 17:21:29

Duplicate pages due to comments, how to remove?

A site without an engine, made in html and php. Such a problem, there are comments under the articles, after each comment, duplicate pages are added to the Google search index. For example, the page address is site.com/cagegory/page, a comment is added and the page address changes to site.com/cagegory/page ?id=9243247632144.91 Duplicates
are created by analogy with wordpress, only in this case the ?id identifier.
I have already talked on the forums, implemented the following recommendations and rules, but duplicates are still being created. After editing, the duplicates disappeared from Yandex and Bing, but Google does not listen to any.
I added the following lines to the robots.txt file:
Disallow: /*?id=*
Disallow: /*?id=
Disallow: /*?*
Disallow: /*?
In .htaccess I wrote:
RewriteEngine On
RewriteCond %{QUERY_STRING} ^id=
RewriteRule (.+) /$1? [R=301,L]
Google in webmaster's google panel automatically created URL parameters for id , where these addresses were not allowed to be crawled. The problem is not solved, what to do? I can’t put the canonical meta tag on duplicates, since they don’t physically exist because there is no engine. Comments generate duplicates.

Answer the question

In order to leave comments, you need to log in

5 answer(s)
I
Ilya Beloborodov, 2016-07-20
@kowap

if they disappeared from Yandex and Bing, then they will disappear from Google, but it’s not known when

A
Adamos, 2016-07-20
@Adamos

Make a sitemap - automatically, specifying the omission of such keys in the script settings.
And feed it to the search engines.
Well, the fact that the link to the same page goes through the get-parameter, and not through # - is sloppy...

D
Den Belzarayn, 2016-07-20
@RockerOne

Another such nuance, the old ones do not disappear, and new comments create duplicates. It turns out my solutions do not help solve the problem.

O
olee444, 2016-07-22
@olee444

https://support.google.com/webmasters/topic/458910...
Right column

M
Maxim Artemiev, 2016-07-25
@dan-pro

For some reason, everyone thinks that it will fall out by itself.
And you take it, collect all these duplicate pages from the issuance (parsing by site:), then push everything into the sitemap and send the sitemap for indexing through the "Recrawl" pages tool. Yandex will quickly pick up a sitemap, in which there will be only 301 redirects, and it will throw everything out more quickly.
For some reason, everyone is afraid of this method o_O

Didn't find what you were looking for?

Ask your question

Ask a Question

731 491 924 answers to any question