S
S
Semisonic2012-04-26 18:32:11
Copyright
Semisonic, 2012-04-26 18:32:11

What methods can be used to protect data published on the site from theft?

Suppose I have some kind of online resource like a reference book. This resource encourages (including financially) its users to enter certain data, which it then shows to all other users, having some profit on advertising and other means of monetization.

Again, suppose one day a competing online resource comes along trying to do the same thing. But instead of collecting data on his own, he prefers to steal it completely from my resource, thereby saving on the procedure for collecting and validating data from users.

The question is: are there any effective ways to combat this kind of actions today? Interested in both technical and legal methods.

Thank you.

Answer the question

In order to leave comments, you need to log in

6 answer(s)
F
furyk, 2012-04-26
@furyk

Optimize your site for search engines in advance so that your competitor does not rise above you in the search results.
If there are those who wish, then they will be able to copy the information in any case, no matter how you resist.
So it's better to give up on all these methods and concentrate on making your own site interesting and updated frequently.
Then even if you lose some visitors due to the copied site, there will be nothing to worry about.

W
werdender, 2012-04-26
@werdender

There are no technical methods.
Legally incompetent.

D
dasm32, 2012-04-26
@dasm32

I really did not observe effective technical methods other than watermarks.
And so - try to take any top article from here and google its exact phrases;)

V
Vampiro, 2012-04-26
@Vampiro

Do not offer disconnection from the Internet? (99% way, by the way)
The pictures are written across the image.
With the text, you can test options for replacing Cyrillic with Latin, but this is if you want to run around the courts later. Although in Russia this is still difficult.

Z
Zibx, 2012-04-26
@Zibx

Add information as before, but start displaying it at the moment the search bot enters. In any case, then the content will be considered original. We have a lot of search engines, so you need to choose one priority.

5
5ap, 2012-04-27
@5ap

it is technically difficult to parse large volumes in a short time, at least an unprepared resource can be filled up. And vryatli users request pages more often than N times.
An analysis of requests for a resource is built, distributions by geo, useragent are adjusted (in particular, OS and browser separately), then mathematical expectations are built, by the way, it’s good to make them adaptive with the contribution of certain risks to these very adaptations.
Then protection with a page of the lock with the text, as it was often in contact - by the type of page you often ask.
Lock algorithms can be read in the direction of ddos. Only DDoS is blocked more roughly, and in case of site parsing detection, finer tuning is needed.
Public parsers understand that it is impossible to parse everything at once and parse a bit. You can also parse access logs and make lookups of IPs and their owners, add public crawlers to whitelists.
On the detection of stolen content, you can read in the direction of steganography. Or it is easier to use services to search for unique content.
In general, you can invent a lot, the main thing is to understand that the price of hacking should be more expensive than the implementation of your own and you should not strive for the ideal. So, for example, if you make $1 money with $0.9 effort, then it’s profitable to fake it, but if you make it for $10, then it’s real and you can’t distinguish it from the original, but the expediency is lost.

Didn't find what you were looking for?

Ask your question

Ask a Question

731 491 924 answers to any question