php - Prevent bots from scraping our content / overloading our server -
we planning offer job platform-service firms. have few thousand jobs offer our guest/visitors.
since yesterday noticed our server-load crazy , when checked logs saw had multiple site-request per second different ip addresses. order in pages called indicate same user / bot
we want available public if bots slowing our server massively down or forcing new hardware in trouble.
we displaying our job-content in iframes, encoder like: http://www.tareeinternet.com/scripts/iframe-encoder/
help solve our problem?
or options have? annoying since don't have user-sessions or recurring ip-addreses (i think using proxys switch regulary)
have checked headers recurring data? if they, example, have recurring user-agent can can block those:
- apache:
setenvifnocase user-agent "^wget" bad_bot setenvifnocase user-agent "^emailsiphon" bad_bot setenvifnocase user-agent "^emailwolf" bad_bot <directory "/var/www"> order allow,deny allow deny env=bad_bot </directory>
- code: can check each requests specific header , redirect somewhere else.
Comments
Post a Comment