php - Prevent bots from scraping our content / overloading our server -


we planning offer job platform-service firms. have few thousand jobs offer our guest/visitors.

since yesterday noticed our server-load crazy , when checked logs saw had multiple site-request per second different ip addresses. order in pages called indicate same user / bot

we want available public if bots slowing our server massively down or forcing new hardware in trouble.

we displaying our job-content in iframes, encoder like: http://www.tareeinternet.com/scripts/iframe-encoder/

help solve our problem?

or options have? annoying since don't have user-sessions or recurring ip-addreses (i think using proxys switch regulary)

have checked headers recurring data? if they, example, have recurring user-agent can can block those:

  • apache:

setenvifnocase user-agent "^wget" bad_bot  setenvifnocase user-agent "^emailsiphon" bad_bot  setenvifnocase user-agent "^emailwolf" bad_bot  <directory "/var/www">          order allow,deny          allow          deny env=bad_bot  </directory>

  • code: can check each requests specific header , redirect somewhere else.

Comments

Popular posts from this blog

android - MPAndroidChart - How to add Annotations or images to the chart -

javascript - Add class to another page attribute using URL id - Jquery -

firefox - Where is 'webgl.osmesalib' parameter? -