Skip to content

Anish-addweb/zabbix-webcrawler

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

zabbix-webcrawler

A webcrawler designed to be used with zabbix to check the health of a website

Work in progress

It's early days still; functional in my setup but treat as experimental.

Getting started

  • Download
  • Setup the database ( mysql < schema.sql)
  • mv/cp config.php.example config.php
  • Test with:
    • php crawl.php "your url" scan //Returns how many pages were scanned(run multiple times)
    • php crawl.php "your url" status.links //Will tell you how many urls have been found
    • php crawl.php "your url" status.links //Will tell you home many links have been found
    • php crawl.php "your url" status.deadurls //Will tell you how many urls have returned a status other than 200
    • php crawl.php "your url" status.deadlinks //Will tell you how many links have return a status other than 200

Configure in zabbix

  • Setup zabbix agent - add the following to zabbix_agent.conf or in a new zabbix_agent.d/.conf file.
  • UserParameter=spider[*],/usr/bin/php /crawl.php $1 $2
  • Import the webcrawler template into zabbix (template.xml)
  • Add template to hosts - the template expects the host name to be a domain name. In the future this may change to use a URL macro.

##Setup options You can install this on any* server that has the zabbix_agent installed - just configure the agent interface to point to this server

##Config options TBD

About

A webcrawler designed to be used with zabbix to check the health of a website

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • PHP 100.0%