Micro Web Crawler in PHP & Manticore
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ghost 905af165fb draft webui 1 year ago
src draft webui 1 year ago
.gitignore ignore storage folder 1 year ago
LICENSE Initial commit 1 year ago
README.md draft webui 1 year ago
composer.json draft webui 1 year ago

README.md

Yo! Micro Web Crawler in PHP & Manticore

Next generation of YGGo! project with goal to reduce server requirements and make deployment process simpler

  • Index model changed to the distributed cluster model, and oriented to aggregate search results from different instances trough API
  • Refactored data exchange model with drop all primary keys dependencies
  • Snaps now using tar.gz compression to reduce storage requirements and still supporting remote mirrors, FTP including
  • Minimalism everywhere

Implementation

Engine written in PHP and uses Manticore on backend.

Default build inspired and adapted for Yggdrasil but could be used to make internet search portal.

Components

  • CLI tools for index operations
  • JS-less frontend to make search web portal
  • API tools to make search index distributed

Features

  • MIME-based crawler with flexible filter settings
  • Page snap history with local and remote mirrors support

Install

  1. Install composer, php and manticore
  2. Grab latest Yo version git clone https://github.com/YGGverse/Yo.git
  3. Run composer update inside the project directory
  4. Check src/config.json for any customizations
  5. Make sure storage folder writable
  6. Run indexes init script php src/cli/index/init.php
  7. Add new URL php src/cli/document/add.php URL
  8. Run crawler php src/cli/document/crawl.php
  9. Get search results php src/cli/document/search.php '*'

Web UI

  1. cd src/webui
  2. php -S 127.0.0.1:8080
  3. now open 127.0.0.1:8080 in your browser!

Documentation

CLI

Index

Init

Create initial index

php src/cli/index/init.php [reset]
  • reset - optional, reset existing index

Document

Add
php src/cli/document/add.php URL
  • URL - add new URL to the crawl queue
Crawl
php src/cli/document/crawl.php
php src/cli/document/search.php '@title "*"' [limit]
  • query - required
  • limit - optional search results limit