Menu

the crawler stops

Help
Anonymous
2015-08-03
2015-08-29
  • Anonymous

    Anonymous - 2015-08-03

    Hello,
    I installed the crawler today. It stops always at different pages and never runs to the end (summary). I don't get any error code. My changes in example.php are:
    $crawler->setTrafficLimit(2000000 * 1024);
    $crawler->setStreamTimeout(5);
    $crawler->setConnectionTimeout(10);
    $crawler->setPageLimit(2000);

    What can I do to find out, why it stops? TY

     
  • Anonymous

    Anonymous - 2015-08-03

    Hi,

    Please try this:

    $report = $crawler->getProcessReport();
    if ($report->abort_reason == 1){
    echo "Abort reason: ABORTREASON_PASSEDTHROUGH";
    }else{
    echo "Abort reason: ".$report->abort_reason;
    }

    This http://phpcrawl.cuab.de/classreferences/PHPCrawlerProcessReport/property_detail_tpl_property_abort_reason.htm gives more insight in why the crawler stopped.

     
  • Anonymous

    Anonymous - 2015-08-04

    Hi,

    tried that, but nothing was displayed. I suppose, it stops before reaching that instruction. :-/

     
  • Anonymous

    Anonymous - 2015-08-06

    What is the time limit of PHP?
    I set this in the first lines of my script:

    // It may take a whils to crawl a site ...
    set_time_limit(150000);

    Normally PHP will stop the script after 30 seconds I guess...

     
  • Anonymous

    Anonymous - 2015-08-06

    I did set it like this:
    set_time_limit(2000000);

    It should be enough for about 700 pages?

     
  • Uwe Hunfeld

    Uwe Hunfeld - 2015-08-29

    Hi!

    Sorry for my late answer!

    DId you get it to work meanwhile?
    Or dooes the problem still exist?

     

Anonymous
Anonymous

Add attachments
Cancel





Want the latest updates on software, tech news, and AI?
Get latest updates about software, tech news, and AI from SourceForge directly in your inbox once a month.