I am new to the concept of web crawling, so sorry if my question is stupid. <br />
I want to know if there is a way to redirect the content received from websites crawled with PHPCrawl to a mysql database for later usage. <br />Any help would be appreciated. Thanks.
If you would like to refer to this comment somewhere else in this project, copy and paste the following link:
Anonymous
Anonymous
-
2020-11-13
Post awaiting moderation.
If you would like to refer to this comment somewhere else in this project, copy and paste the following link:
of course, you can do with found sites and all corresponding data whatever you want.
Please take a look at the example-script in the package (or here: http://phpcrawl.cuab.de/example.html).
From line 14 (inside the handlePageData-function), you can access all data delivered by the crawler.
So, if you want to put the content of a found site to a db-table, simply do a database-query inside the handlePageData-function that puts $page_data into a db-table.
Hi,
I am new to the concept of web crawling, so sorry if my question is stupid. <br />
I want to know if there is a way to redirect the content received from websites crawled with PHPCrawl to a mysql database for later usage. <br />Any help would be appreciated. Thanks.
Hi radutoev,
of course, you can do with found sites and all corresponding data whatever you want.
Please take a look at the example-script in the package (or here: http://phpcrawl.cuab.de/example.html).
From line 14 (inside the handlePageData-function), you can access all data delivered by the crawler.
So, if you want to put the content of a found site to a db-table, simply do a database-query inside the handlePageData-function that puts $page_data into a db-table.
For all other available information about the found/crawled site inside the handlePageData-function, just take a look here: http://phpcrawl.cuab.de/classreference.html#handlepagedata
Best regards!