what exactly I want to do is:
**input**: wikipedia xml dump
**output**: a list of triples like this:
<http://dbpedia.org/resource/Lists_of_computer_languages> <http://dbpedia.org/ontology/wikiListOf> <http://dbpedia.org/resource/C_(programming_language)> .
<http://dbpedia.org/resource/Lists_of_computer_languages> <http://dbpedia.org/ontology/wikiListOf> <http://dbpedia.org/resource/Java_(programming_language)> .
...
..
.
<http://dbpedia.org/resource/List_of_XML_markup_languages> <http://dbpedia.org/ontology/wikiListOf> <http://dbpedia.org/resource/AdsML> .
<http://dbpedia.org/resource/List_of_XML_markup_languages> <http://dbpedia.org/ontology/wikiListOf> <http://dbpedia.org/resource/Agricultural_Ontology_Service> .
We have already set up and customised dbpedia extraction framework but I think it would be difficult to configure the framework for extracting this data. I was shocked by the fact that extraction framework does not have any extractors for this !
moved from Bugs to Feature Requests
related to http://stackoverflow.com/questions/4885025/how-can-i-extract-data-of-list-pages-from-wikipedia
We would have to extend TableMapping. Major effort, probably no time for that. Sorry.
But it's open source, so if you find a way to do this, please submit a patch!
Feel free to ask for help on dbpedia-discussion@lists.sourceforge.net