From: Karsten K. <kru...@my...> - 2012-10-25 19:14:29
|
Hello everbody, my new project wikijournals is now online and localized (thanx to "translate" extension). You can find the project at http://wikijournals.info . Wikijournals is a crowdsourcing plattform for articles (from journals and newspapers). The aim is to create a central place for searching newspaper and journals articles. At the moment semantic metadata for articles (f.i. title, author, page, journal, persons, locations) can be collect here. There also exist a facebook page for this project (see https://www.facebook.com/wikijournals) and a blog ( http://wikijournals.blogspot.de/) . This is an opensource project. You can find the source on github at http://swertschak.github.com/wikijournals-server/. I develop this project in my sparetime. It will be great, if anybody has comments, critics, questions or ideas for the project. best regards from Berlin Karsten |
From: Ed S. <Ed....@sa...> - 2012-10-25 19:52:04
|
Here is something you might want to consider: You could create an automated process to harvest news articles (most are now available via RSS), and try to extract the relevant information. Then you could push this into the wiki automatically through the API. I've done a lot of work with the API and automated extraction. Edward Swing Applied Research Technologist Vision Systems + Technology, Inc. 6021 University Boulevard . Suite 360 . Ellicott City . Maryland . 21043 Tel: 410.418.5555 Ext: 919 . Fax: 410.418.8580 Email: Ed....@vs... Web: http://www.vsticorp.com -----Original Message----- From: Karsten Krumrück [mailto:kru...@my...] Sent: Thursday, October 25, 2012 2:47 PM To: Semantic MediaWiki users Subject: [Semediawiki-user] A crowdsourcing platform for articles based an SMW Hello everbody, my new project wikijournals is now online and localized (thanx to "translate" extension). You can find the project at http://wikijournals.info . Wikijournals is a crowdsourcing plattform for articles (from journals and newspapers). The aim is to create a central place for searching newspaper and journals articles. At the moment semantic metadata for articles (f.i. title, author, page, journal, persons, locations) can be collect here. There also exist a facebook page for this project (see https://www.facebook.com/wikijournals) and a blog ( http://wikijournals.blogspot.de/) . This is an opensource project. You can find the source on github at http://swertschak.github.com/wikijournals-server/. I develop this project in my sparetime. It will be great, if anybody has comments, critics, questions or ideas for the project. best regards from Berlin Karsten ------------------------------------------------------------------------------ Everyone hates slow websites. So do we. Make your web apps faster with AppDynamics Download AppDynamics Lite for free today: http://p.sf.net/sfu/appdyn_sfd2d_oct _______________________________________________ Semediawiki-user mailing list Sem...@li... https://lists.sourceforge.net/lists/listinfo/semediawiki-user |
From: emijrp <em...@gm...> - 2012-10-25 23:00:39
|
2012/10/25 Ed Swing <Ed....@sa...> > Here is something you might want to consider: > You could create an automated process to harvest news articles (most are > now available via RSS), and try to extract the relevant information. Then > you could push this into the wiki automatically through the API. > I've done a lot of work with the API and automated extraction. > > Can you provide more info about the process? It may be useful for my wiki. Thanks. > Edward Swing > Applied Research Technologist > Vision Systems + Technology, Inc. > 6021 University Boulevard . Suite 360 . Ellicott City . Maryland . 21043 > Tel: 410.418.5555 Ext: 919 . Fax: 410.418.8580 > Email: Ed....@vs... > Web: http://www.vsticorp.com > > -----Original Message----- > From: Karsten Krumrück [mailto:kru...@my...] > Sent: Thursday, October 25, 2012 2:47 PM > To: Semantic MediaWiki users > Subject: [Semediawiki-user] A crowdsourcing platform for articles based an > SMW > > Hello everbody, > > my new project wikijournals is now online and localized (thanx to > "translate" extension). You can find the project at > http://wikijournals.info . > > Wikijournals is a crowdsourcing plattform for articles (from journals and > newspapers). The aim is to create a central place for searching newspaper > and journals articles. At the moment semantic metadata for articles (f.i. > title, author, page, journal, persons, locations) can be collect here. > > There also exist a facebook page for this project (see > https://www.facebook.com/wikijournals) and a blog ( > http://wikijournals.blogspot.de/) . > > This is an opensource project. You can find the source on github at > http://swertschak.github.com/wikijournals-server/. > > I develop this project in my sparetime. It will be great, if anybody has > comments, critics, questions or ideas for the project. > > best regards from Berlin > Karsten > > ------------------------------------------------------------------------------ > Everyone hates slow websites. So do we. > Make your web apps faster with AppDynamics Download AppDynamics Lite for > free today: > http://p.sf.net/sfu/appdyn_sfd2d_oct > _______________________________________________ > Semediawiki-user mailing list > Sem...@li... > https://lists.sourceforge.net/lists/listinfo/semediawiki-user > > > > > ------------------------------------------------------------------------------ > Everyone hates slow websites. So do we. > Make your web apps faster with AppDynamics > Download AppDynamics Lite for free today: > http://p.sf.net/sfu/appdyn_sfd2d_oct > _______________________________________________ > Semediawiki-user mailing list > Sem...@li... > https://lists.sourceforge.net/lists/listinfo/semediawiki-user > -- Emilio J. Rodríguez-Posada. E-mail: emijrp AT gmail DOT com Pre-doctoral student at the University of Cádiz (Spain) Projects: AVBOT <http://code.google.com/p/avbot/> | StatMediaWiki<http://statmediawiki.forja.rediris.es> | WikiEvidens <http://code.google.com/p/wikievidens/> | WikiPapers<http://wikipapers.referata.com> | WikiTeam <http://code.google.com/p/wikiteam/> Personal website: https://sites.google.com/site/emijrp/ |
From: Michael T. <mic...@gm...> - 2012-10-26 04:06:54
|
On Fri, Oct 26, 2012 at 4:51 AM, Ed Swing <Ed....@sa...> wrote: > Here is something you might want to consider: > You could create an automated process to harvest news articles (most are now available via RSS), and try to extract the relevant information. Then you could push this into the wiki automatically through the API. But isn't the "crowdsourcing" aspect of this project motivated precisely by the occasional failures of such automated processes, and the need for human intervention? (Even Google News can't do it right sometimes, and they have lots of smart programmers to throw at the problem.) It seems to me you'd want to "push to the wiki" only if it were first pushed to some workspace where the human-mediated workflow process can kick in -- to inspect, and, where necessary, refine the results of automation attempts. I've played around a little with Mechanical Turk, trying some jobs. Some of the work seems to be related to patching up bad metadata extractions from news sites and blogs. Anyway, I'd like to make use of some such framework myself, to more efficiently process the Google News notifications I keep running on various keyword combinations, and as a component when I start using SMW for my wiki. I wouldn't mind hearing more about this. Regards, Michael Turner Project Persephone 1-25-33 Takadanobaba Shinjuku-ku Tokyo 169-0075 (+81) 90-5203-8682 tu...@pr... http://www.projectpersephone.org/ "Love does not consist in gazing at each other, but in looking outward together in the same direction." -- Antoine de Saint-Exupéry > -----Original Message----- > From: Karsten Krumrück [mailto:kru...@my...] > Sent: Thursday, October 25, 2012 2:47 PM > To: Semantic MediaWiki users > Subject: [Semediawiki-user] A crowdsourcing platform for articles based an SMW > > Hello everbody, > > my new project wikijournals is now online and localized (thanx to "translate" extension). You can find the project at http://wikijournals.info . > > Wikijournals is a crowdsourcing plattform for articles (from journals and newspapers). The aim is to create a central place for searching newspaper and journals articles. At the moment semantic metadata for articles (f.i. > title, author, page, journal, persons, locations) can be collect here. > > There also exist a facebook page for this project (see > https://www.facebook.com/wikijournals) and a blog ( > http://wikijournals.blogspot.de/) . > > This is an opensource project. You can find the source on github at http://swertschak.github.com/wikijournals-server/. > > I develop this project in my sparetime. It will be great, if anybody has comments, critics, questions or ideas for the project. > > best regards from Berlin > Karsten > ------------------------------------------------------------------------------ > Everyone hates slow websites. So do we. > Make your web apps faster with AppDynamics Download AppDynamics Lite for free today: > http://p.sf.net/sfu/appdyn_sfd2d_oct > _______________________________________________ > Semediawiki-user mailing list > Sem...@li... > https://lists.sourceforge.net/lists/listinfo/semediawiki-user > > > > ------------------------------------------------------------------------------ > Everyone hates slow websites. So do we. > Make your web apps faster with AppDynamics > Download AppDynamics Lite for free today: > http://p.sf.net/sfu/appdyn_sfd2d_oct > _______________________________________________ > Semediawiki-user mailing list > Sem...@li... > https://lists.sourceforge.net/lists/listinfo/semediawiki-user -- Regards, Michael Turner Project Persephone 1-25-33 Takadanobaba Shinjuku-ku Tokyo 169-0075 (+81) 90-5203-8682 tu...@pr... http://www.projectpersephone.org/ "Love does not consist in gazing at each other, but in looking outward together in the same direction." -- Antoine de Saint-Exupéry |
From: Karsten K. <kru...@my...> - 2012-10-26 06:21:03
|
Thanks a lot for all interesting mails. At first, yes if an article exist online and he is available in a common format (f.i. RSS) it should be possible to retrieve the articles automatically by using frameworks. And then the metadata of the article can be evaluate and save into the wiki (for the last step I use pywikipedia). But there are also many articles which exist only offline, in print format. For such articles in my opinion it´s only possible to evaluate the semantic metadata manually, for example with a crowdsourcing platform. And on the other hand many articles which exist online don´t support a common export format like RSS. So for all these exceptions a special parser should be developed. Best regards Karsten Krumrück 2012/10/26 Michael Turner <mic...@gm...> > On Fri, Oct 26, 2012 at 4:51 AM, Ed Swing <Ed....@sa...> wrote: > > Here is something you might want to consider: > > You could create an automated process to harvest news articles (most are > now available via RSS), and try to extract the relevant information. Then > you could push this into the wiki automatically through the API. > > But isn't the "crowdsourcing" aspect of this project motivated > precisely by the occasional failures of such automated processes, and > the need for human intervention? (Even Google News can't do it right > sometimes, and they have lots of smart programmers to throw at the > problem.) It seems to me you'd want to "push to the wiki" only if it > were first pushed to some workspace where the human-mediated workflow > process can kick in -- to inspect, and, where necessary, refine the > results of automation attempts. > > I've played around a little with Mechanical Turk, trying some jobs. > Some of the work seems to be related to patching up bad metadata > extractions from news sites and blogs. > > Anyway, I'd like to make use of some such framework myself, to more > efficiently process the Google News notifications I keep running on > various keyword combinations, and as a component when I start using > SMW for my wiki. I wouldn't mind hearing more about this. > > Regards, > Michael Turner > Project Persephone > 1-25-33 Takadanobaba > Shinjuku-ku Tokyo 169-0075 > (+81) 90-5203-8682 > tu...@pr... > http://www.projectpersephone.org/ > > "Love does not consist in gazing at each other, but in looking outward > together in the same direction." -- Antoine de Saint-Exupéry > > > > -----Original Message----- > > From: Karsten Krumrück [mailto:kru...@my...] > > Sent: Thursday, October 25, 2012 2:47 PM > > To: Semantic MediaWiki users > > Subject: [Semediawiki-user] A crowdsourcing platform for articles based > an SMW > > > > Hello everbody, > > > > my new project wikijournals is now online and localized (thanx to > "translate" extension). You can find the project at > http://wikijournals.info . > > > > Wikijournals is a crowdsourcing plattform for articles (from journals > and newspapers). The aim is to create a central place for searching > newspaper and journals articles. At the moment semantic metadata for > articles (f.i. > > title, author, page, journal, persons, locations) can be collect here. > > > > There also exist a facebook page for this project (see > > https://www.facebook.com/wikijournals) and a blog ( > > http://wikijournals.blogspot.de/) . > > > > This is an opensource project. You can find the source on github at > http://swertschak.github.com/wikijournals-server/. > > > > I develop this project in my sparetime. It will be great, if anybody has > comments, critics, questions or ideas for the project. > > > > best regards from Berlin > > Karsten > > > ------------------------------------------------------------------------------ > > Everyone hates slow websites. So do we. > > Make your web apps faster with AppDynamics Download AppDynamics Lite for > free today: > > http://p.sf.net/sfu/appdyn_sfd2d_oct > > _______________________________________________ > > Semediawiki-user mailing list > > Sem...@li... > > https://lists.sourceforge.net/lists/listinfo/semediawiki-user > > > > > > > > > ------------------------------------------------------------------------------ > > Everyone hates slow websites. So do we. > > Make your web apps faster with AppDynamics > > Download AppDynamics Lite for free today: > > http://p.sf.net/sfu/appdyn_sfd2d_oct > > _______________________________________________ > > Semediawiki-user mailing list > > Sem...@li... > > https://lists.sourceforge.net/lists/listinfo/semediawiki-user > > > > -- > Regards, > Michael Turner > Project Persephone > 1-25-33 Takadanobaba > Shinjuku-ku Tokyo 169-0075 > (+81) 90-5203-8682 > tu...@pr... > http://www.projectpersephone.org/ > > "Love does not consist in gazing at each other, but in looking outward > together in the same direction." -- Antoine de Saint-Exupéry > > > ------------------------------------------------------------------------------ > Everyone hates slow websites. So do we. > Make your web apps faster with AppDynamics > Download AppDynamics Lite for free today: > http://p.sf.net/sfu/appdyn_sfd2d_oct > _______________________________________________ > Semediawiki-user mailing list > Sem...@li... > https://lists.sourceforge.net/lists/listinfo/semediawiki-user > |