You can subscribe to this list here.
2010 |
Jan
|
Feb
|
Mar
|
Apr
|
May
|
Jun
(1) |
Jul
(8) |
Aug
(7) |
Sep
(1) |
Oct
(8) |
Nov
(1) |
Dec
(13) |
---|---|---|---|---|---|---|---|---|---|---|---|---|
2011 |
Jan
(11) |
Feb
(1) |
Mar
(13) |
Apr
(27) |
May
(31) |
Jun
(8) |
Jul
(11) |
Aug
(12) |
Sep
(6) |
Oct
(11) |
Nov
(37) |
Dec
(6) |
2012 |
Jan
(13) |
Feb
(55) |
Mar
(78) |
Apr
(60) |
May
(199) |
Jun
(46) |
Jul
(33) |
Aug
(17) |
Sep
(23) |
Oct
(6) |
Nov
(18) |
Dec
(8) |
2013 |
Jan
(16) |
Feb
(69) |
Mar
(50) |
Apr
(48) |
May
(23) |
Jun
(66) |
Jul
(92) |
Aug
(48) |
Sep
(25) |
Oct
(20) |
Nov
(44) |
Dec
(52) |
2014 |
Jan
(9) |
Feb
(24) |
Mar
(33) |
Apr
(24) |
May
(20) |
Jun
(20) |
Jul
(19) |
Aug
(17) |
Sep
(16) |
Oct
(11) |
Nov
(3) |
Dec
(10) |
2015 |
Jan
(12) |
Feb
(14) |
Mar
(34) |
Apr
(16) |
May
(12) |
Jun
(14) |
Jul
(11) |
Aug
(12) |
Sep
(24) |
Oct
(17) |
Nov
(11) |
Dec
(14) |
2016 |
Jan
(27) |
Feb
(9) |
Mar
(23) |
Apr
(17) |
May
(25) |
Jun
(11) |
Jul
(17) |
Aug
(14) |
Sep
(12) |
Oct
(20) |
Nov
(18) |
Dec
(1) |
2017 |
Jan
(7) |
Feb
(1) |
Mar
(6) |
Apr
(1) |
May
(2) |
Jun
(2) |
Jul
(2) |
Aug
(1) |
Sep
(4) |
Oct
(1) |
Nov
|
Dec
(2) |
2018 |
Jan
(1) |
Feb
(1) |
Mar
(1) |
Apr
|
May
|
Jun
|
Jul
|
Aug
(3) |
Sep
(3) |
Oct
(3) |
Nov
|
Dec
|
2019 |
Jan
(1) |
Feb
|
Mar
|
Apr
(2) |
May
(6) |
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
|
2020 |
Jan
|
Feb
|
Mar
(2) |
Apr
(1) |
May
|
Jun
|
Jul
|
Aug
(1) |
Sep
|
Oct
|
Nov
|
Dec
|
2021 |
Jan
|
Feb
|
Mar
|
Apr
|
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
(1) |
Nov
|
Dec
|
2022 |
Jan
|
Feb
|
Mar
|
Apr
|
May
|
Jun
|
Jul
|
Aug
(1) |
Sep
|
Oct
|
Nov
|
Dec
|
From: Michael Z. <mic...@li...> - 2022-08-23 09:25:48
|
Dear colleagues, Following the request of various authors (who are still on vacations), we’ve extended the deadline of COGALEX by one week. The new deadline is: 02-Sept-2022 Hoping that this will allow you to submit your paper under better conditions. 'Bon courage' Michael Zock ============ Call for papers for COGALEX-VII The 7th International Workshop on Cognitive Aspects of the Lexicon https://sites.google.com/view/cogalexvii2022/home co-located with AACL-IJCNLP 2022 https://www.aacl2022.org/home Taipei, Taiwan Submission Deadline: 02-Sept-2022 Workshop date : Date: 24-Nov-2022 Key words: Dictionary, (Mental)Lexicon, Brain, Cognition, Neuroscience, Computational Linguistics; Corpus Linguistics, Complex graphs, Navigation Meeting Description: COGALEX is a workshop devoted to the cognitive aspects of the lexicon. While in the past, it has always been co-located with COLING, this time it will be hosted by AACL-IJCNLP 2022, at the NTUH International Convention Center, Taipei, Taiwan (https://www.aacl2022.org/home). The accepted papers will be published as proceedings appearing in the ACL anthology. The goal of COGALEX is to provide a snapshot of the current state of the art in the different disciplines —lexicography, psycholinguistics, neuroscience— dealing words, their organization (lexicon) and usage (for example: navigation in a hybrid conceptual-lexical resource; word production and analysis). The approach being deliberately cross-disciplinary. In sum, we solicit original and unpublished work related to the cognitive aspects of the lexicon. For details, see: https://sites.google.com/view/cogalexvii2022/home Short papers can be up to 4 pages in length and long papers up to 8 pages. Both submission formats can have an unlimited number of pages for references. All submissions must follow the ACL stylesheet. We don’t accept submissions that consist only of an abstract. The submissions must be anonymous and they will be peer-reviewed by our program committee. The peer review is double blinded. Papers must be submitted via SoftConf by September 02, 2022. Submission page: https://softconf.com/aacl2022/CogALex-VII/user/scmd.cgi?scmd=submitPaperCustom&pageid=0 At least one of the authors of an accepted paper must register for the main conference and present the paper. Accepted papers (short and long) will be published in the workshop proceedings that will appear in the ACL Anthology. Accepted papers will also be given an additional page to address the reviewers’ comments. The length of a camera-ready submission can then be 4 pages for a short paper and 8 for a long paper with an unlimited number of pages for references. We consider to invite the authors of the accepted papers to submit an extended version of their workshop paper for a special issue. Important dates * Paper submission (full and short): September 02, 2022 * Notification of acceptance: October 02, 2022 * Camera ready deadline: October 20, 2022 * Workshop: November 24, 2022 Workshop organizers : * Michael Zock (CNRS, LIS, Aix-Marseille University, Marseille, France) * Emmanuele Chersoni (The Hong Kong Polytechnic University, Hong Kong, China) * Yu-Yin Hsu (The Hong Kong Polytechnic University, Hong Kong, China) * Enrico Santus (Bayer, Whippany, NJ, 07981, USA) For specific requests or information Please send an e-mail to cog...@gm..., or to Michael Zock (mic...@li...), Emmanuele Chersoni (emm...@gm...) -- Michael ZOCK Emeritus Research Director CNRS LIS UMR 7020 (Group TALEP) Aix Marseille Université 163 avenue de Luminy - case 901 13288 Marseille / France Mail: mic...@li... <mailto:mic...@li...> Tel.: +33 (0)6 51.70.97.22 Secr.: +33 (0)4.86.09.04.60 http://pageperso.lif.univ-mrs.fr/~michael.zock/ <http://pageperso.lif.univ-mrs.fr/%7Emichael.zock/> |
From: Sebastian H. <hel...@in...> - 2021-10-29 14:28:22
|
Dear Martynas, please resubmit this to the issue tracker in the hosting template: https://github.com/dbpedia/extraction-framework/issues/new/choose Normally, the hosting category gets handled faster (days/weeks) than data issues (weeks/months). LG, Sebastian BlueMail for Android herunterladen Am 29. Okt. 2021, 13:06, um 13:06, "Martynas Jusevičius" <mar...@at...> schrieb: >No response to this? How is the server supposed to know that http:// >and https:// resources are the same? In RDF they are not. > >Another example, this time requesting https://: > >$ curl -L -OJ -H "Accept: text/turtle" >https://dbpedia.org/resource/Copenhagen >% Total % Received % Xferd Average Speed Time Time Time >Current > Dload Upload Total Spent Left Speed >0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- > 0 >100 153 100 153 0 0 725 0 --:--:-- --:--:-- >--:--:-- 725 >100 675k 100 675k 0 0 1139k 0 --:--:-- --:--:-- >--:--:-- 3235k >curl: Saved to filename 'sparql_2021-10-29_10-31-22Z.ttl' > >$ cat sparql_2021-10-29_10-31-22Z.ttl >@prefix dbo: <http://dbpedia.org/ontology/> . >@prefix dbr: <http://dbpedia.org/resource/> . >dbr:Vivi_Bach dbo:birthPlace dbr:Copenhagen . >... > >On Thu, Oct 7, 2021 at 11:20 AM Martynas Jusevičius ><mar...@at...> wrote: >> >> Hi, >> >> We have noticed a URI mismatch on DBPedia which breaks Linked Data >clients. >> >> The server forces https:// URLs: >> >> $ curl -I -H "Accept: text/turtle" >http://dbpedia.org/resource/Copenhagen >> HTTP/1.1 303 See Other >> Server: nginx/1.18.0 >> Date: Thu, 07 Oct 2021 09:11:29 GMT >> Content-Type: text/html >> Content-Length: 153 >> Connection: keep-alive >> Location: https://dbpedia.org/resource/Copenhagen >> Access-Control-Allow-Credentials: true >> Access-Control-Allow-Methods: HEAD, GET, POST, OPTIONS >> Access-Control-Allow-Headers: >> >Depth,DNT,X-CustomHeader,Keep-Alive,User-Agent,X-Requested-With,If-Modified-Since,Cache-Control,Content-Type,Accept-Encoding >> >> But the returned RDF data contains http:// URIs: >> >> $ curl -o - https://dbpedia.org/data/Copenhagen.ttl >> @prefix dbo: <http://dbpedia.org/ontology/> . >> @prefix dbr: <http://dbpedia.org/resource/> . >> ><http://dbpedia.org/resource/2011\u201312_West_Ham_United_F.C._season> >> dbo:wikiPageWikiLink dbr:Copenhagen . >> <http://dbpedia.org/resource/AEK_Athens_F.C._in_European_football> >> dbo:wikiPageWikiLink dbr:Copenhagen . >> dbr:Adform dbo:wikiPageWikiLink dbr:Copenhagen . >> dbr:Helena_Paparizou dbo:wikiPageWikiLink dbr:Copenhagen . >> dbr:MS_Jutlandia dbo:wikiPageWikiLink dbr:Copenhagen . >> ... >> >> Can you please align the URLs in the data so that the resources are >> actually self-describing? >> >> Thanks. >> >> Martynas >> atomgraph.com > > >_______________________________________________ >DBpedia-discussion mailing list >DBp...@li... >https://lists.sourceforge.net/lists/listinfo/dbpedia-discussion |
From: Michael Z. <mic...@li...> - 2020-08-04 22:47:16
|
*CogALex-VI*(*Cog*nitive *A*spects of the *Lex*icon) workshop co-located with COLING 2020 (12/12/2020) Paper submission *deadline*: September 1st, 2020 https://sites.google.com/view/cogalex-2020 *1 BACKGROUND * Supporting us in many tasks (thinking, searching, memorising and communicating) words are important. Hence, one may wonder how to build tools supporting its learning and usage (access/navigation). Alas the answer is not quite as straightforward as it may seem. It depends on various factors: the questioner's background (lexicography, psychology, computer science), the task (production/reception), and the material support (hardware). Words in books, computers and the human brain are not the same. Obviously, being aware of this, different communities have focused on different issues —(dictionary building; creation of navigational tools; representation and organisation of words; time course for accessing a word, etc.)— yet, their views and respective goals have changed considerably over time. Rather than considering the lexicon as a static entity, where discrete units (words) are organised alphabetically (database view), dictionaries are now viewed dynamically, i.e., as lexical graphs, whose entities are linked in various ways (topical relations; associations) and whose weight links may vary over time. While lexicographers view words as products (holistic entities), psychologists and neuroscientists view them as processes (decomposition), involving various steps or layers (representations) between an input and an output. Computational linguists have their own ways to look at words, and their proposals have also changed quite a bit during the last decade. Discrete count-based vector representations have successively been replaced by continuous vectors (i.e., word embeddings) and then by language-model-based contextualised representations. These latter are more powerful than any of the other forms, as they are able to account for context ambiguity, outperforming the static models (including word-embeddings) in a broad range of tasks. As one can see, different communities look at words from different angles, which can be an asset, as complementary views may help us to broaden and deepen our understanding of this fundamental cognitive resource. Yet, this diversity of perspectives can also a problem, in particular if the field is rapidly moving on, as in our case. Hence it becomes harder and harder for everyone, including experts, to remain fully informed about the latest changes (state of the art). This is one of the reasons why we organise this workshop. More precisely, our goal is not only to keep people informed without getting them crushed by the information glut, but also to help them to perceive clearly what is new, relevant, hence important. Last, but not least, we would like to connect people from different communities in the hope that this may help them to gain new insights or inspirations. * 2 SCOPE and TOPICS* This workshop is about possible enhancements of lexical resources (representation, organisation of the data, etc.). To allow for this we invite researchers to submit their contributions. The idea is to discuss the limitations of existing resources and to explore possible enhancements that take into account the users’ and the engineers' needs (computational aspects). Also, given the success of the shared task devoted to the corpus-based identification of semantic relations (CogALex-V., 2016), we propose anotheredition byadding this time a multilingual component. Our special focus will be on paradigmatic semantic relations, such as synonymy, antonymy and hypernymy, which are notoriously difficult to be distinguished by the classical word embedding models. For this workshop we solicit papers including but not limited to the following topics, each of which can be considered from various points of view: linguistics (lexicography, computational- or corpus linguistics), neuro- or psycholinguistics (tip-of-the-tongue problem, word associations), network-related sciences (vector-based approaches, graph theory, small-world problem), and so on. *Organization, i.e. structure of the lexicon * - Micro- and macrostructure of the lexicon; - Indexical categories (taxonomies, thesaurus-like topical structures, etc.); - Map of the lexicon (topology) and relations between words (word associations). *The meaning of words and how to reveal it* - Lexical representation (holistic, decomposed); - Meaning representation (concept based, primitives); - Distributional semantics (count models, neural embeddings, etc. ) ***Analysis of the conceptual input given by a dictionary user* - What information do language producers typically provide when looking for a word (terms, relations)? - What kind of relational information do they give: typed or untyped relations? - Which relations are typically used? *Methods for crafting dictionaries or indexes * - Manual, automatic or collaborative building of dictionaries and indexes (crowdsourcing, serious games, etc.); - Extraction of associations from corpora to build semantic networks supporting navigation; - (Semi-) automatic induction of the link type (e.g., synonym, hypernym, meronym, ...). ***Creation of new types of dictionaries** * - Concept dictionary; - Dictionary of larger segments than words (clauses, phrasal elements); - Dictionary of patterns or concept-patterns; - Dictionary of syllables. ***Dictionary access*(navigation and search strategies), interface issues - Search based on sound (rhymes), meaning or contextually related words (associations); - Determination of appropriate search space based on the user’s cognitive state (information available at the onset: query) and meta-knowledge (knowledge concerning the relationship between the input and the target word), ... - Identification of typical word access strategies (navigational patterns) used by people; - Interface problems, data visualisation. *3 WORKSHOP SUBMISSIONS* The workshop features two tracks: - A regular research track, where the submissions must be substantially original. For details, see: https://sites.google.com/view/cogalex-2020/home/submissions - A shared task track, with submissions consisting of system description papers. For details see : https://sites.google.com/view/cogalex-2020/home/shared-task https://sites.google.com/site/cogalexvisharedtask/ ** *4 IMPORTANT DATES* Workshop papers - Paper submission deadline: September 1, 2020 - Notification of acceptance: October 10, 2020 - Camera-ready papers due: October 25, 2020 Shared task - Release of development data : August 1st, 2020 - Release of test data : September 1st, 2020 - Announcement of winners : October 1st, 2020 - Shared task papers due: October20, 2020 ** *5 INVITED SPEAKER* Alex Arenas (http://deim.urv.cat/~alexandre.arenas/ <http://deim.urv.cat/%7Ealexandre.arenas/>) Alephsys Lab, Computer Science & Mathematics, Universidad Rovira i Virgili, 43007 Tarragona, Spain *6 WORKSHOP ORGANISERS* - Michael Zock (LIS, CNRS, AMU, Marseille, France) - Alessandro Lenci (Comput. Linguistics Laboratory, University of Pisa, Italy) - Enrico Santus (Bayer, Whippany, NJ, 07981, USA) - Emmanuele Chersoni (Hong Kong Polytechnic University, Hong Kong, China) *7 PROGRAM COMMITTEE* see : https://sites.google.com/view/cogalex-2020/home/programme-committee *8 CONTACTS* For general questions, please get in touch with Michael Zock (mic...@li... <mailto:mic...@li...>). Concerning the shared task, contact Rong Xiang (csr...@co... <mailto:csr...@co...>)or Emmanuele Chersoni (emm...@gm...) <mailto:emm...@gm...> |
From: Michael Z. <mic...@li...> - 2020-04-25 12:40:12
|
============================================================================ CogALex-VI, a Workshop on "Cognitive Aspects of the Lexicon" co-located with COLING 2020 Please pay attention to the newdeadlines, and apologies for cross-postings.============================================================================ Call for Papers ** ****C*ogALex* *Cognitive Aspectsof the Lexicon* *Workshop*co-located with *COLING* (28th International Conference on Computational Linguistics), Barcelona, Spain, *December 12, 2020 * Paper submission deadline: August 15, 2020 camera-ready version : October 20, 2020 For latest information always look here https://sites.google.com/view/cogalex-2020 *1 Background * Supporting us in many tasks (/thinking, searching/, /memorizing/ and /communicating)/ words are important. Hence, one may wonder how to build tools supporting their learning and usage (access/navigation). Alas the answer is not quite as straightforward as it may seem. It depends on various factors: the questioner's /background/ (lexicography, psychology, computer science), the /task /(production/reception), and the /material support/ (hardware). Words in books, computers and the human brain are not the same. Obviously, being aware of this, different communities have focused on different issues —(dictionary building; creation of navigational tools; representation and organization of words; time course for accessing a word, etc.)— yet, their views and respective goals have changed considerably over time. Rather than considering the lexicon as a static entity, where discrete units (words) are organized alphabetically (database view), dictionaries are now viewed dynamically, i.e., as lexical graphs, whose entities are linked in various ways (topical relations; associations) and whose weight links may vary over time. While lexicographers view words as products (holistic entities), psychologists and neuroscientists view them as processes (decomposition), involving various steps or layers (representations) between an input and an output. Computational linguists have their own ways to look at words, and their proposals have also changed quite a bit during the last decade./Discrete count-based vector representations /have successively been replaced by /continuous vectors/ (i.e., word embeddings) and then by /language-model-based contextualized representations/. These latter are more powerful than any of the other forms, as they are able to account for context ambiguity, outperforming the static models (including word-embeddings) in a broad range of tasks. As one can see, different communities look at words from different angles, which can be an asset, as complementary views may help us to broaden and deepen our understanding of this fundamental cognitive resource. Yet, this diversity of perspectives can also a problem, in particular if the field is rapidly moving on, as in our case. Hence it becomes harder and harder for everyone, including experts, to remain fully informed about the latest changes (state of the art). This is one of the reasons why we organize this workshop. More precisely, our goal is not only to keep people informed without getting them crushed by the information glut, but also to help them to perceive clearly what is new, relevant, hence important. Last, but not least, we would like to connect people from different communities in the hope that this may help them to gain new insights or inspirations. *2 Scope and Topics* This workshop is about possible /enhancements/ of /lexical resources/ (representation, organization of the data, etc.). To allow for this we invite researchers to submit their contributions. The idea is to discuss the limitations of existing resources and to explore possible enhancements that take into account the users’ and the engineers' needs (computational aspects). Also, just like in the past we propose again a 'shared task'. This time the goal is to provide a common benchmark for testing lexical representations for the automatic identification of lexical semantic relations (synonymy, antonymy, hypernymy, part-whole meronymy) in various languages (English, Chinese, and so on). For this workshop we solicit papers including but not limited to the following *topics*, each of which can be considered from various points of view: /linguistics /(lexicography, computational- or corpus linguistics), /neuro/- or /psycholinguistics/ (tip-of-the-tongue problem, word associations), /network/-related sciences (vector-based approaches, graph theory, small-world problem), and so on. 1 Organization, i.e. structure of the lexicon * Micro- and macrostructure of the lexicon; * Indexical categories (taxonomies, thesaurus-like topical structures, etc.); * Map of the lexicon (topology) and relations between words (word associations). 2 The meaning of words and how to reveal it * Lexical representation (holistic, decomposed); * Meaning representation (concept based, primitives); * Distributional semantics (count models, neural embeddings, etc. ) 3 Analysis of the conceptual input given by a dictionary user * What information do language producers typically provide when looking for a word (terms, relations)? * What kind of relational information do they give: typed or untyped relations? * Which relations are typically used? 4 Methods for crafting dictionaries or indexes * Manual, automatic or collaborative building of dictionaries and indexes (crowdsourcing, serious games, etc.); * Extraction of associations from corpora to build semantic networks supporting navigation; * (Semi-) automatic induction of the link type (e.g., synonym, hypernym, meronym, ...). 5 Creation of new types of dictionaries * Concept dictionary; * Dictionary of larger segments than words (clauses, phrasal elements); * Dictionary of patterns or concept-patterns; * Dictionary of syllables. 6 Dictionary access (navigation and search strategies), interface issues * Search based on sound (rhymes), meaning or contextually related words (associations); * Determination of appropriate search space based on the user’s cognitive state (information available at the onset: query) and meta-knowledge (knowledge concerning the relationship between the input and the target word), ... * Identification of typical word access strategies (navigational patterns) used by people; * Interface problems, data visualization. *3 Workshop Submissions* ** The workshop features two tracks: * A*regular research track*, where the submissions must be substantially original. * A *shared task track*, with submissions consisting of system description papers. The regular research track submissions should follow one of the 2 formats: * *Long papers*(9 content pages + references) should report on solid and finished research including new experimental results, resources and/or techniques. * *Short papers*(4 content pages + references) should report on small experiments, focused contributions, ongoing research, negative results and/or philosophical discussion. Submissions must be anonymized, conform to the style sheet of COLING (https://coling2020.org/pages/call_for_papers), and be submitted via their website (https://www.softconf.com/coling2020/CogALex/ <https://www.google.com/url?q=https%3A%2F%2Fwww.softconf.com%2Fcoling2020%2FCogALex%2F&sa=D&sntz=1&usg=AFQjCNEAUzHo0vS4_MVdi-nlx1qIxrCpUA>). While some papers may be accepted only as posters, in the proceedings no distinction will be made between them and full papers. *4 Important Dates* *Workshop papers* * Paper *submission deadline*: August15, 2020 * Notification of acceptance: September 20, 2020 * Camera-ready papers due: October20, 2020 * Workshop date: December 12, 2020** ** ***Shared task* * Release of development data : August1st, 2020 * Release of test data : September 1st, 2020 * Announcement of winners : October 1st, 2020 * Shared task papers due: October20, 2020 *5 Invited Speaker*** //*Alex Arenas*//(http://deim.urv.cat/~alexandre.arenas/ <http://deim.urv.cat/%7Ealexandre.arenas/>) Alephsys Lab, Computer Science & Mathematics, Universidad Rovira i Virgili, 43007 Tarragona, Spain *6 Workshop Organizers* * Michael Zock (LIS, CNRS, Aix-Marseille University, Marseille, France) * Alessandro Lenci (Comput. Linguistics Laboratory, University of Pisa, Italy) * Enrico Santus (MIT Computer Science & AI Lab, Boston, USA) * Emmanuele Chersoni (Hong Kong Polytechnic University, Hong Kong, China) * 7 Program Committee* see : https://sites.google.com/view/cogalex-2020/home/programme-committee * 8 Contacts* For *general questions*, please get in touch with /Michael Zock/ e-mail:mic...@li... <mailto:mic...@li...> Homepage: http://pageperso.lif.univ-mrs.fr/~michael.zock/ <http://pageperso.lif.univ-mrs.fr/%7Emichael.zock/> Concerning the*shared task*,**please contact* * Enrico Santus (es...@gm... <mailto:es...@gm...>), or Emmanuele Chersoni (emm...@gm... <mailto:emm...@gm...>) ---------------------------------------- Michael ZOCK Directeur de Recherche Émérite LIS UMR CNRS 7020 (Groupe TALEP) Aix Marseille Université 163 Avenue de Luminy - case 901 F-13288 Marseille/France Mail: mic...@li... Tel.: +33 (0)4.86.09.06.85 Secr.: +33 (0)4.86.09.04.60 +33 (0)4.86.09.06.75 http://pageperso.lif.univ-mrs.fr/~michael.zock/<http://pageperso.lif.univ-mrs.fr/%7Emichael.zock/> ------------------------------------------ |
From: Brooke S. <br...@in...> - 2020-03-19 19:49:28
|
No clue what you are trying to ask here. On Thu, Mar 19, 2020 at 1:05 PM Alberto Moya Loustaunau < alb...@gm...> wrote: > Hi, I'd like to rebuild some snapshots of DBpedia live. I found the > changesets here http://downloads.dbpedia.org/live/changesets, but I > don't know where to find a reference dump to start with. The URI > http://live.dbpedia.org/dumps/ is unavailable. > Also, I have a doubt with the meaning of the files XXX.clear.nt.gz and > XXX.reinserted.nt.gz. > > Thanks in advance, > Alberto > _______________________________________________ > DBpedia-developers mailing list > DBp...@li... > https://lists.sourceforge.net/lists/listinfo/dbpedia-developers > -- Brooke Spina, Co-Founder Mobile: +1 917 972 6787 8000 Avalon Blvd STE 200; Alpharetta, GA 30009 CONFIDENTIALITY NOTICE The content of this email is confidential and intended for the recipient specified in message only. If you have received this by mistake, please inform us by an email reply and then delete the message. It is forbidden to copy, forward, or in any way reveal the contents of this message to anyone. Views and opinions included in this email belong to their author and do not necessarily mirror the views and opinions of the company. The company will not take any liability for statements made by the author (defamatory clauses, infringe, or authorized infringement of any legal right.) Furthermore employees are fully responsible for the content of their emails. The integrity and security of this email cannot be guaranteed. Therefore, the sender will not be held liable for any damage caused by the message. |
From: Alberto M. L. <alb...@gm...> - 2020-03-19 17:05:44
|
Hi, I'd like to rebuild some snapshots of DBpedia live. I found the changesets here http://downloads.dbpedia.org/live/changesets, but I don't know where to find a reference dump to start with. The URI http://live.dbpedia.org/dumps/ is unavailable. Also, I have a doubt with the meaning of the files XXX.clear.nt.gz and XXX.reinserted.nt.gz. Thanks in advance, Alberto |
From: Sebastian H. <hel...@in...> - 2019-05-17 15:14:46
|
Dear Almir, strategic restructuring is progressing well. Before we were satisficing with nobody being happy now we are maximizing 1. business fitness and 2. community building separately. For Live we focused on reliability which will hopefully conclude next week after we do the productive install. Now the community option is to self-host a triple store based on the free changesets or use the public endpoint and we hope to get more endpoints up hosted by chapters, libraries, etc. On the business side we are going to offer this: # Standard DBpedia and DBpedia-Live mirror on AWS, maintained by OpenLink https://aws.amazon.com/marketplace/pp/B012DSCFEK (should be available again soon) # Contact point to establish a customer relationship important for customizations, support, more features, different access modalities. For now you can contact me, as I am setting up this structure at the moment. -- Sebastian On 17.05.19 15:54, Almir Vardo wrote: > Hi all, > > I am Almir Vardo and working as Data Engineer at Finit, currently > working for client from USA. I need to extract knowledge for up to > 10millions of entities from DBpedia live and point and it is really > slow. Also I will need to > sync that knowledge in the future. > > For that I need stable DBpedia endpoint and do anyone has option to > sell us access to the dbpedia synced endpoint or to sell synced > virtuoso database? > > Thanks, Almir > > > -- > Almir Vardo, Bsc of Electrical Engineering > tel: +387 62-263-253 > web: https://ba.linkedin.com/in/almir-vardo-84a1b986 > <https://almirvardo.wordpress.com> > e-mail: alm...@gm... <mailto:alm...@gm...> > > > _______________________________________________ > DBpedia-developers mailing list > DBp...@li... > https://lists.sourceforge.net/lists/listinfo/dbpedia-developers -- All the best, Sebastian Hellmann Director of Knowledge Integration and Linked Data Technologies (KILT) Competence Center at the Institute for Applied Informatics (InfAI) at Leipzig University Executive Director of the DBpedia Association Projects: http://dbpedia.org, http://nlp2rdf.org, http://linguistics.okfn.org, https://www.w3.org/community/ld4lt <http://www.w3.org/community/ld4lt> Homepage: http://aksw.org/SebastianHellmann Research Group: http://aksw.org |
From: Sebastian H. <hel...@in...> - 2019-05-17 13:56:55
|
Hi Martin, On 17.05.19 15:46, Martin Dominguez wrote: > Thanks a lot, Sebastian for your answer. > I have one remaining question, > for 2017 and 2019, there are note files? I don't understand this question. We are quite happy that we are finally able to stabilize and speed up regular mapping releases. So it will get better in the future. > is there any way to generate this, in > http://live.dbpedia.org/changesets/, there is an hourly changeset, > there is not any way to obtain a type and mapping files for the > desired date? I don't know your use case, but theoretically you could produce arbitrary time slices and snapshots of mappingbased/instance types from the changesets, right? Please post, if you are going to implement this. Note that we are doing a server wipe and fresh install of live.dbpedia.org next week, so the changesets will be unavailable for some days. -- Sebastian > > Cheers > Martín > > > El mar., 14 may. 2019 a las 14:46, Sebastian Hellmann > (<hel...@in... > <mailto:hel...@in...>>) escribió: > > Hi Martin, > > the data you arer looking for is here: > > https://databus.dbpedia.org/dbpedia/mappings > > specifically: > > https://databus.dbpedia.org/dbpedia/mappings/instance-types > > https://databus.dbpedia.org/dbpedia/mappings/mappingbased-objects-uncleaned > > the query for instance-types latest version on > https://databus.dbpedia.org/repo/sparql > > PREFIX dataid: <http://dataid.dbpedia.org/ns/core#> > <http://dataid.dbpedia.org/ns/core#> > PREFIX dataid-cv: <http://dataid.dbpedia.org/ns/cv#> > <http://dataid.dbpedia.org/ns/cv#> > PREFIX dct: <http://purl.org/dc/terms/> <http://purl.org/dc/terms/> > PREFIX dcat: <http://www.w3.org/ns/dcat#> <http://www.w3.org/ns/dcat#> > > # Get all files > SELECT DISTINCT ?file WHERE { > ?dataset dataid:artifact > <https://databus.dbpedia.org/dbpedia/mappings/instance-types> > <https://databus.dbpedia.org/dbpedia/mappings/instance-types> . > ?dataset dcat:distribution ?distribution . > ?dataset dct:hasVersion ?latestVersion . > { > SELECT (max(?version) as ?latestVersion) WHERE { > ?dataset dataid:artifact > <https://databus.dbpedia.org/dbpedia/mappings/instance-types> > <https://databus.dbpedia.org/dbpedia/mappings/instance-types> . > ?dataset dct:hasVersion ?version . > ?dataset dcat:distribution ?distribution . > ?distribution dataid:contentVariant > 'en'^^<http://www.w3.org/2001/XMLSchema#string> > <http://www.w3.org/2001/XMLSchema#string> . > ?distribution dataid:contentVariant > 'transitive'^^<http://www.w3.org/2001/XMLSchema#string> > <http://www.w3.org/2001/XMLSchema#string> . > } > } > ?distribution dataid:contentVariant > 'en'^^<http://www.w3.org/2001/XMLSchema#string> > <http://www.w3.org/2001/XMLSchema#string> . > ?distribution dataid:contentVariant > 'transitive'^^<http://www.w3.org/2001/XMLSchema#string> > <http://www.w3.org/2001/XMLSchema#string> . > ?distribution dcat:downloadURL ?file . > } > > > query can be built on the website, but note that these are > unreleased due to: > > - website is not working 100% > > - small encoding problem of the releases duer to RDF parser > > All the best, > > Sebastian > > > On 14.05.19 15:26, Martin Dominguez wrote: >> Hi all, >> I am Martín Domínguez, a researcher from Universidad de >> Córdoba, Argentina. We make researching by using DBpedia dumps, >> and the latest ones, in the format we use, are available for 2016 >> (https://wiki.dbpedia.org/develop/datasets/downloads-2016-10). >> We specifically use the "Instance Types" and "Mapping-based >> properties" files, which are available for 2016 and previous dates. >> Actually, we found DBpedia Live, and I found the last dumps and >> the changesets described in >> (https://wiki.dbpedia.org/online-access/DBpediaLive#4%20Important%20Pointers). >> After reading this documentation, it is possible to have a new >> Dump up to date, but, is it possible only extracting dumps for >> "Instance Types" and "Mapping-based properties"?. >> We also need to obtain those files not for the latest update, but >> also for a given previous date, it is possible? >> >> regards! >> Martín >> >> >> _______________________________________________ >> DBpedia-developers mailing list >> DBp...@li... <mailto:DBp...@li...> >> https://lists.sourceforge.net/lists/listinfo/dbpedia-developers > -- > All the best, > Sebastian Hellmann > > Director of Knowledge Integration and Linked Data Technologies > (KILT) Competence Center > at the Institute for Applied Informatics (InfAI) at Leipzig University > Executive Director of the DBpedia Association > Projects: http://dbpedia.org, http://nlp2rdf.org, > http://linguistics.okfn.org, https://www.w3.org/community/ld4lt > <http://www.w3.org/community/ld4lt> > Homepage: http://aksw.org/SebastianHellmann > Research Group: http://aksw.org > -- All the best, Sebastian Hellmann Director of Knowledge Integration and Linked Data Technologies (KILT) Competence Center at the Institute for Applied Informatics (InfAI) at Leipzig University Executive Director of the DBpedia Association Projects: http://dbpedia.org, http://nlp2rdf.org, http://linguistics.okfn.org, https://www.w3.org/community/ld4lt <http://www.w3.org/community/ld4lt> Homepage: http://aksw.org/SebastianHellmann Research Group: http://aksw.org |
From: Almir V. <alm...@gm...> - 2019-05-17 13:54:17
|
Hi all, I am Almir Vardo and working as Data Engineer at Finit, currently working for client from USA. I need to extract knowledge for up to 10millions of entities from DBpedia live and point and it is really slow. Also I will need to sync that knowledge in the future. For that I need stable DBpedia endpoint and do anyone has option to sell us access to the dbpedia synced endpoint or to sell synced virtuoso database? Thanks, Almir -- Almir Vardo, Bsc of Electrical Engineering tel: +387 62-263-253 web: https://ba.linkedin.com/in/almir-vardo-84a1b986 <https://almirvardo.wordpress.com> e-mail: alm...@gm... |
From: Martin D. <mar...@gm...> - 2019-05-17 13:47:05
|
Thanks a lot, Sebastian for your answer. I have one remaining question, for 2017 and 2019, there are note files? is there any way to generate this, in http://live.dbpedia.org/changesets/, there is an hourly changeset, there is not any way to obtain a type and mapping files for the desired date? Cheers Martín El mar., 14 may. 2019 a las 14:46, Sebastian Hellmann (< hel...@in...>) escribió: > Hi Martin, > > the data you arer looking for is here: > > https://databus.dbpedia.org/dbpedia/mappings > > specifically: > > https://databus.dbpedia.org/dbpedia/mappings/instance-types > > https://databus.dbpedia.org/dbpedia/mappings/mappingbased-objects-uncleaned > > the query for instance-types latest version on > https://databus.dbpedia.org/repo/sparql > > PREFIX dataid: <http://dataid.dbpedia.org/ns/core#> > <http://dataid.dbpedia.org/ns/core#> > PREFIX dataid-cv: <http://dataid.dbpedia.org/ns/cv#> > <http://dataid.dbpedia.org/ns/cv#> > PREFIX dct: <http://purl.org/dc/terms/> <http://purl.org/dc/terms/> > PREFIX dcat: <http://www.w3.org/ns/dcat#> <http://www.w3.org/ns/dcat#> > > # Get all files > SELECT DISTINCT ?file WHERE { > ?dataset dataid:artifact > <https://databus.dbpedia.org/dbpedia/mappings/instance-types> > <https://databus.dbpedia.org/dbpedia/mappings/instance-types> . > ?dataset dcat:distribution ?distribution . > ?dataset dct:hasVersion ?latestVersion . > { > SELECT (max(?version) as ?latestVersion) WHERE { > ?dataset dataid:artifact > <https://databus.dbpedia.org/dbpedia/mappings/instance-types> > <https://databus.dbpedia.org/dbpedia/mappings/instance-types> . > ?dataset dct:hasVersion ?version . > ?dataset dcat:distribution ?distribution . > ?distribution dataid:contentVariant 'en'^^ > <http://www.w3.org/2001/XMLSchema#string> > <http://www.w3.org/2001/XMLSchema#string> . > ?distribution dataid:contentVariant 'transitive'^^ > <http://www.w3.org/2001/XMLSchema#string> > <http://www.w3.org/2001/XMLSchema#string> . > } > } > ?distribution dataid:contentVariant 'en'^^ > <http://www.w3.org/2001/XMLSchema#string> > <http://www.w3.org/2001/XMLSchema#string> . > ?distribution dataid:contentVariant 'transitive'^^ > <http://www.w3.org/2001/XMLSchema#string> > <http://www.w3.org/2001/XMLSchema#string> . > ?distribution dcat:downloadURL ?file . > } > > > query can be built on the website, but note that these are unreleased due > to: > > - website is not working 100% > > - small encoding problem of the releases duer to RDF parser > > All the best, > > Sebastian > > > On 14.05.19 15:26, Martin Dominguez wrote: > > Hi all, > I am Martín Domínguez, a researcher from Universidad de Córdoba, > Argentina. We make researching by using DBpedia dumps, and the latest ones, > in the format we use, are available for 2016 ( > https://wiki.dbpedia.org/develop/datasets/downloads-2016-10). > We specifically use the "Instance Types" and "Mapping-based properties" > files, which are available for 2016 and previous dates. > Actually, we found DBpedia Live, and I found the last dumps and the > changesets described in ( > https://wiki.dbpedia.org/online-access/DBpediaLive#4%20Important%20Pointers > ). > After reading this documentation, it is possible to have a new Dump up to > date, but, is it possible only extracting dumps for "Instance Types" and > "Mapping-based properties"?. > We also need to obtain those files not for the latest update, but also for > a given previous date, it is possible? > > regards! > Martín > > > _______________________________________________ > DBpedia-developers mailing lis...@li...https://lists.sourceforge.net/lists/listinfo/dbpedia-developers > > -- > All the best, > Sebastian Hellmann > > Director of Knowledge Integration and Linked Data Technologies (KILT) > Competence Center > at the Institute for Applied Informatics (InfAI) at Leipzig University > Executive Director of the DBpedia Association > Projects: http://dbpedia.org, http://nlp2rdf.org, > http://linguistics.okfn.org, https://www.w3.org/community/ld4lt > <http://www.w3.org/community/ld4lt> > Homepage: http://aksw.org/SebastianHellmann > Research Group: http://aksw.org > |
From: Sebastian H. <hel...@in...> - 2019-05-14 17:46:26
|
Hi Martin, the data you arer looking for is here: https://databus.dbpedia.org/dbpedia/mappings specifically: https://databus.dbpedia.org/dbpedia/mappings/instance-types https://databus.dbpedia.org/dbpedia/mappings/mappingbased-objects-uncleaned the query for instance-types latest version on https://databus.dbpedia.org/repo/sparql PREFIX dataid: <http://dataid.dbpedia.org/ns/core#> PREFIX dataid-cv: <http://dataid.dbpedia.org/ns/cv#> PREFIX dct: <http://purl.org/dc/terms/> PREFIX dcat: <http://www.w3.org/ns/dcat#> # Get all files SELECT DISTINCT ?file WHERE { ?dataset dataid:artifact <https://databus.dbpedia.org/dbpedia/mappings/instance-types> . ?dataset dcat:distribution ?distribution . ?dataset dct:hasVersion ?latestVersion . { SELECT (max(?version) as ?latestVersion) WHERE { ?dataset dataid:artifact <https://databus.dbpedia.org/dbpedia/mappings/instance-types> . ?dataset dct:hasVersion ?version . ?dataset dcat:distribution ?distribution . ?distribution dataid:contentVariant 'en'^^<http://www.w3.org/2001/XMLSchema#string> . ?distribution dataid:contentVariant 'transitive'^^<http://www.w3.org/2001/XMLSchema#string> . } } ?distribution dataid:contentVariant 'en'^^<http://www.w3.org/2001/XMLSchema#string> . ?distribution dataid:contentVariant 'transitive'^^<http://www.w3.org/2001/XMLSchema#string> . ?distribution dcat:downloadURL ?file . } query can be built on the website, but note that these are unreleased due to: - website is not working 100% - small encoding problem of the releases duer to RDF parser All the best, Sebastian On 14.05.19 15:26, Martin Dominguez wrote: > Hi all, > I am Martín Domínguez, a researcher from Universidad de Córdoba, > Argentina. We make researching by using DBpedia dumps, and the latest > ones, in the format we use, are available for 2016 > (https://wiki.dbpedia.org/develop/datasets/downloads-2016-10). > We specifically use the "Instance Types" and "Mapping-based > properties" files, which are available for 2016 and previous dates. > Actually, we found DBpedia Live, and I found the last dumps and the > changesets described in > (https://wiki.dbpedia.org/online-access/DBpediaLive#4%20Important%20Pointers). > After reading this documentation, it is possible to have a new Dump up > to date, but, is it possible only extracting dumps for "Instance > Types" and "Mapping-based properties"?. > We also need to obtain those files not for the latest update, but also > for a given previous date, it is possible? > > regards! > Martín > > > _______________________________________________ > DBpedia-developers mailing list > DBp...@li... > https://lists.sourceforge.net/lists/listinfo/dbpedia-developers -- All the best, Sebastian Hellmann Director of Knowledge Integration and Linked Data Technologies (KILT) Competence Center at the Institute for Applied Informatics (InfAI) at Leipzig University Executive Director of the DBpedia Association Projects: http://dbpedia.org, http://nlp2rdf.org, http://linguistics.okfn.org, https://www.w3.org/community/ld4lt <http://www.w3.org/community/ld4lt> Homepage: http://aksw.org/SebastianHellmann Research Group: http://aksw.org |
From: Martin D. <mar...@gm...> - 2019-05-14 13:27:12
|
Hi all, I am Martín Domínguez, a researcher from Universidad de Córdoba, Argentina. We make researching by using DBpedia dumps, and the latest ones, in the format we use, are available for 2016 ( https://wiki.dbpedia.org/develop/datasets/downloads-2016-10). We specifically use the "Instance Types" and "Mapping-based properties" files, which are available for 2016 and previous dates. Actually, we found DBpedia Live, and I found the last dumps and the changesets described in ( https://wiki.dbpedia.org/online-access/DBpediaLive#4%20Important%20Pointers ). After reading this documentation, it is possible to have a new Dump up to date, but, is it possible only extracting dumps for "Instance Types" and "Mapping-based properties"?. We also need to obtain those files not for the latest update, but also for a given previous date, it is possible? regards! Martín |
From: Nitin J. <nit...@gm...> - 2019-01-01 14:43:18
|
Hi, I can't find Index as mentioned here https://github.com/dbpedia/lookup Link provided to get index is http://downloads.dbpedia-spotlight.org/dbpedia_lookup/ is giving 404 Best Regards, Nitin Jha, Aruelam Software Solutions Pvt. Ltd. www.epitomecoders.com Contact No- +91-8750626011 Skype Id - nitin_jha17 [image: cid:image003.gif@01CF4286.8C89DD90]*Please consider the environment before printing this E-mail* Join the campaign at http://thinkBeforePrinting.org <http://thinkbeforeprinting.org/> |
From: Sebastian H. <hel...@in...> - 2018-10-24 11:57:38
|
Hi Elmahdi, work is being done at the moment. There was also a thread here: https://sourceforge.net/p/dbpedia/mailman/message/36413657/ @Lena, Johannes: maybe we can update http://dev.dbpedia.org/ with more details already All the best, Sebastian On 18.10.18 13:54, Elmahdi Korfed wrote: > Hi everyone! > > I think that DBpedia-live doesn't work because the stream/v1 API > ("RCStream") has been deprecated since July 2017 and was no updated in > the extraction-framework. > You can see the error message in this example: > https://codepen.io/Krinkle/pen/laucI?editors=001 > > Maybe we have to update this code source file RCStreamFeeder.java > (GitHub) > <https://github.com/dbpedia/extraction-framework/blob/65d7549dd8d750ee55c995db14c61aa0f0e86e26/live/src/main/java/org/dbpedia/extraction/live/feeder/RCStreamFeeder.java> and > replace 'https://stream.wikimedia.org/rc' by > 'https://stream.wikimedia.org/v2/stream/recentchange'. > An example here: https://codepen.io/ottomata/pen/VKNyEw/?editors=0010 > > Do think that is the good way? or is not working like this? > > Thank you > > > _______________________________________________ > DBpedia-developers mailing list > DBp...@li... > https://lists.sourceforge.net/lists/listinfo/dbpedia-developers -- All the best, Sebastian Hellmann Director of Knowledge Integration and Linked Data Technologies (KILT) Competence Center at the Institute for Applied Informatics (InfAI) at Leipzig University Executive Director of the DBpedia Association Projects: http://dbpedia.org, http://nlp2rdf.org, http://linguistics.okfn.org, https://www.w3.org/community/ld4lt <http://www.w3.org/community/ld4lt> Homepage: http://aksw.org/SebastianHellmann Research Group: http://aksw.org |
From: Elmahdi K. <elm...@in...> - 2018-10-18 11:54:42
|
Hi everyone! I think that DBpedia-live doesn't work because the stream/v1 API ("RCStream") has been deprecated since July 2017 and was no updated in the extraction-framework. You can see the error message in this example: https://codepen.io/Krinkle/pen/laucI?editors=001 Maybe we have to update this code source file [ https://github.com/dbpedia/extraction-framework/blob/65d7549dd8d750ee55c995db14c61aa0f0e86e26/live/src/main/java/org/dbpedia/extraction/live/feeder/RCStreamFeeder.java | RCStreamFeeder.java (GitHub) ] and replace 'https://stream.wikimedia.org/rc' by 'https://stream.wikimedia.org/v2/stream/recentchange'. An example here: https://codepen.io/ottomata/pen/VKNyEw/?editors=0010 Do think that is the good way? or is not working like this? Thank you |
From: Elmahdi K. <elm...@in...> - 2018-10-04 13:49:12
|
Hello, I'm trying to extract data with extraction.wikidata.properties from this dataset: http://dumps.wikimedia.your.org/wikidatawiki/20180801/ About extraction file, is from github: [ https://github.com/dbpedia/extraction-framework/blob/master/dump/extraction.wikidata.properties | https://github.com/dbpedia/extraction-framework/blob/master/dump/extraction.wikidata.properties ] With this parameters: - languages=wikidata - extractors= - extractors.wikidata=.WikidataR2RExtractor,.WikidataLLExtractor,.WikidataReferenceExtractor,.WikidataAliasExtractor,.WikidataLabelExtractor,.WikidataNameSpaceSameAsExtractor,.WikidataPropertyExtractor,.WikidataLabelExtractor,.WikidataDescriptionExtractor When I launched the extraction: ../run extraction extraction.wikidata.properties I've got some errors exception like this: Exception; wikidata; Main Extraction at 114:19.380s for 124 datasets; Main Extraction failed for instance [ http://wikidata.dbpedia.org/resource/Q445358 | http://wikidata.dbpedia.org/resource/Q445358 ] : Unexpected token (END_OBJECT), expected FIELD_NAME: missing property 'type' that is to contain type id (for class org.wikidata.wdtk.datamodel.json.jackson.JacksonTermedStatementDocument) Also somes <BAD URI: IRI validation failded for: null://null1262011> Do you know if this extraction's version works or if I forgot something in extractors config file to fix this? Thank you very much Elmahdi |
From: sonfack - Y. <sss...@ya...> - 2018-09-24 21:24:42
|
Hi to every one I have a question: When will look at DBPedia as Knowledge representation, is it under the TBox or the Abox ? If it is under the Abox which Ontology does it use ? thanks |
From: sonfack - Y. <sss...@ya...> - 2018-09-24 00:29:23
|
Hi to all I am in the same case. I have tried DBPedia with Java using Apache Jena link: https://github.com/sonfack/spring_jena_dbpedia Actually working with Python too. Please how can someone bring his contributions Thanks On 24/09/2018 00:56, Amandeep Srivastava wrote: > Hi all, > > I'm new to this open source world, and want to contribute to the > DBpedia organization. > > I know Python, Java, JS, Scala, C++ and Machine Learning. I've worked > with WIkidata's RDF data dump earlier. Can you please guide me and > point me to the project that I can start contributing to. > > Anything new and completely unrelated to my skills would be fine too. > Learning something new isn't bad afterall;) > > Thanks. > > -- > Regards, > Aman > > > > > _______________________________________________ > DBpedia-developers mailing list > DBp...@li... > https://lists.sourceforge.net/lists/listinfo/dbpedia-developers |
From: Amandeep S. <ama...@gm...> - 2018-09-23 16:57:22
|
Hi all, I'm new to this open source world, and want to contribute to the DBpedia organization. I know Python, Java, JS, Scala, C++ and Machine Learning. I've worked with WIkidata's RDF data dump earlier. Can you please guide me and point me to the project that I can start contributing to. Anything new and completely unrelated to my skills would be fine too. Learning something new isn't bad afterall;) Thanks. -- Regards, Aman |
From: Robert B. <rob...@li...> - 2018-08-29 13:20:53
|
Hi Elmahdi, These Log-Messages look pretty normal. Nothing should be broken. - AVERTISSEMENT: Language not found: cbk-zam. To extract this language, please edit the addonLanguage.json in core This is only relevant if you want to extract from the cbk-zam language. - INFOS: Will extract redirects from source for commons wiki, could not load cache file This is normal, if you extract from a language for the first time. It generates a file called template-redirects.obj that will be used for the extraction. That process takes some time and could throw a few warning messages, that you can safely ignore. After the template-redirects.obj file is generated the extraction itself should start and you should see files appear in your specified data directory next to the input file. With your configuration that would be: /Users/macbookpro/Documents/web_pro/github/dbpedia/extraction-framework/dump/extraction-dump/2018-08/commonswiki/20180801/ The whole process can take a few hours for big languages like commons, en or fr, but no more input from you is required. Hope that this was helpful, Robert Bielinski ________________________________ Von: Sebastian Hellmann <hel...@in...> Gesendet: Dienstag, 28. August 2018 10:19 An: Elmahdi Korfed; DBp...@li... Cc: Fabien Gandon; micbuffa; Robert Bielinski Betreff: Re: [DBpedia-developers] Hi DBpedians, where the extracted files are stored? Hi Elmahdi, @Robert: could you have a look at this email? We established to do the first release of the "Generic" DBpedia Core module yesterday, it contains all the files you can find here: http://downloads.dbpedia.org/repo/lts/generic-spark/ Since we are establishing more frequent releases now, we split up the publishing into LTS for long term and then "dev" for things we will eventually delete. The most important change is: - clearer release and versioning methodology - metadata provided We would hope that you would also join in with some datasets. By the way, we moved most of the communications in the "#releasea" channel on slack All the best, Sebastian On 24.08.2018 16:54, Elmahdi Korfed wrote: Hi everyone, I'm working on an updated version of DBpediaFR chapter 2018 and I just want to know where the extracted files are stored. Some explications from beginning: I downloaded : - dbpedia/extraction-framework from github<https://github.com/dbpedia/extraction-framework/> - commons + fr + wiki dump 2018-08's version from the dumps wikimedia website<http://dumps.wikimedia.your.org/frwiki/20180801/> (source: **-pages-articles.xml.bz2) Now I would like to extract, first, commonswiki. To do that, I configured 2 files: => "extraction.commons.properties" (content of file): source=pages-articles.xml.bz2 require-download-complete=false languages=commons extractors= extractors.commons=.MappingExtractor,.ContributorExtractor,.TemplateParameterExtractor,.FileTypeExtractor,.GalleryExtractor,.ImageAnnotationExtractor,.CommonsKMLExtractor,.DBpediaResourceExtractor copyrightCheck=false => "universal.properties" (content of file): dbpedia-version=2018-08 base-dir=/Users/macbookpro/Documents/web_pro/github/dbpedia/extraction-framework/dump/extraction-dump/2018-08 log-dir=/Users/macbookpro/Documents/web_pro/github/dbpedia/extraction-framework/dump/extraction-data/2018-08 wiki-name=wiki source=pages-articles.xml.bz2 parallel-processes=4 ontology=../ontology.xml mappings=../mappings uri-policy.iri=generic:en format.ttl.bz2=turtle-triples After that, I launched these command: - cd extraction-frameworkd/dump - ../clean-install-run extraction extraction.commons.properties Now I have some messages like this: - AVERTISSEMENT: Language not found: cbk-zam. To extract this language, please edit the addonLanguage.json in core. - INFOS: Will extract redirects from source for commons wiki, could not load cache file '/Users/macbookpro/Documents/web_pro/github/dbpedia/extraction-framework/dump/extraction-dump/2018-08/commonswiki/20180801/commonswiki-20180801-template-redirects.obj': java.io.FileNotFoundException: /Users/macbookpro/Documents/web_pro/github/dbpedia/extraction-framework/dump/extraction-dump/2018-08/commonswiki/20180801/commonswiki-20180801-template-redirects.obj (No such file or directory) - AVERTISSEMENT: wrong redirect. page: [title=UNC;ns=0/Main/;language:wiki=commons,locale=en]. - found by dbpedia: [title=University of North Carolina at Chapel Hill;ns=0/Main/;language:wiki=commons,locale=en]. - found by wikipedia: [title=University of North Carolina at Chapel Hill;ns=0/Main/;language:wiki=commons,locale=en] It's seem ok right? Do you know if I just have to wait for the extraction to finish to see the extracted files? Because I need to storage files in VirtuosoDB Thank you for your help |
From: Sebastian H. <hel...@in...> - 2018-08-28 08:19:41
|
Hi Elmahdi, @Robert: could you have a look at this email? We established to do the first release of the "Generic" DBpedia Core module yesterday, it contains all the files you can find here: http://downloads.dbpedia.org/repo/lts/generic-spark/ Since we are establishing more frequent releases now, we split up the publishing into LTS for long term and then "dev" for things we will eventually delete. The most important change is: - clearer release and versioning methodology - metadata provided We would hope that you would also join in with some datasets. By the way, we moved most of the communications in the "#releasea" channel on slack All the best, Sebastian On 24.08.2018 16:54, Elmahdi Korfed wrote: > Hi everyone, > > I'm working on an updated version of DBpediaFR chapter 2018 and I just > want to know where the extracted files are stored. > Some explications from beginning: > > I downloaded : > - dbpedia/extraction-framework from github > <https://github.com/dbpedia/extraction-framework/> > - commons + fr + wiki dump 2018-08's version from the dumps wikimedia > website <http://dumps.wikimedia.your.org/frwiki/20180801/> (source: > **-pages-articles.xml.bz2) > > Now I would like to extract, first, commonswiki. > To do that, I configured 2 files: > > => "extraction.commons.properties" (content of file): > > source=pages-articles.xml.bz2 > require-download-complete=false > languages=commons > extractors= > extractors.commons=.MappingExtractor,.ContributorExtractor,.TemplateParameterExtractor,.FileTypeExtractor,.GalleryExtractor,.ImageAnnotationExtractor,.CommonsKMLExtractor,.DBpediaResourceExtractor > copyrightCheck=false > > => "universal.properties" (content of file): > dbpedia-version=2018-08 > base-dir=/Users/macbookpro/Documents/web_pro/github/dbpedia/extraction-framework/dump/extraction-dump/2018-08 > log-dir=/Users/macbookpro/Documents/web_pro/github/dbpedia/extraction-framework/dump/extraction-data/2018-08 > wiki-name=wiki > source=pages-articles.xml.bz2 > parallel-processes=4 > ontology=../ontology.xml > mappings=../mappings > uri-policy.iri=generic:en > format.ttl.bz2=turtle-triples > > > After that, I launched these command: > - cd extraction-frameworkd/dump > - ../clean-install-run extraction extraction.commons.properties > > Now I have some messages like this: > - AVERTISSEMENT: Language not found: cbk-zam. To extract this > language, please edit the addonLanguage.json in core. > - INFOS: Will extract redirects from source for commons wiki, could > not load cache file > '/Users/macbookpro/Documents/web_pro/github/dbpedia/extraction-framework/dump/extraction-dump/2018-08/commonswiki/20180801/commonswiki-20180801-template-redirects.obj': > java.io.FileNotFoundException: > /Users/macbookpro/Documents/web_pro/github/dbpedia/extraction-framework/dump/extraction-dump/2018-08/commonswiki/20180801/commonswiki-20180801-template-redirects.obj > (No such file or directory) > > - AVERTISSEMENT: wrong redirect. page: > [title=UNC;ns=0/Main/;language:wiki=commons,locale=en]. > - found by dbpedia: [title=University of North Carolina at Chapel > Hill;ns=0/Main/;language:wiki=commons,locale=en]. > - found by wikipedia: [title=University of North Carolina at Chapel > Hill;ns=0/Main/;language:wiki=commons,locale=en] > > It's seem ok right? > Do you know if I just have to wait for the extraction to finish to see > the extracted files? Because I need to storage files in VirtuosoDB > > Thank you for your help > > > ------------------------------------------------------------------------------ > Check out the vibrant tech community on one of the world's most > engaging tech sites, Slashdot.org! http://sdm.link/slashdot > > > _______________________________________________ > DBpedia-developers mailing list > DBp...@li... > https://lists.sourceforge.net/lists/listinfo/dbpedia-developers -- All the best, Sebastian Hellmann Director of Knowledge Integration and Linked Data Technologies (KILT) Competence Center at the Institute for Applied Informatics (InfAI) at Leipzig University Executive Director of the DBpedia Association Projects: http://dbpedia.org, http://nlp2rdf.org, http://linguistics.okfn.org, https://www.w3.org/community/ld4lt <http://www.w3.org/community/ld4lt> Homepage: http://aksw.org/SebastianHellmann Research Group: http://aksw.org |
From: Elmahdi K. <elm...@in...> - 2018-08-24 14:54:37
|
Hi everyone, I'm working on an updated version of DBpediaFR chapter 2018 and I just want to know where the extracted files are stored. Some explications from beginning: I downloaded : - dbpedia/extraction-framework from [ https://github.com/dbpedia/extraction-framework/ | github ] - commons + fr + wiki dump 2018-08's version from [ http://dumps.wikimedia.your.org/frwiki/20180801/ | the dumps wikimedia website ] (source: **-pages-articles.xml.bz2) Now I would like to extract, first, commonswiki. To do that, I configured 2 files: => "extraction.commons.properties" (content of file): source=pages-articles.xml.bz2 require-download-complete=false languages=commons extractors= extractors.commons=.MappingExtractor,.ContributorExtractor,.TemplateParameterExtractor,.FileTypeExtractor,.GalleryExtractor,.ImageAnnotationExtractor,.CommonsKMLExtractor,.DBpediaResourceExtractor copyrightCheck=false => "universal.properties" (content of file): dbpedia-version=2018-08 base-dir=/Users/macbookpro/Documents/web_pro/github/dbpedia/extraction-framework/dump/extraction-dump/2018-08 log-dir=/Users/macbookpro/Documents/web_pro/github/dbpedia/extraction-framework/dump/extraction-data/2018-08 wiki-name=wiki source=pages-articles.xml.bz2 parallel-processes=4 ontology=../ontology.xml mappings=../mappings uri-policy.iri=generic:en format.ttl.bz2=turtle-triples After that, I launched these command: - cd extraction-frameworkd/dump - ../clean-install-run extraction extraction.commons.properties Now I have some messages like this: - AVERTISSEMENT: Language not found: cbk-zam. To extract this language, please edit the addonLanguage.json in core. - INFOS: Will extract redirects from source for commons wiki, could not load cache file '/Users/macbookpro/Documents/web_pro/github/dbpedia/extraction-framework/dump/extraction-dump/2018-08/commonswiki/20180801/commonswiki-20180801-template-redirects.obj': java.io.FileNotFoundException: /Users/macbookpro/Documents/web_pro/github/dbpedia/extraction-framework/dump/extraction-dump/2018-08/commonswiki/20180801/commonswiki-20180801-template-redirects.obj (No such file or directory) - AVERTISSEMENT: wrong redirect. page: [title=UNC;ns=0/Main/;language:wiki=commons,locale=en]. - found by dbpedia: [title=University of North Carolina at Chapel Hill;ns=0/Main/;language:wiki=commons,locale=en]. - found by wikipedia: [title=University of North Carolina at Chapel Hill;ns=0/Main/;language:wiki=commons,locale=en] It's seem ok right? Do you know if I just have to wait for the extraction to finish to see the extracted files? Because I need to storage files in VirtuosoDB Thank you for your help |
From: Niketa <ni...@gm...> - 2018-03-01 22:08:44
|
Call for Papers Apologies for cross-posting. Kindly help to distribute this final CFP to your mailing list. -- The 14th International Conference on Information Assurance and Security (IAS ’18) -- http://www.mirlabs.net/ias18 http://www.mirlabs.org/ias18 Indexed by: SCOPUS, Thomson ISI Web of Science, DBLP etc. History of IAS series: http://www.mirlabs.net/ias18/previous.php IAS 2018: Scopus Proceedings All accepted and registered papers will be published in AISC Series of Springer, indexed in ISI Proceedings, EI-Compendex, DBLP, SCOPUS, Google Scholar and Springerlink. Proceedings will be made available during the conference. Expanded versions of selected papers will be published in special issues of internationally referred journals (indexed by SCI) and edited volumes. ** Important Dates ** (Extended) ---------------------------- Paper submission due: September 30, 2018 Notification of paper acceptance: October 31, 2018 Registration and Final manuscript due: November 15, 2018 Conference: December 13 - 15, 2018 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ About IAS'18: ------------------------- The International Conference on Information Assurance and Security (IAS) is a major international conference aims to bring together researchers, practitioners, developers, and policy makers involved in multiple disciplines of information security and assurance to exchange ideas and to learn the latest development in this important field. Information assurance and security has become an important research issue in the networked and distributed information sharing environments. Finding effective ways to protect information systems, networks and sensitive data within the critical information infrastructure is challenging even with the most advanced technology and trained professionals. All accepted and registered papers will be included in the conference proceedings to expected be published by Springer. Topics (not limited to): -------------------------------- Information Assurance, Security Mechanisms, Methodologies and Models Authentication and Identity Management Authorization and Access Control Trust Negotiation, Establishment and Management Anonymity and User Privacy Data Integrity and Privacy Network Security Operating System Security Database Security Intrusion Detection Security Attacks Security Oriented System Design Security and Performance trade-off Security Management and Strategy Security Verification, Evaluations and Measurements Secure Software Technologies New Ideas and Paradigms for Security Cryptography Cryptographic Protocols Key Management and Recovery Secure System Architectures and Security Application Web Services Security Grid Security Ubiquitous Computing Security Mobile Agent Security Internet Security Intellectual Property Protection E-Commerce Security E-Government Security E-Health Security Home System Security Sensor Network Security Ad hoc network security Biometrics Security and Applications Secure Hardware and Smartcards Image Engineering, Multimedia Signal Processing and Communication Security Multimedia Security Multimedia Forensic Digital Watermarking and DRM Communication Security Biometrics Information Fusion Image Registration Image Mosaic Image Indexing and Retrieval Image and Video Coding Multiscale Geometric Analysis Motion Detection and Tracing Feature Extraction 3G Communication Embedded System Design ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Submission Guidelines: -------------------------------- Submission of paper should be made through the submission page from the conference web page. Please refer to the conference website for guidelines to prepare your manuscript. Paper format templates: http://www.springer.com/series/11156 IAS’18 Submission Link: https://easychair.org/conferences/?conf=ias2018 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ * Organizing Committee * ---------------------------------- General Chairs: Ana Maria Madureira, Instituto Superior de Engenharia do Porto, Portugal Ajith Abraham, Machine Intelligence Research Labs, USA Technical Committee (Please refer website): http://www.mirlabs.net/ias17/committees.php For technical contact: ------------------------------ Ajith Abraham Email: aji...@ie... |
From: Sandra P. <pr...@in...> - 2018-02-06 14:27:21
|
Hi everyone, this just a little reminder, that our monthly Developer Telco is scheduled for tomorrow, 2PM, CET. Among others, the focus of this week's session will be Google Summer of Code. You can check the topics and access in our Agenda/minutes: https://docs.google.com/document/d/1ZEhOv9Cp2LyBcyFc2J7PncsCu5brov3_qRxKleMtq60/edit#heading=h.37lyh9bjyf87<https://exch.informatik.uni-leipzig.de/owa/redir.aspx?C=wlMPWRaju6Jnr2KMXuTzLZceQDVOS2XzhayCvnB-wO8dZz92bW3VCA..&URL=https%3a%2f%2fdocs.google.com%2fdocument%2fd%2f1ZEhOv9Cp2LyBcyFc2J7PncsCu5brov3_qRxKleMtq60%2fedit%23heading%3dh.37lyh9bjyf87> Best Sandra Sandra Prätor, Mag. Artium DBpedia Association phone: +49 341 97 32355 e-mail: pr...@in... Institute for Applied Informatics (InfAI) adjunct Institute at the University of Leipzig Hainstraße 11 | 04109 Leipzig Registergericht: AG Leipzig | Registernummer: VR 4342 http://www.infai.org<http://www.infai.org/> |
From: Magnus K. <kn...@in...> - 2018-01-03 09:57:07
|
Hi everyone, A happy new year 2018 to all of you. Due to the past holidays, there is likely not much to discuss for the DevTelco. If someone of you has topics that he wants to discuss please add them to the minutes document [https://docs.google.com/document/d/1ZEhOv9Cp2LyBcyFc2J7PncsCu5brov3_qRxKleMtq60/edit#heading=h.37lyh9bjyf87]. If there are no topics upcoming, I’d suggest to cancel the telco today and talk to each other on Wed 07/02/2018, Time: 02:00 PM CEST. All the best, Magnus -- Magnus Knuth Universität Leipzig Institut für Informatik Abt. Betriebliche Informationssysteme, AKSW/KILT Augustusplatz 10 04109 Leipzig DE mail: kn...@in... tel: +49 177 3277537 webID: http://magnus.13mm.de/ |