You can subscribe to this list here.
| 2011 |
Jan
|
Feb
|
Mar
|
Apr
|
May
|
Jun
|
Jul
|
Aug
(8) |
Sep
(4) |
Oct
(1) |
Nov
(5) |
Dec
(16) |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 2012 |
Jan
(4) |
Feb
|
Mar
(14) |
Apr
|
May
|
Jun
(1) |
Jul
(5) |
Aug
|
Sep
|
Oct
|
Nov
|
Dec
(19) |
| 2013 |
Jan
|
Feb
(1) |
Mar
(1) |
Apr
(3) |
May
|
Jun
|
Jul
|
Aug
(6) |
Sep
(3) |
Oct
|
Nov
|
Dec
|
| 2014 |
Jan
|
Feb
|
Mar
(6) |
Apr
|
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
(1) |
Nov
(13) |
Dec
(1) |
| 2016 |
Jan
|
Feb
|
Mar
(1) |
Apr
|
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
|
| 2017 |
Jan
|
Feb
|
Mar
|
Apr
|
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
(1) |
Nov
|
Dec
|
| 2018 |
Jan
|
Feb
|
Mar
|
Apr
(1) |
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
|
| 2019 |
Jan
(1) |
Feb
|
Mar
|
Apr
|
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
|
|
From: Joe W. <jo...@gm...> - 2011-11-02 16:48:26
|
Hi Chris, I'm glad you found the workshop valuable. Sorry for my slow reply! The structure for your data makes sense to me. Admittedly I haven't done an analogous project with annotated bibliographies. I'd invite others to may have suggestions about ways to structure annotated bibliographies in TEI. Also, you might consult http://tei-l.markmail.org/search/?q=bibliographies for recent discussions on doing bibliographies in TEI). And TEI-L would be another good place to ask. Let me focus, then, on the question of how eXist's indexing technologies might factor into the way you store and query your TEI; along the way I'll touch on the elements vs. attributes question. Your example entry is very well suited for rapid searches using eXist's structural, range, and lucene full text indexes. Indexes speed up queries, because rather than "brute-force" scanning through your documents (potentially many gigabytes on disk) to find the answer, eXist can find the answer indexes (much smaller, and likely to fit in RAM; in-memory index searches the fastest). Structural Index: eXist's structural index keeps track of every element and attribute and their structural placements in the database. This makes pure XPath-based queries very fast, e.g. //tei:bibl would quickly return all of the bibl elements in the database. Or //tei:date[not(@when)] would quickly return all of the date elements that do not have a @when attribute. The structural index is always on by default. Other indexes must be manually configured and applied by you: Range Index: A range index stores all of the values of a specific element or attribute, and it greatly speeds up queries on the values of an element or attribute. Range indexes are typed as string, integer, date, year, etc. For example, if you wanted to search for entries in the 1990s, you could apply a year-based range index to the @when attribute, speeding up queries like: //tei:bibl[tei:date/@when ge 1990 and tei:date/@when lt 2000] If you wanted to query all entries from a specific journal, you could apply string-based range indexes to (1) the @level attribute, since that distinguishes between different title types and (2) the title element: //tei:title[@level eq 'j'][. eq 'Some Journal'] Notice a pattern here: whenever you filter an expression with a predicate (the bit in square brackets) that uses comparisons (equals, less than, greater than), you can most likely apply a range index. Think of a range index as a very literal dictionary of the exact values. Notice also that we have been using plain, pure XPath and XQuery here. Your queries simply get faster by virtue of the built-in structural index and the user-specified range indexes. Lucene Full Text Index: Whereas a range index is very literal and enables queries of the full value of an element or attribute, a full text index does a lot more work behind the scenes: it identifies "words" in the text in an element or attribute (typically, by treating a space as the thing that separates words) and stores those words in an index; this process is called "tokenization". Full text indexes also let you search with wildcards like * or ?. So it makes sense to apply full text indexes in cases where you have many words. I wouldn't apply a full text index to the @level attribute in your example, since that contains codes (e.g., "a" and "j"); nor would I apply it to the @when attribute, since that contains years. But I would think about applying one to the title elements, or to the note elements. For example, if you want to search titles that have some words, you can apply a lucene index to the "note" element use this kind of query: //tei:bibl[ft:query(tei:note, 'emotion')] I'll leave it as an exercise for you to think about whether you might ever want a range index and a full text index on the same element, e.g., title. (There's no reason you can't.) To configure and apply these indexes above, you would need to create a collection.xconf file and place it in the /db/system/config/db... collection corresponding to your own data's collection in the database (e.g., for data stored in /db/myapp, you would put the index configuration file in /db/system/config/db/myapp/collection.xconf): <collection xmlns="http://exist-db.org/collection-config/1.0"> <index xmlns:tei="http://www.tei-c.org/ns/1.0"> <!-- Disable the legacy full text index --> <fulltext default="none" attributes="false"/> <!-- Lucene index configuration --> <lucene> <!-- The standard analyzer will ignore stopwords like 'the', 'and' --> <analyzer class="org.apache.lucene.analysis.standard.StandardAnalyzer"/> <!-- Whitespace analyzer includes stopwords like 'the', 'and' --> <!--analyzer class="org.apache.lucene.analysis.WhitespaceAnalyzer"/--> <text qname="tei:note"/> <text qname="tei:title"/> </lucene> <!-- Range index configuration --> <create qname="@level" type="xs:string"/> <create qname="title" type="xs:string"/> <create qname="@when" type="xs:date"/> <!-- Note on @when: "eXist can only use a range index if all values within that index are valid instances of the defined index type. So every date has to be an xs:date and if there's just one exception, the index will no longer be used. --> </index> </collection> There's quite a bit that you can do to customize your full text search indexes with eXist, including using Lucene query syntax, expressing your query in XML, selecting various Lucene analyzers (with purpose-built and/or language-specific features). Here are some links about these topics: 1. Configuring Database Indexes: http://exist-db.org/indexing.html 2. Tips on Writing Queries: http://exist-db.org/tuning.html#d1973e562 3. Lucene-based Full Text Index: http://exist-db.org/lucene.html 4. Lucene query syntax: http://lucene.apache.org/java/2_9_3/queryparsersyntax.html Nothing I've discussed mandates that you store your data in elements vs. attributes. That said, I believe that there is currently a limitation on applying Lucene full text indexes to attribute data in eXist (no such restriction on range indexes though). Perhaps one of the core eXist developers will comment on that. I hope this answer, belated as it may be, helps. Cheers, Joe On Tue, Oct 18, 2011 at 5:59 PM, Christopher Thomson <chr...@ca...> wrote: > Hello, > > First of all I wanted to say Joe's introductory workshop on eXist at > Oxford this year was really valuable, and although it's taken me a while > to post something to this list, I think it's a great idea and I hope > there are others out there interested in getting started with eXist. For > my own part, I'd like to gather some feedback on a project I'm working > on, a digital edition of an annotated bibliography that has existed for > a few years in print form. It's a modest project, and I'm using it to > learn XQuery and half a dozen other things as I go along :) > > At present I have created some sample TEI XML files, and have used the > <particDesc> in the TEI header to record biographical information about > each author. The body of each entry contains bibliographic references > in a number of categories, and I've marked these up using divs for the > categories and <bibl> elements sitting as list items within. > > I'm aiming to use elements rather than attributes to hold the data > wherever possible, as my limited understanding of full text search > suggests this is a good idea for creating indexes. However, some data > is held as attributes, as in the sample div below. Does this make it > any more difficult to search/index? > > Using eXist, I've managed to produce a very basic local hosted website > to view some of my sample TEI. Any advice on this general approach to > creating an annotated bibliography would be most welcome, as would any > resources or examples I should consult. > > Best regards, > Chris > > > <div type="autobiographical"> > <head>Autobiographical articles</head> > <list> > <item> > <bibl> > <author>Person, Ann</author> > <title level="a">Reflections</title> > <title level="j">Some Journal</title> > <biblScope type="vol">1</biblScope> > <date when="1992">1992</date> > <biblScope type="pp">3-4</biblScope> > </bibl> > <note>Annotation and summary of text ...</note> > </item> > </list> > </div> |
|
From: Christopher T. <chr...@ca...> - 2011-10-18 22:00:13
|
Hello, First of all I wanted to say Joe's introductory workshop on eXist at Oxford this year was really valuable, and although it's taken me a while to post something to this list, I think it's a great idea and I hope there are others out there interested in getting started with eXist. For my own part, I'd like to gather some feedback on a project I'm working on, a digital edition of an annotated bibliography that has existed for a few years in print form. It's a modest project, and I'm using it to learn XQuery and half a dozen other things as I go along :) At present I have created some sample TEI XML files, and have used the <particDesc> in the TEI header to record biographical information about each author. The body of each entry contains bibliographic references in a number of categories, and I've marked these up using divs for the categories and <bibl> elements sitting as list items within. I'm aiming to use elements rather than attributes to hold the data wherever possible, as my limited understanding of full text search suggests this is a good idea for creating indexes. However, some data is held as attributes, as in the sample div below. Does this make it any more difficult to search/index? Using eXist, I've managed to produce a very basic local hosted website to view some of my sample TEI. Any advice on this general approach to creating an annotated bibliography would be most welcome, as would any resources or examples I should consult. Best regards, Chris <div type="autobiographical"> <head>Autobiographical articles</head> <list> <item> <bibl> <author>Person, Ann</author> <title level="a">Reflections</title> <title level="j">Some Journal</title> <biblScope type="vol">1</biblScope> <date when="1992">1992</date> <biblScope type="pp">3-4</biblScope> </bibl> <note>Annotation and summary of text ...</note> </item> </list> </div> |
|
From: Ron V. d. B. <ron...@ka...> - 2011-09-09 09:00:36
|
Hi Wolfgang, On donderdag 8 september 2011 17:47:06, Wolfgang Meier wrote: > > Instead of just highlighting the match, we could tag all preceding and > following tokens without much additional cost. This could probably > also include the relative position of the token to the match, so you > would end up with something like<context pos="-1">...</context> > <match>...</match> <context pos="+1">...</context>. > This sounds great! Such a 'native' segmentation would undoubtedly perform much faster. Additionally, I guess it would facilitate further interaction with such collocation data as well. For example, if a collocation table shows that "great" occurs at position 3 after the search term "eXist", I can imagine that users would want a link from there to "exact proximity searches", where "eXist" occurs exactly 3 words before "great". That's something the Lucene search syntax doesn't support, does it? > > I suppose Lucene does store the total number of words per indexed > document somewhere (it should be relevant for computing weights), so > we could add a function to retrieve it. > Ditto: would be very useful! > P.S.: I plan to integrate your improved version of the kwic module. I > just wanted to test it on some of my existing apps first to see if it > breaks backwards compatibility or not. That's nice to hear. Please make sure to test the version at <http://www.kantl.be/ctb/download/kwic.xql>, which has some improvements and fixes some dumb errors, compared to the one I posted on eXist-open. Kind regards, Ron |
|
From: Wolfgang M. <wol...@ex...> - 2011-09-08 15:47:22
|
Hi, > With eXist's util:expand() function for 'materializing' search results > in their context nodes, it is possible to build nice KWIC displays (as > aptly illustrated in eXist's own KWIC display module). Recently [1] I > explored ways of linguistically exploiting such KWIC searches (mostly by > adding sorting possibilities for the left and right contexts). One step > further would be the construction of collocation tables for all words > occurring in in certain contexts of search words. Yet, I think I am > stumbling on eXist's current limitations here, since what would make > such data really useful, information on the (relative) occurrence of > words at certain context positions, is currently not available in eXist > (and would be prohibitively expensive to compute). After reading your post recently, I thought about how we could better support collocation tables and similar features. One possibility would be to provide additional information when expanding the search results: to find the full text match position, eXist always needs to tokenize the text again by passing it through Lucene's analyzer. Instead of just highlighting the match, we could tag all preceding and following tokens without much additional cost. This could probably also include the relative position of the token to the match, so you would end up with something like <context pos="-1">...</context> <match>...</match> <context pos="+1">...</context>. > Actually, this is > also the case for a simple 'frequency list': while the util:index-keys() > function does allow one to construct a list of all indexed terms, with > the number of their absolute occurrences, I think it is ratios linguists > are interested in most. That would require additional information on the > total number of words occurring in the collection being queried. I suppose Lucene does store the total number of words per indexed document somewhere (it should be relevant for computing weights), so we could add a function to retrieve it. Wolfgang P.S.: I plan to integrate your improved version of the kwic module. I just wanted to test it on some of my existing apps first to see if it breaks backwards compatibility or not. |
|
From: Ron V. d. B. <ron...@ka...> - 2011-09-08 15:02:26
|
Hi all, Inspired by Joe's nice summary of eXist's openness, I would like to take the opportunity to both tell a bit of how I would *like* to use eXist, and ask how other people are using eXist for linguistic applications. I'm very fond of eXist's comprehensive scope for developing powerful web XML/XSLT/XQuery applications in one single environment. (Though I still live in the Cocoon+eXist days, I have experimented with eXist's MVC framework and was delighted to see how well and painlessly it can replace most of Cocoon's sitemap pipelining functionality.) I am using eXist for a number of projects, ranging from critical edition interfaces of TEI encoded texts to interfaces for TEI encoded linguistic corpora. In doing so, I sometimes feel myself pushing the limits of what's possisble/sensible with XQuery, though mostly eXist does scale up well for my not-too-large text collections. There's one of such more 'experimental' uses of eXist I would like to focus on here. With eXist's util:expand() function for 'materializing' search results in their context nodes, it is possible to build nice KWIC displays (as aptly illustrated in eXist's own KWIC display module). Recently [1] I explored ways of linguistically exploiting such KWIC searches (mostly by adding sorting possibilities for the left and right contexts). One step further would be the construction of collocation tables for all words occurring in in certain contexts of search words. Yet, I think I am stumbling on eXist's current limitations here, since what would make such data really useful, information on the (relative) occurrence of words at certain context positions, is currently not available in eXist (and would be prohibitively expensive to compute). Actually, this is also the case for a simple 'frequency list': while the util:index-keys() function does allow one to construct a list of all indexed terms, with the number of their absolute occurrences, I think it is ratios linguists are interested in most. That would require additional information on the total number of words occurring in the collection being queried. I hope I am not being too specific (or vague, on the contrary) in sketching this particular use of XQuery/eXist. Are there good use cases of employing eXist for more linguistically-oriented tasks out there? Or is there perhaps a possibility in specific Lucene analyzers (I'm not too Lucene-proficient, but I appreciate that eXist's pluggable indexer architecture might ) for linguistic / statistical analyses? Kind regards, Ron [1] <http://rvdb.wordpress.com/2011/07/20/from-kwic-display-to-kwicer-processing-with-exist/> -- Ron Van den Branden Wetenschappelijk attaché / Senior Researcher Reviews Editor LLC. The Journal of Digital Scholarship in the Humanities Centrum voor Teksteditie en Bronnenstudie - CTB (KANTL) Centre for Scholarly Editing and Document Studies Koninklijke Academie voor Nederlandse Taal- en Letterkunde Royal Academy of Dutch Language and Literature Koningstraat 18 / b-9000 Gent / Belgium tel: +32 9 265 93 51 / fax: +32 9 265 93 49 E-mail : ron...@ka... http://www.kantl.be/ctb |
|
From: Joe W. <jo...@gm...> - 2011-09-08 12:47:23
|
Hi all, One thing worth highlighting about Peter's work is how he has used eXist-db as a hub of his research data: He is importing data from SQL databases (MediaWiki), RDF data from Zotero, and MS Access databases. By importing this data into eXist-db, he is able to enrich what he can do with his TEI data. And he is able to "export" his combined data into useful forms: HTML for web views, PDF for dissertation publishing, and NodeXL for data visualization. eXist's ability to import and house data from a variety of sources is really remarkable. You can ignore this ability and use it for querying just your TEI, of course, but there are many hooks for integrating data from lots of other sources. A lot of this flexibility is due to eXist's openness to incorporating libraries from other open source projects that facilitate import and export: - eXist can generate charts - using the built-in JFreeChart library. See http://atomic.exist-db.org/blogs/dizzzz/JFreeChart - eXist can generate PDFs - using the built-in Apache FOP framework (commercial XSL-FO libraries can be supplied instead of FOP) - eXist can scale images and get metadata about them - see http://demo.exist-db.org/exist/functions/image - eXist can make SQL queries and can connect to FTP and SVN servers - eXist is adding the Apache Tika framework for importing PDF, Microsoft Office files (and the formats listed at http://tika.apache.org/0.9/formats.html) -- making all of these formats full text searchable and queryable with XQuery Of course with the growth in RESTful API services such as Google Charts, Captcha, and OpenCalais, not all features need to come out of the box in eXist -- you can just use eXist's HTTP client to send data off to the API and get your results back. A word on Jens' mention of eXist version 1.5: The current release version of eXist is 1.4.1, but eXist is adding new features to its "bleeding edge" development version, 1.5. Such features include the ability to install "app"-like packages such as Jens's Tamboti project, the Apache Tika module, more sophisticated Lucene indexers, etc. This new version isn't available as a convenient installer yet, but it can be downloaded and used -- if you're willing to tinker with Subversion. If you haven't heard of Subversion, then it's probably worth waiting for an official release version, but you're welcome to dive in. See http://exist.sourceforge.net/building.html#svn. I, for one, am eager for the release of the current development version, because I have already adapted the Punch demo from the Oxford Summer School eXist workshop into an easily installable "app" package. It'll make the installation step a one-click process, instead of the manual process it is now. (For more info on eXist's package and repository format, see http://demo.exist-db.org/repo/repo.xml.) So there's exciting stuff coming down the pike! Again, you can just stick with querying your TEI -- but it's nice to know that when you need to reach out to integrate other data sources or generate new forms of output, lots of solutions are there for you to make use of. Cheers, Joe |
|
From: Jens Ø. P. <oe...@gm...> - 2011-08-31 13:44:47
|
Hi Peter, Quite amazing what one week in a summer school can do for your research! From within an installation of eXist 1.5, you can download and install a bibliographical tool built on the MODS format from the Library of Congress <http://www.loc.gov/standards/mods/>. The app is called Tamboti Metadata Framework, and it is installed through eXist's Package Repository. It is actually more than a bibliographical database - we are now enlarging it for display of images. It should be easy to set up links from your database to Tamboti. Tamboti also comes with a Zotero unAPI support, i.e. you can import to Zotero from an icon in the navigation toolbar. Since Zotero exports in MODS format, it should be easy for your to integrate your bibliography. You might want to wait until the Zotero version 3 is released (which will happen soon - the beta is out), since it contains a lot of fixes to MODS export. Tamboti also contains a full MODS editor. Visualisation is also something I have begun to think about. There are so many different SVG- and Javascript-based solutions to choose between. Perhaps I will get together a summary of the options …. It really should not be necessary to export your data to Excel to generate a visualisation! Are you thinking of possibilities to edit your XML files while they are inside the eXist database? I would also like to inform you about another summer school, also held in Oxford, the XML Summer School, September 18 — 23, 2011; see <http://xmlsummerschool.com/>. Cheers, Jens On Aug 31, 2011, at 11:58 AM, Peter Watson wrote: > Hi Joe and Others > > As a graduate of the 2010 TEI Summer School this list is a really welcome development for me. My own project is a personal one supporting my DPhil research on a set of twelfth and thirteenth century land deeds for a family living in the depths of the English Midlands. I looked for software that would allow me to place the deeds in a database using a single copy of the transcribed texts for all purposes. TEI plus eXist-db does this for me. I have developed the following functions, apart from the html rendering which came from the summer school material that Joe provided: > TEI tagging of names, dates and places allows me to list and search the database in a variety of structured ways. > PDF printing from an fo transform based on XQuery, quite similar to rendering in HTML. > Visualization of social network data is under development using NodeXL, externally at present, though long term I would like to be able to integrate visualization into eXist. > Integration with my bibliography based on Zotero. At present this uses an RDF copy of my Zotero database saved into eXist so is not completely real time. Ffrom eXist I can also gain access to the Zotero data via SQLite, on which Zotero is based, though again this uses a copy of the Zotero data. > I have also transformedand imported an XML copy of a much larger database of C12 &C13 charters from an Access database which allow these to be accessed much more easily whilst also retaining the ability to search the data that Access allows. > I have the utilities to enable me to search an earlier database I developed based on MediaWiki, which sits on an SQL database. > I'd be very keen to learn of any one else who is doing something similar to this. At the moment all of this is on my laptop. Providing access to what is a prototype is some way down the list of priorities, unless there is a simple solution to getting the data hosted. > > Best wishes > > Peter > > > > > On 30/08/2011 23:40, Joe Wicentowski wrote:Hello all, Today I sent out an announcement about this list to the broader TEI community via TEI-L. Welcome to all who have just joined! To kick the discussion here off, I think it might be useful for those new to eXist-db to learn about some current TEI-based projects that employ eXist-db in some way. There is a list of some eXist-db projects on the TEI wiki: http://wiki.tei-c.org/index.php/EXist#Sample_implementations. The project I work on (also listed on the wiki) is the Office of the Historian website (http://history.state.gov/). The entire website is driven by eXist-db, it has been in production since March 2009, and it is hosted on Amazon EC2. The functions of eXist-db are used throughout the site, and all article- and book-length content on the site is encoded in TEI. All pages are rendered into HTML by eXist-db on the fly as they are requested. Perhaps one of the most interesting portions from a TEI perspective is how we use TEI person and term tags to drive filtered glossary listings for the documentary edition as the heart of our site, the Foreign Relations of the United States series. See, for example, the right sidebar labelled "Persons" and "Abbreviations & Terms" on http://history.state.gov/historicaldocuments/frus1969-76v18/d1. The information shown in these sidebars comes from each book's glossaries. The site-wide full text search is driven by eXist-db's lucene index. We also use eXist-db to drive our content management system (accessible only to internal users). Please feel free to respond with the URL of your eXist-db / TEI project, and a brief description of your project. Feel free to add your projects to the wiki link above too if it isn't already there. Cheers, Joe Sent from my iPad ------------------------------------------------------------------------------ Special Offer -- Download ArcSight Logger for FREE! Finally, a world-class log management solution at an even better price-free! And you'll get a free "Love Thy Logs" t-shirt when you download Logger. Secure your free ArcSight Logger TODAY! http://p.sf.net/sfu/arcsisghtdev2dev _______________________________________________ eXist-TEIXML mailing list eXi...@li... https://lists.sourceforge.net/lists/listinfo/exist-teixml >> > > > -- > Peter Watson > ------------------------------------------------------------------------------ > Special Offer -- Download ArcSight Logger for FREE! > Finally, a world-class log management solution at an even better > price-free! And you'll get a free "Love Thy Logs" t-shirt when you > download Logger. Secure your free ArcSight Logger TODAY! > http://p.sf.net/sfu/arcsisghtdev2dev_______________________________________________ > eXist-TEIXML mailing list > eXi...@li... > https://lists.sourceforge.net/lists/listinfo/exist-teixml |
|
From: Peter W. <plw...@bl...> - 2011-08-31 10:17:21
|
Hi Joe and Others
As a graduate of the 2010 TEI Summer School this list is a really
welcome development for me. My own project is a personal one supporting
my DPhil research on a set of twelfth and thirteenth century land deeds
for a family living in the depths of the English Midlands. I looked for
software that would allow me to place the deeds in a database using a
single copy of the transcribed texts for all purposes. TEI plus
eXist-db does this for me. I have developed the following functions,
apart from the html rendering which came from the summer school material
that Joe provided:
* TEI tagging of names, dates and places allows me to list and search
the database in a variety of structured ways.
* PDF printing from an fo transform based on XQuery, quite similar to
rendering in HTML.
* Visualization of social network data is under development using
NodeXL, externally at present, though long term I would like to be
able to integrate visualization into eXist.
* Integration with my bibliography based on Zotero. At present this
uses an RDF copy of my Zotero database saved into eXist so is not
completely real time. Ffrom eXist I can also gain access to the
Zotero data via SQLite, on which Zotero is based, though again this
uses a copy of the Zotero data.
* I have also transformedand imported an XML copy of a much larger
database of C12 &C13 charters from an Access database which allow
these to be accessed much more easily whilst also retaining the
ability to search the data that Access allows.
* I have the utilities to enable me to search an earlier database I
developed based on MediaWiki, which sits on an SQL database.
I'd be very keen to learn of any one else who is doing something similar
to this. At the moment all of this is on my laptop. Providing access to
what is a prototype is some way down the list of priorities, unless
there is a simple solution to getting the data hosted.
Best wishes
Peter
On 30/08/2011 23:40, Joe Wicentowski wrote:Hello all, Today I sent out
an announcement about this list to the broader TEI community via TEI-L.
Welcome to all who have just joined! To kick the discussion here off, I
think it might be useful for those new to eXist-db to learn about some
current TEI-based projects that employ eXist-db in some way. There is a
list of some eXist-db projects on the TEI wiki:
http://wiki.tei-c.org/index.php/EXist#Sample_implementations. The
project I work on (also listed on the wiki) is the Office of the
Historian website (http://history.state.gov/). The entire website is
driven by eXist-db, it has been in production since March 2009, and it
is hosted on Amazon EC2. The functions of eXist-db are used throughout
the site, and all article- and book-length content on the site is
encoded in TEI. All pages are rendered into HTML by eXist-db on the fly
as they are requested. Perhaps one of the most interesting portions from
a TEI perspective is how we use TEI person and term tags to drive
filtered glossary listings for the documentary edition as the heart of
our site, the Foreign Relations of the United States series. See, for
example, the right sidebar labelled "Persons" and "Abbreviations &
Terms" on
http://history.state.gov/historicaldocuments/frus1969-76v18/d1. The
information shown in these sidebars comes from each book's glossaries.
The site-wide full text search is driven by eXist-db's lucene index. We
also use eXist-db to drive our content management system (accessible
only to internal users). Please feel free to respond with the URL of
your eXist-db / TEI project, and a brief description of your project.
Feel free to add your projects to the wiki link above too if it isn't
already there. Cheers, Joe Sent from my iPad
------------------------------------------------------------------------------
Special Offer -- Download ArcSight Logger for FREE! Finally, a
world-class log management solution at an even better price-free! And
you'll get a free "Love Thy Logs" t-shirt when you download Logger.
Secure your free ArcSight Logger TODAY!
http://p.sf.net/sfu/arcsisghtdev2dev
_______________________________________________ eXist-TEIXML mailing
list eXi...@li...
https://lists.sourceforge.net/lists/listinfo/exist-teixml
--
Peter Watson
|
|
From: Joe W. <jo...@gm...> - 2011-08-30 22:40:54
|
Hello all, Today I sent out an announcement about this list to the broader TEI community via TEI-L. Welcome to all who have just joined! To kick the discussion here off, I think it might be useful for those new to eXist-db to learn about some current TEI-based projects that employ eXist-db in some way. There is a list of some eXist-db projects on the TEI wiki: http://wiki.tei-c.org/index.php/EXist#Sample_implementations. The project I work on (also listed on the wiki) is the Office of the Historian website (http://history.state.gov/). The entire website is driven by eXist-db, it has been in production since March 2009, and it is hosted on Amazon EC2. The functions of eXist-db are used throughout the site, and all article- and book-length content on the site is encoded in TEI. All pages are rendered into HTML by eXist-db on the fly as they are requested. Perhaps one of the most interesting portions from a TEI perspective is how we use TEI person and term tags to drive filtered glossary listings for the documentary edition as the heart of our site, the Foreign Relations of the United States series. See, for example, the right sidebar labelled "Persons" and "Abbreviations & Terms" on http://history.state.gov/historicaldocuments/frus1969-76v18/d1. The information shown in these sidebars comes from each book's glossaries. The site-wide full text search is driven by eXist-db's lucene index. We also use eXist-db to drive our content management system (accessible only to internal users). Please feel free to respond with the URL of your eXist-db / TEI project, and a brief description of your project. Feel free to add your projects to the wiki link above too if it isn't already there. Cheers, Joe Sent from my iPad |
|
From: Jens Ø. P. <oe...@gm...> - 2011-08-30 14:14:43
|
Hi, oXygen 13 has come out, with (as usual) a number of useful new features, this time focussed on user interface improvements, many of which are relevant to eXist users. - See the full description at <http://www.oxygenxml.com/#new-version>. The Data Source Explorer view, which we use to see inside an eXist database, now features drag and drop capabilities between different databases. One can now drag files and folders from one's operating system, that is, from Windows Explorer and Mac OS Finder into eXist (drag and drop does not work the other way round), thereby copying resources into eXIst. One can also drag and drop files and collections between different collections inside eXist, even across different eXist instances. Note that this does not amount to copying, but to moving, that is, the resource is deleted at the origin (a dialog pops up, asking you to confirm your decision to move), however, if you drag and then press Ctrl, the resource is copied, not moved. The distinction between copy and move it perhaps not implemented the way I would like, but these changes makes working with eXist considerably easier - WebDAV (or eXist's Java Admin Client) is still needed to make backups, but had oXygen 13 been out during the Summer School, we could have skipped quite a few steps, just dragging Joe's files inside eXist! One can also drag and drop from the Project Tree. This also amounts to copying, not moving. In the Data Source Explorer one is also supposed to be able to use keyboard shortcuts to copy, cut and paste in the usual way, but this I cannot confirm on Mac OS (and I am happy that this is so, for I believe this is rather dangerous). There are also a few TEI-specific changes, among them built-in transformation scenarios to convert TEI documents to and from DOCX. I have not been able to determine who is eligible for upgrades, but I have a version 12 license from April and I am able to run version 13 with it. I assume there will be a mail about upgrades from oXygen soon. As usual, you can request a 30-day trial. There is also a 30% discount on upgrades until September 15, 2011: <http://www.oxygenxml.com/buy.html#_upgrade>, but note that this does not affect Academic Licenses, which is probably what most of us have. Cheers, Jens |
|
From: Joe W. <jo...@gm...> - 2011-08-24 12:57:38
|
Hi all, One important bit of news since the seminar at digital.humanities@oxford is that eXist 1.4.1 has been released. Compared to the 1.4.0 version of eXist that we used in our seminar, the new version is improved in many ways -- belying its small ".0.1" bump in version number. You can read more about the release at http://atomic.exist-db.org/blogs/eXist/Release141. I strongly recommend upgrading to the new version. Cheers, Joe |
|
From: Joe W. <jo...@gm...> - 2011-08-23 19:27:01
|
Hi all, Welcome to everyone who has joined since the announcement about this list went out today to the digital.humanities@oxford (#dhox) participants. Before long we'll also welcome participants via other lists and channels, but until then, for the first week or so, I wanted to ensure that all of you who attended #dhox are able to send any questions you have. As this morning's email from dhox said, the slides and sample files from my presentation have been posted at http://digital.humanities.ox.ac.uk/DHSS2011/sessions.html#xmldb. Please feel free to download the files and send any comments or questions. Cheers, Joe |
|
From: Joe W. <jo...@gm...> - 2011-08-09 03:06:12
|
Hello, and welcome to the eXist-TEIXML mailing list! Inspired by the "Introduction to XML Databases" seminar at the recent Digital.Humanities@Oxford Summer School 2011 (http://digital.humanities.ox.ac.uk/DHSS2011/sessions.html#xmldb), this mailing list aims to fill the need for a discussion list for the community of users of TEI and eXist-db. There is a growing community of digital humanists who have learned TEI to encode their documents. They want to create dynamic, searchable websites to display and interact with their documents and share them with others. eXist-db, a free, open source XML database, offers an excellent platform for doing this. In fact, eXist-db is already popular in the TEI community, and TEI is popular in the eXist-db community. Both communities have their own mailing lists, but they cover a broad range of topics. This list exists to cover the intersection of eXist-db, TEI, and digital humanities. Thanks to Wolfgang Meier and Sourceforge for graciously hosting this mailing list. Joseph Wicentowski and Jens Østergaard Petersen, moderators |
|
From: Joe W. <jo...@gm...> - 2011-08-09 02:11:19
|
Hello, and welcome to the eXist-TEIXML mailing list! Inspired by the "Introduction to XML Databases" seminar at the recent Digital.Humanities@Oxford Summer School 2011 (http://digital.humanities.ox.ac.uk/DHSS2011/sessions.html#xmldb), this mailing list aims to fill the need for a discussion list for the community of users of TEI and eXist-db. There is a growing community of digital humanists who have learned TEI to encode their documents. They want to create dynamic, searchable websites to display and interact with their documents and share them with others. eXist-db, a free, open source XML database, offers an excellent platform for doing this. In fact, eXist-db is already popular in the TEI community, and TEI is popular in the eXist-db community. Both communities have their own mailing lists, but they cover a broad range of topics. This list exists to cover the intersection of eXist-db, TEI, and digital humanities. Thanks to Wolfgang Meier and Sourceforge for graciously hosting this mailing list. Joseph Wicentowski and Jens Østergaard Petersen, moderators |