1.       The 150 GB consists of a few—multiple, but not many—datasets

2.       These are supplementary files for an ETD

3.       I have no idea what they actually are yet, other than that they are a hard drive that the graduate college folks should be bringing over soon

4.       Storage in DSpace is mainly for preservation, but the data may be critical to reading and understanding the dissertation






From: Benjamin Ryan [mailto:benjamin.ryan@manchester.ac.uk]
Sent: Thursday, August 30, 2012 2:19 PM
To: Ingram, William A; dspace-tech@lists.sourceforge.net
Subject: RE: [Dspace-tech] Ingesting large data set


My 2p worth,

1.       What does the 150Gb consist of – one data set, multiple data sets (that may be related e.g. time/geographic location)

2.       How would someone use this data set – at 150Gb I  would assume (hope) offline processing

3.       There are ways for GIS, Geo-spatial, Time-Series data that you can implement a “browse” functionality by “slicing/tiling” but very dependent on the data and what you want to ask about the data.

4.       Is your Dspace system a long-term archival storage with a focus on preservation?


Sorry for no answers, but willing to discuss questions.





Dr Ben Ryan
Jorum Technical Coordinator (Services)

5.12 Roscoe Building
The University of Manchester
Oxford Road
M13 9PL
Tel: 0160 275 6039


From: Ingram, William A [mailto:wingram2@illinois.edu]
Sent: 30 August 2012 16:54
To: dspace-tech@lists.sourceforge.net
Subject: [Dspace-tech] Ingesting large data set


I apologize if a similar questuon has been answered in a prior thread. 


We have a student needing to submit a 150 GB data set into DSpace. Is this even possible? Are there any tips or workarounds I should try?