http://www.ijdc.net/issue/feed International Journal of Digital Curation 2021-08-16T15:25:58+01:00 IJDC Editorial Team ijdc@mlist.is.ed.ac.uk Open Journal Systems <p>The IJDC publishes pre-prints, research papers, general articles and editorials on digital curation, research data management and related issues. &nbsp;It complements the International Conference on Digital Curation (IDCC) and includes selected proceedings as Conference Papers.</p> http://www.ijdc.net/article/view/742 Identifying Opportunities for Collective Curation During Archaeological Excavations 2021-07-13T03:38:50+01:00 Ixchel Faniel fanieli@oclc.org Anne Austin austinan@umsl.edu Sarah Whitcher Kansa sarahkansa@gmail.com Eric Kansa ekansa@alexandriaarchive.org Jennifer Jacobs jejacobs@mac.com Phoebe France phoebe.france@gmail.com <p>Archaeological excavations are comprised of interdisciplinary teams that create, manage, and share data as they unearth and analyse material culture. These team-based settings are ripe for collective curation during these data lifecycle stages. However, findings from four excavation sites show that the data interdisciplinary teams create are not well integrated. Knowing this, we recommended opportunities for collective curation to improve use and reuse of the data within and outside of the team.</p> 2021-04-18T22:33:52+01:00 ##submission.copyrightStatement## http://www.ijdc.net/article/view/735 Cross-tier Web Programming for Curated Databases: a Case Study 2021-07-13T03:38:50+01:00 Simon Fowler simon.fowler@glasgow.ac.uk Simon Harding simon.harding@igmm.ed.ac.uk Joanna Sharman sjsy@novonordisk.com James Cheney jcheney@inf.ed.ac.uk <p>Curated databases have become important sources of information across several scientific disciplines, and as the result of manual work of experts, often become important reference works. Features such as provenance tracking, archiving, and data citation are widely regarded as important features for the curated databases, but implementing such features is challenging, and small database projects often lack the resources to do so.</p> <p>A scientific database application is not just the relational database itself, but also an ecosystem of web applications to display the data, and applications which allow data curation. Supporting advanced curation features requires changing all of these components, and there is currently no way to provide such capabilities in a reusable way.</p> <p>Cross-tier programming languages allow developers to write a web application in a single, uniform language. Consequently, database queries and updates can be written in the same language as the rest of the program, and it should be possible to provide curation features via program transformations. As a step towards this goal, it is important to establish that realistic curated databases can be implemented in a cross-tier programming language.</p> <p>In this article, we describe such a case study: reimplementing the web frontend of a realworld scientific database, the IUPHAR/BPS Guide to PHARMACOLOGY (GtoPdb), in the Links cross-tier programming language. We show how programming language features such as language-integrated query simplify the development process, and rule out common errors. Through an automated functional correctness evaluation, we show that the Links implementation correctly implements the functionality of the official version. Through a comparative performance evaluation, we show that the Links implementation performs fewer database queries, while the time neededto handle the queries is comparable to the official Java version. Furthermore, while there is some overhead to using Links because of its comparative immaturity compared to Java, the Links version is usable as a proof-of-concept case study of cross-tier programming for curated databases.</p> 2021-04-19T23:10:48+01:00 ##submission.copyrightStatement## http://www.ijdc.net/article/view/746 Understanding the Data Management Plan as a Boundary Object through a Multi-stakeholder perspective 2021-07-13T03:38:48+01:00 Live Kvale live.kvale@oslomet.no Nils Pharo nils.pharo@oslomet.no <div class="WordSection1"> <p class="Abstract">A three-phase Delphi study was used to investigate an emerging community for research data management in Norway and their understanding and application of data management plans (DMPs). The findings reveal visions of what the DMP should be as well as different practice approaches, yet the stakeholders present common goals. This paper discusses the different perspectives on the DMP by applying Star and Griesemer’s theory of boundary objects (Star &amp; Griesemer, 1989). The debate on what the DMP is and the findings presented are relevant to all research communities currently implementing DMP procedures and requirements. The current discussions about DMPs tend to be distant from the active researchers and limited to the needs of funders and institutions rather than to the usefulness for researchers. By analysing the DMP as a boundary object, plastic and adaptable yet with a robust identity (Star &amp; Griesemer, 1989), and by translating between worlds where collaboration on data sharing can take place we expand the perspectives and include all stakeholders. An understanding of the DMP as a boundary object can shift the focus from shaping a DMP which fulfils funders’ requirements to enabling collaboration on data management and sharing across domains using standardised forms.</p> </div> 2021-07-04T22:16:55+01:00 ##submission.copyrightStatement## http://www.ijdc.net/article/view/684 Doctoral Students' Educational Needs in Research Data Management: Perceived Importance and Current Competencies 2021-08-11T21:30:28+01:00 Jukka Rantasaari jukka.rantasaari@utu.fi <div class="WordSection1"> <p class="Abstract">Sound research data management (RDM) competencies are elementary tools used by researchers to ensure integrated, reliable, and re-usable data, and to produce high quality research results. In this study, 35 doctoral students and faculty members were asked to self-rate or rate doctoral students’ current RDM competencies and rate the importance of these competencies. Structured interviews were conducted, using close-ended and open-ended questions, covering research data lifecycle phases such as collection, storing, organization, documentation, processing, analysis, preservation, and data sharing. The quantitative analysis of the respondents’ answers indicated a wide gap between doctoral students’ rated/self-rated current competencies and the rated importance of these competencies. In conclusion, two major educational needs were identified in the qualitative analysis of the interviews: to improve and standardize data management planning, including awareness of the intellectual property and agreements issues affecting data processing and sharing; and to improve and standardize data documenting and describing, not only for the researcher themself but especially for data preservation, sharing, and re-using. Hence the study informs the development of RDM education for doctoral students.</p> </div> 2021-08-09T16:33:03+01:00 ##submission.copyrightStatement## http://www.ijdc.net/article/view/689 Futureproofing Visual Effects 2021-08-16T15:25:58+01:00 Evanthia Samaras evanthia.samaras@student.uts.edu.au <p class="Abstract">Digital visual effects (VFX), including computer animation, have become a commonplace feature of contemporary episodic and film production projects. Using various commercial applications and bespoke tools, VFX artists craft digital objects (known as “assets”) to create visual elements such as characters and environments, which are composited together and output as shots.</p> <p class="Abstract">While the shots that make up the finished film or television (TV) episode are maintained and preserved within purpose-built digital asset management systems and repositories by the studios commissioning the projects; the wider VFX network currently has no consistent guidelines nor requirements around the digital curation of VFX digital assets and records. This includes a lack of guidance about how to effectively futureproof digital VFX and preserve it for the long-term.</p> <p class="Abstract">In this paper I provide a case study – a single shot from a 3D animation short film – to illustrate the complexities of digital VFX assets and records and the pipeline environments whence they are generated. I also draw from data collected from interviews with over 20 professional VFX practitioners from award-winning VFX companies, and I undertake socio-technical analysis of VFX using actor-network theory. I explain how high data volumes of digital information, rapid technology progression and dependencies on software pose significant preservation challenges.</p> <p>In addition, I outline that by conducting holistic appraisal, selection and disposal activities across their entire digital collections, and by continuing to develop and adopt open formats; the VFX industry has improved capability to preserve first-hand evidence of their work in years to come.</p> 2021-08-15T23:01:26+01:00 ##submission.copyrightStatement## http://www.ijdc.net/article/view/678 Assessment, Usability, and Sociocultural Impacts of DataONE 2021-07-13T03:38:50+01:00 Robert J. Sandusky sandusky@uic.edu Suzie Allard sallard@utk.edu Lynn Baird a.whyte@ed.ac.uk Leah Cannon a.whyte@ed.ac.uk Kevin Crowston crowston@syr.edu Amy Forrester cci@utk.edu Bruce Grant bwgrant@widener.edu Rachael Hu Rachael.Hu@ucop.edu Robert Olendorf rko5039@psu.edu Danielle Pollock danielle.pollock@simmons.edu Alison Specht a.specht@uq.edu.au Carol Tenopir ctenopir@utk.edu Rachel Volentine uxlab@utk.edu <p class="Abstract">DataONE, funded from 2009-2019 by the U.S. National Science Foundation, is an early example of a large-scale project that built both a cyberinfrastructure and culture of data discovery, sharing, and reuse. DataONE used a Working Group model, where a diverse group of participants collaborated on targeted research and development activities to achieve broader project goals. This article summarizes the work carried out by two of DataONE’s working groups: Usability &amp; Assessment (2009-2019) and Sociocultural Issues (2009-2014). The activities of these working groups provide a unique longitudinal look at how scientists, librarians, and other key stakeholders engaged in convergence research to identify and analyze practices around research data management through the development of boundary objects, an iterative assessment program, and reflection. Members of the working groups disseminated their findings widely in papers, presentations, and datasets, reaching international audiences through publications in 25 different journals and presentations to over 5,000 people at interdisciplinary venues. The working groups helped inform the DataONE cyberinfrastructure and influenced the evolving data management landscape. By studying working groups over time, the paper also presents lessons learned about the working group model for global large-scale projects that bring together participants from&nbsp;multiple disciplines and communities in convergence research.</p> 2021-04-18T22:10:55+01:00 ##submission.copyrightStatement## http://www.ijdc.net/article/view/592 Improving the Usability of Organizational Data Systems 2021-07-13T03:38:49+01:00 Chung-Yi Hou hou@ucar.edu Matthew S. Mayernik mayernik@ucar.edu <p>For research data repositories, web interfaces are usually the primary, if not the only, method that data users have to interact with repository systems. Data users often search, discover, understand, access, and sometimes use data directly through repository web interfaces. Given that sub-par user interfaces can reduce the ability of users to locate, obtain, and use data, it is important to consider how repositories’ web interfaces can be evaluated and improved in order to ensure useful and successful user interactions. This paper discusses how usability assessment techniques are being applied to improve the functioning of data repository interfaces at the National Center for Atmospheric Research (NCAR). At NCAR, a new suite of data system tools is being developed and collectively called the NCAR Digital Asset Services Hub (DASH). Usability evaluation techniques have been used throughout the NCAR DASH design and implementation cycles in order to ensure that the systems work well together for the intended user base. By applying user study, paper prototype, competitive analysis, journey mapping, and heuristic evaluation, the NCAR DASH Search and Repository experiences provide examples for how data systems can benefit from usability principles and techniques. Integrating usability principles and techniques into repository system design and implementation workflows helps to optimize the systems’ overall user experience.</p> 2021-05-18T22:37:05+01:00 ##submission.copyrightStatement## http://www.ijdc.net/article/view/741 Leveraging Existing Technology: Developing a Trusted Digital Repository for the U.S. Geological Survey 2021-07-13T03:38:48+01:00 Vivian B. Hutchison vhutchison@usgs.gov Tamar Norkin tnorkin@usgs.gov Madison L. Langseth mlangseth@usgs.gov Drew A. Ignizio dignizio@usgs.gov Lisa S. Zolly lisa_zolly@usgs.gov Ricardo McClees-Funinan vhutchison@usgs.gov Amanda Liford aliford@usgs.gov <div class="WordSection1"> <p class="Abstract">As Federal Government agencies in the United States pivot to increase access to scientific data (Sheehan, 2016), the U.S. Geological Survey (USGS) has made substantial progress (Kriesberg et al., 2017). USGS authors are required to make federally funded data publicly available in an approved data repository (USGS, 2016b). This type of public data product, known as a USGS data release, serves as a method for publishing reviewed and approved data. In this paper, we present major milestones in the approach the USGS took to transition an existing technology platform to a Trusted Digital Repository. We describe both the technical and the non-technical actions that contributed to a successful outcome.We highlight how initial workflows revealed patterns that were later automated, and the ways in which assessments and user feedback influenced design and implementation. The paper concludes with lessons learned, such as the importance of a community of practice, application programming interface (API)-driven technologies, iterative development, and user-centered design. This paper is intended to offer a potential roadmap for organizations pursuing similar goals.</p> </div> <p>&nbsp;</p> 2021-07-11T23:00:28+01:00 ##submission.copyrightStatement## http://www.ijdc.net/article/view/674 Data Curation, Fisheries, and Ecosystem-based Management: the Case Study of the Pecheker Database 2021-07-13T03:38:49+01:00 Alexis Martin alexis.martin@mnhn.fr Charlotte Chazeau alexis.martin@mnhn.fr Nicolas Gasco alexis.martin@mnhn.fr Guy Duhamel alexis.martin@mnhn.fr Patrice Pruvost alexis.martin@mnhn.fr <div class="WordSection1"> <p class="Abstract">The scientific monitoring of the Southern Ocean French fishing industry is based on the use the Pecheker database. Pecheker is dedicated to the digital curation of the data collected on field by scientific observers and which analysis allows the scientists of the Muséum national d’Histoire naturelle institution to provide guidelines and advice for the regulation of the fishing activity, the protection of the fish stocks and the protection of the marine ecosystems. The template of Pecheker has been developed to make the database adapted to the ecosystem-based management concept. Considering the global context of biodiversity erosion, this modern approach of management aims to take account of the environmental background of the fisheries to ensure their sustainable development. Completeness and high quality of the raw data is a key element for an ecosystem-based management database such as Pecheker. Here, we present the development of this database as a case study of fisheries data curation to be shared with the readers. Full code to deploy a database based on the Pecheker template is provided in supplementary materials. Considering the success factors we could identify, we propose a discussion about how the community could build a global fisheries information system based on a network of small databases including interoperability standards.</p> </div> 2021-06-07T21:44:06+01:00 ##submission.copyrightStatement## http://www.ijdc.net/article/view/650 Scaling by Optimising: Modularisation of Data Curation Services in Growing Organisations 2021-07-13T03:38:50+01:00 Hagen Peukert hagen.peukert@uni-hamburg.de <p class="Abstract">After a century of theorising and applying management practices, we are in the middle of entering a new stage in management science: digital management. The management of digital data submerges in traditional functions of management and, at the same time, continues to recreate viable solutions and conceptualisations in its established fields, e.g. research data management. Yet, one can observe bilateral synergies and mutual enrichment of traditional and data management practices in all fields. The paper at hand addresses a case in point, in which new and old management practices amalgamate to meet a steadily, in part characterised by leaps and bounds, increasing demand of data curation services in academic institutions. The idea of modularisation, as known from software engineering, is applied to data curation workflows so that economies of scale and scope can be used. While scaling refers to both management science and data science, optimising is understood in the traditional managerial sense, that is, with respect to the cost function. By means of a situation analysis describing how data curation services were applied from one department to the entire institution and an analysis of the factors of influence, a method of modularisation is outlined that converges to an optimal state of curation workflows.</p> 2021-04-26T22:48:45+01:00 ##submission.copyrightStatement## http://www.ijdc.net/article/view/770 How Long Can We Build It? Ensuring Usability of a Scientific Code Base 2021-07-13T03:38:49+01:00 Klaus Rechert klaus.rechert@rz.uni-freiburg.de Jurek Oberhauser jurek.oberhauser@gmx.net Rafael Gieschke rafael.gieschke@rz.uni-freiburg.de <p>Software and in particular source code became an important component of scientific publications and henceforth is now subject of research data management.&nbsp; Maintaining source code such that it remains a usable and a valuable scientific contribution is and remains a huge task. Not all code contributions can be actively maintained forever. Eventually, there will be a significant backlog of legacy source-code. In this article we analyse the requirements for applying the concept of long-term reusability to source code. We use simple case study to identify gaps and provide a technical infrastructure based on emulator to support automated builds of historic software in form of source code.</p> <p>&nbsp;</p> 2021-05-17T12:21:39+01:00 ##submission.copyrightStatement##