Herbert Van de Sompel
Anastasia Dimou
Sarven Capadisli
Staff at academic institutions are regularly called upon to produce and maintain lists of scholarly publications authored by individuals ranging from principal investigators to people affiliated with other institutions such as alumni and residents. This work tends to be done on an ad hoc basis and is time-consuming, especially when profiled individuals have common names. Often, feedback from the authors themselves is not adequately captured in some central location and repurposed for future requests. ReCiter is a highly accurate, machine learning system for inferring which publications in PubMed a given person has authored. ReCiter includes a Java application, a DynamoDB-hosted database, and a set of RESTful microservices which collectively allow institutions to maintain accurate and up-to-date author publication lists for thousands of people. This software is optimized for disambiguating authorship in PubMed and, optionally, Scopus. ReCiter rapidly and accurately identifies articles, including those at previous affiliations, by a given person. It does this by leveraging institutionally maintained identity data (e.g., departments, relationships, email addresses, year of degree, etc.) With the more complete and efficient searches that result from combining these types of data, individuals at institutions can save time and be more productive. Running ReCiter daily, one can ensure that the desired users are the first to learn when a new publication has appeared in PubMed. ReCiter is freely available and open source under the Apache 2.0 license. https://github.com/wcmc-its/ReCiter At the end of this workshop, participants will: * Understand the problem of publication management and author disambiguation * See a demo of ReCiter and Publications Manager as it is used at Weill Cornell * Understand at a high level the service-based architecture of ReCiter * Learn how to install ReCiter to run on a local machine * Learn how to easily install ReCiter on Amazon Web Services (AWS) and all its dependencies using a ready-to-use CloudFormation template * Learn how to configure ReCiter In order to maximize the utility of the workshop, we ask that participants complete several tasks ahead of time including install several dependencies, create a free AWS account, and request an API key for PubMed. Additionally, if you wish to try ReCiter against profiles from your institution. For more information, view here: https://shorturl.at/STW14
Paul Albert , Weill Cornell Medicine
Sarbajit Dutta , Weill Cornell Medicine
The Web of Science (WoS) is a trusted source for publication and citation metadata of scholarly works dating back to 1900. The multidisciplinary database covers all areas of science, as well as social sciences, and the arts and humanities. WoS is comprised of works published in over 20,000 journals, as well as books and conferences. In 2019, the Clarivate Web of Science Group will release a new RESTful API that makes accessing and reusing citation metadata easier than ever. In this workshop, participants will be introduced to the new WoS APIs, the metadata available, and the new API registration process. Workshop participants will also gain hands-on experience using two Python script libraries, wos2vivo and incites2vivo. wos2vivo is an open source Python library for easily querying the Web of Science for your institution’s publications in bulk and transforming the data into VIVO-compatible linked data. incites2vivo will add indicator flags from InCites, such as if a publication is a Hot Paper, Industry Collaboration, International Collaboration, or Open Access. Lastly, workshop attendees will learn how to embed dynamically updated citation counts for their publications on their VIVO. This technical workshop is appropriate for both beginning and advanced users. Please bring a laptop with Python installed. While a subscription is required for access to the Web of Science, all participants will be provided with temporary API credentials for the workshop.
View presentationBenjamin Gross , Web of Science Group
What makes current research information (CRIS) systems different from other research information systems used around the world? And what does this mean for VIVO? This workshop introduces CRISs to VIVO users in both the US, Europe and beyond. We’ll compare institutional goals for managing research information and show how they drive the selection of tools, platforms and systems as well as public implementations of these platforms. We’ll explore a number of VIVO implementations that consume and display CRIS data brilliantly. And we’ll get an update on the collaboration between euroCRIS and the VIVO community to map and load CERIF data to VIVO. Workshop agenda: *. Introduction to CRISs: ** How CRISs differ from other systems like repositories and grants systems (pre-award and post-award) ** CRIS use cases in Europe (for reporting, for management support, for generating profiles and CV’s, for managing and archiving research data, and more) *. Pairing VIVO and CRISs: the benefits and challenges of creating and maintaining VIVOs based on institutional CRIS implementations *. Promoting and facilitating interoperability: using standards to make the exchange of data between CRIS and VIVO easier and more efficient. Update on: the CERIF2VIVO mapping project.
View presentations: Link 1 Link 2Pablo de Castro , University of Strathclyde
Anna Guillaumet , SIGMA
Michele Minnielli , DuraSpace
Ed Simons , Radboud University Nijmegen
Julia Trimmer , Duke University
VIVO is an open source tool based on linked open data concepts for connecting and publishing a wide range of research information within and across institutions. The goal of this workshop is to introduce new community members to the VIVO project and software. The workshop will consist of three sections: 1. A summary of VIVO’s history and what it does, 2. How it works, and 3. Where the project is heading. Part 1 will include a background of the VIVO project, how institutions and organizations are currently using it, how institutional stakeholders are involved, what benefits it offers to researchers, to institutions, and to the global community. Part 2 will include a high level discussion about how VIVO works and introduce the concepts of Resource Description Framework, Ontology, Vitro, Triple Pattern Fragments, how VIVO is managed, and how to feed downstream systems. Finally, part 3 will introduce you to next-gen VIVO initiatives such as, decoupling of the architecture, the next version of the ontology, the VIVO Scholar, VIVO Combine, the internationalization efforts, and the VIVO Search. You’ll learn how to find the right resources as a new VIVO implementer including data sources, team members, governance models, and support structures. This workshop brings best practices and “lessons learned” from mature VIVO projects to new implementations. We’ll help you craft your messages to different stakeholders, so you’ll leave this workshop knowing how to talk about VIVO to everyone from your provost to faculty members to web developers.
View presentationBenjamin Gross , Web of Science Group
Violeta Ilik , Columbia University
Michael Conlon , VIVO Project
Violeta Ilik , Columbia University
Michael Conlon , VIVO Project
Nataša Popović , University of Montenegro
Ruben Verborgh , Gent University
The invitation to present a keynote at the VIVO Conference and the goal of the VIVO platform, as stated on the DuraSpace site, to create an integrated record of the scholarly work of an organisation reminded me of various efforts that I have been involved in over the past years that had similar goals. EgoSystem (2014) attempted to gather information about postdocs that had left the organisation, leaving little or no contact details behind. Autoload (2017), an operational service, discovers papers by organisational researchers in order to upload them in the institutional repository. myresearch.institute (2018), an experiment that is still in progress, discovers artefacts that researchers deposit in web productivity portals and subsequently archives them. More recently, I have been involved in thinking about the future of NARCIS, a portal that provides an overview of research productivity in The Netherlands. The approach taken in all these efforts share a characteristic motivated by a desire to devise scalable and sustainable solutions: let machines rather than humans do the work. In this talk, I will provide an overview of these efforts, their motivations, the challenges involved, and the nature of success (if any).
View presentationHerbert Van de Sompel Chief Innovation Officer, Data Archiving and Networked Services
Dr. Herbert Van de Sompel graduated in Mathematics and Computer Science at Ghent University (Belgium), and in 2000 obtained a Ph.D. in Communication Science there. He is currently Chief Innovation Officer at Data Archiving and Networked Services (DANS) in The Netherlands. He has previously held positions as head of Library Automation at Ghent University, Visiting Professor in Computer Science at Cornell University, Director of e-Strategy and Programmes at the British Library, and information scientist at the Research Library of the Los Alamos National Laboratory where he was the team leader of the Prototyping Team. Herbert has played a major role in creating the Open Archives Initiative Protocol for Metadata Harvesting (OAI-PMH), the Open Archives Initiative Object Reuse & Exchange specifications (OAI-ORE), the OpenURL Framework for Context-Sensitive Services (ANSI/NISO Z39.88-2004), the SFX linking server, the bX scholarly recommender service, info URI (RFC 4452), Open Annotation (W3C Community Group specification), ResourceSync (ANSI/NISO Z39.99-2014), Memento "time travel for the Web" (RFC 7089), Robust Links, and Signposting the Scholarly Web.
In this talk we discuss the characteristics of Linked Data-based resource discovery and its limitations in finding content by direct processing of information resources -- in comparison to the solution provided through content metadata supported by knowledge organization systems (KOS) such as thesauri, classification and subject descriptors. KOS are traditional information discovery tools that determine the meaning and control ambiguities of language, hence they are often referred to as controlled vocabularies. They are used by libraries as well as by publishers and bookshops. However, most of KOSs are designed for and used in traditional information environments and are often not readily accessible by programs. Semantic technologies such as linked data offer solutions for expressing KOS in a more formalized and machine-understandable way. They provide a way of uniquely identifying and contextualizing semantically meaningful units irrespective of their possible linguistic or symbolic representations. This “unique identification” (URI) is the key element of linked data technology: anything that can be identified can be linked. The publishing of KOSs as linked data has become the most important form of sharing and using controlled vocabularies in the Web environment. This is also a solution for accessing the meaning and knowledge stored in the collections indexed by KOS (both directly or indirectly). As more and more KOSs are being published as linked data and more and more collection metadata containing KOS concepts join the linked data cloud, some obstacles to linking collection metadata and KOSs have become more obvious. Human knowledge is in constant flux and KOSs develop over time to embrace new terminology and new fields of knowledge. These changes affect unique identifiers used in KOS and consequently all links between KOS and resource collections.
Ronald Siebes , DANS
Aida Slavic , UDC Consortium
Andrea Scharnhorst , Royal Netherlands Academy of Arts and Sciences
Conferences are an essential part of scholarly communication. However, like researchers and organizations they suffer from the disambiguation problem, when the same acronym or the conference name refers to very different conferences. In 2017, Crossref and DataCite started a working group on conference and project identifiers. The group includes various publishers, A&I service providers, and other interested stakeholders. The group participants have drafted the metadata specification and gathered the feedback from the community. In this talk, we would like to update the VIVO participants with where we stand with the PIDs for conferences, conference series and Crossmark for proceedings and are inviting the broader community to comment. Read the CrossRef post for more info about the group: https://www.crossref.org/working-groups/conferences-projects/
View presentationAliaksandr Birukou , Springer Nature
Patricia Feeney , Crossref
The VIVO Scholar Task Force is creating a new, read-only front-end for VIVO. Come hear an update about the work on VIVO Scholar so far. Task force representatives will demo components and answer questions. * Learn how five universities have worked together to reach the current stage of VIVO Scholar. * Review the new profile and search pages. * Watch the quick and easy GraphQL queries. * See how sharing data makes your VIVO essential. For more info on VIVO Scholar, see the Task Force page on the VIVO wiki.
Paul Albert , Weill Cornell Medicine
Sarbajit Dutta , Weill Cornell Medicine
Julia Trimmer , Duke University
Greg Burton , Duke University
Don Elsborg , University of Colorado Boulder
Hans Harlacher , Duke University
Damaris Murry , Duke University
Robert Nelson , Duke University
Ralph O'Flinn , The University of Alabama at Birmingham
Richard Outten , Duke University
Harry Thakkar , Duke University
Jim Wood , Duke University
Alex Viggio , University of Colorado Boulder
There are numerous sources of metadata regarding research activity that Clinical and Translational Science Award (CTSA) hubs currently duplicate effort in acquiring, linking and analyzing. The Science of Translational Science (SciTS) project provides a shared data platform for hubs to collaboratively manage these resources, and avoid redundant effort. In addition to the shared resources, participating CTSA hubs are provided private schemas for their own use, as well as support in integrating these resources into their local environments. This project builds upon multiple components completed in the first phase of the Center for Data to Health (CD2H), specifically: a) data aggregation and indexing work of research profiles and their ingest into and improvements to CTSAsearch by Iowa (http://labs.cd2h.org/search/facetSearch.jsp); b) NCATS 4DM, a map of translational science; and c) metadata requirements analysis and ingest of from a number of other CD2H and CTSA projects, including educational resources from DIAMOND and N-lighten, development resources from GitHub, and data resources from DataMed (bioCADDIE) and DataCite. This work also builds on other related work on data sources, workflows, and reporting from the SciTS team, including entity extraction from the acknowledgement sections of PubMed Central papers, disambiguated PubMed authorship, ORCiD data and integrations, NIH RePORT, Federal RePORTER, and other data sources and tools. Early activities for this project include: * Configuration of a core warehouse instance * Population of the warehouse from the above-mentioned sources * Configuration of local schemas for each CTSA hub and other interested parties * Creation for example solutions for ingest/extraction using JDBC, GraphQL, SPARQL, and tools such as teiid (an open source data federation platform).
David Eichmann , University of Iowa
Kristi Holmes , Northwestern University
Prizes are important indicators of esteem in research, and they deserve a persistent primary record of their own. * Award citation information is needed throughout the sector, all the time To name a few examples, institutions aggregate prizes from their alumni over time to build a story about the minds they have educated, and how welcoming their research environment is to support creativity. Prizes are built into university rankings and accreditation processes. To tell these stories easily, award citation information needs to be easily available. * Award citations should be richly described records An award citation is more than just a date, award, and link to a person and awarding body. A citation links to the research that it acknowledges. Upon acceptance award, often an occasional speech is recorded. The best way to capture award citations in all of the richness they deserve is to establish normative metadata practices based around the minting of a persistent identifier. * Award citations are the historical signposts through which society understands research progress. These signposts deserve a permanent digital record. * Creating transparency around on prizes can help improve research culture At their best, prizes recognise a diversity of research achievement in society from literature to physics and everything in between. It has also been observed that prizes are being awarded to a concentrated set of elite researchers. By making prize awardee information more discoverable, more informed decisions about what prizes to award, and who to award them to can be made. * The flow of prize information through the research systems is currently significantly hampered. It needs fixing. Wikidata is perhaps the best secondary source of prize information. Consider how it gets there. What information does it loose along the way? A significant amount work could be reduced by building information flows around the authority that persistent records provide. To begin to address these issues, we have built an open source awards publishing reference implementation. This implementation is based on on xPub - a Journal submission and peer review platform from the Collaborative Knowledge Foundation. Finalised award records are then published to figshare, with associated metadata pushed to wikidata.
Simon Porter , Digital Science
CRIS systems are becoming a mandatory element of University Information ecosystems. CRIS provides significant amenities for reporting routines for research councils, national authorities, employee evaluation, academic degrees communities and provide valuable insights for university management. Properly implemented (in terms of Organization) CRIS system becomes a source of complete and reliable data going far beyond projects and publications thus providing more profiling capabilities than for instance crowd-based systems like ResearchGate. Developing OMEGA-PSIR CRIS system at Warsaw University of Technology, which is now free and Open System used in 20+ Universities in Poland we thought about research visibility at the same priority as reporting. By applying Text Mining, Artificial Intelligence, external ontologies but also organizational regimes, we obtained a robust profiling systems that allows searching for experts, modelling and visualising research teams, intelligent discipline matching, multi-purpose and multi-source researchers rankings, and many others. We would like to share our experiences in building a robust RPS system that relies on complete and reliable data, basing on 20+ deployments of Omega-PSIR Open Software in Poland
Jakub Koperwas , Warsaw University of Technology
Łukasz Skonieczny , Warsaw University of Technology
Wacław Struk , Warsaw University of Technology
Henryk Rybiński , Warsaw University of Technology
The Semantic Web captures knowledge by making, among others, research data discoverable, accessible and understandable in the long term by increasing their sherability, extensibility and reusability. However, the process of extracting, structuring, and organizing knowledge from one or multiple heterogeneous data sources to construct knowledge-intensive systems has proven to be easier said than done. During this talk, I will elaborate on knowledge graph generation from heterogeneous data sources, as well as on the assessment of not only the knowledge graphs but also the rules that generate a knowledge graph and on the refinement of both rules and knowledge graphs.
Anastasia Dimou Senior post-doctoral researcher, IDLab, Ghent University
Dr. Anastasia Dimou is a senior/post-doctoral researcher at imec and IDLab, Ghent University. Anastasia joined the IDLab research group in February 2013. Her research interests include Linked Data Generation and Publication, Data Quality and Integration, Knowledge Representation and Management. As part of her research, she investigated a uniform language to describe rules for generating high quality Linked Data from multiple heterogeneous data sources. Anastasia currently conducts research on automated Linked Data generation and publication workflows, data validation and quality assessment, query-answering and knowledge integration from Big stream data. Her research activities are applied in different domains, such as Internet of Things (IoT), manufacturing, media and advertisement and led to the development of the RML tool chain. She is involved in different national, bilaterals, and EU projects, authored several peer-reviewed publications presented at prominent conferences and journals such as ESWC, ISWC, JWS and SWJ, participated in several PCs and co-organized tutorials and workshops.
In early 2018, the VIVO Leadership group brought together parties from across the broader VIVO community to Duke University to discuss critical aspects of VIVO as both a product and a community. At the meeting, a number of working groups were created to do deeper work on a set of focus areas to help inform the VIVO leadership in taking steps toward the future growth of VIVO. One group was tasked with understanding the current perception of VIVO's governance and structure from effectiveness, to openness and inclusivity, to make recommendations to the VIVO Leadership group concerning key strengths to preserve and challenges that needed to be addressed. This session will report on the results of a survey run by the Governance and Structure Working Group in late 2018. We will engage with the audience to get reactions to the results, ensure that interpretations make sense to the wider group and discuss the next stages in finalising publication of the overall report of the committee.
Julia Trimmer , Duke University
Michael Conlon , VIVO Project
Kristi Holmes , Northwestern University
Daniel W Hook , Digital Science
Dean B Krafft , Cornell University
Mark P Newton , Boston University
Converis supports comprehensive tracking and support across the entire research lifecycle, including grant, award, and publication management. Built-in integrations with the Web of Science and InCites make it easy to manage your organization’s research activity. Converis is a powerful upstream tool to manage a confluence of data for a VIVO site. In this presentation, we will present improvements made to the Converis application in Converis 6, as well as improvements in the works. We will also discuss enhancements to the Converis to VIVO connector that allows VIVO to be automatically fed by an upstream Converis data source.
Benjamin Gross , Web of Science Group
Miguel Garcia , Web of Science Group
Your VIVO needs updated information. * Updated data means more traffic: visits and pageviews. * More traffic encourages faculty to update their information. Does this sound familiar? If you’re wondering how to build up this positive feedback loop for your institution’s VIVO, join us to review Duke University’s best practices for creating a “buzz” around Scholars@Duke. We’ll talk about: * Crafting effective messages * Choosing the right communications channels * Hosting events that attract faculty * Boosting SEO (or trying to) * Learning from users and measuring success And we'll give updates on our plans to launch a redesigned site, create a video series to build awareness, and improve strategic planning through analytics.
Julia Trimmer , Duke University
Hans Harlacher , Duke University
Lamont Cannon , Duke University
At the Technical University of Denmark VIVO is used as an internal research analytics platform providing the university’s researchers and research managers with easy-to-use analytical views on research output, impact, collaboration, etc. VIVO – combined with a good data source – provides a platform for high-quality, high-specificity and high-integrity analytics under the university’s own control. In the latest release, Web of Science data is supplemented with advanced indicators from InCites, and the triple store with a relational database management system and an elaborate caching system to achieve attractive response times. The WoS data is enhanced by local mapping/normalization of research department names, enabling department-level analytics. The DTU RAP is developed in collaboration between the university, the IT-consultants Vox Novitas and Ontocale and Clarivate Analytics, producer of WoS and InCites. All software is available as open source for anyone with similar needs and similar access to data.
Benjamin Gross , Web of Science Group
Christina Steensboe , Technical University of Denmark
Karen Hytteballe Ibanez , Technical University of Denmark
Mogens Sandfaer , Technical University of Denmark
Franck Falcoz ,
Brian Lowe , Ontocale
Texas A&M University Libraries has been using VIVO in production since 2015. In that time, we have come up with many creative solutions to meet the needs of our users. In early 2019, we began developing a replacement front end for the VIVO interface to formally address campus demands here at Texas A&M University. Initial requirements: * Align the technology stack as much as possible with the existing VIVO stack to assist with implementation by others if they choose especially smaller libraries. * The majority of the front end is customizable by others. * Read only UI. No updating back to the triple store. * All data is retrieved via a REST API endpoint using Spring Data for Apache Solr. * 100% Search Engine Optimization. IE: A person / crawler can disable JavaScript and still have the same experience. Server side, and Client side rendering if needed. A demo of current work can be found here. https://demos.library.tamu.edu/scholars-ui/
William S Welling , Texas A&M University
Dong Joon Lee , Texas A&M University
Douglas Hahn , Texas A&M University
Jason Savell , Texas A&M University
Kevin A Day , Texas A&M University
Ethel Mejia , Texas A&M University
Bruce E Herbert , Texas A&M University
VIVO at Osnabrück University is used to present scholarly activity especially third-party funded research projects to the interested public to improve transparency in research funding. In the early stage of development, the VIVO content was limited to research project key facts (e.g. abstract, related/selected publications, external project links, keywords) mainly extracted from the operational finance systems that is not intended to be used for presentation purposes. As VIVO strengths lay in searching, browsing and visualizing scholar activity for the wider public, the searchable content should not be limited to complex academic language as it is used in project abstracts or publications, a nonacademic person would not search for. Science goes Public - For more than 10 years the event series “Osnabrück Knowledge Forum” tries to make research more accessible. Every year, public is called upon to challenge Osnabrück University Professors by sending questions about any kinds of science topics to be answered within 4 minutes lightning talks in a precise, clear and entertaining way in front of the interested public. This »evening of knowledge» and its findings are showcased in a variety of media such as film sequences on YouTube or as an image brochure for University guests (PDF). All these formats do not connect content to researchers as VIVO connects researchers, projects or organizations with each other. In this presentation, we will show how social media content could be used to enrich research projects, researcher profils and research topics to make VIVO (even more) vivid!
Sonja Schulze , Osnabrück University
Dominik Feldschnieders , Osnabrück University
Kathrin Schnieders , Osnabrück University
Manuel Schwarz , Osnabrück University
Marco Seegers , Osnabrück University
How to best disseminate one’s research and get credit for one’s work? How to best and fairly assess the quality and impact of a given individual, group, or institution’s research? These are questions with which many are struggling, from individual researchers to departments, to a global world of research institutions. Recently, the Faculty Senate and University Libraries surveyed the faculty of our large, public research university to explore their perspective on these questions and more. In this presentation we present a summary of results from 501 respondents (out of 4451 faculty in total) representing different types of faculty (both within and outside of tenured and tenure-track positions), at different ranks, and from different disciplines. Results shared will indicate trends within the faculty on topics such as, the current most commonly used profile systems (top 5: Google Scholar, ORCID ID, LinkedIn, Elements (internal system), and ResearchGate); which profile systems are used most for: networking and connecting with colleagues (top 3: LinkedIn, Twitter, and ResearchGate), tracking research impact metrics (top 3: Google Scholar, ORCID, Elements (internal system)), showcasing one’s work to increase visibility (top 3:Google Scholar, ResearchGate, self-published sites); what types of research metrics are relied on (top 3: journal reputation (separate from impact factor), number of publications, and citation counts to individual works); the perceived fairness of evaluation by level of review (e.g., department, college, and university levels) and how they differ; and summaries of qualitative responses to questions such as why faculty rely on certain profile systems or research metrics, and perspectives on how fair research evaluation could be accomplished, within or across disciplines. Results will be summarized at the institutional level with breakout analysis of results from some disciplinary fields or other subsets. For us, these results from faculty across a range of disciplines will help inform institutional policy and practice discussions about research tracking and evaluation, such as a responsible research assessment policy. Results will also inform our in-process implementation of an institutional researcher profiles systems, and training offerings on disseminating research and assessing its impact. As movements such as DORA (Declaration on Research Assessment) and the Leiden Manifesto for Research Metrics demonstrate, faculty, institutions, and funders are re-examining the way metrics are used and methods for demonstrating impact. This presentation on a university-wide survey that includes summary data and the survey questions used offers an example that could be adapted and repeated elsewhere to gauge current practices and faculty perspectives on how to change or move forward with research assessment across a range of disciplines and levels within a large, research institution.
View presentationRachel A Miles , Virginia Tech
Amanda Mac Donald , Virginia Tech
Nathaniel D Porter , Virginia Tech
Virginia Pannabecker , Virginia Tech
Jim A Kuypers , Virginia Tech
Dong Joon Lee , Texas A&M University
Terrie Wheeler , Weill Cornell Medicine
What are some of the sociotechnical constraints and the effects of contemporary scholarly communication? How can we appropriate the Open Web Platform to facilitate an actor-centric scholarly ecosystem? In this talk, we discuss designing decentralised and socially-aware systems as well as their effects and artifacts.
View presentationSarven Capadisli Researcher, University of Bonn and TIB, Hannover
Sarven Capadisli is currently writing his PhD thesis with University of Bonn, and researches with TIB, Hannover. His research involves the Linked Research initiative and dokieli (a clientside editor for decentralised article publishing, annotations and social interactions).
Texas A&M has strategic initiatives focused on academic reputation and interdisciplinary research. The Scholars@TAMU team used campus needs associated with these initiatives to drive the evolution of the Texas A&M implementation of VIVO and associated library services. Our focus on meeting campus needs has led to strong faculty engagement with Scholars@TAMU as well as Provost and Dean buy-in for our programs. We discuss the strategic framework that guided Scholars@TAMU development and implementation. This framework may help provide a roadmap for future VIVO development as well as VIVO implementations at diverse universities.
Dong Joon Lee , Texas A&M University
Douglas Hahn , Texas A&M University
Ethel Mejia , Texas A&M University
Bruce E Herbert , Texas A&M University
The OPERA project (Open Research Analytics) is developing a pilot VIVO with data for all 8 Danish universities. The key data come from the Dimensions database, but data from the Danish national bibliometric and open access indicators are also integrated. This pilot VIVO will facilitate national research analytics including its dimensions of openness – using open concepts, open software and data that are as open as possible. This joint experimental platform will be used to explore aspects such as: publication output and impact, Open Science efforts, funding patterns, subject fields, gender patterns and both established and potential collaboration patterns within and outside Denmark as well as with partners in industry. Network analyses and visualizations will be integrated in the VIVO platform to complement more traditional analytics and visual elements and to support new ways of perceiving numbers, patterns and potentials. The Dimensions VIVO is a collaboration between Danish universities, the IT consultants Vox Novitas and Ontocale and Digital Science, the producer of Dimensions.
View presentationSimon Porter , Digital Science
Christina Steensboe , Technical University of Denmark
Franck Falcoz ,
Brian Lowe , Ontocale
Karen H Ibanez , Technical University of Denmark
Nikoline D Lauridsen , Technical University of Denmark
Marianne Gauffriau , Technical University of Denmark
Adrian Price , Technical University of Denmark
Anne L Høj , Technical University of Denmark
Kirsten K Kruuse , Technical University of Denmark
The Smithsonian Institution began planning a VIVO implementation in 2015 and in August 2018, launched Smithsonian Profiles to the public. The Smithsonian is a research institution that comprises a network of 19 museums, 8 research centers, and the National Zoo. In addition to public programs and exhibits, staff conduct research in a wide range of domains, with over 700 scholars and their sponsored fellows authoring over 2500 publications every year. For over 10 years, the Smithsonian Libraries has been collecting and managing these publications using a home-grown system called Smithsonian Research Online, which now feeds into Smithsonian Profiles. This presentation will discuss Smithsonian Profiles and its role at the Smithsonian Institution, and will touch on the data systems it interacts with both internally and externally. There will also be a focus on the challenges presented by implementing VIVO in a non-university setting, including policy and accessibility considerations, and issues with defining identifying eligible from a pool of hundreds.
Kristina Heinricy , Smithsonian Libraries
Alvin Hutchinson , Smithsonian Libraries
Suzanne Pilsk , Smithsonian Libraries
VIVO by default offers some data visualizations, which might be limited or lacking user needs in the scope of analytics and reporting. Often institutes using VIVO has to rely on manual information gathering for basic analytics and reporting despite the required information being present in their VIVO instance. The inclusion of Elasticsearch driver, for indexing, in VIVO enables institutions to directly use the data from VIVO for custom applications. The open-source tool Kibana, which is a plugin for Elasticsearch, is a platform for building curated visualizations and dashboards on the data in Elasticsearch indexes. This presentation highlights how custom fields from VIVO could be indexed in Elasticsearch and how interactive dashboards could be created in Kibana for data analytics and reporting purposes. Furthermore, the possibility of including curated dashboards from Kibana into VIVO would also be discussed.
Qazi Asim Ijaz Ahmad , Technische Informationsbibliothek (TIB) û German National Library of Science and Technology
Graham Triggs , Technische Informationsbibliothek (TIB) û German National Library of Science and Technology
Christian Hauschke , Technische Informationsbibliothek (TIB) û German National Library of Science and Technology
The growing complexity of the digital research environments and specially with the advance of the OpenScience that includes not only the OpenAccess, but OpenData, and the need to make the data FAIR, has highlighted the importance of evolving the way we search and visualize the research results with powerful tools. In this way, when the scientific information is correctly stored in the CRIS, it must also be easy to search and find, and must guide the discovery of the information. This is one of the goals of the SIGMA Strategic Plan for the Research area that, since the beginning of 2018, is working to improve the search tools and results for the scientific production of the researchers of an institution. This project began by evaluating the best ontology to translate the current data model for SIGMA CRIS, to a semantic model. To do this, we analyzed some of the semantic engines that exist and finally we decided to test the VIVO solution. We found that VIVO ontology fits more than 80% with the Spanish model for Research and took into account that VIVO is an OpenSource software supported by members and an ontology to represent the scholarship that is used by relevant universities, highly positioned in the international rankings and, finally, we also value the large community of members behind VIVO and Duraspace. For these reasons, SIGMA decided to join the DuraSpace community at the beginning of 2018, participating in the governance of the VIVO project. During this time SIGMA has collaborated with VIVO to define the 2019 Roadmap of the product, having now a calendar of sprints scheduled to carry out what was agreed in the Product direction. This calendar it’s aligned by the SIGMA side with its strategic plan. So, we have now an adapted SIGMA ontology based on the VIVO ontology and some tools that we will show in this presentation as examples of the VIVO ontology and tools, adapted to the Spanish science model.
View presentationAnna Guillaumet , SIGMA
A couple of years ago at Brown University we wrote a new frontend for the public interface of our VIVO installation. This new frontend provided a modern look and feel for searching and displaying information on Researchers at Brown[1] and was well received by researchers and the public in general. This year we are focusing on adding an editing interface to this application to allow researchers to easily add or update the information on their VIVO profile including publications, research statements, collaborators, appointments, and so on. In this presentation we'll show the general architecture of our VIVO installation, including the different services and applications that interact with it, we'll discuss the challenges during the development of the new editing features, and some of the gains that our approach has given us. The new editing interface is built on top of the Ruby on Rails application that we built for the public interface a couple of years ago. This application interacts with a typical REST API service that communicates with VIVO to submit the changes to the triplestore. One of the lessons that we've learned during our years using VIVO has been that we can preserve Linked Data in the backend (as VIVO natively does) while at the same time provide traditional REST API endpoints to allow other applications, written in a variety of programming languages, to consume and in this case update the information. This approach has the advantage that we can run sophisticated SPARQL queries against the triplestore (for example to generate networks graphs of collaborators to power visualizations) while at the same time isolate client applications written in Python and Ruby from the triplestore and RDF complexities, and instead, expose the data to those applications in ways that they can easily consume, for example via a REST API passing JSON back and forth. [1] https://vivo.brown.edu/
View presentationHector Correa , Brown University
Steven Mc Cauley , Brown University
For more than a decade, VIVO sites have been creating semantic data regarding scholarship that could be used to change how scholarly work is found and how expertise is assessed and compared. Previous work has attempted to centrally collect and normalize semantic data for search purposes. Other effort has used federated search across sites to provide simple access to profiles. Can we now consider how best to create a semantic cross-site search capability? Panelists will discuss the following questions: What is semantic search and how might it differ from other search paradigms? Should the approach be centralized, in which semantic data is brought together to a single provider of search functionality, decentralized in which data remains at rest and search functionality is localized, or should other approaches be considered? What are the roles for ontology, data and software provisioning in semantic search? How might technologies such as TPF, GraphQL, Schema.org, Solid, and others be leveraged? What is needed to create a semantic cross-site search capability for VIVO?
Herbert Van de Sompel Chief Innovation Officer, Data Archiving and Networked Services
Dr. Herbert Van de Sompel graduated in Mathematics and Computer Science at Ghent University (Belgium), and in 2000 obtained a Ph.D. in Communication Science there. He is currently Chief Innovation Officer at Data Archiving and Networked Services (DANS) in The Netherlands. He has previously held positions as head of Library Automation at Ghent University, Visiting Professor in Computer Science at Cornell University, Director of e-Strategy and Programmes at the British Library, and information scientist at the Research Library of the Los Alamos National Laboratory where he was the team leader of the Prototyping Team. Herbert has played a major role in creating the Open Archives Initiative Protocol for Metadata Harvesting (OAI-PMH), the Open Archives Initiative Object Reuse & Exchange specifications (OAI-ORE), the OpenURL Framework for Context-Sensitive Services (ANSI/NISO Z39.88-2004), the SFX linking server, the bX scholarly recommender service, info URI (RFC 4452), Open Annotation (W3C Community Group specification), ResourceSync (ANSI/NISO Z39.99-2014), Memento "time travel for the Web" (RFC 7089), Robust Links, and Signposting the Scholarly Web.
Sarven Capadisli Researcher, University of Bonn and TIB, Hannover
Sarven Capadisli is currently writing his PhD thesis with University of Bonn, and researches with TIB, Hannover. His research involves the Linked Research initiative and dokieli (a clientside editor for decentralised article publishing, annotations and social interactions).
Violeta Ilik , Columbia University
Michael Conlon , VIVO Project
Ruben Verborgh , Gent University
The diversity, energy and innovation in the VIVO community is inspirational. Community initiatives are strong, as are contributions back to the core application. One of the VIVO project's primary objectives is to channel, where appropriate, community effort into the core application towards agreed upon outcomes. We are delighted to say that this has been happening! This presentation will not attempt to detail all of the technical efforts over the past year, but will highlight a range of advancements and milestones accomplished since the previous VIVO conference. In the context of reviewing the year’s activity, this session is also intended to solicit feedback from attendees on technical and community initiatives and processes. At the end of 2018, the VIVO Leadership Group collected input from the community and created a "Statement of VIVO's Product Direction for 2019". This statement details four strategic initiatives: * Modernize the presentation layer of VIVO * Decouple the architecture * VIVO Combine * VIVO Search Following the publication of this statement, an architecturally-minded team representing distinct VIVO stakeholder constituencies was gathered for the purpose of developing architectural approaches required to address the direction of the project. The primary goal of the two-day face-to-face meeting was to assess and document a plan for improving the VIVO application architecture towards enabling and realizing the technical efforts defined in the "Statement of VIVO's Product Direction for 2019". This presentation will include a detailed status of the completed as well as planned development activities following from the decisions of the architectural meeting.
View presentationBenjamin Gross , Web of Science Group
Brian Lowe , Ontocale
Andrew Woods , DuraSpace
Duke University's VIVO-based implementation, Scholars@Duke, has become an essential tool for the maintenance and dissemination of scholarly work within our institution. In addition to VIVO, Scholars@Duke includes user-friendly editing options that make it easy for collaborators to co-maintain a single scholarly public record. Before any records are sent to VIVO, our in-house editor, Profile Manager, along with Symplectic Elements, work together to create workflows that manage the rights and display preferences of multiple collaborators on a single record. In this presentation, I'll go through some of the solutions we've implemented regarding attribution, individual privacy concerns, conflicts in display preferences, and representing a project over time. I'll give examples from publication, artistic work, course, grant, and advisee records. I'll also give suggestions for weighing the benefits of shared records against its complexities.
Damaris Murry , Duke University
The VIVO community has long promoted the value of modeling the domain of scholarship in a form that is independent of a particular software implementation. The VIVO ontology is the product of a collaborative effort to define a shared understanding of the semantics behind the complex graphs of scholarly activity that different research networking systems might choose to process internally in different ways. As research networking software projects such as VIVO begin to develop the next generation of decoupled, dynamic and responsive user interfaces, there is an opportunity to consider a similar kind of collaborative modeling approach to define robust UIs whose behavior can be reasoned about and tested separately from a concrete software implementation. Statecharts were first described by David Harel in 1987 as an extension of finite state machines and state diagrams[1], but have recently gained traction in the web UI development community[2]. With UI statecharts, the representation of the features, behavior, and possible effects of different UI interactions is decoupled from the code that actually implements the behavior. This can lead to a number of potential benefits, such as validating that software properly implements community requirements, opening up aspects of development to contributors who may not be experts in particular web UI frameworks, and automatically generating more robust, testable and bug-free code. In this presentation we will examine the principles of UI statecharts and consider their application to community development of research networking software. REFERENCES [1] Harel, David. Statecharts: A Visual Formalism For Complex Systems. (1987). Science of Computer Programming, 8(3), 231-274. DOI: 10.1016/0167-6423(87)90035-9. [2] „Welcome to the world of statecharts.” https://statecharts.github.io/ . Retrieved 29 April 2019.
View presentationAndrei Tudor , Ontocale
Research information management systems (RIMSs) use different approaches to collecting and curating research identity information: manual curation by information professionals or users, automated data mining and curation scripts (aka bots); and some combination of the above. Assuring the quality of information is one of the critical ethical issues of information systems. Although data curation by professionals usually produces the highest quality results, it is costly and may not be scalable. RIMSs may not have enough resources to control the quality of large-scale information often batch harvested and aggregated from the Web and various databases of different scope and coverage. RIMSs are in great need of researchers to contribute and curate their research identity data. This presentation reports the findings of a collaborative study about researcher participation in RIMSs. The presenters of this study developed a theoretical framework for researcher participation in RIMSs (Stvilia, Wu, & Lee, 2019). The framework is grounded in empirical research and can guide the design of RIMSs by defining typologies of researcher activities in RIMSs, related motivations, levels of participation, and metadata profiles. RIMS managers and scholarly communications librarians can use the framework to assemble RIMS service and metadata profiles that are tailored to the researcher’s context. Likewise, the framework can guide the construction of communication messages personalized to the researcher’s priorities and her or his motivations for engaging in a specific activity, which will enhance the researcher’s engagement with the RIMS. In addition, this presentation discusses how the framework can be operationalized in practice using the case of Scholars@TAMU, a VIVO-based RIMS at Texas A&M University. Reference Stvilia, B., Wu, S., & Lee, D. J. (2019). A framework for researcher participation in research information management systems. The Journal of Academic Librarianship. 45(3), 195-202. doi:10.1016/j.acalib.2019.02.014
Dong Joon Lee , Texas A&M University
Besiki Stvilia , Florida State University
Shuheng Wu , City University of New York
We see the first VIVO conference to be held in Europe as a very positive sign of the worldwide attention the VIVO platform has attained. In this regard, we strongly believe a complete VIVO internationalization (i18n) is critical to reach a major adoption at an international level. We are certain it is a major driver of growth for VIVO, as many institutions need to have support for languages other than English. The VIVO i18n internationalization task force has established the roadmap to achieve this goal. However, a critical mass of stakeholders is essential for it to become a development priority. This presentation is meant as an exchange to gather the community and create a synergistic involvement in the internationalization of the VIVO platform.
View presentationRachid Belkouch , Université du Québec à Montréal
Pierre Roberge , Université du Québec à Montréal
The EU funded COURAGE project collected the methods and memories for cultural opposition in the socialist era (cc. 1950-1990), and built a registry for preserving data about culture as a form of opposition. The outcomes of the project include learning material, online exhibition, a public browseable interface and an open SPARQL endpoint. The heart of the software environment is a Vitro instance, which provides easily re-usable and connected data for the specific extensions implementing the virtual exhibition, the learning platform and other satellite services. Data was input and curated by historians, social scientists and other researchers from humanities. The Vitro code base had to be extended and sometimes overriden to comply with the requirements of the project. For example, editing rights had to be granted based on the edited data context, and a workflow for quality management had to be added to the system. The analysis of collected data has been helped by various statistical pages using SPARQL queries. Although the project has ended, we are still catching up with ideas to connect our registry with more and more services in the field.
View presentationAndrás Micsik , Institute for Computer Science and Control, Hungarian Academy of Sciences
Tamás Felker , Institute for Computer Science and Control, Hungarian Academy of Sciences
A second version of the VIVO ontology. We propose to develop a consistent, sufficient, BFO based ontology for representing scholarship. By consistent, we mean the ontology uses a single approach to representation. Using a single approach, we expect to simplify the ontology -- patterns are reused and complexity reduced. By sufficient, we mean we cover the domain of scholarship at the level necessary to represent and use information about scholarship. The ontology is informed by its applications. By BFO-based, we mean we commit to an approach to representation based on the Basic Formal Ontology. The approach is well-understood and well-adopted in the ontology community. The domain of the ontology is well-defined and stable. Why a new ontology, and why now? The original work on the VIVO ontology began in 2007 at Cornell. The 2009 NIH grant significantly expanded the ontology. The 2013 CTSA Connect effort significantly re-engineered the ontology attempting to bring the ontology to standards current at the time, and introducing BFO as an upper level ontology, but the effort was never completed, and introduction of the new ontology (VIVO version 1.6) was not accompanied by sufficient tooling, training, and time to manage the community change. Since 2013, the ontology has essentially been frozen. As development seeks to create an interface between the ontology and the presentation software, there is an opportunity to create an ontology that is independent of the software and can be mapped to it. Benefits of a new ontology. The new ontology will: Add to our ability to represent all of scholarship, including the arts, peer review, new research outputs, research impact, and global needs Adopt current ontological best practice including tooling, OBO Principles, focus on the domain of scholarship and expertise, and use only those ontologies that are aligned Use simple, consistent representations supporting ontological reasoning Be appropriate for use by any project seeking to build and use research graphs How and when. A new ontology could be developed in three phases by the Ontology Interest Group of the VIVO Project, working in collaboration with other projects, ontologists, developers, and community members. All are welcome to join the effort. A second phase would be necessary for refinement, testing, and tooling for adoption. A third phase is need for community change management, mapping to presentation data structures, testing, and training. The existing ontology (version 1.x) will continue to be supported indefinitely.
View presentationVioleta Ilik , Columbia University
Michael Conlon , VIVO Project
Brian Lowe , Ontocale
Christian Hauschke , Technische Informationsbibliothek (TIB) û German National Library of Science and Technology
Marijane White , Oregon Health and Science University
Muhammad Javed , Mastercard
Naomi Braun , University of Florida
The rich semantic data captured in a VIVO instance (or any other application built on a Vitro core) presents a fantastic opportunity to surface knowledge about connected resources. However, the query interfaces within VIVO (or Vitro) are quite limited - other than a full text search, information is presented only in ways that have been baked into the UI. A SPARQL endpoint can be enabled for richer semantic queries, but this comes at a cost - the user needs to understand, or have access to a library of pre-written SPARQL queries, you can't combine queries (a SELECT over a CONSTRUCTed source), and there is a risk of badly written SPARQL queries having an impact on the system performance. With the Vitro Query Tool, we build upon the work of Cornell, using their DataDistributor as an API for storing and running a library of queries. To make this more accessible, we initially created a user interface allowing authorised users the ability to view and create queries using the DataDistributor building blocks, before extending this with the ability to schedule the execution of queries and distribute the results (e.g. via email). In utilising the existing DataDistributor, not only do we provide a means of creating a library of queries and reports for users to execute or receive, but it also allows the data to be exposed via API endpoints that can be ingested by other applications or used by visualisations within the application.
View presentationQazi Asim Ijaz Ahmad , Technische Informationsbibliothek (TIB) û German National Library of Science and Technology
Graham Triggs , Technische Informationsbibliothek (TIB) û German National Library of Science and Technology
Christian Hauschke , Technische Informationsbibliothek (TIB) û German National Library of Science and Technology
Staff at academic institutions are regularly called upon to produce and maintain lists of scholarly publications authored by individuals ranging from principal investigators to people affiliated with other institutions such as alumni and residents. This work tends to be done on an ad hoc basis and is time-consuming, especially when profiled individuals have common names. Often, feedback from the authors themselves is not adequately captured in some central location and repurposed for future requests. ORCID is sometimes touted as a silver bullet to the problem of author disambiguation and publications management, but it does not fully meet institutional needs. ReCiter is a highly accurate, open source, machine learning system for inferring which publications in PubMed a given person has authored. It rapidly and accurately identifies articles, including those at previous affiliations, by a given person. It does this by leveraging institutionally maintained identity data (e.g., departments, relationships, email addresses, year of degree, etc.) With the more complete and efficient searches that result from combining these types of data, individuals at institutions can save time and be more productive. Running ReCiter daily, one can ensure that the desired users are the first to learn when a new publication has appeared in PubMed. We envision ReCiter becoming part of a larger publications management ecosystem, one which performs the following tasks: * harvests data from external sources * synthesizes and deduplicates records across multiple systems * allows administrator to correct metadata and/or disambiguate records (e.g., articles) * makes machine learning-based suggestions re: author/creator identity of records * transforms and standardizes data * collects feedback on author identity through a simple and intuitive user interface * syndicates records to a variety of target systems Much of the infrastructure for this work has been accomplished. In July, Weill Cornell's development team launched a CloudFormation template, which allows for easy install of all the dependencies for this system on the Amazon Web Services infrastructure. When realized, this system will allow institutions to meet a range of sophisticated needs from reporting to populating scholarly profiles.
View presentationPaul Albert , Weill Cornell Medicine
Sarbajit Dutta , Weill Cornell Medicine
Michael Bales , Weill Cornell Medicine
Jie Lin , Weill Cornell Medicine
Terrie Wheeler , Weill Cornell Medicine
In the context of the BMBF-funded project ROSI (Reference Implementation for Open Scientometric Indicators), visualisations are being developed and their impact on researchers investigated. For this purpose, only data from open data sources will be collected via persistent identifiers of persons, documents or organisations. Furthermore, a reference implementation based on VIVO will be developed in which various indicators with data from open sources such as Crossref Event Data or Wikidata will be visualized. Interviews and workshops with researchers will be conducted to gather their requirements concerning scientometric data. Findings from the interviews and workshops will be incorporated into the reference implementation in an iterative process. The selection of the indicators, the data aggregation levels and data visualisation types will be addressed. A Registry of Scientometric Data Sources is created as a by-product in which APIs with data relevant for the purposes of scientometrics are captured and made findable. This poster describes the project and presents the current status of the project results.
Svantje Lilienthal, Christian Hauschke, Grischa Fraumann
In this poster we present our activities aimed at using the Research Core Dataset (KDSF) - a national German standard for reporting - in the VIVO context. In recent times, a constant interest in implementing KDSF into various types of Current Research Information Systems (CRIS) among German research institutions can be observed. At the TIB, an non-public VIVO for KDSF-compliant reporting is being developed. The scope of the activities around KDSF in VIVO covers the alignment of KDSF and VIVO data models, the implementation of additional datasets to meet KDSF requirements, definition of data entry workflows as well as the development of a reporting component.The data models were aligned for both data input and data export, with VIVO ontology being reused as far as possible. Furthermore, KDSF-compliant data recording requiers annotation of entities with subjects from the classification of the German Federal Office of Statistics. To enable the usage of the classification in VIVO, we have converted it into a SKOS concept scheme and made it available on a Skosmos server - readable for both humans and machines. Due to the heterogeneity of institutional data sources and formats a number of individually customized workflows for automated data ingest and update are necessary. The reporting component - the Vitro Query Tool - allows reusing, sharing and scheduling of SPARQL queries for reporting.
Tatiana Walther, Christian Hauschke, Ina Blümel
In 2017 the TIB presented a prototype of a publicly available institutional VIVO at the CeBIT computer fair. In the meantime, numerous enhancements and customisations have been made to adapt the system to the needs of the institution. The measures include some changes to the software architecture, such as the improved password encryption already demonstrated in 2018. However, a number of improvements have also been made to the workflows and user interface. In this presentation we would like to introduce the TIB VIVO, called TIB-FIS-Discovery, including the interface based on the Tenderfoot theme, the import and claiming mechanisms adapted from OpenVIVO for publications via Pubmed-ID and DOI as well as the improvements made so far with regard to the internationalization of VIVO. Finally, we would like to point out the desiderata we consider necessary for the success of VIVO in the TIB of comparable institutions. This includes improved multiple language support including a multilingual editor for logged-in VIVO users, simplified workflows for data integration from heterogeneous data sources, and associated generic mechanisms for claiming entities.
Christian Hauschke, Graham Triggs, Tatiana Walther, Qazi Asim Ijaz Ahmad
AI2 is an accelerator for health informatics innovation. We work with academic centers and industry partners to create communities of innovation, along with industry-academia partnerships in order to accelerate the adoption of informatics innovation. We do this at a physical campus consisting of an accelerator lab and “living lab” with real-world clinical space designed for modeling the clinics of future, and at a virtual campus at AI2.world, which facilitates the sharing of solutions and applications. Our collaborations focus on patient care, engagement, and outcomes. This poster presentation describes the creation and launch of this initiative, focusing on the AI2.world collaboration. Today, healthcare data comes from many sources beyond the electronic health record, including fitness and personal health devices, social media, and genomics sources. This wealth of data is rich with opportunities for innovation, but often requires a wide range of domain expertise to be used effectively. Collaboration and validation are more important than ever. AI2.world facilitates collaboration and validation by providing online access to tools and technologies, datasets, and domain expertise to collaborators who cannot readily visit the physical campus. We are creating a cloud-based model and platform for informatics services and analytics that can be used for research and development of new patient care technologies. At the same time, we are creating a community of innovators from academic centers, healthcare institutions, and industry partners, using a social learning model and crowdsourcing. Social learning and crowdsourcing approaches will be important in training future employees as new skills will be required, driven by advances in informatics, analytics, NLP, pattern recognition and artificial intelligence. For this reason, AI2.world was built using the Alicanto™ social learning platform developed at the Division of Clinical Informatics, Beth Israel Deaconess Medical Center. The platform includes support for a multimedia library that can accommodate video from multiple sources, images and animations, documents, interactive medical calculators, and can integrate with LTI-compliant online learning tools. The site also supports self-paced online courses incorporating materials from the site library and quizzes with multiple choice, true/false and other question types. Public and private groups can be created by community members, with document sharing, threaded discussions and video chat. The goal of AI2 is to help innovators turn health data into information, insight, and innovation quickly and efficiently. It will be a community of practice that facilitates both learning and global multidisciplinary communication on how to implement best practices in informatics and artificial intelligence in healthcare. We have only begun to understand the opportunities in healthcare, research, and industry. AI2.world is a platform for exponential thinking in this expanding universe.
Barbara Rapchak, BS, Vice President Academic Innovation, Intelligent Medical Objects, Inc., Frank Naeymi-Rad, PhD, PhD, MS, MBA. Founder, Chairman of the Board, Intelligent Medical Objects, Inc., Yuri Quintana, PhD, Director of Global Health Informatics, Division of Clinical Informatics, Beth Israel Deaconess Medical Center, Charles Safran MD MS FACMI, Fellowship Director, Professor of Medicine, Chief, Division of Clinical Informatics, Beth Israel Deaconess Medical Center
Metabolomics is the scientific study of metabolites present within an organism, cell, or tissue. A metabolite is the intermediate end product of metabolism. The term metabolite is usually restricted to small molecules. Metabolomic studies involve the identification of metabolites in biological samples, often by processes involving mass spectroscopy. Datasets from such studies may involve the identification of several thousand compounds leaving tens of thousands unidentified. These data sets are valuable for scientific reuse. The National Institutes of Health in the United States has established a National Metabolomics Data Repository known as the Metabolomics Workbench (www.metabolicsworkbench.org) to provide access to datasets resulting from NIH-funded work. In Europe, the European Molecular Biology Laboratory also provides a data registry. VIVO is being used to create investigator-centric metadata regarding datasets and publications in metabolomics. The initial work is focused on investigators participating in the NIH Common Fund Metabolomics Program. Future work could include other metabolomics investigators. The VIVO for metabolomics will provide metadata to the consortium web site http://metabolomics.info which provides information of interest to all metabolomics investigators, as well as a Triple Pattern Fragments (TPF) endpoint for discovery of metadata regarding metabolomic investigations. The work will improve the findability of metabolomics datasets, as each will be an entity with a profile page, discoverable via popular search engines. Each entry will link directly to the metabolomics workbench, improving accessibility. Much work remains in the metabolomics community on interoperability between repositories -- a long term goal is to support the use of a wide range of metabolomic software tools on the widest possible range of metabolomics datasets. This will require coordination across repositories and development and implementation of common representations. Significant additional work remains on reuse to develop and implement standards for the representation of identification information, quantification, and naming of metabolites.
Michael Conlon, Kevin S. Hanson, Taeber Rapczak, Naomi Braun, Christopher P. Barnes
The VIVO ontology is focused on the domain of scholarship. But to represent scholarship we need terms and relations from other domains. For example, VIVO needs to refer to organizations – people have positions at universities, journals have publishers, memberships are offered by associations and societies. Universities, publishers, associations, and societies are organizations. Representing organizations is important for VIVO, but beyond the domain of scholarship. We need to refer to representations by those who have organizations as their domain and focus the VIVO ontology on the domain of scholarship. VIVO was very early to the world of ontologies. When VIVO began as a semantic application in 2007, there were few good ontologies, few were actively maintained, few were focused on defined domains, and few were developed with consistent principles. Foundries (collections of ontologies with common principles) were emerging. Finding ontologies that could be relied on to build the VIVO ontology was challenging. Organizations, Time, Locations, Research Administration, Research Resources, Journals, Languages, Academic Degrees, and Concept vocabulary, are all potentially separable from VIVO and useful beyond their use in representing scholarship. Some of these domains of representation have ontologies that might merit reuse. Others do not. Other domains of representation such as teaching, service, research impact, awards, and mentoring may not be candidates for separate ontologies at this time. Additional specialty domains need also be considered. These include national vocabularies and taxonomies, as well as those of academic disciplines such as agriculture, the performing arts, and medical research. In creating a version 2 of the VIVO ontology, we seek to identify domains that can be represented and maintained outside of the VIVO ontology, and identify ontologies of those domains that can be reused. In some cases we will not be able to reuse the domain ontology directly, but rather refer to terms in the domain ontology. In this poster we will present an overview of the domains commonly encountered in representing scholarship with analysis and recommendations regarding how each may be treated in version 2 of the VIVO ontology.
Michael Conlon, Violeta Ilik, Brian Lowe, Christian Hauschke, Marijane White, Muhammad Javed, Naomi Braun
View posterUniversities come in many different shapes and sizes. Using the Dimensions API to extract coauthorship networks it is possible to build network diagrams that illustrate the collaborative shape of an institution. By shading researchers by the colour of their most frequently used field of research, and sizing their nodes by the number of career publications, these diagrams can also communicate discipline focus. Placed side by side, these networks highlight similarities and differences. Using the Dimensions API, we implemented a repeatable script to extract a collaboration network for an array of institutions identified by their GRID id (http://grid.ac). We then processed these networks using the Gephi toolkit. This allowed us to apply a consistent colour scheme to all graphs, with and identical layout properties. The resultant network diagrams were then assembled into a single poster using Overleaf, with github integration from Gigantum. The code for this project will be available on Gigantum.com
Simon Porter, Jared Watts
InCites is a benchmarking and analytics tool built upon best-in-class Web of Science data that enables comprehensive insight into your organization’s performance. Currently in beta release, Clarivate has recently introduced the “My Organization” module for InCites that enables users to refine the InCites dataset using their own verified faculty and disambiguation data. InCites with My Organization delivers standardized researcher, team & departmental reporting to increase relevance of bibliometric data for all users. Leveraging the data produced from existing workflows users have developed to build their VIVO databases, MyOrg subscribers can easily populate their MyOrg module database with VIVO data using open-source Python code available on GitHub at https://github.com/Clarivate-SAR.
Benjamin Gross, Miguel Garcia
View posterThis poster documents how we made our "What does a University look like?" poster. It details: How we used the Dimensions API to build demographics profiles of University Researchers How we used Gigantum to manage our Jupyter notebooks, and analysis environment How we integrated Figshare to create data assets, with providence links back to the public Gigantum code How we integrated Gigantum with Overleaf to push network diagrams to into our latex templates
Simon Porter, Jared Watts
The Danish project OPERA, Open Research Analytics, is a joint collaboration between Danish Universities as well as international partners. The project is exploring new approaches to best practice research analytics that should be useful, relevant and responsible and furthermore acknowledging Open Science endeavors across research areas. The OPERA project is work in progress and already showing promising results. Some of the fundamental elements in the project is the VIVO platforms showcasing university specific advanced research analytics and a national VIVO platform populated with Dimensions data as well as complementing metadata underlining the Open Science elements, advanced network analysis and modules exploring the collaboration potential of Danish universities. The poster will highlight the halfway results and stipulate the next steps and expected outcomes.
Mogens Sandfær, Nikoline Dohm Lauridsen, Christina Steensboe, Karen Sofie Hytteballe Ibanez, Poul Melchiorsen, David Budtz, Birger Larsen
While Linked Open Data promised automatic and obvious connections between structured data, interoperability remains a thorny social, not technical, issue. It is essentially an information model that relies on the removal of silos and requires interdepartmental collaboration. An intra-university knowledge graph operationalizes the network of relationships on which innovation is built. We are proposing to develop a Linked Open Data prototype that creates and surfaces relationships between the people, publications, and other high level metadata residing within the CU Experts semantic web application with departmental domain specific metadata sites which contain datasets, equipment, or observable phenomena? The problem space is very large and domains within campus currently have structured and unstructured metadata. In order to achieve a goal that demonstrates the value of having these cross departmental relationships, this project is scoped to work a specific institute within our University which already has domain specific metadata in space sciences. We want to demonstrate, using metadata, how CU Boulder was involved with sending a spacecraft to every planet in the solar system.
Don Elsborg, Anne Wilson, Katie Mika, Erik Radio, Matt Ramey, Alex Viggio
View posterThe Citec project funded by RANEPA provides open citation content data parsed from research papers in PDF available at RePEc and Socionet research information system. One of current activities of the Cirtec project is a design of author-oriented services, data and indicators based on citation data parsed from three groups of research papers linked with a specific author. These three groups of papers for each author are: (1) own papers of a specific author, which available at RePEc by linkages from an author profile in RePEc Author Service; (2) papers cited by this author, which can be collected using CitEc citation relationships to the author papers available at RePEc; (3) citing papers for this author, which available by links from CitEc citation relationships available at RePEc.
Sergey Parinov
One of the aims of Open Science is to foster reproducibility and reusability of research outputs. Hence, Open Science covers the whole research lifecycle to create links between the different steps and processes by securing the provenance of data and by enabling retractability of raw and processed data (incl. versioned data) at any time during and after a project ends, while documenting and informing about the processes undertaken to derive results. A Data Management Plan (DMP) contains information concerning the research processes, the data utilized and generated throughout a project’s lifetime, and the policies surrounding them. Still lacks automated solutions and workflows that connect the DMP to the actual data where they are stored and link to other useful information such as publications and funding information, thus enabling the creation of coherent/complete research entities. In an open and FAIR research ecosystem that is currently being realised, information linking between research processes and research outputs is essential. OpenDMP is the joint effort of OpenAIRE and EUDAT to deliver an open platform for Data Management Planning that addresses those demands/ issues and assumes no barriers for its use and adoption. It does so by applying common standards for machine-actionable DMPs as defined by the global research data community of RDA and by communicating and consulting with researchers, research communities and funders to better reflect on their needs. The demo presentation will cover the OpenDMP platform that provides the scientific community with a free, open and flexible instrument for the handling, validation and dissemination of their data management plans, in line with policies settled by a variety of funding agencies, institutions and other initiatives. OpenDMP work was inspired by the lack of a platform that may adapt to any and all requirements that may be present in a data management planning environment. It aims to combine the flexibility of adhoc questionnaires with the power of a structured information system while capturing the semantics of a data management plan, and the rigidness of funding and/or validation rules. It also attempts to embrace the need for collaboration around the structuring of a data management plan, its evolution and sharing. Moreover, it emphasizes on achieving interoperability and enabling actionability of data management plans, that will allow systems to not only validate, but also to automate tedious, currently user driven, activities. In the demo section, the audience will be offered: A very short introduction on the concept of data management and its relation to open science policies and best practices. An overview of key concepts of the data management planning, as incorporated in OpenDMP, with a focus on FAIR DMPs. Those will cover projects, funders, dataset profiles, data management plans, dataset descriptions, contributors etc. An overview of the key features of OpenDMP that differentiate it from other solutions that existed, exist or emerge. A brief overview of the technology that allows OpenDMP to be modern, attractive, performant and open. A walk through the systems’ functionality with hands on access to the system, that will unveil all main project features to the audience. A preview of the plans of the implementation team for latest developments towards interoperable, actionable data management planning. Further information of how to get access and exploit the free and open services of OpenDMP, in any of the supported manners, be it institutional, national or following other scopes.
Georgios Kakaletris, Diamantis Tziotzios, Ioannis Kalyvas, Elli Papadopoulou
Auto-generated charts from declarative (json) code open the path to having a common language for collaboration. The implications are not just theoretical. This demo showcases how a feature planned with Statecharts but coded with React.js can be easily adapted into plain Javascript while retaining the majority of the original code. This demo intends to offer a glimpse of how Statecharts can enable smooth collaboration between code contributors whose institutions have adopted different technologies.
Andrei Tudor
The CU Experts (VIVO) implementation team at CU Boulder has experienced tremendous organizational change in 2019. It finds itself in the midst of an evolving institutional landscape, unprecedented since the creation of the Faculty Information System that powers CU Experts more than 20 years ago. The original team is now split across two separate departments, with one foot remaining on the academic side of the university and the other newly planted on the business side. Many of the groups and systems that we are interacting with are also new or quickly evolving, with administrators across campus pushing for innovation. Continuing to work together as colleagues, we recently engaged with collaborators from campus research institutes and libraries to submit two innovation grant proposals which have been accepted. This view of CU Boulder’s landscape also reminds us that CU Experts, which has been in production since 2010, is no longer the new kid on the block. How might our VIVO implementation evolve, with input from thought leaders in the VIVO community, to best support our evolving institution? This poster hopes to engage attendees from other institutions that have implemented VIVO or similar solutions, or are evaluating an implementation strategy.
Alex Viggio, Don Elsborg, Vance Howard, Matt Ramey