Publications
Peer-reviewed scientific publications
1. Corcho, O., Ekaputra, F. J., Heibi, I., Jonquet, C., Micsik, A., Peroni, S., & Storti, E. (2024). A maturity model for catalogues of semantic artefacts .In: Sci Data 11, 479 (2024). https://doi.org/10.1038/s41597-024-03185-4. Also available in Open Access at: https://doi.org/10.48550/arXiv.2305.06746
Abstract
This work presents a maturity model for assessing catalogues of semantic artefacts, one of the keystones that permit semantic interoperability of systems. We defined the dimensions and related features to include in the maturity model by analysing the current literature and existing catalogues of semantic artefacts provided by experts. In addition, we assessed 26 different catalogues to demonstrate the effectiveness of the maturity model, which includes 12 different dimensions (Metadata, Openness, Quality, Availability, Statistics, PID, Governance, Community, Sustainability, Technology, Transparency, and Assessment) and 43 related features (or sub-criteria) associated with these dimensions. Such a maturity model is one of the first attempts to provide recommendations for governance and processes for preserving and maintaining semantic artefacts and helps assess/address interoperability challenges.
2. Rizzetto, E., Peroni, S. (2024). Mapping bibliographic metadata collections: the case of OpenCitations Meta and OpenAlex. In: CEUR Wokshop Proceedings, vol 3643, 20th Conference on Information and Research Science Connecting to Digital and Library Science (IRCDL 2024), Bressanone, Italy. https://ceur-ws.org/Vol-3643/paper15.pdf. Also available in Open Access at https://arxiv.org/abs/2312.16523.
Abstract
This study describes the methodology and analyses the results of the process of mapping entities between two large open bibliographic metadata collections, OpenCitations Meta and OpenAlex. The primary objective of this mapping is to integrate OpenAlex internal identifiers into the existing metadata of bibliographic resources in OpenCitations Meta, thereby interlinking and aligning these collections. Furthermore, analysing the output of the mapping provides a unique perspective on the consistency and accuracy of bibliographic metadata, offering a valuable tool for identifying potential inconsistencies in the processed data.
3. Massari, A., Mariani, F., Heibi, I., Peroni, S., Shotton, D. (2024). OpenCitations Meta. In : Quantitative Science Studies 1-26. https://doi.org/10.1162/qss_a_00292. Also available in Open Access at https://arxiv.org/abs/2306.16191.
Abstract
OpenCitations Meta is a new database for open bibliographic metadata of scholarly publications involved in the citations indexed by the OpenCitations infrastructure, adhering to Open Science principles and published under a CC0 license to promote maximum reuse. It presently incorporates bibliographic metadata for publications recorded in Crossref, DataCite, and PubMed, making it the largest bibliographic metadata source using Semantic Web technologies. It assigns new globally persistent identifiers (PIDs), known as OpenCitations Meta Identifiers (OMIDs) to all bibliographic resources, enabling it both to disambiguate publications described using different external PIDS (e.g., a DOI in Crossref and a PMID in PubMed) and to handle citations involving publications lacking external PIDs. By hosting bibliographic metadata internally, OpenCitations Meta eliminates its former reliance on API calls to external resources and thus enhances performance in response to user queries. Its automated data curation, following the OpenCitations Data Model, includes deduplication, error correction, metadata enrichment, and full provenance tracking, ensuring transparency and traceability of data and bolstering confidence in data integrity, a feature unparalleled in other bibliographic databases. Its commitment to Semantic Web standards ensures superior interoperability compared to other machine-readable formats, with availability via a SPARQL endpoint, REST APIs, and data dumps.
4. Moretti, A., Soricetti, M., Heibi, I., Massari, A., Peroni, S., & Rizzetto, E. (2024). The Integration of the Japan Link Center’s Bibliographic Data into OpenCitations - The production of bibliographic and citation data structured according to the OpenCitations Data Model, originating from an Anglo-Japanese dataset. In: Journal of Open Humanities Data, 10(1), p. 21. https://doi.org/10.5334/johd.178.
Abstract
This article presents OpenCitations’ main data collections: the unified index of citation data (OpenCitations Index), and the bibliographic data corpus (OpenCitations Meta) in view of the integration of a new dataset provided by the Japan Link Center (JaLC). Based on a computational analysis of the titles of the publications performed in October 2023, 8.6% of the bibliographic metadata stored in OpenCitations Meta are not in English. Nevertheless, the ingestion of an Anglo-Japanese dataset represents the first opportunity to test the soundness of a language-agnostic metadata crosswalk process for collecting data from multilingual sources, aiming to preserve bibliodiversity and to minimize information loss considering the constraints imposed by the OpenCitations data model, which does not allow the acceptance of multiple values in different translations for the same metadata field. The JaLC dataset is set to join OpenCitations’ collections in November 2023, and it will be made available in RDF, CSV, and SCHOLIX formats. Data will be produced using open-source software and provided under a CC0 license via API services, web browsing interfaces, Figshare data dumps, and SPARQL endpoints, ensuring high interoperability, reuse, and semantic exploitation.
5. Koloveas, P., Chatzopoulos, S., Tryfonopoulos, C., Vergoulis, T. (2023). BIP! NDR (NoDoiRefs): A Dataset of Citations from Papers Without DOIs in Computer Science Conferences and Workshops. In: Alonso, O., Cousijn, H., Silvello, G., Marrero, M., Teixeira Lopes, C., Marchesin, S. (eds) Linking Theory and Practice of Digital Libraries. TPDL 2023. Lecture Notes in Computer Science, vol 14241. Springer, Cham. https://doi.org/10.1007/978-3-031-43849-3_9. Also available in Open Access at https://arxiv.org/abs/2307.12794.
Abstract
In the field of Computer Science, conference and workshop papers serve as important contributions, carrying substantial weight in research assessment processes, compared to other disciplines. However, a considerable number of these papers are not assigned a Digital Object Identifier (DOI), hence their citations are not reported in widely used citation datasets like OpenCitations and Crossref, raising limitations to citation analysis. While the Microsoft Academic Graph (MAG) previously addressed this issue by providing substantial coverage, its discontinuation has created a void in available data. BIP! NDR aims to alleviate this issue and enhance the research assessment processes within the field of Computer Science. To accomplish this, it leverages a workflow that identifies and retrieves Open Science papers lacking DOIs from the DBLP Corpus, and by performing text analysis, it extracts citation information directly from their full text. The current version of the dataset contains more than 510K citations made by approximately 60K open access Computer Science conference or workshop papers that, according to DBLP, do not have a DOI.
6. Chatzopoulos, S., Vichos, K., Kanellos, I., Vergoulis, T. (2023). Piloting Topic-Aware Research Impact Assessment Features in BIP! Services. In: Pesquita, C., et al. The Semantic Web: ESWC 2023 Satellite Events. ESWC 2023. Lecture Notes in Computer Science, vol 13998. Springer, Cham. https://doi.org/10.1007/978-3-031-43458-7_15. Also available in Open Access at https://arxiv.org/abs/2305.06047.
Abstract
Various research activities rely on citation-based impact indicators. However these indicators are usually globally computed, hindering their proper interpretation in applications like research assessment and knowledge discovery. In this work, we advocate for the use of topic-aware categorical impact indicators, to alleviate the aforementioned problem. In addition, we extend BIP! Services to support those indicators and showcase their benefits in real-world research activities.
7. Santos, E.A.d., Peroni, S. and Mucheroni, M.L. (2023). An analysis of citing and referencing habits across all scholarly disciplines: approaches and trends in bibliographic referencing and citing practices. In: Journal of Documentation, Vol. 79 No. 7, pp. 196-224. https://doi.org/10.1108/JD-10-2022-0234. Also available in Open Access at https://doi.org/10.48550/arXiv.2202.08469.
Abstract
Purpose: In this study, the authors want to identify current possible causes for citing and referencing errors in scholarly literature to compare if something changed from the snapshot provided by Sweetland in his 1989 paper.
Design/Methdology/Approach: The authors analysed reference elements, i.e. bibliographic references, mentions, quotations and respective in-text reference pointers, from 729 articles published in 147 journals across the 27 subject areas.
Findings: The outcomes of the analysis pointed out that bibliographic errors have been perpetuated for decades and that their possible causes have increased, despite the encouraged use of technological facilities, i.e. the reference managers.
Originality/value: As far as the authors know, the study is the best recent available analysis of errors in referencing and citing practices in the literature since Sweetland (1989).
Media articles and blog posts
- Scientific papers are not just the number of papers and citations, Đorđević, A., Savić, S., Pozitron - 32. University of Belgrade - Faculty of Chemistry, 32, 1-26, March 2024 (In Serbian)
- Toward open research information - Introducing the Information & Openness focal area at CWTS, Anli, Z., Tatum, C., Waltman, L., Leiden Madtrics, January 2024.
- Don't forget the social dimension of research evaluation, Amanatidis, A., Provost, L., Research Europe (print edition), January 2024.
- Community discussion on Research Assessment Reform in Social Sciences and Humanities, Delmazo, C., OPERAS blog, December 2023.
- Engaging with research communities to advance research assessment, Provost, L., Italian Open Science Portal, December 2023 (in Italian).
- Research(er) assessment that considers Open Science, Amanatidis, A., Leiden Madtrics, September 2023.
- GraspOS: Responsible assessment of scientic research and Open Science, Djordjevic, A., Pozitron (pages 60-61), August 2023 (in Serbian).
want to know
more?