“In this module we present the proposal and budget for an open source library to generate CrossRef DOI XML. We imagine a world where people don’t think twice about creating DOIs, and integrate changes with confidence. The proposed libraries could be extended to improve confidence in generating XML for DataCite in web applications as well. This project is unfunded at the time of publication and we are looking for support to realise this mission….”
Category Archives: oa.xml
R= Making it easy to generate CrossRef XML with confidence
“In this module we present the proposal and budget for an open source library to generate CrossRef DOI XML. We imagine a world where people don’t think twice about creating DOIs, and integrate changes with confidence. The proposed libraries could be extended to improve confidence in generating XML for DataCite in web applications as well. This project is unfunded at the time of publication and we are looking for support to realise this mission….”
When XML Marks the Spot: Machine-readable journal articles for discovery and preservation
“If you work with a campus-based journal program and you’re looking to expand the readership and reputation of the articles you publish, adding them to relevant archives and indexes (A&Is) presents a treasure trove of opportunities. A&Is serve as valuable content distribution networks, and inclusion in selective ones is a signal of research quality. You may have heard about XML, one of the primary machine-readable formats academic databases use to ingest content, and wonder if that’s something you need to reach your archiving and indexing goals.
This free webinar, co-hosted by Scholastica, UOregon Libraries, and the GWU Masters in Publishing program, will offer a crash course in the benefits of XML production and use cases, including:
What XML is and the different types required or preferred by academic indexes and archives (with an overview of JATS)
How producing metadata and/or full-text articles in XML can unlock discovery and archiving opportunities with examples
Additional benefits of XML for journal accessibility as well as publishing program and professional development
When XML is needed and when it may not be the best use of journal resources
Ways you can produce XML, including an overview of Scholastica’s production service…”
Welcome to the Single Source Publishing Community | The Single Source Publishing Community (SSPC) is a network stakeholders from the Open Science community that are interested in Single Source Publishing (SSP) for scholarly purposes – developing open-source software and advocacy.
“The Single Source Publishing Community (SSPC) is a network of stakeholders from the Open Science community that are interested in Single Source Publishing (SSP) for scholarly purposes – developing open-source software and advocacy.”
OS-APS: Open Source Academic Publishing Suite
“OS-APS enables XML-based workflows for media-neutral publishing (e.g. Open Access) without technical expertise and cost-intensive XML editing and content management systems. Corporate design can be controlled via existing typesetting templates or in detail with a template development kit. We plan to use and extend existing open source components, such as OJS, OMP and Pandoc, and make all results open source.”
An XML-Based Migration from Digital Commons to Open Journal Systems
Abstract: The Oregon Library Association has produced its peer-reviewed journal, the OLA Quarterly (OLAQ), since 1995, and OLAQ was published in Digital Commons beginning in 2014. When the host institution undertook to move away from Bepress, their new repository solution was no longer a good match for OLAQ. Oregon State University and University of Oregon agreed to move the journal into their joint instance of Open Journal Systems (OJS), and a small team from OSU Libraries carried out the migration project. The OSU project team declined to use PKP’s existing migration plugin for a number of reasons, instead pursuing a metadata-centered migration pipeline from Digital Commons to OJS. We used custom XSLT to convert tabular data exported from Bepress into PKP’s Native XML schema, which we imported using the OJS Native XML Plugin. This approach provided a high degree of control over the journal’s metadata and a robust ability to test and make adjustments along the way. The article discusses the development of the transformation stylesheet, the metadata mapping and cleanup work involved, as well as advantages and limitations of using this migration strategy.
JCORE upgrades to JATS 1.3 – Highwire Press
“HighWire’s journals hosting solution JCORE can now transform XML to industry standard Journal Article Tag Suite (JATS) XML version 1.3, the latest version. This means that publishers using the platform can now seamlessly comply with the specifications of this interoperable standard.
The British Medical Journal (BMJ) is the first JCORE publisher to offer the JATS 1.3 download for all of their OA content which includes over 57,000 articles. This option may be attractive to BMJ and other publishers as the ability to download full text in a machine readable format is one of the strong recommendations made by Coalition S within the Plan S Principles and Implementation guidance for publishers. …”
4 Best Practices for Open Access (OA) Journal Publishing
“1) The Fundamental One?—?Implement Open Access (OA) Metadata for Interoperability…
2) The Essential One?—?Add Your Journals To Scholarly Indexes…
3) Midway To Success?—?Archive Your Open Access Journal…
4) The Never-Ending Cycle?—?Promote and Distribute Your Research Like Your Life Depends On It! …”
Streamlined peer review and PubMed-ready XML: How Spartan Medical Research Journal is using Scholastica to grow
“When SMRJ was started, the editors used email and Word docs to track peer review, and they published all articles in PDF format. However, with the journal continuing to expand, the editors realized they were in need of an easier way to track submissions and a new publishing system to improve the journal’s online reading experience and chances of being added to relevant indexes. As a result, Chief Editor William Corser and Assistant Editor Sam Wisniewski began searching for publishing tools and services, focused on three key areas: streamlining peer review, modernizing the journal’s website, and producing XML for all articles.
After considering different options, Corser and Wisniewski chose to use Scholastica’s peer review and open access publishing software, as well as Scholastica’s typesetting service to produce PDF, HTML, and XML article files. Since making the switch, they’ve found that peer review is smoother for editors and authors and they’re making strides towards reaching their article discovery and indexing goals….”
An XML Repository of All bioRxiv Articles is Now Available for Text and Data Mining
“bioRxiv and medRxiv provide free and unrestricted access to all articles posted on their servers. We believe this should apply not only to human readers but also to machine analysis of the content. A growing variety of resources have been created to facilitate this access.
bioRxiv and medRxiv metadata are made available via a number of dedicated RSS feeds and APIs. Simplified summary statistics covering the content and usage are also available. For bioRxiv, this information is available here’
Bulk access to the full text of bioRxiv articles for the purposes of text and data mining (TDM) is available via a dedicated Amazon S3 resource. Click here for details of this TDM resource and how to access it….”
What is MEI?
“The Music Encoding Initiative (MEI) is a 21st century community-driven open-source effort to define guidelines for encoding musical documents in a machine-readable structure.
It brings together specialists from various music research communities, including technologists, librarians, historians, and theorists in a common effort to discuss and define best practices for representing a broad range of musical documents and structures. The results of these discussions are then formalized into the MEI schema, a core set of rules for recording physical and intellectual characteristics of music notation documents expressed as an eXtensible Markup Language (XML) schema. This schema is developed and maintained by the MEI Technical Team….”
New business models for the open research agenda | Research Information
“The rise of preprints and the move towards universal open access are potential threats to traditional business models in scholarly publishing, writes Phil Gooch
Publishers have started responding to the latter with transformative agreements[1], but if authors can simply upload their research to a preprint server for immediate dissemination, comment and review, why submit to a traditional journal at all? Some journals are addressing this by offering authors frictionless submission direct from the preprint server. This tackles two problems at once: easing authors’ frustrations with existing journal submission systems[2], and providing a more direct route from the raw preprint to the richly linked, multiformat version of record that readers demand and accessibility standards require….
Dissemination of early-stage research as mobile-unfriendly PDF is arguably a technological step backwards. If preprints are here to stay, the reading experience needs to be improved. A number of vendors have developed native XML or LaTeX authoring environments which enable dissemination in richer formats….”
DOAJ to add Crossref compatibility – News Service
“In a series of metadata improvements, publishers will be able to upload XML in the Crossref format to us from 18th February 2020.
In 2018, we asked our publishers what would make their interaction with DOAJ easier and 46% said that they would like us to accept Crossref XML. Today we only accept XML formatted to our proprietary DOAJ format….”
CORE users can now read articles directly on our site – CORE
“We are happy to announce the release of CORE Reader, which provides a seamless experience for users wishing to read papers hosted by CORE. In this post, we provide an overview of what is new and we encourage you to follow this development as new functionalities in the reader are on our roadmap….
At the beginning of this project, there was a reflection that most open access services do not yet provide a rich user experience for reading research papers. Determined to change this, we originally started looking at whether CORE could render research papers as HTML, as has recently become trendy across publisher platforms. While such rendering remains to be one of the ultimate goals, we realised that this could only be achieved for a small fraction of documents in CORE. More specifically, those that the data provider offers in machine readable formats, such as LaTeX or JATS XML. While we want to encourage more repositories to support such formats (and this remains to be a Plan S recommendation), we wanted to improve the reading experience for all of our users across all of our content….”
Guest post: a technical update from our development team – News Service
“Here are some major bits of work that we have carried out:
Enhancements to our historical data management system. We track all changes to the body of publicly available objects (Journals and Articles) and we have a better process for handling that.
Introduced a more advanced testing framework for the source code. As DOAJ gains more features, the code becomes larger and more complex. To ensure that it is properly tested for before going into production, we have started to use parameterised testing on the core components. This allows us to carry out broader and deeper testing to ensure the system is defect free.
A weekly data dump of the entire public dataset (Journals and Articles) which is freely downloadable.
A major data cleanup on articles: a few tens of thousands of duplicates, from historical data or sneaking in through validation loopholes, were identified and removed. We closed the loopholes and cleaned up the data.
A complete new hardware infrastructure, using Cloudflare. This resulted in the significant increase in stability mentioned above and allows us to cope with our growing data set (increasing at a rate of around 750,000 records per year at this point).
And here are some projects we have been working on which you will see come into effect over the next few weeks:
A completely new search front-end. It looks very similar to the old one, but with some major improvements under-the-hood (more powerful, more responsive, more accessible), and gives us the capability to build better, cooler interfaces in the future.
Support for Crossref XML as an article upload format. In the future this may also be extended to the API and we may also integrate directly with Crossref to harvest articles for you. We support the current Crossref schema (4.7) and we will be supporting new versions as they come along….”