- Researchers
- Librarians
- Publishers
- The benefits of Open Access publishing
- Publishing, copyright, and Open Access
- Publisher policies on self-archiving
- Open Access Journals
- Open Access Monographs
- Scholarly societies and Open Access publishing
- University presses and Open Access publishing
- Open Access publishing in the humanities
- Open Access publishing in the social sciences
- Funder policies on Open Access publishing
- Administrators
- Public
- Students
Monthly Archives: August 2010
Testing Jan Velterop’s Hunch About Green and Gold Open Access
Comparing Green and Gold
Yassine Gargouri
&
Stevan Harnad
Cognition/Communication Laboratory
Cognitive Sciences Institute
Universitè du Québec à Montréal
SUMMARY: Velterop (2010) has conjectured that more articles are being made Open Access (OA) by publishing them in an OA journal (“Gold OA”) than by publishing them in a conventional journal and self-archiving them (“Green OA”), even where self-archiving is mandatory. Of our sample of 11,801 articles published 2002-2008 by authors at four institutions that mandate self-archiving, 65.6% were self-archived, as required (63.2% Green only, 2.4% both Green and Gold). For 42,395 keyword-matched, non-mandated control articles, the percentage OA was 21.9% Green and 1.5% Gold. Velterop?s conjecture is the wrongest of all precisely where OA is mandated.
Jan Velterop has posted his hunch that of the overall percentage of articles published annually today most will prove to be Gold OA journal articles, once one separates from the articles that are classified as self-archived Green OA those of them that also happen to be published in Gold OA journals:
JV: ?Is anyone? aware of credible research that shows how many articles (in the last 5 years, say), outside physics and the Arxiv preprint servers, have been made available with OA exclusively via ‘green’ archiving in repositories, and how many were made available with OA directly (‘gold’) by the publishers (author-side paid or not)?
?The ‘gold’ OA ones may of course also be available in repositories, but shouldn’t be counted for this purpose, as their OA status is not due to them being ‘green’ OA.
?It is my hunch (to be verified or falsified) that publishers (the ‘gold’ road) have actually done more to bring OA about than repositories, even where mandated (the ‘green’ road).?— J. Velterop, American Scientist Open Access Forum, 25 August 2010
The results turn out to go strongly contrary to Velterop?s hunch.
Our ongoing project is comparing citation counts for mandated Green OA articles with those for non-mandated Green OA articles, all published in journals indexed by the Thompson/Reuters ISI database (science and social-science/humanities). (We use only the ISI-indexed sample because the citation counts for our comparisons between OA and non-OA are all derived from ISI.)
The four mandated institutions were Southampton University (ECS), Minho, Queensland University of Technology and CERN.
Out of our total set of 11,801 mandated, self-archived OA articles, we first set aside all those (279) articles that had been published in Gold OA journals (i.e., the journals in the DOAJ-indexed subset of ISI-indexed journals) because we were primarily interested in testing the OA citation advantage, which is based on comparing the citation counts of OA articles versus non-OA articles published in the same journal and year. (This can only be done in non-OA journals, because OA journals have no non-OA articles.) This left only the Green OA articles published in non-Gold journals.
We then extracted, as control articles for each article in this purely Green OA subset, 10 keyword-matched articles published in the same journal and year. The total number of articles in this control sample for the years 2002-2008 was 41,755. (Our preprint for PloS, Gargouri et al. 2010, covers a somewhat smaller, earlier period: 2002-2006, with 20,982 control articles.)
Next we used a robot to check what percentage of these unmandated control articles was OA (freely accessible on the web).
Of our total set of 11,801 mandated, self-archived articles, 279 articles (2.4%) had been published in the 63 Gold OA journals (2.6%) among the 2,391 ISI-indexed journals in which the authors from our four mandated institutions had published in 2002-2008. Both these estimates of percent Gold OA are about half as big as the total 5% proportion for Gold OA journals among all ISI-indexed journals (active in the past 10 years). To be conservative, we can use the higher figure of 5% as a first estimate of the Gold OA contribution to total OA among all ISI-indexed journals.
Now, in our sample, we find that out of the total number of articles published in ISI-indexed journals by authors from our four mandated institutions between 2002-2008 (11,801 articles), about 65.6% of them (7,736 articles) had indeed been made Green OA through self-archiving by their authors, as mandated (7,457 or 63.2% Green only, and 279 or 2.4% both Green and Gold).
In contrast, for our 42,395 keyword-matched, non-mandated control articles, the percentage OA was 23.4% (21.9% Green and 1.5% Gold).
Björk et al?s (2010) corresponding finding [Table 3] for their ISI sample (1282 articles for 2008 alone, calculated in 2009), was 20.6% total OA (14% Green plus 6.6% Gold). (For an extended sample that also included non-ISI journals it was 11.9% Green plus 8.5% Gold.)
The variance is probably due to different discipline blends in the samples (see Björk et al’s Figure 4, where Gold exceeds Green in bio-medicine), but whichever overall results one chooses ? whether our 21.9% Green and 1.5% Gold or Björk et al?s 14% Gold and 6.6% Green (or even their extended 11.9% Green and 8.5% Gold), the figures fail to bear out Velterop?s hunch that:
Moreover (and this is really the most important point of all), Velterop’s hunch is the wrongest of all precisely where OA is mandated, for there the percent Green is over 60%, and headed toward 100%. That is the real power of Green OA mandates.
Gargouri, Y., Hajjem, C., Lariviere, V., Gingras, Y., Brody, T., Carr, L. and Harnad, S. (2010) Self-Selected or Mandated, Open Access Increases Citation Impact for Higher Quality Research. PLOS ONE 10(5)
Björk B-C, Welling P, Laakso M, Majlender P, Hedlund T, et al. (2010) Open Access to the Scientific Journal Literature: Situation 2009. PLOS ONE 5(6): e11273.
Of downloads and lemmings
Following is my contribution to a recent discussion on the American Scientist Open Access Forum on the topic of downloads:
What the institutional repository does with respect to download counts, is to direct traffic to the university’s site. This will increase the university’s web presence and hence emerging institutional web impact assessments, likely to be of increasing importance in years to come.
Having said that, I would also like to point out that there are serious dangers to scholarship that come with over-emphasis on the numbers. What is popular from a scholarly perspective at one point in time is not necessarily what is important.
One way to think of this: imagine that we humans are like a group of lemmings rushing madly towards a cliff (given the climate crisis and our limited attention to this, I would argue that this is a reasonable comparison). Any lemming that says (or writes) about – how to get to the cliff even faster – is likely to be well-heeded (and cited, if it is an academic lemming). On the other hand, the scholar who looks ahead and sees the cliff and shouts off (or writes up for a peer-reviewed lemming paper): “Hey! Cliff ahead! Should be change direction? ” may not get much attention immediately. (Later on, after the early birds have gone over the cliff, could be a different story).
Another important point: one real danger of usage statistics is the potential for usage-based pricing. If we go this route, it is just a matter of time before some of us impose limits on reading. If the undergraduate research project becomes a cost item, there will be a strong incentive to limit research and/or eliminate research projects. When one copy of an article can easily serve anyone, anywhere, it would be a shame to go this route. (Thanks to Andrew Odlyzko for pointing to this danger).
For a more in-depth look at this topic, please see my book chapter, “The implications of usage statistics as an economic factor in scholarly communications”, OA copy in the SFU IR at:
http://ir.lib.sfu.ca/handle/1892/1639
Comment: if there is one single possibility that tipped the balance for me to become an open access advocate, it is the spectre of usage-based pricing and the evil that this entails, at least for scholarship.
Wanted: a don’t-be-evil online music store
Update August 23: http://www.emusic.com/ has been suggested. Interesting – DRM-free music, but on the other hand an ongoing monthly financial commitment, which I don’t like, with emusic reserving the right to change the terms and conditions at any time. Not sure about this, but definitely much better than itunes.
Much as I like the idea of the convenience of buying music online (whether for download or CDs), I can’t bring myself to sign the itunes agreement. This agreement says that I can only use what I buy in Canada, and apple reserves the right to use spy technology to enforce this. Nor am I keen on purchasing through Amazon, since I don’t see an easy way to filter out music produced by companies that employ evil sue-their-customers-even-kids techniques and/or who advocate for evil copyright protection. Perhaps an Amazon “no-RIAA members” button would suffice?
Given the popularity of the international Pirate Parties, maybe it’s not just me?
On the other hand, maybe I should just create my own music.
Open access advocacy for the procrastinator
Best excuse for procastinating I’ve come up with in a long time: what do you mean, this won’t be open access? Suitable for authors of all toll access publishers.
Worlds of differences between publishers (economics 101)
Many librarians have never studied economics. Now that we are playing a key role in transforming scholarly communication, we need to know a little about the topic, at least as it applies to the stuff we work with. This is true whether we’re in library management, collections, or at the front lines working with faculty on scholarly communication issues. This post is the first in a series designed to provide an easy reading introduction to librarians on this topic and for this purpose, called Economics 101.
Today’s topic introduces the concept of the worlds of differences that exist between publishers. I will be using one of my favorite examples – since it is both so extreme, and from LIS. Emerald’s Library Management, with a subscription price of Euro 11819 according to Ulrich’s (that’s 15,138 USD by today’s conversion rate), costs about 190 times as much as ACRL’s College and Research Libraries (similar in number of peer-reviewed articles per year), at a maximum non-member non-US rate of 80 USD. While this is an extreme example, it nonetheless illustrates an important point: the journals of the for-profit commercial sector can cost not only more than the not-for-profits, the commercials often cost a very great deal more.
This is important, because if we would like to see a healthy, affordable future for scholarly communication, we need to understand and support the often small, not-for-profit publishers, whether they are our own, as in the case of C&RL, or help our faculty members to see the importance of supporting such journals in their own disciplines.
Following are a few ways to think about this:
- When it comes time for renewal, budgets are tight, and of course we want publishers to keep their prices down. But does it make sense to ask the same sacrifices of a publisher of a journal that costs only $80 and another very similar journal that costs more than $15,000?
- If we thought about the important work that ACRL does and how low that $80 subscription ($70 in the U.S., by the way) is, and asked ACRL to double the price for C&RL, and negotiated with Emerald for a 1% decrease in the cost for Library Management, we would save money (1% of $15,000 is $150 – $80 for doubling the cost of C&RL). We would save $70.
- Or, why not ask Emerald to be just a bit more reasonable, and halve the price of Library Management to just $7,500 U.S.? This would save us enough money to pay for subscriptions to 94 journals like C&RL.
The irony is, that when we explain the dire straights of library budgets to publishers, it is the ACRLs of this world who are the first to step up and assure us that they will keep prices flat.
A simple way to keep this difference in mind in the field: when comparing journals or publishers and their pricing changes over the years, keep in mind the actual price – not just the percentage of increase or decrease. A 1% increase to C&RL is $.80; a 1% increase to Library Management is $150, almost double the full subscription cost of C&RL.
This post is the first in the series Economics 101. While the examples here are subscription-based journals, overall the purpose of understanding the economics is transitioning to open access. What’s the connection? Publishers that have never gouged anyone, but have rather always kept their prices reasonable, often (like C&RL) have a well-deserved reputation for high quality. If they depend on library subscription dollars – let’s help them to make the transition to open access. When I talk to people involved with many of these journals, what I hear is that many of them would love to make such a transition – they just need a little help, and if librarians step up to help them, many would welcome the help.
While Waiting for Mandates: The Fruits of Tireless Advocacy at the Open University
Open Research Online:
A self-archiving success story.
Smith, Colin; Yates, Christopher and Chudasama, Sheila (2010)
5th International Conference on Open Repositories 6-9 July 2010, Madrid, Spain.Colin Smith [CS] et al:
“…Much debate exists in the literature, on listservs, and in the blogosphere as to whether a successful and sustainable repository can be achieved solely through advocacy, management, and development, or whether this is only likely to happen if an institutional mandate is introduced. A much quoted figure is that a non-mandated repository is only likely to capture around 15% of its institution’s research output, and at the very most (for an Incentivised Repository) 30% (Harnad, 2009).
”In this poster, we use the example of Open Research Online – the research repository of the Open University – to show that dedicated management and active development and advocacy of an institutional repository can lead to very successful results under the self-archiving model, in this case capturing regularly an estimated 60% of peer-reviewed journal output. Also demonstrated is the significant rise in full text (i.e. fully open access) items in the repository since the implementation of this approach.”
Note Added:
”To clarify [for Andrew Adams]… it is not 60% of the estimated refereed journal output of the OU that is deposited in full text format; a proportion is metadata-only items. Looking at my data over the last year, the split is around 56/44 (full text/metadata-only).“
SH:
Is that 34% [56% x 60%] of OU’s yearly research article output? How did you do that estimate? (We used Thompson/Reuters Web of Science data to estimate staff annual article output.)
CS:
“I look at it monthly based on Web of Science (WoS) data only (one month and six months after being indexed by WoS), but then I look at it on an annual basis based on WoS and Scopus data combined. I carry out the monthly check manually (i.e. by subscribing to WoS alerts and checking if they’ve been included in ORO), but for the annual analysis I got my developer to write a script which cross-checks DOIs and paper titles from de-duplicated WoS and Scopus data.
”In case it wasn’t clear, the poster was not about saying that mandates aren’t needed. It was more about demonstrating that if you have to hang around waiting for one to be implemented, or if you are at an institution where your superiors are against introducing one, encouraging results can still be achieved“.
SH:
Would you be willing to give me permission to post it, along with your response, above? — Of course, what would be even more helpful would be if I could also add a passage in your own words about what needs to be done at an institution that does not yet have a mandate, in order to achieve a deposit rate as good as OU’s [56% x 60% = 34%]. It would give institutions and IR managers something concrete to do while they are trying to get a deposit mandate adopted!
CS:
”Ok, here are a few thoughts…
On the advocacy front, I would say giving people actual real-world examples/demonstrations of the benefits of green OA. If you know you are talking to, say, the Biology Department about the repository, prepare in advance some impressive Google searches featuring some journal articles already deposited by Biologists from your institution. I normally give them two scenarios to think about:
”(1) Imagine an academic who knows the title of the paper she wants to read… she’s tried her Library and found her institution doesn’t subscribe to the journal… so she plugs it into Google… hey presto, there it is, in the top 10 hits, in our repository… now she clicks through… wonderful, the full text is available… she goes on to cite the paper… as might many other people from around the world who find the paper in a similar way, and so on…
”(2) Imagine an academic conducting a general literature research… most turn straight to Google these days… they plug in “spinal cord repair research” [or whatever search you have prepared in advance]… and there you go, a paper from your department in the repository in the top 10 Google hits, and so on.
”Demonstrating the power of the repository leaves a lasting memory in the minds of the people you are addressing, rather than just a bunch of woolly statements that send your audience to sleep.
”The other thing is what I call “individual advocacy”, which basically means taking the time to communicate and educate individuals at your institution about open access, copyright etc., rather than relying solely on getting yourself on the bill of department or research meetings. No one likes being in meetings, and so having to spend another half an hour listening to someone from the Library bang on about the repository doesn’t always work, unless you capture their attention with good examples, as mentioned above. However, keeping an eye on what’s being published by your institution, and then getting in touch with an academic and talking to them about the research their group is working on… and have they thought about where they are going to publish… and do they know what rights they retain when publishing in that journal… and have they thought about open access etc. etc., I find tends to get you results. And then, of course, when you start to make some inroads in this way, the message spreads outwards from there.
”Moving away from advocacy, I think it’s really important to have a technical developer onboard, or at least access to one that can make the improvements to the repository you as the manager want and know your academic community needs. As part of this, embedding data feeds from the repository in as many other institutional systems as possible across campus is a real winner. For example, one of our faculties is currently rebuilding its entire website and, in the process, is using data from the repository to build publications lists for its staff and research groups. So, if academics in this faculty do not keep the repository up to date, then the publication lists in their staff pages and research group pages on their faculty web pages will also be out of date.
”I could go on, but I imagine this is enough for you…” — Colin Smith
Cultural Studies: for-profit journals cost up to 10 times more than not-for-profit
Striphas analyzes the relative costs of journals from different publishers in cultural studies, and finds that the journals produced by commercial publishers cost from 50% to 1000% more (10 times the cost) than the not-for-profit university press journals, with the average price at Sage the highest at $853. Striphas’ analysis on why addresses added value to the publisher databases (bells and whistles for searching), but notes that the main reason is because publishers can charge this much, because of the inelastic market. The section on alienation in this article is useful for those working on author’s rights.
Citation: Striphas, T. (2010). Acknowledged goods: Cultural studies and the politics of academic journal publishing. Communication and Critical/Cultural Studies, 7(1), 3.
Authors’ Drafts, Publishers’ Versions-of-Record, Digital Preservation, Open Access and Institutional Repositories
Commentary on Richard Poynder’s
“Preserving the Scholarly Record:
Interview with digital preservation specialist Neil Beagrie”
The trouble with universities (or nations) treating digital preservation (which is a genuine problem, and a genuine responsibility) as a single generic problem — covering all the university’s (or nation’s) “digital output,” whether published or unpublished, OA or non-OA — is not only that adding an additional preservation cost and burden where it is not yet needed (by conflating Green OA self-archiving mandates with “preservation mandates” and their funding demands) makes it even harder to get a Green OA self-archiving mandate adopted at all. But taking an indiscriminate, scattershot approach to the preservation problem also disserves the digital preservation agenda itself.
As usual, what is needed is to sort out and understand the actual contingencies, and then to implement the priorities, clearly and explicitly, in the requisite causal order. The priorities here are to focus university (or national) preservation efforts and funds on what needs to be preserved today. And — as far as universities’ own institutional repositories (IRs) are concerned — that does not include the publisher’s official version-of-record for that university’s (or nation’s) journal article output. Preserving those versions-of-record is a matter to be worked out among deposit libraries and the publishers and institutional subscribers of the journals in question. Each university’s own IR is for providing OA to its own authors’ final, refereed drafts of those articles, in order to make them accessible to those users worldwide who do not have subscription access to the version-of-record. The author’s draft does indeed need preservation too, but that’s not the same preservation problem as the problem of preserving the published version-of-record (nor is it the same document!).
Perhaps one day universal Green OA mandates will cause journal subscriptions to become unsustainable, because the worldwide users of journal articles will be fully satisfied with just the author’s final drafts rather than needing the publisher’s version-of-record, and hence journal subscriptions will be cancelled. If and when we ever reach that point, the version-of-record will no longer be produced by the publisher, because the authors’ drafts will effectively become the version-of-record. Journal publishers will then convert to Gold OA publishing, with what remains of the cost of publication paid for by institutions, per individual article published, out of their windfall subscription cancellation savings. (Some of those savings can then also be devoted to digital preservation of the institutional version-of-record.)
But conflating the (nonexistent) need to pay for this hypothetical future contingency today (when we still have next to no OA or OA mandates, and subscriptions are still going strong) with either universities’ (or nations’) digital preservation agenda or their OA IR agenda is not only incoherent but counterproductive.
Let’s keep the agendas distinct: IRs can archive many different kinds of content. Let’s work to preserve all IR content, of course, but let’s not mistake that IR preservation function for journal article preservation or OA.
For journal articles, worry about preserving the version-of-record — and that has nothing to do with what is being deposited in IRs today.
For OA, worry about mandating deposit of the author’s version — and that has nothing to do with digital preservation of the version-of-record.
Nor should the need to mandate depositing the author’s version be in any way hamstrung with extra expenses that concern the publish’s version-of-record, or the university’s IR, or OA. (Exactly the same thing is true, mutatis mutandis, at the national preservation level, insofar as journal articles are concerned: A journal’s contents do not all come from one institution, nor from one nation.)
And, while we’re at it, let’s also keep university (or national) funding of Gold OA publishing costs distinct from the Green OA mandating agenda too. First things first. Needlessly over-reaching (for Gold OA funds or preservation funds) simply delays getting what is already fully within universities’ (and nations’) grasps — which is the newfound (but mostly unused) potential to provide OA to the authors’ drafts of all their refereed journal articles by requiring them to be deposited in their OA IRs (not by reforming journal publishing, nor by solving the digital preservation problem).
Stevan Harnad
American Scientist Open Access Forum
LIBER Keynote: The role of the research library in an emerging global public sphere (LIBER keynote)
The webcast of my keynote speech at the LIBER 2010 Conference, The role of the research library in an emerging public sphere, is now available.
Marking and Tagging the Public Domain
I am cribbing significant amounts of this post from a Creative Commons blogpost about tagging the public domain. Attribution is to Diane Peters for the stuff I’ve incorporated 🙂
The big news is that, 18 months since we launched CC0 1.0, our public domain waiver that allows rights holders to place a work as nearly as possible into the public domain, worldwide…it’s been a success. CC0 has proven a valuable tool for governments, scientists, data providers, providers of bibliographic data, and many others throughout world. CC0 has been used by the pharmaceutical industry giant GSK as well as by the emerging open data leader Sage Bionetworks (disclosure – I’m on the Board of Sage – though not of GSK!).
At the time we published CC0, we made note of a second public domain tool under development — a tool that would make it easy for people to tag and find content already in the public domain. That tool, our new “Public Domain Mark” is now published for comment.
The PDM allows works already in the public domain to be marked and tagged in a way that clearly communicates the work’s PD status, and allows it to be easily discoverable. The PDM is not a legal instrument like CC0 or our licenses — it can only be used to label a work with information about its public domain copyright status, not change a work’s current status under copyright. However, just like CC0 and our licenses, PDM has a metadata-supported deed and is machine readable, allowing works tagged with PDM to be findable on the Internet. (Please note that the example used on the sample deed is purely hypothetical at the moment.)
We are also releasing for public comment general purpose norms — voluntary guidelines or “pleases” that providers and curators of PD materials may request be followed when a PD work they have marked is thereafter used by others. Our PDM deed as well as an upcoming enhanced CC0 deed will support norms in addition to citation metadata, which will allow a user to easily cite the author or provider of the work through copy-paste HTML.
This is absolutely critical to science, because it addresses at last the biggest reason that people mis-use copyright licenses on uncopyrightable materials and data sets: the confusion of the legal right of attribution in copyright with the academic and professional norm of citation of one’s efforts. Making it easy to cite, regardless of the law, is one of the keys to making the public domain something that we can construct through individual private choice at scale, not just by getting governments to adopt.
The public comment period will close on Wednesday, August 18th. Why so short? For starters, PDM is not a legal tool in the same sense our licenses and CC0 are legally operative — no legal rights are being surrendered or affected, and there is no accompanying legal code to finesse. Just as importantly, however, we believe that having the mark used soon rather than later will allow early adopters to provide us with invaluable feedback on actual implementations, which will allow us to improve the marking tool in the future.
The primary venue for submitting comments and discussing the tool is the cc-licenses mailing list. We look forward to hearing from you!
There are a lot of fascinating projects around how to do the non-legal work of data. The Sage Commons has seen a bunch of them come together, but in this context I want to call out – the SageCite project driven by UKOLN, the University of Manchester, and the British Library – which is going to develop and test an entire framework for citation, not attribution, using bioinformatics as a test case.
My own hope is that by making citation inside Creative Commons legal tools that work on the public domain a cut-and-paste process, we can facilitate the emergence of frameworks like SageCite so that the legal aspects fade away on the data sets and databases themselves, and the focus can be on the more complex network models of complex adaptive systems. And I’m tremendously excited to see members of the community leveraging the Sage project to do independent, crucial work on the topic of citation. Like Wikipedia, Sage won’t work unless it is something that we all own together and work on for our own reasons.
This is only still the beginning of really open data – public domain data – that complies with the Panton Principles. Creative Commons has spent six long years studying the open data issue, and rolling out policy and tools and technologies that make it possible for end users from the Dutch government to the Polar Information Commons to create their own open data systems.
We still have to avoid the siren song of property rights on data, and of license proliferation. But it’s starting to feel like momentum is gaining on public domain data, and for the Creative Commons tools that make it a reality. Making citation one-click, and making it easy to tag and mark the public domain, is part of that momentum. Please help us by commenting on the tools, and by promoting their use when you run across any open data project where the terms are unclear.
Read the comments on this post…
Marking and Tagging the Public Domain
I am cribbing significant amounts of this post from a Creative Commons blogpost about tagging the public domain. Attribution is to Diane Peters for the stuff I’ve incorporated
The big news is that, 18 months since we launched CC0 1.0, our public domain waiver that allows rights holders to place a work as nearly as possible into the public domain, worldwide…it’s been a success. CC0 has proven a valuable tool for governments, scientists, data providers, providers of bibliographic data, and many others throughout world. CC0 has been used by the pharmaceutical industry giant GSK as well as by the emerging open data leader Sage Bionetworks (disclosure – I’m on the Board of Sage – though not of GSK!).
At the time we published CC0, we made note of a second public domain tool under development — a tool that would make it easy for people to tag and find content already in the public domain. That tool, our new “Public Domain Mark” is now published for comment.
The PDM allows works already in the public domain to be marked and tagged in a way that clearly communicates the work’s PD status, and allows it to be easily discoverable. The PDM is not a legal instrument like CC0 or our licenses — it can only be used to label a work with information about its public domain copyright status, not change a work’s current status under copyright. However, just like CC0 and our licenses, PDM has a metadata-supported deed and is machine readable, allowing works tagged with PDM to be findable on the Internet. (Please note that the example used on the sample deed is purely hypothetical at the moment.)
We are also releasing for public comment general purpose norms — voluntary guidelines or “pleases” that providers and curators of PD materials may request be followed when a PD work they have marked is thereafter used by others. Our PDM deed as well as an upcoming enhanced CC0 deed will support norms in addition to citation metadata, which will allow a user to easily cite the author or provider of the work through copy-paste HTML.
This is absolutely critical to science, because it addresses at last the biggest reason that people mis-use copyright licenses on uncopyrightable materials and data sets: the confusion of the legal right of attribution in copyright with the academic and professional norm of citation of one’s efforts. Making it easy to cite, regardless of the law, is one of the keys to making the public domain something that we can construct through individual private choice at scale, not just by getting governments to adopt.
The public comment period will close on Wednesday, August 18th. Why so short? For starters, PDM is not a legal tool in the same sense our licenses and CC0 are legally operative — no legal rights are being surrendered or affected, and there is no accompanying legal code to finesse. Just as importantly, however, we believe that having the mark used soon rather than later will allow early adopters to provide us with invaluable feedback on actual implementations, which will allow us to improve the marking tool in the future.
The primary venue for submitting comments and discussing the tool is the cc-licenses mailing list. We look forward to hearing from you!
There are a lot of fascinating projects around how to do the non-legal work of data. The Sage Commons has seen a bunch of them come together, but in this context I want to call out – the SageCite project driven by UKOLN, the University of Manchester, and the British Library – which is going to develop and test an entire framework for citation, not attribution, using bioinformatics as a test case.
My own hope is that by making citation inside Creative Commons legal tools that work on the public domain a cut-and-paste process, we can facilitate the emergence of frameworks like SageCite so that the legal aspects fade away on the data sets and databases themselves, and the focus can be on the more complex network models of complex adaptive systems. And I’m tremendously excited to see members of the community leveraging the Sage project to do independent, crucial work on the topic of citation. Like Wikipedia, Sage won’t work unless it is something that we all own together and work on for our own reasons.
This is only still the beginning of really open data – public domain data – that complies with the Panton Principles. Creative Commons has spent six long years studying the open data issue, and rolling out policy and tools and technologies that make it possible for end users from the Dutch government to the Polar Information Commons to create their own open data systems.
We still have to avoid the siren song of property rights on data, and of license proliferation. But it’s starting to feel like momentum is gaining on public domain data, and for the Creative Commons tools that make it a reality. Making citation one-click, and making it easy to tag and mark the public domain, is part of that momentum. Please help us by commenting on the tools, and by promoting their use when you run across any open data project where the terms are unclear.
Equity for Open-Access Journal Publishing
Compact for Open-Access Publishing Equity
Mismeasure of the Needs of Research, Researchers, Universities and Funders
Phil Davis continues his inexplicable preoccupation with what’s best for the publishing industry, at the expense of what is best for research, researchers, their institutions, their funders, and the tax-paying public that funds them — the ones by and for whom the research funded.
Phil’s latest point is that the low uptake by authors of committed Gold OA funds indicates that authors don’t really want OA.
But authors themselves have responded quite clearly, and repeatedly, in Alma Swan’s international surveys, that they do need, want, and value OA. Yet they also state that they will only provide OA to their own writings if their institutions and funders require (i.e., mandate) them to provide it.
(There is definitely a paradox here, but it is not resolved by simply assuming that authors don’t really want OA! It’s rather more complicated than that. Authors would not publish much, either, if their institutions and funders did not require them to “publish or perish.” And without that, where would the publishing industry be?)
I have dubbed the condition “Zeno’s Paralysis,” which is the fact that for a variety of reasons (38 at last count) — all groundless and easily defeasible but relentlessly recurrent nonetheless, including worries about copyright, worries about getting published, worries about peer review, and even worries about the time and effort it might require to provide OA — most authors will not provide OA spontaneously. And the cure is not only known, but has already been administered by over 150 institutions and funders: Mandate OA.
And the only form of OA that institutions and funders can mandate is Green OA self-archiving of all peer-reviewed journal articles, immediately upon acceptance for publication.
Apart from that, all they can do is provide some of their scarce funds (largely tied up in paying for subscriptions) to pay for a little Gold OA publishing. The uptake for that is even lower than for unmandated Green OA self-archiving, but that’s certainly not evidence against Alma Swan’s survey findings about what authors want, and what it will take to get them to provide it (and Arthur Sale’s data confirming that authors actually do as they say they will do, if mandated).
Ironically, the book Phil cites — Stephen Jay Gould‘s “Mismeasure of Man” — was also a critique, some of it well-taken, but likewise tainted by an ideological bias. In Gould’s case the ideological bias was far, far more noble than Phil’s parochial bias in favor of publishers’ interests and the status quo: Gould was an idealistic marxist whose bias was against evidence indicating that there are genetic differences between human populations. In the core chapter of that book — “The Real Error of Cyril Burt” — Gould tried to show that psychometricians had made a methodological and conceptual error in interpreting the “G” factor — a strong positive common factor that kept being observed among multiple, diverse psychometric tests across diverse populations — as if it were an empirical finding, whereas (according to Gould) it was merely an inevitable result of their method of analysis: the “positive-definite manifold” that invariably results from the construction of a lot of positively correlated tests. (What Gould missed was that there was nothing about the analytic method that guaranteed that the joint correlations should all be positive, and certainly not that the resultant G factor should turn out, as it did, to be correlated with genetic differences between populations.)
The consensus in the field of psychometrics and biometrics today is that in this case the real error was Gould’s.