This posting is sponsored by the Library Publishing SIG and published in cooperation with the ARL Section. Members of the Library Publishing SIG reach out to library publishers and invite them to respond to a series of questions.
This post features Dave Ghamandi who is the Open Publishing Librarian at the University of Virginia and Managing Editor, Aperio.
This article touches upon the problem of inequalities in academia resulting from neoliberal capitalism and existing publishing policy and discusses its possible consequences. Building on the author’s own experiences as a researcher working on linguistic minorities and as an academic administrator, it explores how power relations work in parts of the scientific world situated on the peripheries of the Western “centre” – via the neoliberal economy, access to funding and international recognition. Publishing in high-status, English-language journals requires “non-centre” academics to adopt Western conventions of publishing, including in the style of reasoning, the structure of the text, and preferring references from the Anglo-American academic tradition. Only by complying can such researchers secure a place in academia and further their careers. However difficult it may be, the author argues, the hegemony of Western-model knowledge construction may only be questioned from inside, by the “centre” academics.
Nature 26 Oct 2021--Catalogue of billions of phrases from 107 million papers could ease computerized searching of the literature. Catalogue of billions of phrases from 107 million papers could ease computerized searching of the literature.
In a project that could unlock the world’s research papers for easier computerized analysis, an American technologist [Carl Malamud]has released online a gigantic index of the words and short phrases contained in more than 100 million journal articles — including many paywalled papers.
The catalogue, which was released on 7 October and is free to use, holds tables of more than 355 billion words and sentence fragments listed next to the articles in which they appear. It is an effort to help scientists use software to glean insights from published work even if they have no legal access to the underlying papers, says its creator, Carl Malamud. He released the files under the auspices of Public Resource, a non-profit corporation in Sebastopol, California that he founded.
Malamud says that because his index doesn’t contain the full text of articles, but only sentence snippets up to five words long, releasing it does not breach publishers' copyright restrictions on the re-use of paywalled articles. However, one legal expert says that publishers might question the legality of how Malamud created the index in the first place.
Nature, July 2019. -- A giant data store quietly being built in India could free vast swathes of science for computer analysis — but is it legal? A giant data store quietly being built in India could free vast swathes of science for computer analysis —but is it legal?
Over the past year, Malamud has — without asking publishers — teamed up with Indian researchers to build a gigantic store of text and images extracted from 73 million journal articles dating from 1847 up to the present day. The cache, which is still being created, will be kept on a 576-terabyte storage facility at Jawaharlal Nehru University (JNU) in New Delhi. “This is not every journal article ever written, but it’s a lot,” Malamud says. It’s comparable to the size of the core collection in the Web of Science database, for instance. Malamud and his JNU collaborator, bioinformatician Andrew Lynn, call their facility the JNU data depot.