Edit: A summary of the recent Open Science session at the Berlin Colloquium on Internet and Society with talks from Constanze Engelbrecht, Pasco Bilic and Christoph Lutz has been posted by the Humboldt Institute’s Benedikt Fecher (german version, english version). The text below is a more general discussion of how Open Science can be defined.
One area of research at the Alexander von Humboldt Institute is Open Science, an emerging term used to describe new ways of conducting research and communicating its results through the Internet. There is no single definition of what constitutes Open Science (and one could argue there doesn’t really need to be), but in this blog entry I want to point to attempts to define the term by prominent scientists and activists, and discuss some of the limitations of these definitions. I’ll summarize my observations in the form of five questions that suggest a direction that future research into Open Science could take.
Open Science: a few working definitions
Michael Nielsen – a prominent scientist and author on Open Science whose name pops up invariently when discussing the topic – provides this very comprehensive definition in a post to the Open Science mailing list:
“Open science is the idea that scientific knowledge of all kinds should be openly shared as early as is practical in the discovery process.”
In the same vein, Peter Murray-Rust, a professor in molecular chemistry and Open Access advocate, provides another definition (also through the OKFN’s open science mailing list):
“In a full open science process the major part of the research would have been posted openly and would potentially have been available to people outside the research group both for reading and comment.”
(Also see this interview if you want a more detailed exposition).
Finally, Jean Claude Bradley, also a professor in chemistry, provides a definition of what he calles Open Notebook Science, a very similar approach:
“[In Open Notebook Science] there is a URL to a laboratory notebook that is freely available and indexed on common search engines. It does not necessarily have to look like a paper notebook but it is essential that all of the information available to the researchers to make their conclusions is equally available to the rest of the world.”
(Here’s a presentation summarizing his approach, Open Notebook Science. A similar view is articulated by M. Fabiana Kubke and Daniel Mietchen in this video, though they prefer the term Open Research.)
From natural philosophy to science
One thing that these different definitions have in common is the way in which they frame science. In English, the word science has come to denote primarily the natural sciences (traditionally physics and chemistry, more recently also biology and life sciences). The history of the term is long and complex (check out the Wikipedia entry), but as a result of language change, a wide range of disciplines are considered not to be part of the sciences, but instead belong to the social sciences and Humanities.
Why does this matter? The above definitions are very closely tailored to the methods and organizational structures of the natural sciences. They assume that research is conducted in a research group (Murray-Rust) that works primarily in a laboratory and whose members record the steps of an experimental process in a lab notebook (Bradley), following a sequence of more or less clearly-structured steps that can be summarized as “the discovery process” (Nielsen).
Research processes in other fields strongly differ from this approach, not just in the Humanities (where there is frequently no research group, and data is of varying relevance), but also in the social sciences (where there is generally no laboratory, and data frequently comes from human subjects, rather than technical instruments such as radio telescopes or DNA sequencers). Beyond just using different tools, the instruments also shape the assumptions of their users about the world, and about what they do. Sociologist Karin Knorr-Cetina points to this difference in the title of her book Epistemic Cultures, and similar observations have been made in Bruno Latour & Steve Woolgar’s Laboratory Life: The Construction of Scientific Facts. One crucial aspect of this is how data is conceptualized in the different disciplinary perspectives, and, related to this, how notions differ regarding what openness means.
Openness beyond open access to publications
Openness can be defined in a variety of ways. Not all information that is available online is open in a technical sense – just think about proprietary file formats that make it difficult to share and re-use data. Technical openness does not equal legal openness, a problem that is also on the institute’s agenda.
Open Access – the technical and legal accessibility of scholarly publications via the Internet – is widely regarded to benefit both science and the public at large. In the traditional publishing model access to research results in scholarly monographs and journals is available to subscribers only (usually institutional subscribers, in other words, libraries). The Open Access model shifts the costs, sometimes to authors (who pay a fee to publish) or to publishing funds or other institutional actors. The Budapest and Berlin Declarations on Open Access specify under which provisions publications are truly Open Access, rather than just somehow accessible. Open Access has a range of benefits, from reducing costs and providing access to scientists at small universities and in developing countries, to increasing transparency and raising scholarly impact. Models based on author fees, such as the one utilized by PLoS, are increasingly common and make Open Access economically feasible.
There is broad consensus that Open Access is a first step, but that it’s not enough. Many scientists, such as the ones cited above, call for research data also to be made available more broadly. Sharing research data, instead of packaging data and analysis together in scholarly articles, could enable new forms of research that are much more complementary than current practices, which tend to emphasize positive outcomes (experiments that worked) over negative ones (those that didn’t), despite the fact that negative outcomes can greatly contribute to better understanding a problem.
Making openness count
The barriers to achieving a more open environment in regards to research data aren’t primarily technical or legal, but cultural. Research has always been based on the open dissemination of knowledge (just take the history of the Philosophical Transactions, considered by most to be the oldest scientific journal), but it is also very closely tied to the formats in which knowledge is stored and disseminated, such as books, journal articles, and conference papers, which tend to take on a valorizing role, rather than being just arbitrary containers of scholarly information. Many scholars, regardless of their field, see themselves in the business of publishing books, articles, and papers just as much as they consider themselves to be in the business of doing research. While the technology behind scholarly publishing has changed dramatically, the concepts have not changed. Because institutionalized academia is incentive-driven and highly competetive, collective goals (a more efficient approach to knowledge production) are trumped by individual ones (more highly-ranked publications = more funding and promotions for the individual researcher).
Institutional academia is no longer the only place where research happens. Increasingly, there is (if latently) competition from crowdsourcing platforms that facilitate collaborative knowledge creation (and, more open, problem solving) outside of institutional contexts. Depending on how you define the process of knowledge production, examples include both Wikipedia and projects such as the #SciFund Challenge. The approach to knowledge production in these environments seems to focus on knowledge recombination and remixing at the moment, but it appears plausible that more sophisticated models could arise in the future. Whether the hybrid communities of knowledge production have a potential to displace established institutional academia remains to be seen. Rather, such communities could blossom in those areas where traditional academia fails to deliver.
But even inside institutional academia, the time seems ripe for more openness beyond making publications and data available to other academics. Social media makes it possible for scholars to both communicate with their peers and engage with the public more directly — though they are still hesitant to do either at the moment. Public visibility is not as high on the agenda of most researchers as one might expect, because academic success is largely the result of peer, not popular evaluation.
Redefining scholarly impact
This may change as new, more open measurements of scholarly impact enter the mainstream. Measuring and evaluating the impact and quality of publicly-funded research has been a key political interest for decades. While frameworks exist for conducting large and complex evaluations (Research Assessment Exercises in the UK, Exzellenzinitiative in Germany) the metrics used to evaluate the performance of researchers are generally criticized as too one-dimensional. This criticism applies in particular to measuerments that indicate the quality of publications such as Thompson Reuters’ Impact Factor (IF). A confluence of measures (downloads, views, incoming links) could change the current, extremely one-sided approach to evaluation and make it more holistic, generating a more nuanced picture of scholarly performance.
Questions for research into Open Science
The following questions reflect some of the issues raised by “open” approaches to science and scholarship. They are by no means the only ones, als the Open Science project description on the pages of the HIIG highlights, but reflect my personal take.
- How can Open Science be conceptualized in ways that reach beyond the paradigm of the natural sciences? In other words, what should Open Humanities and Open Social Sciences look like?
- How do different types of data (recorded by machines, created by human subjects, classified and categorized by experts) and diverse methods used for interacting with it (close reading, qualitative analysis, hermeneutics, statistical approaches, data mining, machine learning) impact knowledge creation and what are their respective potentials for openness in the sense described by Nielsen, Murray-Rust and Bradley? What are limits to openness, e.g. for ethical, economic and political reasons?
- What are features of academic openness beyond open access (e.g. availability of data, talks, teaching materials, social media media presence, public outreach activities) and how do they apply differently to different disciplines?
- How can the above-mentioned features be used for a facetted, holistic evaluation of scholarly impact that goes beyond a single metric (in other words, that measures visibility, transparency and participation in both scientific and public contexts)?
- What is the relationship between institutionalized academia and hybrid virtual communities and platforms? Are they competitive or complementary? How do their approaches to knowledge production and the incentives they offer to the individual differ?
I’ve already shared this bit of personal news with a few friends and colleagues, but I thought I’d blog about it as well — especially since I’m woefully behind on my Iron Blogger schedule.
After a fairly long time in the making, I have been awarded a three-year research grant from the Deutsche Forschungsgemeinschaft (DFG) for the project Networking, visibility, information: a study of digital genres of scholarly communication and the motives of their users (summary in German on the DFG’s site). The project investigates new forms of scholarly communication (especially blogging and Twitter) and their role for academia. My key concerns are usage motives, i.e. why scholars use blogs and Twitter, and how these motives correspond with usage practices (how they blog and tweet), rather than how many researchers use these channels of communication or what makes them refrain from using them (see this blog post and the study mentioned in it for that kind of work). My main methods will be qualitative interviews with a sample of 20-25 blogging and/or tweeting academics, along with in-depth content analysis of the material they post in these channels over a prolonged period (>1 year). Identifying usage patterns and relating them to the participants’ narrative about their use will be another key objective. Ultimately, I hope to find a (tentative) answer to the question what role blogs and Twitter may play for the future of digital scholarship, and whether they will remain a niche phenomenon or become mainstream over time.
The project follows up on my work on corporate blogging and connects strongly to what we have been doing at the Junior Researchers Group “Science and the Internet” over the past year, but the focus on interviews should result in a more user-centric analysis. As someone who has been doing (applied) linguistic analysis to make inferences about social processes, I feel much more comfortable actually talking to the people I want to study, rather than just crunching numbers on how they tweet. Big data social science research is obviously and understandably en vogue these days, but I hope to find a good synergy between qualitative and quantitative approaches in my project.
My new institutional home for the next three years will be the Berlin School of Library and Information Science at Humboldt University. I’m grateful to Michael Seadle for supporting my project and really look forward to working with my new colleagues at IBI (that’s the German acronym, which, as far as I can tell, is preferred to its more entertaining English equivalent). I also look forward to working with colleagues from the Alexander von Humboldt Institute for Internet and Society (HIIG) where I’m currently supporting the project Regulation Watch. Finally, I plan to keep in close contact with the colleagues in Düsseldorf, both at the Junior Researchers Group and the Department of English Language and Linguistics, where I have learned virtually everything I know about being a researcher. I am especially indebted to Dieter Stein for his enduring support and for his contagious enthusiasm for all aspects of scholarship.
Sic itur ad astra!
For an overview of previous work I’ve done in this direction, have a look at my publications.
An interesting issue — especially to the library and information science community — that Google’s Max Senges raised at the Berlin Symposium on Internet and Society (#bsis11) was how the impact of the instiute’s research could be measured. HIIG’s mission is not just to produce excellent scholarship, but also to foster a meaningful dialog with a wide range of stakeholders beyond academia in relation to the issues that the institute investigates.
This approach has a number of implications that I want to briefly address. My views are my own, but I consider this an exciting test case for a modern, digital form of science evaluation. I believe three things can serve to make the institute’s research as transparent as possible:
- primary research results (i.e. papers) should be Open Access,
- journalistic contributions (essays, interviews, public speaking) beyond academic publications should be encouraged,
- communication of research via social media (blogs, Twitter) should be encouraged.
Open Access is of key importance
David Drummond emphasized the importance of Open Access in his speech at the Institute’s inauguration. A plausible step to make Open Access part of the institute’s culture could be to sign the Berlin Declaration and set up a dedicated repository of institute publications. HIIG could encourage its researchers to publish in gold road Open Access journals such as those listed in the DOAJ and encourage use of a green road approach par the SHERPA/Romeo list in the remaining cases. It could further encourage the use of Creative Commons or similar licences for scholarly publications.
Journalism and engagement with the general public
The public has a considerable interest in the issues investigated at HIIG and accordingly talking with and through traditional media channels will be of great importance. This should not merely be considered a form of marketing, but rather a form of dialog that will allow HIIG to fulfill its obligation to the public to act as an informed voice in civic debate around issues such as privacy and net neutrality. Engagement with the public via essays, interviews, public speaking and similar activities should be considered part of the institute members’ impact.
Social media’s role for science communication
The institute could consider social media as a central avenue of engaging with a wider public and recognize the willingness to use it accordingly. Scholarly blogging, for example, should be considered as part of a member’s research output instead of being regarded as a chiefly private enterprise. Social media activity cannot supplant traditional scholarly publishing, but it can serve to conduct conversations around research, get the attention of non-academics, and point to formal publications, among other things.
So how could this be implemented? The first and second points — making primary research results available and promoting journalistic contributions — are already standard practice elsewhere. The third is a little more tricky. Should it be important how many friends a researcher has on Facebook, or followers on Twitter (assuming he/she is even on these platforms)? Such an approach would be much too simplistic, but perhaps something a little more nuanced could be tried. How about encouraging the use of the #hiig (hash)tag wherever possible and continuously tracking the results? The institute could run its own blog — this may or may not work well, given that many contributors might already have their own one — or a blog planet, a site that just aggregates material from existing blogs that is #hiig-tagged.
These are just general ideas, but eventually they could coalesce into a framework for evaluating HIIG’s impact beyond purely scholarly (and faulty) forms of measurement such as the impact factor.
I’m currently relaxing at HIIG HQ, watching the staff make final preparations for the Institute’s formal inauguration, which will take place at 5pm today at Humboldt University’s Audimax (do drop by if you’re in the area, even if you haven’t been formally invited). I thought I’d share two statements on the launch of the Institute from Google, which were posted today and yesterday.
“Interaktion von Internet, Forschung und Gesellschaft verstehen” (in German)
David Drummond, VP Google, in German newspaper DIE ZEIT
Launching and Internet & Society Research Institute
Max Senges, Google Policy, Google European Public Policy Blog
I’m heading to Berlin on an early morning train, among other things for next week’s Berlin Symposium on Internet and Society (#hiig2010). The program is available here and should catch your attention if you’re in Internet research. Be sure to give me a shout if you’re coming and want to have drinks some time.
The Symposium kicks off with the formal inauguration of the newly founded Alexander von Humboldt Institute for Internet and Society (HIIG). This is the official name of what has so far been referred to by most people in the field as “the Google Institute” since the plan to launch it was publicly announced by Eric Schmidt in February.
I’m involved with the Institute as a project associate, which means that I’ll be working on one specific issue (a platform dubbed Regulation Watch — more on that soon) for the next few months. I’m excited to be part of an inspiring and highly interdisciplinary team of people who are all studying the Internet’s effect on culture and society in one way or another, which is an especially exciting prospect if you’ve been more or less on your own with your research interest in this area for the better part of your career.
As I’ve been meaning to get back to blogging anyway, I’ve decided to post updates on what’s happening at HIIG (or “hig”, rhyming with “twig”, as I’ve decided to call my new employer in spoken English) on a semi-regular basis. Next week’s inuaguration and symposium will be covered here with occasional short updates, news flashs and comments, as well as links to stuff other people have posted.
The institute’s mission is both to conduct research and to engage in an ongoing dialogue with the general public, an idea that is very much in accord with the vision of it’s patron. Alexander von Humboldt was a scientist, explorer, diplomat and, frankly, somewhat of a crazy person for trying things that most of his contemporaries considered both mad and futile. His research resonated with society and frequently stirred controversy. He challenged widely held beliefs about the world and was unwavering in his dedication to shedding light on scientific truth beyond superstition. I’m excited by HIIG’s aim to do something similar for our time’s unchartered continent — the net — and look forward to contributing to this goal.