[Event] HackYourPhD US wrap-up event on G+

The folks behind HackYourPhD (see last post) are wrapping up their US tour with an online event highlighting their work over the summer, interviewing key players in open science.

I’ll be joining the Hangout, along with William Gunn (Mendeley), Brian Glanz (Open Science Federation), Jeff Spies (Center for Open Science / Open Science Framework), Charles Fracchia and Adam Marblestone (Biobright), and Ann Lam and Elan L. Ohayon (Green Neuroscience Lab).

Do tune in (starts at 1pm EST).

Mozilla Science Lab and HackYourPhD

Mozilla Science Lab and HackYourPhD

A few months ago, I heard about a crowdfunded project called “HackYourPHD” – a community and resource building effort led by a researcher in France to collect information and run interviews with a diverse cast of characters in the open science world. The founder, Celya Gruson-Daniel (a neuroscientist by training), had become frustrated by many of the analog practices and mindsets in science, and decided to set out to crowdfund a trip through North America to do research on open research.

I had the pleasure of chatting with Celya last week at Columbia University, as she passed through New York on her nationwide tour. You can listen to the audio of our interview linked above, as well as listen to her interviews with other key folks in this space here: https://soundcloud.com/hackyourphd/

Intro to Open Data on P2PU

The second of P2PU’s Open Science course videos is now online, this one focusing on “Open Data”. I was delighted to be able to join this module, alongside course-master Billy Meinke, Heather Piwowar from Impact Story, Ross Mounce from OKFN and Wouter van de Bos from Max Planck. You can check out more of the video modules here: https://p2pu.org/en/courses/5/content/1370/.

Have a watch, ping us at @MozillaScience with any questions (or leave them in the comments), and if you’re feeling *really* creative, try remixing / splicing the video using some of the Webmaker tools:

https://webmaker.org/en-US/tools/#popcorn-maker

And let us know what you come up with!

The life sciences would seem, on the surface, ideal for open source. It’s a world built on disclosure – whether publication or patent – it doesn’t count until you tell the world. It’s a world where the knowledge itself snaps together in a fashion that looks eerily like a wiki, where one person only makes a small set of edits in an experiment that establishes a new fact. And it’s a world where the penalty for redundancy is high – no one in their right mind wants to spend scarce research dollars on a problem that has been solved already, a lead that is a dead end, a target guaranteed to lead to side effects.

John Wilbanks in his recent Xconomy piece, “Understanding Open Science”.

The life scienc…

[Experiment] Exploring code review for science

Over the years, science has evolved many quality control mechanisms, the best known of which is probably peer review. This first appeared nearly 300 years ago, and was initially put into place to allow trusted peers to scrutinize new findings, ideas, and societal implications as part of the publication process.

Here at the Mozilla Science Lab, we’re interested in finding ways of pushing the limits to what we think of as “science on the web” and instilling better practices for open, reproducible research. We aim to do this through community building, skills training (like Software Carpentry), and building prototypes and running pilots with other groups and organizations.

Our latest experiment is exploring a model for code review in science. As research becomes increasingly dependent on computation, we want to see whether the code review practices that are now commonplace in open source software development will also work in research.

Why code review?

The code used to produce the results in a paper is not usually reviewed when a paper is published, beyond a basic “sense-check” (or, as a colleague put it to me, “plausibility versus defensibility, not reproducibility”).  As code follows the trajectory of data in being integrated and recognized as a first-class research object, we want to work with the scientific community to figure out how scientists can and should check each other’s computational methods without increasing the time and cost of reviews unsustainably.

How this will work

The staff at PLOS Computational Biology helped us identify a set of already-published papers that include snippets of code in MATLAB, Python, R, and C++ that may be suitable for code review. We’ve put the code in front of a group of Mozilla engineers who regularly do code reviews for Firefox, mobile and for other programs.

These engineers aren’t scientists (though some may be interested in the underlying research).  Instead, they are skilled professionals who don’t have the domain expertise of the scientists, who in turn are not full-time software developers.

Over the course of the next month, each Mozilla engineer will review code from one or two papers. Their reviews, and their insights into the review process, will be captured in ReviewBoard, so that we can have a core collation point for the findings. Reviewers will have the ability to flip through the rest of the paper should they so choose, but this is really about the code snippet itself.

Once they’ve finished reviewing, we’ll analyze the findings, look for any patterns or interesting outliers, as well as examine how the actual commentary on the code itself was done, with the aim of using that as a model for code review in the future.

What we’re testing

This experiment is a means to explore the following:

– What does code review from a software engineer outside of academia look like? How do they approach the task?

– To what extent is domain knowledge needed to do a successful code review? Is the code parseable by someone outside of that discipline?

– What lessons can be learned about code review, possibly to influence and enhance traditional peer review?

– Does this process surface issues around best practice in writing software and code? If so, what are those issues?

– Following the review, how useful is the markup to the author? Does this feedback help them in their work? How can we change those norms?

We’ll be writing more about this experiment as the reviews continue, and hope to have our analysis completed next month. It’s just a first step, but we believe that this could help change the way researchers, administrators, publishers and funders think of code in scholarly research, and possibly unveil some deep-seated issues surrounding how code is recognized, maintained and shared in the scientific community.

 —

Have a question or would like to join in on the conversation? Drop us a line at [email protected] or add your comments here. We’d love to hear your thoughts. Also, have an idea you’d like explored. Let us know.

And special thanks to our Mozilla volunteers, Marian Petre from the Open University, Mike Hoye from Mozilla and Greg Wilson from Software Carpentry.

Intro to Open Science on P2PU: Watch, engage, remix.

The first of P2PU’s Open Science course videos is now online. (I’ll be joining to speak on Open Data in a few weeks.)

Have a watch, ping us at @MozillaScience with any questions (or leave them in the comments), and if you’re feeling *really* creative, try remixing / splicing the video using some of the Webmaker tools:
https://webmaker.org/en-US/tools/#popcorn-maker

And let us know what you come up with!

Session call open for this year’s #MozFest

The Mozilla Festival (or “MozFest” as we affectionately call it) is coming to London October 25-27, and planning is in full swing. This year we’ll have an entire track on “Science and the Web” and we want you to help us shape the program.

MozFest is where many of Mozilla’s best and most innovative ideas come from, bringing together over 1,000 inventors, hackers, creatives (and now those from the research community) to share ideas and explore how we can forge the future of the web together.

This  year’s focus is on the mentors, catalysts and change agents who care  about the using the web to make a difference in their respective fields  and communities. Individual sessions are organized under the following  themes:
  • Build Webmaker Together: The web is wild. Co-design its future with hackable projects, new memes and creativity remixed with digital tools.
  • Connect Your City: Champion digital making and bring together local communities.
  • Look Who’s Watching: Learn how to control who gets your data. Help others protect their privacy and develop long-term solutions to tracking.
  • Make the Web Physical: Take sensors, actuators and more to meld the virtual with the actual and make the web work your way.
  • Open Games: Ready, Set, Go! Join a community of game makers who use the web as a platform to build open games.
  • Science and the Web: Transform how we use the web–a scientist’s invention–to explore, experiment and build on each other’s research.
  • Skills and Badges:  Challenge  conventional skillsharing. Recognize and verify learning in  new ways to increase opportunities and make the most of the web.
  • Source Code for Journalism: Hack the news: learn, teach, and make journalism that’s native to the open web.
  • Teach the Web: Let’s teach the world the web. Discover how to inspire learners and spread digital literacy with hands-on making.
  • Webmaking for Mobile: Become a maker in the booming world of the mobile web.

For the “Science and the Web” track, we’re looking for sessions that show how the Web can (/and is) transforming science, and help introduce the MozFest crowd to some of the folks and organizations leading the charge. This isn’t your average conference, where you present to a group. Preference is given to hands-on and collaborative sessions (in terms of groups involved as well as the session content itself). Or, as we say “Less yack, more hack.” 🙂

A few areas to explore as a starter, in case you’re looking for ideas:

  • Hacking digital scholarship: How can we use the web to push the limits of how we share knowledge in the sciences?
  • Data sharing: Openness is key to advancing discovery. What role does data play and what cool things can we do with open data in research?
  • Citizen Science and engagement: From LHC@Home and Zooniverse to things like Microryza, how can we use the web to make research more accessible and use the wisdom of the crowd?
  • Tools for better web science: Come show off your open tool to help change the way we do science.
  • Badges / Altmetrics for research: How can we use new forms of assigning credit to research to facilitate sharing, collaboration and interoperability?
  • Code and data literacy: Hands-on training with tools or other practices to help us better enable a community of digital researchers.

This is just a start, and feel free to stray from this list. We’re currently pulling together the program, and the session call is open through the end of August for you to submit your ideas. Check out last year’s program, and get those proposals in! To submit, fill out this short form (and be sure to select “Science and the Web” track.)

We look forward to your submissions, and don’t forget to get your tickets!