SXSWi: New Webtech and Science Publishing: (Re)Constructing the Scientific Article
Many SXSW vets recommend attending a panel outside of your comfort zone. This was that one. I also was covering it for one of our projects at work. Here’s the abstract:
New publishing technologies challenge the traditional structure of peer-reviewed scientific journals. For hundreds of years the “article” has been the primary vehicle for conveying scientific information – but semantic markup, tagging, and wiki are reconstructing scientific publications into a flexible and evolving concept. This panel will look at the social and legal implications of “Web 2.0″ and “Semantic Web” as they impact science and scientific knowledge.
Some of this was a bit over my head, but I’m just going to toss up my notes and then follow up with my impressions.
Panelists:
- Moderator: John Wilbanks, Creative Commons
- Matt Cockerill, BioMed Central
- Melissa Hagemann, Open Society Institute
- Timo Hannay, Nature Publishing Group
- Amit Kapoor, Topaz/PLoS One
Notes:
- We used to communicate scientific information by sending it to someone and having them publish it
- Definitions:
- Timo: Web 2.0:
- Multi-faceted concept
- Architecture by participation (network effects, harnessing collective intelligence)
- The Long Tail
- Open APIs, Mashups (Open Data and Open Access)
- Interface Issues (AJAX, web apps)
- Development process (perpetual beta)
- Web as a platform
- Matt: The Semantic Web
- Web pages no longer electronic paper
- Robots/machines reading them along with humans
- Defining standards for allowing computers to read information
- Flood of data, more info that humans can possibly read
- Need semantic web tools to sift through it
- Help scientists publish work in a way that captures semantic structure
- Melissa: Open Access
- Free online availability of scholarly content
- Defined 5 years ago – Budapest open access initiative
- Open Access Journals are freely available online and don’t rely on subscriptions for revenue
- Apply combinations of new business models
- Article processing fee
- Fee paid by researcher or research grant
- Article processing fee
- Many researcher agencies are mandating Open Access for their researchers
- Timo: Web 2.0:
- Measure impact of article by citations in other articles (much different than hyperlinking)
- Citation specifics have been a monopoly, still publish the definitive numbers
- Starting to become available from other sources (Google Scholar)
- Scientists don’t blog or comment very much
- Those things don’t get credit for them
- Very competitive
- Connotea: Del.icio.us for scientists
- Standards are being developed for web usage of scientific sites
- Citation-based metrics—takes too long to get the data
- Ancient history by the time everything is published
- Citation specifics have been a monopoly, still publish the definitive numbers
- Amount of info that has to be dealt with is absolutely amazing
- Lack of standards, lack of formats
- Semantic Web played huge role
- It is an evolving concept
- Start with a concept, modify it over time
- Text annotations
- Attach a note to an article that says specific research is wrong
- You can later search for all opposing theories
- Trying to put out a release every two weeks
- At about version number 0.5 now
- All journals now (whether they like it or not) are databases
- What is missing is the standards for the different types of data to be interwoven
- No way BioMed science can move as quickly as it could if data is not open
- OpenAccess makes this possible
- Facilitates exchange of information
- Nature relies on reader subscriptions
- Nature rejects 92% of papers submitted
- NatureJobs : Scientific career service
- Web 2.0 science tools don’t generate any revenue yet (like Connotea)
- History of policy pressures from government
- OpenAccess – government funders didn’t want to pay for information twice
- One funder provided a grant of 250,000 pounds to pay for research then he himself couldn’t access it
- Wondered how much impact the research will have if nobody can access it
- Semantic Web
- Everything needs a permanent URI
- What is the vocabulary of the community?
- Bottom up AND top down approach to establish vocabularies
- Wikipedia approaches have great value because otherwise nothing will ever happen
- Bookmarking can become much more than keeping a list of favorite articles
- As someone bookmarks and tags articles of interest, they can assemble a collection of scientific knowledge that is intresting and trustworthy
- Can be used to recommend other research
- As someone bookmarks and tags articles of interest, they can assemble a collection of scientific knowledge that is intresting and trustworthy
- RDF and ontologies have their place, Tagging and Microformats have their place
- Tagging gives noisier data
- Microformats are not the same as a formal ontology, but can be very useful
- Defining things like what a gene is should be left to scientists
- Things like defining an author can be done by the community
- Question: Total divergence of solving the same problem (Tagging vs. Semantic Web)
- Tagging: Assign keywords, let the Googles of the world figure it out
- Semantic Web: Each document filled with tons of data
- Amit: Tagging is a way of classifying objects
- Tagging is a small subset of what the Semantic Web is all about
- Tagging is the first step
- The cost of putting in semantic web info is very high
- Who is going to bear that cost? The author? The distributor?
- Great second question from Amy, but a bit over my head. Must get from podcast when it comes out. The gist was how about how user-generated content can feed The Semantic Web.
- Very effective ontologies have been extracted from user generated Wikipedia content
I work with many scientists that publish papers, but a lot of this is still so very foreign to me. I personally have a lot of faith in “Web 2.0″ tools such as tags and Microformats. I struggle a bit with some of this Semantic Web stuff.
I guess if I was trying to record my life by using Semantic Web technologies, I’d have to somehow go back and record everything I’ve ever said, where I said it, who I said it to, what they said back, what other thoughts I had without speaking them aloud, etc. If I use “Web 2.0″ technologies to record my life, I’d have my blog posts, Flickr photos, Del.icio.us links, YouTube videos, Twitter tweets, GMail, and any other social tools—plus the tags, timestamps, and (perhaps) GeoCodes that go with each. I could go a step further and do a desktop crawl for files, local photos (I haven’t thrown all 2500+ photos of Ella into Flickr), perhaps POP email, timestamps, folder names, etc.
Would this second approach be perfect? No. But could it be done? Yes. To me, the Semantic Web just feels too much like a dream that is unachievable. The panelists were definitely giving me that vibe, too. There is already so much scientific information out there. You can’t go back and put all this meta info on it. And for all the information that’s coming out, it’s hard to force people to follow a strict process of extracting certain data for the Semantic Web purposes. You can, however, track how that information floats around the web, how it is tagged, sorted, commented, annotated, etc.
One thing that didn’t sit well with me is the fact that scientists won’t blog because there’s no “credit” for that. They need to release their info in an environment that can be tallied up to rank them against their “competition.” Well, does writing a book directly help a blogger’s Technorati ranking? No, not directly. But with that book out there, more discussion about that person will ensue. Even though links are not being made to the book, the book is causing more discussion. I would like to think that a scientist that shares by blogging and commenting would gain notoriety in other ways. Perhaps not, but I’d like to think that.
But what do I know? I’m just a web hippie. I know many Semantic Web folks turn up their noses at things like Microformats. But to me, that feels a bit too much like someone that complains about small size of the in-ground pools that one person is digging as he is trying to dig an ocean. Only problem is, he’s been digging for ten years already and there’s no swimming in sight.
thanks for the liveblogging. we’ll post trackbacks to this on the CC and SC blogs when the podcast link comes out, and i’ll make a few comments back to your response to the panel once i’ve had some time to digest and think clearly.
Thanks, John. I look forward to hearing your comments for someone who works with cognitive scientists, but doesn’t come from that background. It was an intriguing talk as I’ve often heard about this struggle with getting Semantic Web initiatives off the ground.
[...] Just read a blog post with notes from the panel, Web 2.0 and Semantic Web: The Impact on Scientific Publishing, that took place at the SXSW meeting. One of the note headings stated, scientists don’t blog or comment very much, and a discussion had taken place that this might be due to that there is nothing in it for scientists. We simply don’t get any credit for blogging. [...]
[...] like I missed a good panel at SXSW 2007 on slightly on this topic, discussed in this blog entry (SXSWi: New Webtech and Science Publishing: (Re)Constructing the Scientific Article), and their is a plethora of data on the microformats wiki. Just thought I’d send out a [...]
Melissa is a n ass. How can anyone take here seriously. She’s a nasty bitch who tries to rid herself of perople who are threats. Melissa is an information guru. Shows how stupid you are!
I just had to let this comment through because I have never been called an idiot on my blog before.
I just took notes, dude. Never called anyone a guru. I’m guessing you don’t like Melissa though!