Signing this petition is something that you can do now to help make taxpayer-funded research accessible to taxpayers.
Tag Archives: academia
BibTeX bleg
I have written a little standalone script in python that parses a LaTeX file with \cite{} commands and the relevant BibTeX file and produces:
- formatted HTML suitable for dropping into your homepage
- individual .bib files for each paper
- linking to archival versions via a DOI reference or URL
- linking to a local pdf via the
local-urlfield
The point was to make updating the publications on your homepage just epsilon more difficult that updating a BibTeX file/your CV. Of course, this is moot for people who have other folks to update their pages, but for us plebes, it could save a little hassle.
Clearly you could customize the output format to your needs. However, at the moment it’s not very robust (or efficient, or pretty). I’d like to test it out on likely readers of this blog’s personal .bib files to make it useful before sticking it on github. A subset of readers of this blog are likely to be people who might use such a thing, I’d like to know what your .bib files look like. Because BibTeX has a fair bit of variability, I am pretty sure that I did not catch most of the corner cases in my regexps.
So if you are interested, please do send me a representative file to my TTIC address. Thanks a ton!
Fair use for excerpts
Via Inside Higher Ed I learned about a case in which university presses brought a suit against Georgia State over fair use for excerpts in course readers and online course materials:
Her challenge, she writes, is to determine what size excerpts are “small enough” to justify fair use. Here, after reviewing a range of decisions, Evans settles on 10 percent of a book (or one chapter of a book) as an appropriate measure, allowing professors enough substance to offer students, while not effectively making a large portion of the book available.
I guess this is how sausage is made — 10 percent seems like a nice round number, let’s go with that one. By the way, that’s 10 percent including front and back matter, not 10 percent of the text.
It’s a 300+ page decision, but there has been some analysis already.
The value-add from Elsevier?
I got an email today from Elsevier:
It is our pleasure to inform you that your publication has been cited in a journal published by Elsevier.
Through this unique service we hope we can offer you valuable information, and make you aware of publications in your research area.
The service is called CiteAlert. It sends you an email every time you’re cited!
Clearly, it’s little touches like this that justify the price gouging and subscription bundling. Kind of like the little chocolate on your pillow at the expensive hotel.
Juking the stats in academic publishing
I heard recently of a case where someone got a paper back with revisions requested, and a deadline for said revisions. They ended up asking for a week extension, but then the journal said they would have to do a fresh submission and redo the whole review cycle. I found this baffling — but then that person pointed out that the journal has built a reputation on fast turnaround times, and so to keep their “sub-to-pub” numbers low, they don’t want to give any extensions to the authors. It’s better to do a resubmission than to continue with the same “paper ID” in the system.
This is a classic example of juking the stats:
I just got a rejection from KDD 2012 which smacks of the same ominous reasoning:
We try to notify authors once a decision on a submission is concretely made, and hope that the early notifications can reduce the average review turn-over time.
But the real kicker is that “due to technical constraints” they can’t give us the reviews until May 4th. So I’m not really sure what I am supposed to do with this information — I can’t really start on revisions without the reviews, so this “early notification” thing is really just to make them feel better about themselves, it seems. Or perhaps they can then report that the reviewing was “more efficient.”
In any case, no harm is done, per se. But optimizing metrics like “sub-to-pub” seems to be as misguided as teaching to the test. What do we really want out of our peer review process? Or should we abandon it?
Manuscript Central is annoying
The IEEE Transactions on Information Theory recently transitioned to using Manuscript Central from the old Pareja system, so now all of the IEEE journals for which I review seem to be managed by the same external management system. As a reviewer/author, I have a lot of complaints (small and large) about Manuscript Central:
- Why oh why do I need to disable my popup blocker for your site to work?
- Why can login information not be shared across different IEEE publications? I have a separate account for each journal, with a separate password. Thank goodness I have LastPass, but even that program gets confused sometimes.
- What is the deal with the mandatory subject classifications for papers? One of the “topics” I could pick was “IEEE Transactions on Information Theory.” Really? That’s a topic?
- Why must papers for review be emblazoned with that stupid pale blue “For Peer Review Only” running diagonally across each page? This causes PDF annotations such as highlighting to barf, making paperless reviewing of papers significantly more annoying than it needs to be.
The worst part is that I am sure IEEE could implement a significantly cheaper and just-as-effective system itself, but now each Society is forking over money to Manuscript Central, which as far as I can tell, offers significantly more annoyances for authors and reviewers and is a shoddy product. Perhaps as an editor it’s significantly better (I imagine it is), but it seems like a bad deal overall.
Of course, now I sound curmudgeonly. Get off my lawn!
Do other people like MC? Or do you have other pet peeves?
ICITS Workshop Deadline Extension
(Via Adam Smith) The deadline for submitting workshop papers to the 6th International Conference on Information Theoretic Security (ICITS) has been extended from today to Monday, April 23 (at 3pm Eastern) due to holidays. It’s in Montreal in August, so if you have some recent results that you would like to present in a workshop setting, please send them in. “Papers which broaden the applicability of information-theoretic techniques (say, to areas such as data privacy or anonymity) would also be very welcome.”
ICITS will have two tracks this year : a more CS-style conference with published proceedings, and a workshop (think ITA) track without proceedings where you can present older stuff.
Linkage
The variation of the human body across sports is fascinating (via Matt Tong).
The films of Andrei Tarkovsky (1932-1986) are available for free online (via Zhenya Tumanova).
A paper arguing that systems-CS conference reviews are bad. (via Manu Sridharan)
A watch that’s on Indian Time. Works for other cultures too! (via Harbeer)
Typical review loads
Since becoming faculty at TTI, I’ve started to appreciate better the tensions of service commitments and I can see how many people begin to view reviewing as a chore, a burden they must bear to maintain goodwill in the “community.” Since I work in a few different communities now, I end up reviewing papers from a lot of different areas : information theory and signal processing of course, but also machine learning, security, and networks. There’s been a distinct uptick in my reviewing queue, which I find somewhat alarming.
Looking back, I did a quick calculation and in the almost 6 months I’ve been here, I’ve either finished or committed to reviewing 9 journal papers and 16 conference papers. These numbers don’t really mean too much, because some journal papers are shorter (e.g. a correspondence) and some conference papers are long (40+ pages including supplementary material). Page numbers also don’t really help because of formatting differences. I’m hoping my new iPad (ooh, shiny!) will let me pack in some reviewing time during my commute and stop me from killing so many trees.
However, I have no idea if these numbers are typical. I’ve turned down review requests because I felt like I don’t have enough time as it is. So readers : what’s a typical review load like? Should I just suck it up and accept more reviews?
Note that I’m not asking about what’s “fair” in terms of I submit N papers and therefore should review 3N or something like that. Those games are fine and all, but I really wonder what the distribution of review load is across individuals for a given journal. More on that point later…
Update: I should be clear that being on a PC will clearly cause your review load to go up. I am on 2 PCs but for smaller conferences; having 10+ ISIT reviews would add significantly to one’s total load.
Updated perl script for merging TeX files for ArXiV
Manu Sridharan (blog) left a comment the other day on my old post on my script to merge multiple TeX files (and strip the comments) for posting to ArXiV. He’s created a git repository for it, which seem so much more official and stuff. It’s at:
https://gist.github.com/2175026
Thanks a bunch, Manu!
As a side note, Péter Gács has a de-macro script to eliminate all of your private macros if you’re so inclined.