3 Aprilie 2015 Lasă un comentariu
Yet just last week, a scientist at the Université Joseph Fourier did almost the exact same test—not for the business of animal parks, but for the business of scientific publishing. In collaboration with the Springer publishing house, he released SciDetect, a computer program that spots a particular category of phony papers that has been plaguing computer science publications. (Just last year Springer had to pull 18 fake papers, while another publisher, IEEE, had to pull more than 100.) “SciDetect indicates whether an entire document or its parts are genuine or not,” the press release brags. “The software reports suspicious activity by relying on sensitivity thresholds that can be easily adjusted.” Great news, right? Not so much. For the fake papers it aims to detect—which are generated by an algorithm known as SCIgen—are the shaved mastiffs of the computer science world. They are so obviously fake that nobody who has any business being within 10 feet of a computer science journal would fail to spot them with even the most careless examination. A technological solution is completely unnecessary. So why did Springer spend so much time and effort (and presumably money) to build SciDetect?