Review mills

It is hard to believe – but after research paper mills there are now also review mills

What I eventually found was a Review Mill, a set of 85 very similar review reports in 23 journals published by MDPI (Agronomy, Antibiotics, Applied Sciences, Atoms, Biomimetics, Biomolecules, Cancers, Catalysts, Chemistry, Coatings, Electronics, International Journal of Molecular Sciences, Journal of Clinical Medicine, Journal of Personalized Medicine, Materials, Metals, Molecules, Nutrients, Pathogens, Polymers, Prothesis, Sensors and Water) from August 2022 to October 2023, most of the time with coercive citation, that is, asking authors to “cite recently published articles” which were always co-authored by one or more reviewers of the Review Mill.

 

CC-BY-NC Science Surf accessed 16.01.2026

Parallelized computer code and DNA transcription

At stackexchange there is a super interesting discussion on parallelized computer code and DNA transcription (which is different to the DNA-based molecular programming literature…)

IF : Transcriptional activator; when present a gene will be transcribed. In general there is no termination of events unless the signal is gone; the program ends only with the death of the cell. So the IF statement is always a part of a loop.

WHILE : Transcriptional repressor; gene will be transcribed until repressor is not present.

FUNCTION: There are no equivalents of function calls. All events happen is the same space and there is always a likelihood of interference. One can argue that organelles can act as a compartment that may have a function like properties but they are highly complex and are not just some kind of input-output devices.

GOTO is always dependent on a condition. This can happen in case of certain network connections such as feedforward loops and branched pathways. For example if there is a signalling pathway like this: A → B → C and there is another connection D → C then if somehow D is activated it will directly affect C, making A and B dispensable.

Of course these are completely different concepts. I fully agree with the further stackexchange discussion that

it is the underlying logic that is important and not the statement construct itself and these examples should not be taken as absolute analogies. It is also to be noted that DNA is just a set of instructions and not really a fully functional entity … However, even being just a code it is comparable to a HLL [high level language] code that has to be compiled to execute its functions. See this post too.

Please forget everything you read from Francis Collins about this.

 

CC-BY-NC Science Surf accessed 16.01.2026

When AI results cannot be generalized

There is a new Science paper that shows

A central promise of artificial intelligence (AI) in healthcare is that large datasets can be mined to predict and identify the best course of care for future patients.  … Chekroud et al. showed that machine learning models routinely achieve perfect performance in one dataset even when that dataset is a large international multisite clinical trial … However, when that exact model was tested in truly independent clinical trials, performance fell to chance levels.

This study predicted antipsychotic medication effects for schizophrenia – admittedly not a trivial task due to high individual variability (as there are no extensive pharmacogenetics studies behind). But why did it completely fail? The authors highlight two major points in the introduction and detail three in the discussion

  • models may overfit the data by fitting the random noise of one particular dataset rather than a true signal
  • poor model transportability is expected due to patients, providers, or implementation characteristics that vary across trials
  • in particular patient groups that are too different across trials while this heterogeneity is not covered in the model
  • missing outcomes and covariates like psychosocial information and social determinants of health were not recorded in all studies
  • patient outcomes may be too context-dependent where trials may have subtly important differences in recruiting procedures, inclusion criteria and/or treatment protocols

So are we left now without any clue?

I remember another example of Gigerenzer in  “Click” showing misclassification of chest X rays due to different devices (mobile or stationary) which associates with more or less serious cases (page 128 refers to Zech et al.).  So we need to know the relevant co-factors first.

There is even a first understanding of the black box data shuffling in the neuronal net.  Using LRP  (Layer-wise Relevance Propagation) the recognition by weighting the characteristics of the input data can already be visualized as a heatmap.

 

CC-BY-NC Science Surf accessed 16.01.2026

Data voids and search engines

An interesting Nature editorial reporting a recent study

A study in Nature last month highlights a previously underappreciated aspect of this phenomenon: the existence of data voids, information spaces that lack evidence, into which people searching to check the accuracy of controversial topics can easily fall…
Clearly, copying terms from inaccurate news stories into a search engine reinforces misinformation, making it a poor method for verifying accuracy…
Google does not manually remove content, or de-rank a search result; nor does it moderate or edit content, in the way that social-media sites and publishers do.

So what could be done?

There’s also a body of literature on improving media literacy — including suggestions on more, or better education on discriminating between different sources in search results.

Sure increasing media literacy at the consumer site would be helpful. But letting Google earn all that money without any further curation efforts? The original study found

Here, across five experiments, we present consistent evidence that online search to evaluate the truthfulness of false news articles actually increases the probability of believing them.

So why not putting out red flags? Or de-rank search results?

fake screen shot

 

 

CC-BY-NC Science Surf accessed 16.01.2026

Das Ende der Bachelorarbeit

ist wohl schon eingeleitet zumindest bei der Betriebswirtschaft in Prag, Zitat

Texte, die mit Künstlicher Intelligenz verfasst wurden sind kaum von menschlichen zu unterscheiden. Eine Prüfung sei für Unis deshalb nur sehr schwer möglich, sagt Dekan Hnilica.  “Wir haben andere Teile unseres Studiums, in denen die Studierenden ihre Lernergebnisse oder erwarteten Lernergebnisse nachweisen können. Daher ist die Bachelorarbeit überflüssig.”

 

CC-BY-NC Science Surf accessed 16.01.2026

Scientific integrity is not a weapon

Bill Ackman is threadening Harvard faculty on X

I expect that in the not too distant future AI will target every paper and not only a suspicious table or an image found by chance. Nevertheless using this now as a weapon is immoral and at high risk of false accusations. And , it may even be prosecuted as criminal defamation.

 

Feb 11, 2025

Unfortunately scientific integrity is being used again as personal weapon.  Stefan Weber is making a business from right wing clients to verify doctoral theses. Without doubt, he has excellent technical skills (or at least a Turnitin account) but also completely lost all sense of proportion and direction. See

Back to SPIEGEL yesterday, translated

In some cases, his accusations turned out to be unfounded or less serious than he portrayed them. That’s why he is viewed more critically in Austria. … Until the Föderl-Schmid case, none of this had harmed him much. But for those he accused, it was a different story. Even if the allegations came to nothing, their reputation was tarnished

 

CC-BY-NC Science Surf accessed 16.01.2026

PubPeer – an introduction for Science Journalists

5 tips for using PubPeer to investigate scientific research errors and misconduct

  • Install the PubPeer browser extension
  • Don’t publish a news story simply to point out people have made negative comments
  • Verify all claims made on PubPeer
  • Use caution when describing the likelihood that researcher misconduct happened
  • No matter how small a role a researcher plays in your story, check PubPeer before including them

As an introductory text I can recommend a 5 year old blog entry introducing PubPeer

Battle lines are being drawn on the internet, between the scientific establishment and volunteer vigilantes trying to impose their own vision of the scientific process through “post-publication peer review”.

On one side is the cream of the scientific aristocracy: a professor with a meteoric career trajectory at Imperial College London, one of the best universities in the world, and the top academic publisher, Nature Publishing Group. On the other side: a few anonymous malcontents carping on an obscure web site called PubPeer (welcome to our site!).

 

CC-BY-NC Science Surf accessed 16.01.2026