I will try to have some more comments about the physics when I have time and energy. In the meanwhile:
Entropy in thermodynamics is not actually a hard concept. It's the ratio of the size of a heat flow to the temperature at which that flow is happening. (So, joules per kelvin, if you're using SI units.) See episodes 46 and 47 of The Mechanical Universe for the old-school PBS treatment of the story. The last time I taught thermodynamics for undergraduates, we used Finn's Thermal Physics, for the sophisticated reason that the previous professor used Finn's Thermal Physics.
Entropy in information theory is also not actually that hard of a concept. It's a numerical measure of how spread-out a probability distribution is.
It's relating the two meanings that is tricky and subtle. The big picture is something like this: A microstate is a complete specification of the positions and momenta of all the pieces of a system. We can consider a probability distribution over all the possible microstates, and then do information theory to that. This bridges the two definitions, if we are very careful about it. One thing that trips people up (particularly if they got poisoned by pop-science oversimplifications about "disorder" first) is forgetting the momentum part. We have to consider probabilities, not just for where the pieces are, but also for how they are moving. I suspect that this is among Vopson's many problems. Either he doesn't get it, or he's not capable of writing clearly enough to explain it.
So these two were published in American Institute of Physics Advances, which looks like a serious journal about physics. Does anyone know about it? It occupies a space where I can’t easily find any obvious issues, but I also can’t find anyone saying “ye this is legit”. It claims to be peer-reviewed, and at least isn’t just a place where you dump a PDF and get a DOI in return.
I have never heard of anything important being published there. I think it's the kind of journal where one submits a paper after it has been rejected by one's first and second (and possibly third) choices.
However, after skimming, I can at least say that it doesn’t seem outlandish?
Oh, it's worse than "outlandish". It's nonsensical. He's basically operating at a level of "there's an E in this formula and an E in this other formula, so I will set them equal and declare it revolutionary new physics".
Here's a passage from the second paragraph of the 2023 paper:
The physical entropy of a given system is a measure of all its possible physical microstates compatible with the macrostate, S~Phys~. This is a characteristic of the non-information bearing microstates within the system. Assuming the same system, and assuming that one is able to create N information states within the same physical system (for example, by writing digital bits in it), the effect of creating a number of N information states is to form N additional information microstates superimposed onto the existing physical microstates. These additional microstates are information bearing states, and the additional entropy associated with them is called the entropy of information, S~Info~. We can now define the total entropy of the system as the sum of the initial physical entropy and the newly created entropy of information, S~tot~ = S~Phys~ + S~Info~, showing that the information creation increases the entropy of a given system.
wat
Storing a message in a system doesn't make new microstates. How could it? You're just rearranging the pieces to spell out a message — selecting those microstates that are consistent with that message. Choosing from a list of available options doesn't magically add new options to the list.
"So, professor sir, are you OK with psychologically torturing Black people, or do you just not care?"