I mean, if it produces the correct result, it can't literally be nonsense. In the case of ResNet, the output is computed entirely based on the 7x7x512 representation of the final hidden layer (though of course that layer is computed based on the previous stuff), so all information that went into the networks decision is present there. All 512 filers in the hidden layer mean something.
You've acknowledged that it's not coincidentally given that it was nudged by gradient descent. I don't think I get in what sense you still think it's kind of nonsense.
Anyway, the claim that it's not at all
human-understandable is empirically untrue. ( Although I'd like to point out that, if it were true, this would be even more reason to expect doom from AGI.) My paper is based on the observation that the 512 filters at the end of ResNet do, in fact, activate for human-understandable concepts. The algorithmic contribution was to find clever ways to connect human-made annotations to approximate what the neuron is doing.
Here's an example:
This is from one of the 512 filters and one image, where I've thresholded the numbers, i.e., all of the 7x7 cells where the number is above a certain value are highlighted.
Now, this example is cherry-picked and most of them aren't as crisp. But nonetheless, this neuron is clearly reacting to the tree house. This is a human-understandable concept.
There is also work that goes into the loss function and edits it to encourage the filters to be more human-understandable.