[D] Is RL or Non-RL computer vision more computationally expensive?
At first glance, RL appears more expensive and require orders of magnitude more training steps, however computer vision for object localization and classification may require larger and more fine-tuned architectures given the problems are a bit more mature.
Of the two, which would you say is more computationally demanding? Or is it a tie? Or is it the ever common “it depends” answer?