Asking a computer to ‘tell me about this picture’ poses other problems, though. We do not have HAL 9000, nor any path to it, and we cannot recognise any arbitrary object, but we can make a guess, of varying quality, in quite a lot of categories. So how should the user know what would work, and how does the system know what kind of guess to make? Should this all happen in one app with a general promise, or many apps with specific promises? Should you have a poster mode, a ‘solve this equation’ mode, a date mode, a books mode and a product search mode? Or should you just have mode for ‘wave the phone’s camera at things and something good will probably happen’?
This last is the approach Google is taking with ‘Lens’, which is integrated into the Android camera app next to ‘Portrait’ – point it at things and magic happens. Mostly.
Source: Cameras that understand: portrait mode and Google Lens — Benedict Evans