The appeal of "Where's Walter?" or in the original "Where's Waldo?" actually consists of having to search for what feels like hours before you finally discover Walter. A robot has now brutally destroyed this search fun. The artificial intelligence fed with pictures of Waldo scans the image motif and as soon as it is sure, the latex hand moves to the appropriate place. Cheers to the technical future!
The robot arm is controlled by a Raspberry Pi using the PYARM Python library for the UARM Metal. Once initialized the arm is instructed to extend and take a photo of the canvas below. It then uses OpenCV to find and extract faces from the photo. The faces are sent to the Google Auto ML Vision service which compares each one against the trained Waldo model. If a confident match of 95% (0.95) or higher is found the robot arm is instructed to extend to the coordinates of the matching face and point at it. If there are multiple Waldos in a photo it will point to each one. While only a prototype, the fastest There's Waldo has pointed out a match has been 4.45 seconds which is better than most 5 year olds.