Description
a sequence of gray-coded patterns are projected onto a flat surface. Then, the photodiodes on the robot independently decodes into a location within the projected area.
If I am right, the projector just sends out a gray-coded pattern message (a map), it doesn't know the coordinates of the robots and it doesn't communicate with other parts . And the mini bots will use the photodiodes to get the pattern and decode the pattern to know where it is in the projected area (map). When it is received the command from the receiver, it will compare with the goal and its current coordinate decoded from the projector to adjust its tracks.
How the gray-coded pattern is generated and how mini bots can use photodiode to sense its location, is there any speciality about the photodiode on the bots?