In the article, you provide a good overview of the modules themselves and their general programming. But it's completely unclear how the modules communicate with each other and how to program this communication. For example, there's hardware communication between the ASRPRO and the LCD-ESP32, but how do you make the eyes change their expression immediately after recognizing a phrase? How do the eyes from the MCU-ESP32 even change? What source code is currently being loaded into the device: the regular Movecall Moji or a modified one?