An agent has an acquired knowledge associated with an expert language; this embeds silently what it assumes to be trivial or compressible up to not mentioning it in details.
Another agent is unfamiliar with the previous implicit communication structure and won’t therefore understand the first agent statements.
To fix this, the information needs to be described to the second agent. As the first agent knowledge is more or less distant from the second agent knowledge, some aspects need to be described in a more verbose way, through lower-level language expressions, and implicit knowledge have to be added and conveyed.
This is a tedious task, for the first agent, to:
– find the level of comprehension of the second agent
– review each high-level points that need to be mentioned
– detail those points to the second agent level of comprehension
– get to understand and adjust the non-expert questions
– figure out the suitable answers
– detail those answers to the second agent level of comprehension
– alter every step that needs to be updated
This could be handled by an interface; each agent requires a way that suits them to communicate. It needs to acknowledge for its agent knowledge, and to acquire the features required to communicate to other languages and levels of expertise.
I would even consider that, without the ability to translate what needs to be communicated between level of comprehensions, languages and representations; us, humans, wouldn’t be able to establish a communication more complex than any other creature.