
Multi-Agent Interaction Framework. arXiv e-prints,
page arXiv:2308.12503.
Kottur, S., Moura, J. M., Lee, S., and Batra, D. (2017).
Natural language does not emerge’naturally’in multi-
agent dialog. arXiv preprint arXiv:1706.08502.
Lazaridou, A. and Baroni, M. (2020). Emergent multi-
agent communication in the deep learning era. arXiv
preprint arXiv:2006.02419.
Levenshtein, V. I. et al. (1966). Binary codes capable of cor-
recting deletions, insertions, and reversals. In Soviet
physics doklady, volume 10, pages 707–710. Soviet
Union.
Li, F. and Bowling, M. (2019). Ease-of-teaching and lan-
guage structure from emergent communication.
Li, G., Hammoud, H. A. A. K., Itani, H., Khizbullin, D., and
Ghanem, B. (2023). Camel: Communicative agents
for ”mind” exploration of large scale language model
society.
Park, J. S., O’Brien, J. C., Cai, C. J., Morris, M. R., Liang,
P., and Bernstein, M. S. (2023). Generative agents:
Interactive simulacra of human behavior.
Ren, Y., Guo, S., Labeau, M., Cohen, S. B., and Kirby, S.
(2020). Compositional languages emerge in a neural
iterated learning model.
Ri, R., Ueda, R., and Naradowsky, J. (2023). Emergent
communication with attention.
Rita, M., Strub, F., Grill, J.-B., Pietquin, O., and Dupoux,
E. (2022). On the role of population heterogeneity in
emergent communication.
Słowik, A., Gupta, A., Hamilton, W. L., Jamnik, M.,
Holden, S. B., and Pal, C. (2020). Structural Inductive
Biases in Emergent Communication. arXiv e-prints,
page arXiv:2002.01335.
Tishby, N., Pereira, F. C., and Bialek, W. (2000). The infor-
mation bottleneck method.
Turpin, M., Michael, J., Perez, E., and Bowman, S. R.
(2023). Language models don’t always say what they
think: Unfaithful explanations in chain-of-thought
prompting. arXiv preprint arXiv:2305.04388.
Wang, Z., Simoncelli, E. P., and Bovik, A. C. (2003). Mul-
tiscale structural similarity for image quality assess-
ment. In The Thrity-Seventh Asilomar Conference on
Signals, Systems & Computers, 2003, volume 2, pages
1398–1402. Ieee.
Wu, Q., Bansal, G., Zhang, J., Wu, Y., Li, B., Zhu, E., Jiang,
L., Zhang, X., Zhang, S., Liu, J., Awadallah, A. H.,
White, R. W., Burger, D., and Wang, C. (2023). Au-
togen: Enabling next-gen llm applications via multi-
agent conversation.
Wu, Q., Bansal, G., Zhang, J., Wu, Y., Zhang, S., Zhu, E.,
Li, B., Jiang, L., Zhang, X., and Wang, C. (2023).
AutoGen: Enabling Next-Gen LLM Applications via
Multi-Agent Conversation Framework. arXiv e-prints,
page arXiv:2308.08155.
APPENDIX
Fully Connected Network-Based
Architecture
We created a simple fully connected network-based
auto-encoder architecture for the agents. As shown
in Table 5, the encoder agents receive an image with
a fixed image height and width. The encoder then
transforms the features into a series of fully connected
layers with a fixed hidden dimension size specified
in hidden dim. Finally, the encoder outputs the se-
quence of symbols via the Gumbel SoftMax layer.
As shown in Table 6, the decoder receives a se-
quence of symbols and outputs the image after pass-
ing through a series of fully connected layers with a
fixed hidden dimension of size hidden dim.
Convolution-Based Architecture
We created a convolution-based auto-encoder archi-
tecture for the agents. As shown in Table 7, the en-
coder agents receive an image with a fixed image
height and width. The first two convolutional layers
extract the features in the image using 2D convolu-
tion with a kernel size of 3. Next, we apply the Max-
Pool2D layer with a kernel size of 2. During train-
ing, the dropout function on the output feature with a
probability of 0.25. We flatten the features before ap-
plying the fully connected layer. Next, we apply two
fully connected layers with an optional dropout layer.
Finally, the encoder outputs the sequence of symbols
via the Gumbel SoftMax layer.
As shown in Table 8, The decoder receives a se-
quence of symbols and outputs the image after pass-
ing through a series of fully connected layers with a
fixed hidden dimension of size hidden dim and 9216.
A Framework for Studying Communication Pathways in Machine Learning-Based Agent-to-Agent Communication
125