
Conflict Resolution of Production-Marketing Collaborative Planning 
based on Multi-Agent Self-adaptation Negotiation 
Hao Li, Ting Pang, Yuying Wu and Guorui Jiang   
The Economics and Management school, Beijing University of Technology, 
No. 100, Pingleyuan, Chaoyang District, Beijing, China 
Keywords:  Production-Marketing Collaborative Conflict, Multi-Agent, Self-adaptation Negotiation, RBF Neural 
Network, Q-reinforcement Learning. 
Abstract:  In order to overcome the lack of adaptability and learning ability of traditional negotiation, we regard supply 
chain production-marketing collaborative planning negotiation as the research object, design one 
five-elements negotiation model, adopt a negotiation strategy based on Q-reinforcement learning, and 
optimize the negotiation strategy by the RBF neural network and predict the information of opponent for 
adjusting the concession extent. At last, we give a sample that verifies the negotiation strategy can enhance 
the ability of the negotiation Agents, reduce the negotiation times, and improve the efficiency of resolving 
the conflicts of production-marketing collaborative planning, comparing to the un-optimized 
Q-reinforcement learning.   
1 INTRODUCTION 
In order to meet the market requirements of the 
dynamic changes quickly and timely, the retailer and 
manufacturer on the supply chain establish contract 
early and draw up merchandise procurement plan. 
Now, the production-marketing planning changes 
from the simple trading to the consideration of 
general interests of supply chain. However, because 
the objectives of different enterprise are always 
various, their disagreements and conflicts appear 
usually. 
Distributed Agent technology have characters 
such as interaction, autonomy and learning (Wang, 
2013), it is not restricted by time and space, using 
the negotiation means based on multi-agent in the 
supply chain production-marketing collaborative 
planning, not only resolves conflicts but also solves 
the enterprise-decentralization problem. Many 
scholars have used the Agent technology in the 
negotiation of supply chain. For example,Kumar 
proposed a multi-agent system, selected the best 
supplier by automatic negotiation based on cost, 
distance and quality(Kumar, 2011); Sara used the 
multi-agent system to simulate multi-layer supply 
chain, controlled the inventory and cost by sharing 
information, predicting knowledge (Sara, 2012). In 
order to adapt to environment and the opponent’s 
dynamic information, the scholars in the intelligent 
negotiation field began to introduce self-learning 
mechanism into negotiation. For instance, Cheng 
learned the opponent’s utility function by using 
SVM (Cheng, 2009). Q-reinforcement learning is an 
algorithm which is independent of environment 
model, proposed by Watkins(Watkins, 1992), each 
action of Agent during negotiating has a return 
function value Q, then evaluates the present action of 
Agent and predicts the next action of Agent, 
accomplishes the proposed process by calculating 
Q(Sui, 2010). Many studies have introduced 
Q-reinforcement learning into collaborative 
negotiations (Shen, 2012; Ariel, 2013) to resolve 
conflicts effectively, optimize collaborative effect.   
The studies above all improve the running 
effectiveness of supply chain, but still there are some 
shortages. Now there are fewer studies on supply 
chain negotiation adopting adaptive algorithms; 
self-learning ability and adaptability of negotiation 
Agents are relative poor; most of negotiation 
strategies based on Q-reinforcement learning are not 
self-adaptive, there are lesser studies on adjusting Q 
value depending on opponents’ behavior, and the 
convergence speed is still slow. To solve the 
problems, we propose a multi-agent adaptive 
negotiation method for the problem of supply chain 
production-marketing collaborative planning conflict. 
209
Li H., Pang T., Wu Y. and Jiang G..
Conflict Resolution of Production-marketing Collaborative Planning based on Multi-Agent Self-adaptation Negotiation.
DOI: 10.5220/0004830602090214
In Proceedings of the 6th International Conference on Agents and Artificial Intelligence (ICAART-2014), pages 209-214
ISBN: 978-989-758-016-1
Copyright
c
 2014 SCITEPRESS (Science and Technology Publications, Lda.)