In PowerTAC, all the information is provided to
the broker agent by asynchronous messages. At the
beginning of a game, after the brokers sign in but be-
fore the brokers start playing, each broker receives
the game parameters, the broker identities, the cus-
tomer records, and the default tariffs. During two fic-
tional weeks, no competitor can buy or sell energy in
the market. The information about the bootstrap cus-
tomer data, the bootstrap market data, the bootstrap
weather data, the weather report, and the weather
forecast of these two weeks is also received before the
game starts. The bootstrap market data does not cor-
rectly reflect the future clearing prices because only
a single standard seller exists. Once per time-slot,
the predictor will receive the public information about
the 24 clearing prices of the wholesale market, the
weather report, and the weather forecast. When no
trades occur, neither the message of the wholesale
market clearing data or the wholesale market order
books is provided to the broker agent.
In section 2 a review of different PowerTAC
agents’ approaches is presented, followed by section
3, where the methodology explained. Section 4 shows
the obtained results, along with an analysis of the re-
sults. Finally, section 5 presents the conclusions that
can be drawn from this work, the advantages and lim-
itations of the proposed solution, and aspects for im-
provement in future developments.
2 RELATED WORK
The first PowerTAC competition was in 2012. Since
then, the PowerTAC brokers improve in every edi-
tion. For a complete and holistic view of the competi-
tion, we recommend the work of (Ketter et al., 2020),
which specifies 2020’s competition.
Several agents have their code open source, an ex-
ample of this is the agent SPOT (Chowdhury et al.,
2017), an agent that uses machine learning and suc-
cessfully predicts market prices in a PDA, the Pow-
erTAC wholesale market. Chowdhury et. al used
three machine learning algorithms; a REPTree (De-
cision tree), Linear Regression, and a Multilayer Per-
ceptron (Neural network). They selected some poten-
tial information between the one available in the sim-
ulation at runtime to train a price predictor. Specif-
ically, they used 8 prices from the past bidding, as
recent trading histories reflect the present wholesale
market economy. To predict the energy price in a spe-
cific hour, their models consider the clearing prices
for the previous hour and the price in the matching
time-slot in the past day and week. The rest of their
inputs are weather forecast data, number of partici-
pants in the game, and the moving average prices pre-
dicted by the baseline agent. Besides this, the authors
also investigate the feasibility of using learning strate-
gies to predict the clearing price in the wholesale mar-
ket. The paper demonstrates learning strategies are
promising ways to improve the performance of their
agent, SPOT, in future competitions
¨
Ozdemir and Unland presented some generic data-
driven electricity price forecasting approaches and
prove that weather data can successfully reduce the
electricity price forecasting error up to a certain de-
gree (
¨
Ozdemir and Unland, 2016). Their work uses
additional drivers like weather observation data to
minimize forecasting error. Thoroughly, their hybrid
model firstly makes price predictions based on histor-
ical market-clearing prices. This model alters a sea-
sonal regression model by changing the aged terms
with a belief function. Afterward, those predicted
prices are reassessed by correlating the weather ob-
servations and market-clearing prices.
The Crocodile Agent (Grgi
´
c et al., 2018) was also
very successful by placing third in the finals of Pow-
erTAC 2018. The authors use game theory and the Ef-
ficient Market Hypothesis to model their agent, creat-
ing a complex multi-module agent. Specifically, their
agent contains a Tariff Rate Creator, Smart Tariff Cre-
ator, Tariff Manager, Portfolio Manager, and a Whole-
sale Manager. Their Wholesale Manager uses rein-
forcement learning to minimize a cost-function de-
fined in their work and create bidding strategies.
The agent Maxon (Urban and Conen, 2017) has 4
types of tariffs available. Each of these tariffs is best
suited to different scenarios and is also improved over
time. This agent also makes predictions about how
much energy he’ll need to buy using a multi-linear re-
gression model. Using this prediction, the agent tries
to buy the energy in the wholesale market by placing
orders in different slots in advance, so as to counter
agents trying to monopolize this market.
(Rodr
´
ıguez Gonz
´
alez et al., 2019) has also orga-
nized their agent into well defined modules, namely
Data Management (divided into Customer Data View
and Wholesale Data View), Retail Market (containing
Production Tariff Expert and Consumption Tariff Ex-
pert), and Wholesale Market (featuring the Wholesale
Expert). Regarding the Retail Market, this agent uses
Reinforcement Learning on Markov Decision Pro-
cesses to get two objectives; attract as many producers
as possible and bring in enough consumers to reduce
the energy imbalance in client portfolio. Regarding
the Wholesale Market, the agent uses mathematical
approximations to estimate the price of energy so as
to decide whether to buy, sell, or hold on to energy.
The related work shows a clear interest in mak-
ICAART 2022 - 14th International Conference on Agents and Artificial Intelligence
360