AuthorsMarden, Jason R.
Shamma, Jeff S.
KAUST DepartmentComputer, Electrical and Mathematical Science and Engineering (CEMSE) Division
Electrical and Computer Engineering Program
Permanent link to this recordhttp://hdl.handle.net/10754/668662
MetadataShow full item record
AbstractGame theory is the study of decision problems in which there are multiple decision makers and the quality of a decision maker's choice depends on both that choice and the choices of others. While game theory has been studied predominantly as a modeling paradigm in the mathematical social sciences, there is a strong connection to control systems in that a controller can be viewed as a decision-making entity. Accordingly, game theory is relevant in settings with multiple interacting controllers. This article presents an introduction to game theory, followed by a sampling of results in three specific control theory topics where game theory has played a significant role: ( a) zero-sum games, in which the two competing players are a controller and an adversarial environment; ( b) team games, in which several controllers pursue a common goal but have access to different information; and ( c) distributed control, in which both a game and online adaptive rules are designed to enable distributed interacting subsystems to achieve a collective objective.
CitationMarden, J. R., & Shamma, J. S. (2018). Game Theory and Control. Annual Review of Control, Robotics, and Autonomous Systems, 1(1), 105–134. doi:10.1146/annurev-control-060117-105102
SponsorsThis work was supported by Office of Naval Research grant N00014-15-1-2762, National Science Foundation grant ECCS-1351866, and funding from King Abdullah University of Science and Technology (KAUST).