Cognitive radio networks are more susceptible to jamming attacks due to the nature of unlicensed users accessing the spectrum by performing dynamic spectrum access. In such a context, a natural concern for operators is the resilience of the system. We model such a scenario as one of adversity in the system consisting of a single legitimate (LU) pair and malicious user (MU). The aim of the LU is to maximize throughput of transmissions, while the MU is to minimize the throughput of the LU completely. We present the achievable transmission rate of the LU pair under jamming attacks taking into account mainly on the transmission power per channel. Furthermore, we embed our utility function in a zero-sum matrix game and extend this by employing a fictitious play when both players learn each other’s strategy over time, e.g., such an equilibrium becomes the system’s global operating point. We further extend this to a reinforcement learning (RL) approach, where the LU is given the advantage of incorporating RL methods to maximize its throughput for fixed jamming strategies.

An optimal transmission strategy in zero-sum matrix games under intelligent jamming attacks

Leonardo Goratti;Francesco De Pellegrini;
2017-01-01

Abstract

Cognitive radio networks are more susceptible to jamming attacks due to the nature of unlicensed users accessing the spectrum by performing dynamic spectrum access. In such a context, a natural concern for operators is the resilience of the system. We model such a scenario as one of adversity in the system consisting of a single legitimate (LU) pair and malicious user (MU). The aim of the LU is to maximize throughput of transmissions, while the MU is to minimize the throughput of the LU completely. We present the achievable transmission rate of the LU pair under jamming attacks taking into account mainly on the transmission power per channel. Furthermore, we embed our utility function in a zero-sum matrix game and extend this by employing a fictitious play when both players learn each other’s strategy over time, e.g., such an equilibrium becomes the system’s global operating point. We further extend this to a reinforcement learning (RL) approach, where the LU is given the advantage of incorporating RL methods to maximize its throughput for fixed jamming strategies.
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11582/312307
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
social impact