Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
A New Dual-Input Deep Anomaly Detection Method for Early Faults Warning of Rolling Bearings
Next Article in Special Issue
LazyFrog: Advancing Security and Efficiency in Commercial Wireless Charging with Adaptive Frequency Hopping
Previous Article in Journal
Fault-Tolerant Trust-Based Task Scheduling Algorithm Using Harris Hawks Optimization in Cloud Computing
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Generative Adversarial Network (GAN)-Based Autonomous Penetration Testing for Web Applications

1
6sense Insights Inc., San Francisco, CA 94105, USA
2
School of Computing and Augmented Intelligence, Arizona State University, Tempe, AZ 85281, USA
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(18), 8014; https://doi.org/10.3390/s23188014
Submission received: 23 August 2023 / Revised: 15 September 2023 / Accepted: 19 September 2023 / Published: 21 September 2023

Abstract

:
The web application market has shown rapid growth in recent years. The expansion of Wireless Sensor Networks (WSNs) and the Internet of Things (IoT) has created new web-based communication and sensing frameworks. Current security research utilizes source code analysis and manual exploitation of web applications, to identify security vulnerabilities, such as Cross-Site Scripting (XSS) and SQL Injection, in these emerging fields. The attack samples generated as part of web application penetration testing on sensor networks can be easily blocked, using Web Application Firewalls (WAFs). In this research work, we propose an autonomous penetration testing framework that utilizes Generative Adversarial Networks (GANs). We overcome the limitations of vanilla GANs by using conditional sequence generation. This technique helps in identifying key features for XSS attacks. We trained a generative model based on attack labels and attack features. The attack features were identified using semantic tokenization, and the attack payloads were generated using conditional sequence GAN. The generated attack samples can be used to target web applications protected by WAFs in an automated manner. This model scales well on a large-scale web application platform, and it saves the significant effort invested in manual penetration testing.

1. Introduction

Penetration testing is a method of evaluating the security posture of a network, by launching controlled attacks against crucial network services and users. The goal is to identify and patch the security holes before an attacker discovers them. An attack typically starts by targeting edge sensor devices, and the attacker tries to exploit the known or unknown vulnerabilities present in the network services. The attacker can exploit a vulnerability, to obtain sensitive information or elevated privileges on a machine. The metric for measuring successful attacks is the number of vulnerabilities exploited and the cumulative impact on the network’s Confidentiality, Integrity, or Availability (CIA) as a direct result of successful exploitation [1]. The attack progression depends on the network setup. The attacker can target individual vulnerabilities in isolation if the vulnerabilities are not dependent on one other, i.e., the need to exploit one vulnerability before another. If the network is multi-hop and follows fine-grained access control principles, the penetration tester must compromise multiple vulnerabilities that are dependent on one other.
In the past few years, we have observed a rapid surge in web application tools, technologies, and libraries, like NodeJS, React, AngularJS, and Ruby on Rails [2], being deployed on edge devices, e.g.,the user interface of a smart camera. Naturally, with each web application framework, inherent security vulnerabilities are reported every year [3]. While security researchers invest much time identifying and reporting these vulnerabilities, they need help to keep up with web application vulnerability discoveries manually [4]. Attackers have also invested in using deceptive means for masquerading original attacks [5]. Moreover, the emergence of sophisticated attacks, such as Advanced Persistent Threats (APTs) [6], has increased the need for the identification of attack patterns beyond traditional signature-based attack detection.
The penetration testing market is expected to grow from USD 1718 M in 2020 to USD 4598 M by the year 2025, a compound annual growth rate (CAGR) of 21.8% [7], to address the continuously escalating security challenges. The web application market is expected to reach USD 10.44 B by 2027. Surveys, such as MIT Technology Review, have reported a 3.5 M cybersecurity workforce shortage in 2021 [8]. There is a significant demand to use artificial intelligence (AI)-enabled pentesting techniques to automate and continuously improve pentesting outcomes, which used to be handled by skilled pentesters, who could investigate vulnerabilities in a multi-stage approach. An AI-based detection mechanism for detecting deception attacks has been discussed by Pang et al. [5]. This is similar to the APT attacks that use alternate variations of known attack patterns to deceive the web application firewalls (WAFs). As a result, an AI-enabled penetration test in the real world is similar to addressing an AI planning problem. The reward model aims to obtain the highest possible reward for exploited vulnerabilities, bypassing the security mechanism to defend against the attack variants, e.g., hlWAFs and Intrusion Detection Systems (IDSs).
The computational and storage capacities of sensor networks, which typically struggle with resource constraints, have been significantly improved by the recent merger of cloud computing with WSNs [9]. Incorporation of the IoT in WSNs has even increased the attacks’ surface manifolds. Sensing devices in WSNs normally collect sensor data and transmit them, without much processing, directly to the sink node; however, in IoT networks, sensing devices are more intelligent than WSN nodes [10]. In both scenarios, they eventually leverage the distributed edge network, to store data on the cloud servers. Hence, this exposes another attack surface on the cloud-based databases in WSNs and IoT communication contexts. The Cross-Site Scripting (XSS) attack, Cross-Site Request Forgery (CSRF), and injection-based attacks are a few examples [10,11,12].
Smart manufacturing integrates various technologies, like sensors, the Industrial Internet of Things (IIoT), and Supervisory Control and Data Acquisition (SCADA) production [13]. The latest industry standards drive this—yet, due to the advent of different web technologies, they are exposed to web-based attacks, like XSS. The three-layered structure of the IoT, i.e., the perception layer, the network layer, and the application layer, introduces security issues at multiple layers [14]. Accessibility to data is over a broad spectrum of devices and platforms. Similar to traditional networks, the application layer’s vulnerability to attacks, which varies based on the particular IoT scenario, is the primary security concern, even in sensor networks; hence, there is a need to develop adaptive defense strategies to counter these attacks [12].
Generative Adversarial Network (GAN)-based approaches have been successfully applied to network IDSs in recent years. IDSGAN [15] generates adversarial malicious traffic to attack IDSs, by deceiving and evading detection. The research utilized functional and non-functional features from the NSL-KDD dataset [16], to train a GAN model to fool a BlackBox IDS. The generator was able to fool different detection algorithms, such as Support Vector Machine (SVM) [17], Naive-Bayes classifier [18], Multi-Layer Perceptron (MLP) [19], and Decision Tree (DT) [20], by generating variations of network and host-based attacks, such as User to Root (U2R), Remote to User (R2U), and DDoS attack. The GAN model used by IDSGAN is slow to train. While generating valid samples for fooling an IDS works well, this may not scale well for a web application framework protected by a WAF. The unconditional generative model has no control over the mode of data generation. Model conditioning using additional information allows more targeted data generation [21]. The conditioning is based on the class labels or some of the data. The generator G and discriminator D, two building blocks of a GAN, are conditioned on some extra information, such as a class label y or data modality.
GANs have proven to be an effective approach to generating continuous data, such as images [22]. However, using GANs for generating discrete data or attack sequences, such as XSS and SQLI attack payloads, is challenging. The reason for this inherent limitation is that generation starts with random sampling, followed by a deterministic transform on the model parameters. The gradient loss of D is used to guide G to change the generated value and make it more realistic. In the case of discrete token generation, the slight change approach makes limited sense, because there might be no token in the limited dictionary of the generator. Recent works on sequence generation, such as SeqGAN [23], overcome this limitation by modeling data generation as a stochastic policy in a Reinforcement Learning (RL) setting.
We considered the problem of generating attack payloads that can bypass the signature-based web defense mechanism. The GAN is provided with conditional information on the attack labels. This helps in generating high-quality attack samples. There are some research works that involve the use of fuzzy logic for generating attack samples. The Fuzzy Logic System (FLS), introduced by Shahriar et al. [24], utilizes input from different attack types, described as top threats in Open Worldwide Application Security Project (OWASP) web attacks, and risk assessment models, to generate attack payloads. These payloads can be tested against PHP-based applications, to check the security risk level of different applications. The tokens used in fuzzy logic are often at individual character level. Our framework uses semantic tokenization and a Byte Pair Encoding (BPE) [25] algorithm, to create better tokens, so as to generate logically correct attack sequences.
Moreover, the fuzzy logic scales poorly as the input size increases. The number of variations of discrete tokens can be exponential, in terms of token space. In our semantic tokenization approach, the tokens can map to a constant set of classes, such as tags, script parameters, function body, hyperlinks, etc. This makes the space complexity of the token generation method polynomial, in terms of the maximum length of sequence and the number of attack tokens. Thus, conditional sequencing scales well with an increase in input size, compared to a fuzzy-logic-based approach.
In this research, we utilized conditional sequence generation, to target web application firewalls protecting web applications against application layer attacks, such as XSS, SQL Injection, and Directory Traversal. The semantic knowledge from security experts encoded the data modality required for the attacks. The sequence generation process utilized this information for generating targeted attack payloads. We tested the generated attack payloads against open-source ModSecurity WAFs [26] and AWS WAFs [27]. The payloads generated by the conditional sequencing were able to bypass ModSecurity WAFs and AWS WAFs.
These payloads can help improve the attack signatures in the WAF ruleset. The key contributions of this work are as follows:
  • Conditional sequence generation, by understanding the semantic structure of web attack payloads. The technique helps in improving the training efficiency of the generator and in generating valid attack signatures that can fool the discriminator.
  • Evaluation of generated attack samples on production-grade WAFs. We used ModSecurity and AWS WAFs to test the quality of generated web attack samples. We observed that 8.0% of the attack samples targeting AWS WAF allowed listing and that up to 44% of the samples targeting AWS WAF block listing were able to bypass the rules in place for blocking web attacks.
  • Generating a GAN-based synthetic attack dataset, by training a GAN model on real and fake attack samples. This synthetic data can help to train web application layer defensive devices, such WAFs, against sophisticated attacks, like APT.

2. Related Work

Web attacks, such as XSS, can lead to disruption of confidentiality and availability in a Cyber–Physical System (CPS). Duo et al. [1] modeled a CPS based on time-driven and event-driven cyber attacks. Penetration testing can be considered as an event-driven attack simulation, to detect vulnerabilities in a CPS. Alsaffar et al. [28] conducted a study of different types of XSS attacks, and they proposed a greedy algorithm for the detection of XSS vulnerabilities in web applications. The program only considered a static set of rules for detecting XSS attacks. An automated mechanism to conduct pentesting in a controlled manner is challenging. Several approaches have been used, to formulate pentesting as a planning problem. Lucangeli et al. [29] used Partially Descriptive Domain Modeling (PDDL)-based attack modeling. This approach was limited, since it assumed complete information about the attack states and actions. Attack planning has been modeled as a Partially Observable Markov Decision Process (POMDP) problem by Sarraute et al. [30]. This helps in incorporating uncertainty, such as non-deterministic actions. The POMDP modeling used in this work has been examined in limited experimental settings. As the environment becomes more complex (an increased number of exploits and machines), the runtime of the POMDP solver increases significantly. A reinforcement learning (RL)-based approach to automated pentesting has been considered in research works [31,32]. Schwartz et al. [31] formulated the problem using Markov Decision Process (MDP) modeling. The authors noted that an RL approach was scalable only in small-scale environments. Schwartz et al. [33] improved on their earlier work, by using a modified version of the POMDP. The authors incorporated the defender’s behavior as part of the response to pentesting activities within the model of autonomous pentesting. Ghanem et al. [32] used the POMDP modeling approach. Naturally, the time consumed to conduct pentesting on small-scale networks using this approach is of an order of hours. Tran et al. [34] used multi-agent RL for decomposing action space into smaller subsets, to help conduct pentesting at scale. Zhou et al. [35] used an improved Deep Q-Network (DQN) for addressing issues with sparse rewards, by improving the exploration ability of the neural network. Some other approaches that have been used for autonomous pentesting include using contingency planning to model the problem. Empirical evaluation was conducted in a simulated setting with known vulnerabilities.
Adversarial examples have been used for generating fake images with success. Adversarial networks, such as GANs, use the generative network to generate counterfeit images/samples that fool the discriminate model with a knowledge base of real data samples. GAN-based models have been used in cybersecurity operations, such as password cracking, intrusion detection, and XSS attack payload validation. PassGAN [36] uses a deep learning approach for password guessing. PassGAN uses training on 9.9 million unique leaked passwords, and using a GAN-based password cracking approach produced better password guesses than well-known tools, such as John the Ripper and HashCat. IDSGAN generates malicious traffic records, to attack IDSs by evading detection. The IDSGAN design classifies traffic into functional and non-functional features. The authors altered the non-functional features, to generate adversarial examples for different attack categories, e.g., retaining intrinsic (session-based) and time-based features for a DDoS attack, and modifying content and host-based features. An empirical evaluation showed a low detection rate against attack classification algorithms for the NSL-KDD dataset. Deep Convoluted GAN (DCGAN) has been used by Yang et al. [37], to deal with unbalanced network intrusion data. Zhang et al. [38] used a Monte Carlo Tree Search (MCTS)-based algorithm, to generate adversarial XSS attack samples. The research restricted attack sample modification using predefined rules and used a GAN to optimize the detector and to improve the attack detection rate.
There is a lack of robust attack datasets that can help detect sophisticated attacks, such as APTs  [6,39]. The use of deception-based attacks for some recent datasets, such as DAPT2020 [40] and Unraveled [41], targeted a general class of APT attacks, by simulating the threat vectors used in APT attacks. As the scale of web infrastructure and web technologies expands, it will become difficult for security researchers to generate real attack samples by using attack simulations. This research proposes complementing datasets such as DAPT2020 [40], and Unraveled [41], by generating fake attack data from real attack samples. GANs can create adversarial examples that mimic sophisticated attack techniques, as we have demonstrated in this research. By incorporating these examples into the WAF training data, it is possible to bolster a WAF’s resilience against evasion attempts and to improve its effectiveness against more advanced attacks.

3. Background

The process of penetration testing involves information gathering about the target, such as open ports, service version, Operating System (OS), and using the information to mount targeted attacks against a service. Several tools and techniques help in conducting penetration testing. One key issue in using tools is that the known vulnerabilities limit them. Most of these vulnerabilities have an associated Common Vulnerability Enumeration Identifier (CVE-ID) stored in a Common Vulnerability Scoring System (CVSS) [42]. Some vulnerabilities are left unidentified during the development life-cycle of a product. These vulnerabilities are known as zero-day attacks [43].

3.1. Web Application Attacks

There are several parts of a web application that can be targeted by web application attacks. A typical web application includes a web application protocol, e.g., HTTP/S, server-side functionality, the use of scripts or code to generate dynamic content, application design flaws, authentication, and a data storage mechanism used by the application. Some well-known web vulnerabilities include session hijacking, bypassing authentication, SQL injection attacks, and XSS  [44]. The XSS attacks involve using some aspect of the application’s behavior, to carry out malicious actions against users. These actions include logging user keystrokes, and masquerading user privileges, to carry out unintended actions.
An example of how an attacker can capture the session token of an authenticated user has been provided in Figure 1. An authenticated user who logs into an application is issued a cookie—step 1. The attacker supplies a crafted URL to the user—step 2. The user requests the URL—step 3, and executes the malicious JavaScript returned by the attacker—steps 4 and 5. The malicious script requests the server owned by the attacker, with the user’s session token. In effect, the user’s captured token is supplied to the domain controlled by the attacker, and the user’s session is hijacked—steps 6 and 7. The payloads for such web-based attacks can also cause website defacement and other user actions, such as adding a new user with admin privilege (if the admin’s session has been hijacked).

Defense Mechanisms against Web Attacks

Modern servers and applications use several protection mechanisms to prevent web-based attacks. These techniques include (a) blocking the attacker’s input, based on an attack signature match, (b) input sanitation or encoding, and (c) truncating attack strings to a fixed length, to prevent attackers from injecting malicious scripts. Web applications exposed to the public internet make use of WAFs  [45] with these defense mechanisms, to filter and monitor application layer traffic. The attackers have also adapted to the defensive techniques employed by WAFs.
Figure 2 shows expressions blocked by WAFs. The first attack vector uses script tags to insert the XSS payload. Modern WAFs can block expressions by signature matching, but a crafty attacker can use a dynamic expression to bypass the filters. The attacker can alternatively leverage other scripting platforms that the application server provides, such as Visual Basic (VB), to create a script that can pass through undetected firewall filters. The attacker uses NULL bytes in the second attack vector, to bypass the WAF filter. Other techniques used include event handlers, like onclick and onmouseover, which bypass signature-based WAF filters. Some WAFs limit the script length that can be inserted as payload. The length limits can also be bypassed by using a script being loaded from a remote source, e.g., <script src=http://remote-server/malicious.js></script>. Next, we describe how this process of blocking and bypassing web attacks can be formulated as a two-player zero-sum game and modeled as a GAN.

3.2. Generative Adversarial Networks (GANs)

A GAN defines two neural networks: generator G and discriminator D [46]. In a traditional adversarial network, the data distribution of the generator is defined as p g over data x. A prior input p g ( z ) is used as an input noise variable. The mapping of the input noise to the data space is represented as G ( z ; θ g ) . The generator G is a differential function represented by an MLP with parameter θ g . The second MLP used in this model, called the discriminator, is represented as D ( x ; θ g ) , which outputs a scalar. D ( x ) represents the probability that x came from data rather than from noise p g .
The variable p d a t a ( x ) refers to the original data distribution. E x p d a t a ( x ) is the expectation value function: it means that the expected value of x is assumed to be distributed over p d a t a ( x ) . The value function V ( G , D ) represents a min–max game between the generator and the discriminator. The discriminator is trained to maximize ( m a x D ) the probability of assigning the correct label to the training example l o g D ( x ) . Simultaneously, the generator is trained to minimize ( m i n G ) the function l o g ( 1 D ( G ( z ) ) . In summary, a min–max game with value function V ( G , D ) is defined as
m i n G m a x D V ( D , G ) = E x p d a t a ( x ) [ l o g D ( x ) ] + E z p z ( z ) [ l o g ( 1 D ( G ( z ) ) ) ] .
The initial samples generated by G are not optimal enough to bypass the detection criterion of D, and are rejected by the discriminator. The generator keeps generating the adversarial samples and updating the parameters for the subsequent samples, and the generator learns a better evasion technique, to fool the discriminator.

3.3. GANs for Generating Web Attacks

3.3.1. Motivating Example

GANs can be used to generate simulated attack data, such as malicious input payloads for injection attacks (e.g., SQL injection, XSS) or evasion techniques for bypassing security filters. These simulated attacks can be employed to evaluate the effectiveness of security mechanisms and to identify potential vulnerabilities. In this work, we improved the structure of GAN modeling, by using conditional sequencing. We considered conditional sequence generation as a process of identifying stochastic reinforcement learning policy. The policy rewards are judged on the complete sequence of the attack payload and are passed to intermediate state-action pairs, using the Monte Carlo (MC) search process.
Consider Figure 3. We assume that the provided dataset has known payloads used for XSS attacks. We use a process known as semantic tokenization, which will be elaborated in Section 4.1, to obtain tokens from the initial dataset, which represents different feature values for XSS attacks, e.g., <script> is a tag attribute, while alert(1); is a function body attribute. It is difficult to label all such attributes, so we classify the attributes with no such classification by using the other label. Moreover, we also know that attacks provided by a dataset can be replayed against a malicious web application, to check the validity of the attack. We obtain different results when we replay these attacks against known vulnerable applications, such as DVWA, Gruyere, and OWASP vulnerable web applications. If the attack payload generates a stored, reflected, or domain-based XSS attack, we add the label ok. If there is an error when the payload is replayed against the web application, we add the label error. If nothing happens when the payload is replayed on the vulnerable web page, we add the label fail.
These labels and tokens are passed to the generator, G. Generating conditional semantic sequences starts by generating a random initial state, e.g.,  s 0 = <script>. The next state, s 1 , is selected from the list of available tokens, e.g., the action selects token alert(1); and the model transitions to state s 1 . The state transition is deterministic, based on the action selected, s 0 × a s 1 . In this example, a = alert(1); and  s 2 = <script> alert(1);. The entire sequence { s 1 , s 2 , . . . , s N } is evaluated by the discriminator, to check if the generated sequence is a valid attack. The discriminator is also pre-trained on both valid and invalid attack sequences, since pre-training helps improve the generator’s efficiency. Consider the generated sequence <script>alert(1);<script>: the model achieves a higher reward from the discriminator, because this sequence passes the fitness function test for a valid attack. In case the generated attack sequence is not valid, the model utilizes policy gradient and a Monte Carlo search based on the expected reward from the discriminator model.

3.3.2. GAN for Bypassing a Web Application Firewall

Figure 4 provides a GAN framework for a web application pentest. The generator model uses the web application attack samples from a distribution of attack samples tried and tested as payloads. As shown in Figure 4a, the generator passes the attack payload to the discriminator. The attack sample is validated against a web application, to check if it generates an exploit against the web application. The discriminator model uses known attack samples that have worked on the application, to check if the sample provided by the generator will work on the web application. The classification result is used to classify the attack sample as valid/invalid. The model of the generator and the attack signature database are updated, based on the result.
In order to understand the semantic meaning of using a GAN against a web application and to showcase the practical application of GAN-generated payloads, consider Figure 4b, where an attack payload from a generator is replayed against a web application protected by a WAF. The WAF signature match is used as a criterion to classify an attack as malicious or benign. The attack payloads are tried against the WAF, to check if they are identified as malicious and are blocked. Using a GAN-based attack payload generation and validation mechanism, we can generate payloads that trigger web application vulnerabilities but are not classified by the WAF as malicious. During the subsequent rounds of training for the GAN, the generative model can be updated with improved versions of the attack payloads. This approach will be beneficial for generating valid attack payloads for a large-scale web application platform that is difficult to test by using known attack payloads or a manual pentesting approach. One challenge to the direct use of a GAN for non-image datasets, such as cyber-intrusion detection systems, is that features present in these datasets are discrete. Thus, numeric 0-1 features and non-numeric features are represented using One-Hot Encoding or Dummy Encoding. The dimension expansion used to account for this encoding leads to the problem of vanishing gradient [47]. Chen et al. [48] used a Wasserstein-Distance-based modified training goal to deal with the vanishing gradient problem. The research work used an additional variable Encoder (E) to train the modified GAN network. Another problem with the direct use of a GAN for generating sequential data that represent an attack such as XSS is that a GAN is designed for generating real-valued, continuous data. However, using a GAN to generate a sequence of discrete tokens is challenging. The GAN can give the score/loss for the entire sequence when it has been generated; the measure of fitness for the partially generated sequence is quite difficult. SeqGAN [23] considered sequence generation as a sequential decision making process, and the generative model was treated as an RL agent. The state was generated tokens so far, and the action was the next token in the sequence. The authors used the policy gradient method and employed an MC search to approximate state-action value. In the next section, we explain how we used a SeqGAN framework with conditional token encoding for training a GAN network and generating valid attack payloads.

4. Conditional Attack Sequence Generation

4.1. Attack Payload Tokenization

Tokenization is the process of breaking raw text into small chunks. The tokens can be groups of characters, words, or sentences. The tokens help interpret the meaning of the text, by analyzing the sequence of words (tokens). In text tokenization, the parts of the text that do not add any special meaning to the sentence, such as stop words, are removed. Removing these words from the dictionary reduces the noise and dimension of the feature set. There are different ways to perform tokenization. Some popular techniques include white-space, dictionary-based, rule-based, regular expression (regex)-based and subword-based token generation. Most of these methods suffer from inherent limitations, e.g., limitations on vocabulary size and handling words that are absent in the vocabulary. Techniques such as BPE are used to deal with the Out-Of-Vocabulary (OOV) sequences. It segments OOV as subwords and represents the words in terms of those subwords.
The tokenization method that identifies meaningful attack payload tokens can be applied to a dataset of attack inputs, such as XSS, to identify relevant sub-sequences that can be combined to target the vulnerable web application. Semantic tokenization uses markers (such as tags <>, <script), parameter names (such as href=), function body (such as alert(), common words (such as javascript, VBScript), and special encoding (such as u003c), http/https links. Once the semantic meaning has been assigned to the tokens, the BPE [25], a variant of Huffman Encoding, is applied to the semantically labeled tokens. It uses more embedding or symbols for representing less frequent terms in the corpus.
The semantic tokenization process takes the XSS dataset as input, as shown in Figure 5. The input text is split, based on matching conditions for the markers, such as tags, encoding, function body, and parameter name. The input corpus (XSS dataset) is parsed line by line, and the data are converted to the HTML-rendered format. The rendered data from each are replayed against vulnerable applications. We utilized Burpsuite [49] to replay the initial attack data D and to label each attack payload, based on the result of the HTML code replayed against the web application (ok, note, warn, fail, error). The labels were used as input for the conditional sequential GAN. The tokens from each line were annotated and grouped into frequently occurring symbols, using a BPE algorithm. These were added to the vocabulary of the known tokens. The process was repeated until no new combination of symbols was present. The vocabulary and data labels were passed to the generator, G.

4.2. Conditional Sequencing

The architecture for conditional attack sequence generation has been described in Figure 5. We use the input data from the XSS dataset D a . The data are preprocessed, to extract the semantic tokens. The XSS data are also labeled with the results of attack sequences that are replayed on a vulnerable web application. The label information p and tokens Y 1 : T are passed to the generator, G θ ( Y 1 : T ) . The discriminator is assumed to have inputs from the original dataset χ 1 : N and from some fake data generated from input sequences that failed to generate valid alerts on vulnerable web applications. The discriminator D ϕ ( Y 1 : T , p ) utilizes the policy gradient or the Maximum Likelihood Estimate (MLE) for learning optimal policies for the generation of attack sequences, Q D ϕ G θ . The attack payloads are validated against vulnerable web applications, and the payloads that pass the validation phase are added to the original XSS dataset D a .
We consider the dataset χ 1 : N and the labeling information p as the initial input to the sequence generation process. The generative model is θ -parameterized, and the model parameters can be determined by the data distribution based on labels p. The goal of generator G θ is to produce a sequence Y 1 : T = { y 1 , y 2 , , y T } , such that y t Y , where Y is the vocabulary of the candidate tokens extracted from χ 1 : N . This process can be considered an RL policy generation problem. The policy generation process is a modified version of sequence generation, as discussed in SeqGAN [23], with semantic tokenization and conditional labeling. The policy model for conditional sequencing G θ ( y t | Y 1 : t 1 , p ) is stochastic. The transition between states is deterministic, i.e.,  δ s , s a = 1 , where s = Y 1 : t 1 , s = Y 1 : t , a = y t , and, for all other states, s . The discriminator model D ϕ is ϕ -parameterized for improving the generator, G θ . The model D ϕ ( Y 1 : T | p ) is a problem indicating how likely the sequence is, from the real attack dataset D a . The discriminator is trained by providing positive examples from the attack dataset D a and negative examples from the synthetic dataset. The negative examples are malformed attack payloads that fail the XSS attack test on vulnerable web applications.

Conditional Sequence-Based Attack Generation

The objective of the generator model G θ ( y t | Y 1 : t 1 , p ) is to generate a sequence from the start state s 0 , the model parameters θ , and the attack labels p. As an example, the start state could be one of the semantically labeled tokens, e.g., s 0 = </scrip</script>t>. The goal of the model is to maximize the expected reward R T for the generation of a complete attack sequence, described by Equation (2). The function E [ R T | s 0 , θ , p ] represents the expected reward, given the labels, start state, and model parameters:
J ( θ ) = E [ R T | s 0 , θ , p ] = y Y G θ ( y 1 | s 0 , p ) . Q D ϕ G θ ( s 0 , y 1 ) .
The Q D ϕ G θ ( s , a ) is the action value function for a sequence, i.e., the expected reward accumulated by starting with the initial state s, taking the action a, and following the conditional sequence G θ parameterized by the attack labels. The objective function for the sequence starts from the initial states. It follows the policy to generate a sequence of tokens Y 1 : T = { y 1 , . . , y t , . . y T } that can be considered real attacks when evaluated on the vulnerable web application—Figure 5. The action-value function REINFORCE [50] is used by the discriminator D ϕ ( Y 1 : T n ) for estimating the reward. The reward calculated by the discriminator is for the finished attack sequence. The model captures the fitness of the previous tokens in the attack sequence (prefix) and the resulting future outcomes. The model utilizes an MC search with the roll-out policy G β , to sample T t unknown tokens. The N-time MC search procedure is represented by Equation (3) below:
{ Y 1 : T 1 , , Y 1 : T N } = M C G β ( Y 1 : t ; N | p ) .
The tokens are Y 1 : t n = ( y 1 , , y t ) , and  Y t + 1 : T n is sampled, based on the roll-out policy G β and the current state. The roll-out policy is started from the current state, and run for N times, to obtain the batch output of the attack samples. The roll-out policy for the conditional sequence starts from the current state and runs till the end of the sequence, for N times, to obtain a batch of output samples. As described in Equation (4), this process reduces the variance and obtains a more accurate assessment of the action value:
Q D ϕ G θ ( s = Y 1 : t 1 , a = y t | p ) = 1 N n = 1 N D ϕ ( Y 1 : T n | p ) , Y 1 : T n M C G β ( Y 1 : t ; N | p ) f o r t < T Q D ϕ G θ ( s = Y 1 : t 1 , a = y t | p ) = D ϕ ( Y 1 : t | p ) f o r t = T
The process does not provide intermediate rewards; instead, the function iteratively updates and improves the generative model, starting from s = Y 1 : t . The discriminator is retrained when more realistic attack payloads are generated from the model. In turn, the new discriminator model is used to retrain the generator. The policy-based model optimizes the parameterized policy, to maximize long-term rewards directly.
We describe conditional sequence generation and XSS attack test procedures in Figure 6, and we also provide the detailed Algorithm 1, for the same. The generator G θ , parameterized by attack labels p, is pre-trained on S, using the MLE algorithm. The supervised signal from the pre-trained discriminator helps improve the generator’s efficiency. The generator is conditioned on the attack labels p and trained for g-steps, to generate the sequence Y 1 : T (line 6). The Q-function Q D ϕ G θ is calculated for each step of the generator (line 7). If the current state is represented by s = Y 1 : t 1 and the action is a = y t , the next state is calculated by using the action-value function. The generator is updated, using the policy gradient approach described earlier. The discriminator needs to be re-trained periodically, to improve its performance. The positive examples are provided from training set S, and the negative examples are provided from the failed attack sequences from the generator. The number of positive and negative examples is the same for each d-step in the algorithm. The trained generator is used for XSS attack validation, by replaying the sequences against vulnerable web application lines 18–24. The valid attacks are added to the base initial training set χ 1 : N , to improve the variability of the training data.
Algorithm 1 Conditional Sequence Generation
1:
procedure Conditional Sequence Generation( X 1 : N , p )
2:
    Initialize G θ , D ϕ , p, β θ
3:
     G θ pre-trained using MLE on S, p
4:
    Train D ϕ from positive, negative G θ samples
5:
    Pre-train D ϕ to minimize cross entropy
6:
    for g-steps do Generate Y 1 : T = ( y 1 , . . , y T | p ) G θ
7:
        for t ∈ {1:T} do
8:
           Calculate Q-function Q ( a = y t ; s = Y 1 : t 1 | p )
9:
        end for
10:
      Update generator using policy gradient
11:
    end for
12:
    for d-steps do
13:
        Generate true alerts, false alerts using G θ , S
14:
        Train D ϕ for k-epochs
15:
    end for
16:
end procedure
17:
procedure XSS Attack Test( G θ , S , χ 1 : N )
18:
    for s ∈S, G θ  do
19:
        s ← html_render(s)
20:
        if xss_eval (s) then
21:
           assign_label (s)
22:
           update χ 1 : N , add s
23:
        end if
24:
    end for
25:
end procedure

5. Experimental Evaluation

We used the XSS dataset [51] collected from multiple XSS scanning tools containing the payload data covering different XSS attacks. The dataset covers different features of XSS attacks, such as tags, function body, URL, and encoding.

5.1. Evaluation of Loss for Conditional GANs

We utilized the sample payloads from the XSS dataset [51] to train our GAN model. The discriminator loss consisted of two parts, i.e.,  d _ l o s s 1 and  d _ l o s s 2 . The first loss value detected real attack samples as real, and the second loss detected fake attack samples as fake. On the other hand, the generator loss tried to generate attack samples that were hard for discriminators to detect as real attacks. The generator and discriminators were trained to improve loss functions till convergence was achieved. We observed that our discriminator loss functions decreased as the number of training samples increased, converging to a stable value ∼1.1 × 10 2 (see Figure 7). This meant that our discriminator was more accurate at distinguishing between real and fake attack samples. The value of the loss function for the generator also decreased with time, reaching a minimum value of ∼250 epochs. We observed that there was no further improvement in the loss function of the discriminator. This signified an improvement in the quality of generating attack samples that could fool the discriminator’s ability to detect attack payloads. In summary, the attack samples generated at around 250 epochs could be utilized to test the web application firewall’s effectiveness in detecting attacks.

5.2. Web Application Firewall Bypass

5.2.1. ModSecurity WAF Testing

We utilized the ModSecurity WAF to check the attack payloads generated by different variants of the GAN network. The attacks were first verified over vulnerable web applications and were then replayed against the WAF, to check how many attacks were detected by the rules of the WAF. ModSecurity consists of modules, such as PhantomJS (a headless WebKit with JavaScript API). The module uses WebKit’s browser environment to detect reflected XSS attacks accurately. For instance, XSS attack payloads use partial non-alphanumeric obfuscation. The code
  • <script>eval("aler"+(!![]+[])[+[]])("xss")</script>
can be used to bypass normal XSS detection filters. The PhantomJS conducts execution time analysis within the browser Document Object Model (DOM) after de-obfuscation, to validate the attack payload. Other modules, such as Lua API, allow the security team to hook in external programs that extract HTTP data and pass it to PhantomJS for detection.
We evaluated the effectiveness of a vanilla GAN and a conditional GAN (CGAN) on a vulnerable web application protected by a ModSecurity WAF. For each trial run in Table 1, we randomly selected payloads from the XSS dataset. The test set consisted of payloads generated by the vanilla GAN and the CGAN. The percentages represent the success rate for each version of the GANs, i.e., the number of valid attack samples that were generated after the training of the GAN was finished. These payloads were able to successfully bypass the WAF. During the experiment, we found that 10.37% of the vanilla GAN payloads could bypass the ModSecurity WAF filters in the first batch, whereas for the CGAN only 7.66% of the generated payloads could bypass the WAF. For the second batch, the CGAN performed slightly better than the vanilla GAN (see Table 1). We observed good performance for the CGAN in the fourth batch, i.e., only 0.08% of the vanilla GAN payloads were able to bypass the WAF, whereas 12% of the CGAN payloads were able to bypass the WAF. This meant that the CGAN was more consistent in providing valid payloads across all the trial runs. This was because the CGAN utilized semantic tokenization to understand the structure of the XSS payloads and mimicked the valid payloads closely. In effect, the quality of generated attack samples was better for the CGAN.

5.2.2. AWS WAF Testing

We enabled the AWS WAF to protect a commercial-grade web application. The infrastructure in AWS requires the creation of an application load balancer (ALB) or an API gateway. We utilized an ALB for our setup [52]. We created a custom ruleset, including AWS pre-set rules, to detect and prevent attacks, like URI path inclusion, SQLI, anonymous IP address, and different variants of XSS attacks. The commercial rules from AWS marketplace vendors like Fortinet and F5 were also used as a part of the AWS access control list (ACL). The ACL was attached to the created ALB. In total, the WAF comprised 3000 rules. The attack payloads generated using the conditional GAN were replayed against the AWS WAF, and the results were observed in the AWS WAF management dashboard.
The % Attack Match means the percentage of valid payloads. We observed that 8% of the GAN-generated payloads were able to bypass the rulesets of the AWS WAF (Table 2). This means that the success rate for the CGAN on commercial firewalls is quite low. This was expected, because commercial firewalls utilize a broader set of signatures to detect web attacks. We observed that 44.9% of attack payloads were detected under AWS-managed XSS rules, whereas 44% of payloads were detected using Fortinet-based commercial rules downloaded from the AWS marketplace. The WAF misclassified 3.1% of the attack payloads as SQLI attacks. This indicates that attack signatures from commercial WAFs are prone to minor errors. While the bypass rate was quite low for the AWS WAF, a malicious attack group only requires a few valid signatures to bypass a WAF; thus, security teams can utilize the valid attack signatures to update the WAF rulesets.
 /?saivs.js%20%20/%3E%3Cvideo%3E%3Csource%20%20/%3E
/?%3Ca/src=/%20%3C/img%3E%3Cinput%20’
/?%5Cu0061lert%60%60;
An example of attack payloads that successfully bypassed the AWS WAF can be seen above. The caveat that attackers use to bypass XSS filters includes removing different parts of the attack string and checking if the input is still blocked. Other methods include using alternate means of introducing scripts, such as img tags, event handlers, script pseudo-protocols (such as javascript), and dynamically evaluated styles. The discriminator in a GAN learns these variations after being trained on attack payloads, and, hence, the discriminator can generate attack sequences with no matching attack signatures.

5.3. Comparative Analysis with Existing Research

Our research provides a unique mechanism for learning the signatures configured in a given WAF and new data that can be used for improving attack detection systems. We identified several gaps in the existing research on autonomous web app pentesting. We identified that, except for the POMDP+ model proposed by Schwartz et al. [33], most research fails to incorporate the defender’s perspective when modeling autonomous pentesting. Our GAN model captures the attacker’s and the defender’s perspective through generator and discriminator models. Some other limitations of the existing research include the use of static configurations [28], lack of practical evaluation [34], and showing limited scalability on an extensive network [35]. We compared the convergence rate of a CGAN network to the POMDP model proposed by Schwartz, et al. [31]. The authors conducted training on single and multi-site networks, and convergence took ∼1000 episodes, which was 4× that of the convergence achieved by conditional GAN network in the proposed solution. This is due to a better understanding of attack semantics in our model.

6. Discussion

We discussed the challenges posed by integrating cloud computing with WSNs and incorporating the IoT in these networks. Although there are defense mechanisms, such as commercial WAFs, to identify these security issues, the sophistication of these attacks keeps increasing, including deceptive means used by attackers to masquerade as genuine traffic. Attacks such as APTs require the identification of attack patterns beyond traditional signature-based detection. Penetration testing plays a crucial role in identifying security issues and risks related to the IoT, sensor networks, smart solutions, and web-based vulnerabilities. A significant shortage of cybersecurity professionals has led to a demand for AI-enabled penetration testing techniques.
GAN provides a mechanism to mimic a pentester and a network defender in a two-player zero-sum game. Using GANs to generate discrete data or attack sequences, such as XSS and SQL Injection payloads, is challenging. The generation process can suffer from inherent limitations, such as poor input quality, lack of diversity, and mode collapse, as discussed in SentiGAN [53]. We propose a conditional sequence generation mechanism that utilizes the pentester’s semantic knowledge in the generative model for performing autonomous pentesting against commercial WAFs, such as ModSecurity and AWS WAF.
The practicality of the proposed approach was highlighted by learning the success rate of valid attack payloads on WAFs. Our model achieved fast convergence, due to the semantic encoding of attack tokens. Moreover, the synthetic data generated by the generator upon convergence can help to improve the signature database of the WAFs. Although we achieved some valid payloads, the sequence GAN framework used in our work suffered from the inherent limitation of capturing long-term dependencies between sequences—hence, performing poorly on an AWS WAF. Category-aware generative networks with hierarchical learning models [54] could help overcome some of the limitations present in sequence GANs.

7. Conclusions and Future Work

We propose a GAN-based solution for modeling web application attacks and conducting autonomous pentesting on applications protected by commercial WAFs. Our model utilizes conditional sequence generation, to learn the structure of attack payloads that can bypass WAFs. Experimental evaluation conducted on a ModSecurity WAF and an AWS WAF generated several valid payloads. Our proposed solution tests the rule structure of WAFs, by generating new payloads that are hard to detect, using existing WAF rules. These payloads can, in turn, be utilized to improve the robustness of WAFs and to deal with sophisticated attacks. A limitation of our work is the low bypass rate on commercial-grade WAFs, such as the AWS WAF. We plan to explore alternate GAN versions, to better model the long-term sequences of valid attack payloads. This approach could help to improve our model’s performance on commercial WAFs. We also plan to expand our experimental section attacks, such as CSRF, SQLI, and directory traversal attacks.

Author Contributions

Conceptualization, A.C. and K.J.; Validation, M.Z.; Investigation, A.C. and K.J.; Writing—original draft, A.C. and K.J.; Writing—review & editing, M.Z.; Supervision, M.Z.; Project administration, M.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Science Foundation award #OAC-2126291.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available in a publicly accessible repository that does not issue DOIs Publicly available datasets were analyzed in this study. This data can be found here: https://github.com/payloadbox/xss-payload-list.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Duo, W.; Zhou, M.; Abusorrah, A. A survey of cyber attacks on cyber physical systems: Recent advances and challenges. IEEE/CAA J. Autom. Sin. 2022, 9, 784–800. [Google Scholar] [CrossRef]
  2. Milos Timotic. 9 Web Technologies Every Web Developer Must Know in 2021. Available online: https://tms-outsource.com/blog/posts/web-technologies/ (accessed on 2 October 2021).
  3. Disawal, S.; Suman, U. An Analysis and Classification of Vulnerabilities in Web-Based Application Development. In Proceedings of the 2021 8th International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 17–19 March 2021; pp. 782–785. [Google Scholar]
  4. Chowdhary, A.; Huang, D.; Mahendran, J.S.; Romo, D.; Deng, Y.; Sabur, A. Autonomous security analysis and penetration testing. In Proceedings of the 2020 16th International Conference on Mobility, Sensing and Networking (MSN), Tokyo, Japan, 17–19 December 2020; pp. 508–515. [Google Scholar]
  5. Pang, Z.H.; Fan, L.Z.; Guo, H.; Shi, Y.; Chai, R.; Sun, J.; Liu, G.P. Security of networked control systems subject to deception attacks: A survey. Int. J. Syst. Sci. 2022, 53, 3577–3598. [Google Scholar] [CrossRef]
  6. Alshamrani, A.; Myneni, S.; Chowdhary, A.; Huang, D. A survey on advanced persistent threats: Techniques, solutions, challenges, and research opportunities. IEEE Commun. Surv. Tutor. 2019, 21, 1851–1877. [Google Scholar] [CrossRef]
  7. GlobeNewsWire. Global Penetration Testing Market 2020–2025. Available online: https://www.globenewswire.com/en/news-release/2020/07/10/2060450/28124/en/Global-Penetration-Testing-Market-2020-2025-Increased-Adoption-of-Cloud-based-Penetration-Testing-Presents-Opportunities.html (accessed on 2 May 2021).
  8. Cybersecurity Ventures. Cybersecurity Talent Crunch. Available online: https://cybersecurityventures.com/jobs/ (accessed on 2 July 2021).
  9. Alturki, R.; Alyamani, H.J.; Ikram, M.A.; Rahman, M.A.; Alshehri, M.D.; Khan, F.; Haleem, M. Sensor-cloud architecture: A taxonomy of security issues in cloud-assisted sensor networks. IEEE Access 2021, 9, 89344–89359. [Google Scholar] [CrossRef]
  10. Pundir, S.; Wazid, M.; Singh, D.P.; Das, A.K.; Rodrigues, J.J.; Park, Y. Intrusion detection protocols in wireless sensor networks integrated to Internet of Things deployment: Survey and future challenges. IEEE Access 2019, 8, 3343–3363. [Google Scholar] [CrossRef]
  11. Medeiros, I.; Beatriz, M.; Neves, N.; Correia, M. SEPTIC: Detecting injection attacks and vulnerabilities inside the DBMS. IEEE Trans. Reliab. 2019, 68, 1168–1188. [Google Scholar] [CrossRef]
  12. Mitropoulos, D.; Louridas, P.; Polychronakis, M.; Keromytis, A.D. Defending against web application attacks: Approaches, challenges and implications. IEEE Trans. Dependable Secur. Comput. 2017, 16, 188–203. [Google Scholar] [CrossRef]
  13. Mrabet, H.; Alhomoud, A.; Jemai, A.; Trentesaux, D. A secured industrial Internet-of-things architecture based on blockchain technology and machine learning for sensor access control systems in smart manufacturing. Appl. Sci. 2022, 12, 4641. [Google Scholar] [CrossRef]
  14. Chu, G.; Lisitsa, A. Penetration Testing for Internet of Things and Its Automation. In Proceedings of the 2018 IEEE 20th International Conference on High Performance Computing and Communications, IEEE 16th International Conference on Smart City, IEEE 4th International Conference on Data Science and Systems (HPCC/SmartCity/DSS), Exeter, UK, 28–30 June 2018; pp. 1479–1484. [Google Scholar]
  15. Lin, Z.; Shi, Y.; Xue, Z. Idsgan: Generative adversarial networks for attack generation against intrusion detection. arXiv 2018, arXiv:1809.02077. [Google Scholar]
  16. Revathi, S.; Malathi, A. A detailed analysis on NSL-KDD dataset using various machine learning techniques for intrusion detection. Int. J. Eng. Res. Technol. (IJERT) 2013, 2, 1848–1853. [Google Scholar]
  17. Jakkula, V. Tutorial on Support Vector Machine (SVM); School of EECS, Washington State University: Pullman, WA, USA, 2006; Volume 37. [Google Scholar]
  18. Rish, I. An empirical study of the naive Bayes classifier. In Proceedings of the IJCAI 2001 Workshop on Empirical Methods in Artificial Intelligence, Seattle, WA, USA, 4–6 August 2001; Volume 3, pp. 41–46. [Google Scholar]
  19. Noriega, L. Multilayer Perceptron Tutorial; School of Computing, Staffordshire University: Staffordshire, UK, 2005. [Google Scholar]
  20. Myles, A.J.; Feudale, R.N.; Liu, Y.; Woody, N.A.; Brown, S.D. An introduction to decision tree modeling. J. Chemom. A J. Chemom. Soc. 2004, 18, 275–285. [Google Scholar] [CrossRef]
  21. Dai, B.; Fidler, S.; Urtasun, R.; Lin, D. Towards diverse and natural image descriptions via a conditional gan. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2970–2979. [Google Scholar]
  22. Marra, F.; Gragnaniello, D.; Cozzolino, D.; Verdoliva, L. Detection of gan-generated fake images over social networks. In Proceedings of the 2018 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), Miami, FL, USA, 10–12 April 2018; pp. 384–389. [Google Scholar]
  23. Yu, L.; Zhang, W.; Wang, J.; Yu, Y. Seqgan: Sequence generative adversarial nets with policy gradient. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  24. Shahriar, H.; Haddad, H. Risk assessment of code injection vulnerabilities using fuzzy logic-based system. In Proceedings of the 29th Annual ACM Symposium on Applied Computing, San Francisco, CA, USA, 4–9 February 2014; pp. 1164–1170. [Google Scholar]
  25. Shibata, Y.; Kida, T.; Fukamachi, S.; Takeda, M.; Shinohara, A.; Shinohara, T.; Arikawa, S. Byte Pair Encoding: A Text Compression Scheme that Accelerates Pattern Matching; Kyushu University: Fukuoka, Japan, 1999. [Google Scholar]
  26. Singh, J.J.; Samuel, H.; Zavarsky, P. Impact of paranoia levels on the effectiveness of the modsecurity web application firewall. In Proceedings of the 2018 1st International Conference on Data Intelligence and Security (ICDIS), South Padre Island, TX, USA, 8–10 April 2018; pp. 141–144. [Google Scholar]
  27. Singh, H. Security in Amazon Web Services. In Practical Machine Learning with AWS; Springer: Berlin/Heidelberg, Germany, 2021; pp. 45–62. [Google Scholar]
  28. Alsaffar, M.; Aljaloud, S.; Mohammed, B.A.; Al-Mekhlafi, Z.G.; Almurayziq, T.S.; Alshammari, G.; Alshammari, A. Detection of Web Cross-Site Scripting (XSS) Attacks. Electronics 2022, 11, 2212. [Google Scholar] [CrossRef]
  29. Obes, J.L.; Sarraute, C.; Richarte, G. Attack planning in the real world. arXiv 2013, arXiv:1306.4044. [Google Scholar]
  30. Sarraute, C.; Buffet, O.; Hoffmann, J. Penetration testing== pomdp solving? arXiv 2013, arXiv:1306.4714. [Google Scholar]
  31. Schwartz, J.; Kurniawati, H. Autonomous penetration testing using reinforcement learning. arXiv 2019, arXiv:1905.05965. [Google Scholar]
  32. Ghanem, M.C.; Chen, T.M. Reinforcement learning for efficient network penetration testing. Information 2020, 11, 6. [Google Scholar] [CrossRef]
  33. Schwartz, J.; Kurniawati, H.; El-Mahassni, E. Pomdp + information-decay: Incorporating defender’s behaviour in autonomous penetration testing. In Proceedings of the International Conference on Automated Planning and Scheduling, Nancy, France, 14–19 June 2020; Volume 30, pp. 235–243. [Google Scholar]
  34. Tran, K.; Standen, M.; Kim, J.; Bowman, D.; Richer, T.; Akella, A.; Lin, C.T. Cascaded reinforcement learning agents for large action spaces in autonomous penetration testing. Appl. Sci. 2022, 12, 11265. [Google Scholar] [CrossRef]
  35. Zhou, S.; Liu, J.; Hou, D.; Zhong, X.; Zhang, Y. Autonomous penetration testing based on improved deep q-network. Appl. Sci. 2021, 11, 8823. [Google Scholar] [CrossRef]
  36. Hitaj, B.; Gasti, P.; Ateniese, G.; Perez-Cruz, F. Passgan: A deep learning approach for password guessing. In Proceedings of the International Conference on Applied Cryptography and Network Security, Bogota, Colombia, 5–7 June 2019; Springer: Cham, Switzerland, 2019; pp. 217–237. [Google Scholar]
  37. Yang, J.; Li, T.; Liang, G.; He, W.; Zhao, Y. A simple recurrent unit model based intrusion detection system with DCGAN. IEEE Access 2019, 7, 83286–83296. [Google Scholar] [CrossRef]
  38. Zhang, X.; Zhou, Y.; Pei, S.; Zhuge, J.; Chen, J. Adversarial examples detection for XSS attacks based on generative adversarial networks. IEEE Access 2020, 8, 10989–10996. [Google Scholar] [CrossRef]
  39. Sengupta, S.; Chowdhary, A.; Huang, D.; Kambhampati, S. General sum markov games for strategic detection of advanced persistent threats using moving target defense in cloud networks. In Proceedings of the Decision and Game Theory for Security: 10th International Conference, GameSec 2019, Stockholm, Sweden, 30 October–1 November 2019; Proceedings 10. Springer: Cham, Switzerland, 2019; pp. 492–512. [Google Scholar]
  40. Myneni, S.; Chowdhary, A.; Sabur, A.; Sengupta, S.; Agrawal, G.; Huang, D.; Kang, M. DAPT 2020-constructing a benchmark dataset for advanced persistent threats. In Proceedings of the Deployable Machine Learning for Security Defense: First International Workshop, MLHat 2020, San Diego, CA, USA, 24 August 2020; Proceedings 1. Springer: Cham, Switzerland, 2020; pp. 138–163. [Google Scholar]
  41. Myneni, S.; Jha, K.; Sabur, A.; Agrawal, G.; Deng, Y.; Chowdhary, A.; Huang, D. Unraveled—A semi-synthetic dataset for Advanced Persistent Threats. Comput. Netw. 2023, 227, 109688. [Google Scholar] [CrossRef]
  42. Scarfone, K.; Mell, P. An analysis of CVSS version 2 vulnerability scoring. In Proceedings of the 2009 3rd International Symposium on Empirical Software Engineering and Measurement, Lake Buena Vista, FL, USA, 15–16 October 2009; pp. 516–525. [Google Scholar]
  43. Bilge, L.; Dumitraş, T. Before we knew it: An empirical study of zero-day attacks in the real world. In Proceedings of the 2012 ACM Conference on Computer and Communications Security, Raleigh, NC, USA, 16–18 October 2012; pp. 833–844. [Google Scholar]
  44. Stuttard, D.; Pinto, M. The Web Application Hacker’s Handbook: Finding and Exploiting Security Flaws; John Wiley & Sons: Hoboken, NJ, USA, 2011. [Google Scholar]
  45. Prandl, S.; Lazarescu, M.; Pham, D.S. A study of web application firewall solutions. In Proceedings of the International Conference on Information Systems Security, Kolkata, India, 16–20 December 2015; pp. 501–510. [Google Scholar]
  46. Security Intelligence. Generative Adversarial Networks and Cybersecurity. Available online: https://securityintelligence.com/generative-adversarial-networks-and-cybersecurity-part-1/ (accessed on 2 July 2021).
  47. Hochreiter, S. The vanishing gradient problem during learning recurrent neural nets and problem solutions. Int. J. Uncertain. Fuzziness Knowl. Based Syst. 1998, 6, 107–116. [Google Scholar] [CrossRef]
  48. Chen, H.; Jiang, L. Efficient GAN-based method for cyber-intrusion detection. arXiv 2019, arXiv:1904.02426. [Google Scholar]
  49. Mahajan, A. Burp Suite Essentials; Packt Publishing Ltd.: Birmingham, UK, 2014. [Google Scholar]
  50. Williams, R.J. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Mach. Learn. 1992, 8, 229–256. [Google Scholar] [CrossRef]
  51. Ismail Tasdelen. Payload Box. Available online: https://github.com/payloadbox/xss-payload-list (accessed on 8 October 2021).
  52. AWS. AWS WAF–Web Application Firewall. Available online: https://aws.amazon.com/waf/ (accessed on 8 November 2021).
  53. Wang, K.; Wan, X. SentiGAN: Generating Sentimental Texts via Mixture Adversarial Networks. In Proceedings of the IJCAI, Stockholm, Sweden, 13–19 July 2018; pp. 4446–4452. [Google Scholar]
  54. Liu, Z.; Wang, J.; Liang, Z. Catgan: Category-aware generative adversarial networks with hierarchical evolutionary learning for category text generation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 8425–8432. [Google Scholar]
Figure 1. Cross-Site Scripting (XSS) vulnerability present in an application exploited by a remote attacker.
Figure 1. Cross-Site Scripting (XSS) vulnerability present in an application exploited by a remote attacker.
Sensors 23 08014 g001
Figure 2. Expressions blocked by WAFs and corresponding bypass techniques.
Figure 2. Expressions blocked by WAFs and corresponding bypass techniques.
Sensors 23 08014 g002
Figure 3. Example of conditional sequences generated from semantic tokens.
Figure 3. Example of conditional sequences generated from semantic tokens.
Sensors 23 08014 g003
Figure 4. GAN-based approach for generating attack payloads that bypass web application firewall (WAF) filters.
Figure 4. GAN-based approach for generating attack payloads that bypass web application firewall (WAF) filters.
Sensors 23 08014 g004
Figure 5. Conditional Attack Sequence Generation by semantic tokenization and attack payload validation.
Figure 5. Conditional Attack Sequence Generation by semantic tokenization and attack payload validation.
Sensors 23 08014 g005
Figure 6. Conditional Attack Sequence Generation.
Figure 6. Conditional Attack Sequence Generation.
Sensors 23 08014 g006
Figure 7. The result of GAN training loss for 250 epochs.
Figure 7. The result of GAN training loss for 250 epochs.
Sensors 23 08014 g007
Table 1. Number of successful WAF bypasses, using several variants of GAN.
Table 1. Number of successful WAF bypasses, using several variants of GAN.
Run #Vanilla GANCGAN
110.37%7.66%
27.69%8.04%
317.64%9.08%
40.08%12%
516.19%8.28%
Table 2. GAN-generated XSS attack payloads against AWS WAF.
Table 2. GAN-generated XSS attack payloads against AWS WAF.
Matching Rule% Attack MatchAWS WAF Action
WAF Bypass8.0%ALLOW
AWS-managed XSS44.9%BLOCK
Fortinet XSS Rule44.0%BLOCK
Misclassified3.1%BLOCK
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chowdhary, A.; Jha, K.; Zhao, M. Generative Adversarial Network (GAN)-Based Autonomous Penetration Testing for Web Applications. Sensors 2023, 23, 8014. https://doi.org/10.3390/s23188014

AMA Style

Chowdhary A, Jha K, Zhao M. Generative Adversarial Network (GAN)-Based Autonomous Penetration Testing for Web Applications. Sensors. 2023; 23(18):8014. https://doi.org/10.3390/s23188014

Chicago/Turabian Style

Chowdhary, Ankur, Kritshekhar Jha, and Ming Zhao. 2023. "Generative Adversarial Network (GAN)-Based Autonomous Penetration Testing for Web Applications" Sensors 23, no. 18: 8014. https://doi.org/10.3390/s23188014

APA Style

Chowdhary, A., Jha, K., & Zhao, M. (2023). Generative Adversarial Network (GAN)-Based Autonomous Penetration Testing for Web Applications. Sensors, 23(18), 8014. https://doi.org/10.3390/s23188014

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop