Share this post on:

Listed in Table 1. We’ll describe these evaluation indicators in detail.Appl. Sci. 2021, 11,7 ofFigure 5. BiLSTM framework. Table 1. Information of evaluation metrics. “Auto” and “Human” represent automatic and human evaluations respectively. “Higher” and “Lower” mean the higher/lower the metric, the better a model performs. Metrics Composite score Accomplishment Rate Word Freqency Grammaticality Fluency Naturality Evaluation System Auto Auto Auto Auto (Error Rate) Auto (Perplexity) Human (Naturality Score) Far better Larger Greater Larger Reduced Decrease Larger(1) The attack achievement price is defined as the percentage of samples incorrectly TP-064 Histone Methyltransferase predicted by the target model to the total variety of samples. In this experiment, these samples are all connected towards the universal trigger. The formula is defined as follows S= 1 Ni =( f (t, xi ) = yi ),N(six)where N is definitely the total number of samples, f represents the target model, t represents the universal trigger, xi represents the ith test sample, and yi represents the actual label of xi . (two) We divide it into 4 parts for the quality of triggers, like word 15(S)-15-Methyl Prostaglandin F2�� Epigenetic Reader Domain frequency [29], grammaticality, fluency, and naturality [23]. The average frequency from the words within the trigger is calculated utilizing empirical estimates in the education set of your target classifier.Appl. Sci. 2021, 11,eight ofThe higher the typical frequency of a word, the far more instances the word appears within the instruction set. Grammaticality is measured by adding triggers of your similar number of words to benign text, then making use of a web-based grammar check tool (Grammarly) to obtain the grammatical error price of your sentence. Using the enable of GPT-2 [14], we use Language Model Perplexity (PPL) to measure fluency. Naturalness reflects whether or not an adversarial example is organic and indistinguishable from human-written text. (3) We construct a composite score Q to comprehensively measure the efficiency of our attack strategy. The formula is defined as follows Q = + W – – (7)where S will be the attack good results rate in the trigger, W is definitely the average word frequency with the trigger, G will be the grammatical error price of the trigger, and P is definitely the perplexity from the GPT-2 [14]. W, G, P are all normalized. , , is definitely the coefficient of every parameter, and + + + = 1. In order to balance the weight of every parameter, we set , and to 0.25. The greater the Q score, the much better the attack efficiency. To additional verify that our attack is much more natural than the baseline, we performed a human evaluation study. We deliver 50 pairs of comparative texts. Every single team consists of one trigger and one baseline trigger (with or with no benign text). Workers are asked to opt for a far more natural one, and humans are allowed to choose an uncertain option. For every single instance, we collected 5 different human judgments and calculated the average score. four.four. Attack Results Table 2 shows the results of our attack and baseline [28]. We observe that our attack achieves the highest composite score Q on each of the two datasets, proving the superiority of our model over baselines. For each optimistic and damaging situations, our process has a higher attack results rate. It could be found that the success price of triggers on SST-2 or IMDB information has reached more than 50 . In addition, our system achieved the ideal attack effect around the Bi-LSTM model trained on the SST-2 data set, with a results rate of 80.1 . Comparing the models trained around the two data sets, the conclusion might be drawn: The Bi-LSTM model educated around the SST-2 information set.

Share this post on:

Author: calcimimeticagent