DemoNSF: A Multi-task Demonstration-based Generative Framework for Noisy Slot Filling Task

Guanting Dong1*1{}^{1*}start_FLOATSUPERSCRIPT 1 * end_FLOATSUPERSCRIPT, Tingfeng Hui1*1{}^{1*}start_FLOATSUPERSCRIPT 1 * end_FLOATSUPERSCRIPT, Zhuoma GongQue11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT, Jinxu Zhao11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT, Daichi Guo11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT,
Gang Zhao11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT, Keqing He22{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT, Weiran Xu11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT
11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT
Beijing University of Posts and Telecommunications, Beijing, China
22{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPTMeituan, Beijing, China
{dongguanting,huitingfeng,xuweiran}@bupt.edu.cn
  The first two authors contribute equally. Weiran Xu is the corresponding author.
Abstract

Recently, prompt-based generative frameworks have shown impressive capabilities in sequence labeling tasks. However, in practical dialogue scenarios, relying solely on simplistic templates and traditional corpora presents a challenge for these methods in generalizing to unknown input perturbations. To address this gap, we propose a multi-task demonstration-based generative framework for noisy slot filling, named DemoNSF. Specifically, we introduce three noisy auxiliary tasks, namely noisy recovery (NR), random mask (RM), and hybrid discrimination (HD), to implicitly capture semantic structural information of input perturbations at different granularities. In the downstream main task, we design a noisy demonstration construction strategy for the generative framework, which explicitly incorporates task-specific information and perturbed distribution during training and inference. Experiments on two benchmarks demonstrate that DemoNSF outperforms all baseline methods and achieves strong generalization. Further analysis provides empirical guidance for the practical application of generative frameworks. Our code is released at https://github.com/dongguanting/Demo-NSF.

1 Introduction

The slot filling (SF) task in the goal-oriented dialog system aims to identify task-related slot types in certain domains for understanding user utterances. Recently, traditional discriminative and generative models Liu and Lane (2015, 2016); Goo et al. (2018); Niu et al. (2019); He et al. (2020a); Yan et al. (2021a); Wang et al. (2022b); Hao et al. (2023) have shown remarkable ability in slot filling. Despite their powerful capabilities, the high performance of these models heavily depends on the consistency of data distribution between the training and test sets. When faced with the uncertainty and diversity of human language expression Wu et al. (2021), these perturbations significantly impact the SF model’s generalization ability, thereby hindering its application in practical dialogue scenarios.


Refer to caption
Figure 1: The impact of diverse input perturbations on the slot filling system in real scenarios.

In real dialogue systems, models often encounter a wide range of input perturbations and errors made by humans. As illustrated in Figure 1, users may interact with the dialogue system in ways that deviate from the standard input format and even simplify their queries to convey the same intent, all due to diverse human expression habits. Furthermore, errors originating from the upstream input system may introduce disturbances to the downstream model (e.g. Typos from keyboard input, Speech errors from ASR systems). Existing slot filling models are typically pre-trained and fine-tuned on perturbation-free datasets, leading to decreased performance when confronted with such situations.

Recently, existing studies Wu et al. (2021); Moradi and Samwald (2021a); Gui et al. (2021) have explored the issue of robustness. However, these methods are mainly designed for particular perturbations, limiting generalization ability to unknown perturbations. To capture the noisy semantic structure, PSSAT and CMDA Dong et al. (2022a); Guo et al. (2023) further introduce additional corpus and generative models. Nevertheless, this approach carries the risk of introducing extra noise and increasing computing resource consumption. While Large language models Brown et al. (2020); Touvron et al. (2023); OpenAI (2023) and prompt-based methods Lu et al. (2022); Xie et al. (2022b) have achieved strong performance on information extraction, the exploration of these generative frameworks on diverse input perturbations remains a blank area, hindering their application in realistic task-oriented dialogue systems.

To address this limitation, we propose a multi-task demonstration-based generative framework for noisy slot filling tasks, named DemoNSF. Specifically, we design three noisy auxiliary tasks, namely noisy recovery (NR), random mask (RM), and hybrid discrimination (HD), to boost the performance against input perturbations in different levels. NR aims to capture the mapping relationship between fine-grained noisy and clean data. RM implicitly learns the slot entity distribution of perturbed data during the process of mask infilling. HD assists generative models consider global information while implicitly capturing the semantic characteristics unique to perturbed data. In the downstream process, we formulate the SF task as a sequence-to-sequence generation guided by noisy task demonstrations. In detail, DemoNSF selects a semantically similar example for each query from a noisy candidate pool, converts it into a natural demonstration sentence, and encodes the demonstration along with the input text by integrating noisy semantic information. With the boost of noisy auxiliary tasks and demonstrations, DemoNSF learns the semantic structure of perturbations from both explicit and implicit levels. Our contributions are three-fold:

1) To the best of our knowledge, we are the first to comprehensively investigate the effects of diverse input perturbations on generative frameworks in slot filling tasks and further validate the vulnerability of existing prompt-based generative methods when confronted with different human expressions.

2) We propose a simple but unified multi-task demonstration-based generative framework, which includes three novel noisy auxiliary tasks and a noisy demonstration construction strategy, to enhance the model’s robustness and adaptability to perturbed inputs in real-world dialogue scenarios.

3) Experiments on two benchmarks demonstrate that our method outperforms all baseline methods and achieves strong generalization. The extensive analysis also provides empirical guidance for the practical application of generative frameworks.

Refer to caption
Figure 2: The overall architecture of our proposed approach DemoNSF.

2 Method

In this section, we introduce the overall framework of our proposed DemoNSF. We first briefly describe the problem definition against input perturbations in the slot filling task. Next, we propose three distinctive noisy auxiliary tasks for diverse perturbations. Finally, we present a novel noisy demonstration construction strategy. We will introduce these in the following subsections111Training, and Inference can be found in Appendix.

2.1 Problem Definition

Given an input utterance X={x1,x2,,xN}?subscript?1subscript?2subscript??X=\{x_{1},x_{2},\dots,x_{N}\}italic_X = { italic_x start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , italic_x start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT , … , italic_x start_POSTSUBSCRIPT italic_N end_POSTSUBSCRIPT } and its corresponding slot type set S={s1,,sm}?subscript?1subscript??S=\{s_{1},...,s_{m}\}italic_S = { italic_s start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , … , italic_s start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT }, the slot filling task aims to extract all the entities in X?Xitalic_X. For the noisy slot filling task, we formulate the input perturbation process in the real scenario as [(X,S)=?(X,S)]delimited-[]superscript?superscript????[(X^{\prime},S^{\prime})=\mathcal{P}(X,S)][ ( italic_X start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT , italic_S start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT ) = caligraphic_P ( italic_X , italic_S ) ], The model’s robustness is evaluated on the perturbed test dataset {(X,Y)}superscript?superscript?\{(X^{\prime},Y^{\prime})\}{ ( italic_X start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT , italic_Y start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT ) } but with no access to the input perturbation process ?()?\mathcal{P}(\cdot)caligraphic_P ( ⋅ ) or perturbed data during the training phase. In this paper, We use Dcleansubscript??????D_{clean}italic_D start_POSTSUBSCRIPT italic_c italic_l italic_e italic_a italic_n end_POSTSUBSCRIPT, Daugsubscript????D_{aug}italic_D start_POSTSUBSCRIPT italic_a italic_u italic_g end_POSTSUBSCRIPT, and Dtestsubscript?????D_{test}italic_D start_POSTSUBSCRIPT italic_t italic_e italic_s italic_t end_POSTSUBSCRIPT to denote clean data, augmented data, and test data.

2.2 Multi-level Data Augmentation

Figure 2 demonstrates how we construct our noisy candidate pool using NLPAug Ma (2019), enabling the input utterance of the clean training set into an augmented dataset comprising three distinct levels: character-level, word-level, and sentence-level augmentation. Specifically, at the character level, we incorporate random operations such as character addition, deletion, and substitution within a token, governed by a probability parameter denoted as p?pitalic_p. Moving to the word level, we introduce random word deletion, insertion, and replacement, along with the substitution of words with homophones within a sentence, again governed by the probability parameter p?pitalic_p222p?pitalic_p is an empirical parameter, we set it to 0.3. Furthermore, at the sentence level, we substitute sentences with synonymous alternatives.

2.3 Noisy Auxiliary Tasks

The performance of the noisy slot filling task highly depends on the prior knowledge of the distribution of input perturbations. In this section, we introduce three novel noisy auxiliary tasks:

Noisy Recovery (NR). Given a character-level augmented utterance Xcharaug={x1,x2,,xmaug,,xN}subscriptsuperscript???????subscript?1subscript?2subscriptsuperscript?????subscript??X^{aug}_{char}=\{x_{1},x_{2},...,x^{aug}_{m},\\ ...,x_{N}\}italic_X start_POSTSUPERSCRIPT italic_a italic_u italic_g end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_c italic_h italic_a italic_r end_POSTSUBSCRIPT = { italic_x start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , italic_x start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT , … , italic_x start_POSTSUPERSCRIPT italic_a italic_u italic_g end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT , … , italic_x start_POSTSUBSCRIPT italic_N end_POSTSUBSCRIPT }, where xmaugsubscriptsuperscript?????x^{aug}_{m}italic_x start_POSTSUPERSCRIPT italic_a italic_u italic_g end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT represents the augmented token with character-level augmentation, the objective of the NR task, as illustrated in Figure 2, is to restore Xcharaugsubscriptsuperscript???????X^{aug}_{char}italic_X start_POSTSUPERSCRIPT italic_a italic_u italic_g end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_c italic_h italic_a italic_r end_POSTSUBSCRIPT to its corresponding clean utterance X?Xitalic_X. This task enables the model to capture the mapping relationship between fine-grained input perturbations and their clean equivalents, thereby enhancing the model’s capability to represent fine-grained noisy data. Hence, the loss function can be formulated as:

LNR=1Bj=1Bi=1NCE(Xji,Xchar,jiaug)subscript???1?superscriptsubscript?1?superscriptsubscript?1???subscript???subscriptsuperscript?????????L_{NR}=\frac{1}{B}\sum\limits_{j=1}^{B}\sum\limits_{i=1}^{N}CE(X_{ji},X^{aug}_{char,ji})italic_L start_POSTSUBSCRIPT italic_N italic_R end_POSTSUBSCRIPT = divide start_ARG 1 end_ARG start_ARG italic_B end_ARG ∑ start_POSTSUBSCRIPT italic_j = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_B end_POSTSUPERSCRIPT ∑ start_POSTSUBSCRIPT italic_i = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_N end_POSTSUPERSCRIPT italic_C italic_E ( italic_X start_POSTSUBSCRIPT italic_j italic_i end_POSTSUBSCRIPT , italic_X start_POSTSUPERSCRIPT italic_a italic_u italic_g end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_c italic_h italic_a italic_r , italic_j italic_i end_POSTSUBSCRIPT ) (1)

where B and N denote the batch size and sequence length, respectively.

Random Mask (RM). Inspired by the concept of masked language modeling (MLM) introduced in BERT Devlin et al. (2019), we present our random mask-filling task. Specifically, given an utterance in Daugsubscript????D_{aug}italic_D start_POSTSUBSCRIPT italic_a italic_u italic_g end_POSTSUBSCRIPT, we randomly mask one entity with the special [MASK] symbol, resulting in Xmaskaug={x1aug,,[MASK],,xNaug}subscriptsuperscript????????subscriptsuperscript????1[MASK]subscriptsuperscript?????X^{aug}_{mask}=\{x^{aug}_{1},...,\emph{[MASK]},...,x^{aug}_{N}\}italic_X start_POSTSUPERSCRIPT italic_a italic_u italic_g end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_m italic_a italic_s italic_k end_POSTSUBSCRIPT = { italic_x start_POSTSUPERSCRIPT italic_a italic_u italic_g end_POSTSUPERSCRIPT start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , … , [MASK] , … , italic_x start_POSTSUPERSCRIPT italic_a italic_u italic_g end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_N end_POSTSUBSCRIPT }. We aim to restore the [MASK] token to its original value. The mask-filling procedure enables the model to implicitly incorporate the semantic distribution of slot entities within the perturbed data. Hence, the loss function of the RM task can be defined as:

LRM=1Bj=1BCE(ym,P([MASK]))subscript???1?superscriptsubscript?1???subscript???[MASK]L_{RM}=\frac{1}{B}\sum\limits_{j=1}^{B}CE(y_{m},P(\emph{[MASK]}))italic_L start_POSTSUBSCRIPT italic_R italic_M end_POSTSUBSCRIPT = divide start_ARG 1 end_ARG start_ARG italic_B end_ARG ∑ start_POSTSUBSCRIPT italic_j = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_B end_POSTSUPERSCRIPT italic_C italic_E ( italic_y start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT , italic_P ( [MASK] ) ) (2)

where ymsubscript??y_{m}italic_y start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT denotes the original token and [MASK] represents the logits of [MASK] token.

Hybrid Discrimination (HD). To further address coarse-grained input perturbations, we propose the HD task. In detail, we combine Dcleansubscript??????D_{clean}italic_D start_POSTSUBSCRIPT italic_c italic_l italic_e italic_a italic_n end_POSTSUBSCRIPT and Daugsubscript????D_{aug}italic_D start_POSTSUBSCRIPT italic_a italic_u italic_g end_POSTSUBSCRIPT to create a mixed dataset, denoted as Dmixsubscript????D_{mix}italic_D start_POSTSUBSCRIPT italic_m italic_i italic_x end_POSTSUBSCRIPT. We randomly select utterances from Dmixsubscript????D_{mix}italic_D start_POSTSUBSCRIPT italic_m italic_i italic_x end_POSTSUBSCRIPT and assign distinct labels based on whether the chosen utterance is clean or has different levels of perturbation. As shown in Figure 2, the generative model can implicitly capture the unique semantic distribution of perturbed data while considering global information by discriminating between inputs with and without perturbation. The loss function LHDsubscript???L_{HD}italic_L start_POSTSUBSCRIPT italic_H italic_D end_POSTSUBSCRIPT is the same as LNRsubscript???L_{NR}italic_L start_POSTSUBSCRIPT italic_N italic_R end_POSTSUBSCRIPT.

Therefore, the overall loss function L?Litalic_L is defined as:

L=αLNR+βLRM+γLHD??subscript????subscript????subscript???L=\alpha L_{NR}+\beta L_{RM}+\gamma L_{HD}italic_L = italic_α italic_L start_POSTSUBSCRIPT italic_N italic_R end_POSTSUBSCRIPT + italic_β italic_L start_POSTSUBSCRIPT italic_R italic_M end_POSTSUBSCRIPT + italic_γ italic_L start_POSTSUBSCRIPT italic_H italic_D end_POSTSUBSCRIPT (3)

where α?\alphaitalic_α, β?\betaitalic_β, and γ?\gammaitalic_γ represent the weights of NR, RM, and HD task loss functions, respectively.

Methods Clean Sentence-level Character-level Word-level Perturbed-Avg.
Verbose Paraphrase Simplification Typos Speech
GPT2 95.37 80.52 85.66 82.98 60.19 77.78 77.43
BART 95.28 77.87 82.72 82.95 53.90 73.57 74.20
T5 95.49 81.34 89.13 83.73 62.43 81.13 79.55
BARTNER 94.88 78.00 88.55 85.04 65.37 72.65 77.93
LightNER 95.30 78.85 87.65 84.90 57.68 71.61 76.14
InstructionNER 95.67 81.57 88.45 85.29 65.34 80.13 80.16
DemoNSF(GPT2) 95.66(±plus-or-minus\pm±0.4) 81.95(±plus-or-minus\pm±0.3) 87.63(±plus-or-minus\pm±1.1) 87.02(±plus-or-minus\pm±0.2) 69.75(±plus-or-minus\pm±0.3) 86.31(±plus-or-minus\pm±0.7) 82.53(±plus-or-minus\pm±0.5)
DemoNSF(BART) 95.71(±plus-or-minus\pm±1.3) 78.83(±plus-or-minus\pm±0.7) 88.29(±plus-or-minus\pm±0.8) 86.01(±plus-or-minus\pm±0.7) 65.60(±plus-or-minus\pm±0.3) 82.48(±plus-or-minus\pm±0.4) 80.24(±plus-or-minus\pm±1.1)
DemoNSF(T5) 95.72(±plus-or-minus\pm±0.5) 82.37(±plus-or-minus\pm±1.2) 89.98(±plus-or-minus\pm±1.1) 89.49(±plus-or-minus\pm±0.7) 76.63(±plus-or-minus\pm±0.5) 87.55(±plus-or-minus\pm±0.7) 85.20(±plus-or-minus\pm±0.9)
Table 1: F1 scores with standard deviations under 5 different input perturbations on RADDLE.
Methods Char+Word Char+Sen Word+Sent Char+Word+Sen
Ent.+Sub. Ent.+App. App.+Sub. Ent.+Sub.+App.
BART 58.00 47.27 50.28 38.36
T5 57.44 56.79 73.47 47.93
BARTNER 54.83 49.10 58.92 42.25
LightNER 42.34 35.82 45.44 27.00
InstructionNER 57.87 58.89 74.45 50.75
Ours(BART) 61.27(±plus-or-minus\pm±0.5) 51.26(±plus-or-minus\pm±0.9) 68.72(±plus-or-minus\pm±0.5) 44.27(±plus-or-minus\pm±0.9)
Ours(T5) 63.59(±plus-or-minus\pm±0.3) 63.94(±plus-or-minus\pm±1.2) 77.69(±plus-or-minus\pm±0.7) 55.12(±plus-or-minus\pm±0.3)
Table 2: F1 scores with standard deviations under 4 kinds of mixed perturbations on SNIPS.

2.4 Noisy Demonstration Construction

Different from prior demonstration-based work Min et al. (2022), we select examples s?sitalic_s from Daugsubscript????D_{aug}italic_D start_POSTSUBSCRIPT italic_a italic_u italic_g end_POSTSUBSCRIPT instead of Dcleansubscript??????D_{clean}italic_D start_POSTSUBSCRIPT italic_c italic_l italic_e italic_a italic_n end_POSTSUBSCRIPT for each input X?Xitalic_X to incorporate perturbed semantic information into the model. For retrieval, we employ SBERT Reimers and Gurevych (2019) which independently produces [CLS]delimited-[]???[CLS][ italic_C italic_L italic_S ] embeddings for both X?Xitalic_X and s?sitalic_s, and compute their similarity scores to rank s?sitalic_s. Subsequently, we select the top-k?kitalic_k examples to construct the noisy demonstrations X^^?\hat{X}over^ start_ARG italic_X end_ARG and concatenate them with the input X?Xitalic_X to form the complete input [X^;X]^??[\hat{X};X][ over^ start_ARG italic_X end_ARG ; italic_X ]. Our demonstration template is shown below:

"Demonstrations: [Retrieved Noisy Utterances]. [Text Span] is [Slot Type]. Input Utterance: [Original Input]."

3 Experiment

3.1 Datasets

Based on RADDLE Peng et al. (2021) and SNIPS Coucke et al. (2018), we adopt the evaluation set provided by Dong et al., which includes two different perturbation settings. For single perturbations setting, we include five types of noisy utterances (character-level: Typos, word-level: Speech, and sentence-level: Simplification, Verbose, and Paraphrase) from RADDLE. For mixed perturbations setting, we utilize TextFlint Gui et al. (2021) to introduce character-level perturbation (EntTypos), word-level perturbation (Subword), and sentence-level perturbation (AppendIrr) and combine them to get a mixed perturbations dataset333Due to space limitations, detailed experimental settings (Baselines, Datasets..) can be found in the Appendix B.1..

Methods Clean Sentence-level Character-level Word-level Perturbed-Avg.
Verbose Paraphrase Simplification Typos Speech
Text-davinci-003 43.09 34.26 39.34 38.42 40.12 37.18 38.54
ChatGPT 71.43 40.65 60.00 55.56 65.54 55.56 57.21
ChatGPT + Clean Demos 71.31 61.01 57.81 53.43 65.03 61.71 62.32
ChatGPT + Aug Demos 68.21 65.04 70.56 58.82 73.03 63.77 68.34
ChatGPT + Mixed Demos 76.92 58.73 68.26 58.61 74.19 57.78 65.36
Table 3: The evaluation on ChatGPT under 5 different input perturbations on Raddle.

3.2 Main Results.

Table 1 shows the main results of DemoNSF and comparison baselines under a single perturbation setting. We make the following observations:

(1) When faced with perturbations of different granularities, generative models suffer severe performance degradation, especially in Typos (GPT2: 35.18% | BART: 41.38% | T5: 33.06%) and Speech (GPT2: 17.57% | BART: 21.71% | T5: 14.36%), which indicates that generative models have poor robustness against fine-grained perturbations.

(2) DemoNSF(T5) shows remarkable superiority under different kinds of input perturbations while maintaining the best performance on the clean test data. For the fine-grained perturbations , our method achieves a significant improvement of 11.29% and 7.42% in Typos and Speech compared with InstructionNER. DemonNSF maintains strong performance for coarse-grained perturbations, especially with a 4.2% improvement in Simplification. These results clearly demonstrate that DemonNSF effectively captures the mapping relationship between fine-grained noisy and clean data while also considering the ability to generalize to coarse-grained global semantic perturbations.

(3) DemonNSF is a plug-and-play method that can achieve good results on different backbones. Specifically, we replace the backbone with BART/GPT and also get the good performances compared with the corresponding baseline. The results of the backbone ablation further demonstrate our approach remarkably enhances the robustness of generative models when facing perturbations.

Refer to caption
Figure 3: Performance comparison of different types of demonstrations

3.3 Mixed Perturbations Scenario.

In real dialogue scenarios, mixed perturbations often appear in one utterance simultaneously. To further verify the effectiveness of DemoNSF in more realistic scenarios, we conduct the mixed perturbations experiment. As shown in Table 2, DemoNSF significantly outperforms other baselines in all two-level perturbations, especially achieving over 63% F1-score in fine-grained mixed perturbations. Even with the joint interference of 3 perturbations, DemoNSF can still maintain a 4.37% improvement compared with baseline, which further validates the stability of DemoNSF in challenging settings.

3.4 Impact of Different Demonstrations.

Figure 3 shows the impact of the number of different types of demonstrations under single perturbations. We have the following findings: (1) DemoNSF exhibits a significant performance gain with only two augmented samples while its performance severely decreases as the number increases. This may be because diverse augmented instances can help the model explicitly fuse noisy semantic distribution Xie et al. (2022a) while the sample diversity exceeding a certain threshold may even bring additional noise. (2) Clean demonstrations only bring slightly improves as the number increases, which indicates that clean samples only provide some task general information(e.g. entity distributions, slot-value mapping) for prompting. (3) Retrieved demonstrations from the mixed data pool show a stable performance gain, which further confirms the mutual promotion between noisy semantic distribution and task general information, and provides guidance for the robustness of prompt-based generative models.

3.5 The ICL Evaluation on ChatGPT.

In order to further validate the effectiveness of our noisy demonstration strategy on the large-scale generative framework, we conduct experiments on ChatGPT and Text-davinci-003 Brown et al. (2020). We directly use them to do inference based on in-context learning (ICL) Dong et al. (2022b); Brown et al. (2020); Min et al. (2022) on RADDLE, which means language models make predictions only based on the conditioned demonstration examples without any training samples.

Table 3 illustrates the overall results of ChatGPT under 5 different single perturbations. We draw the following findings: (1) ChatGPT and Text-davinci-003 perform poorly on diverse input perturbations, which far behind the finetune SOTA methods (DemoNSF, Instruction NER) presented in Table 1. The possible reason is that large language models are usually pre-trained on the large-scale general training corpus, making it difficult to adapt well to specific domain perturbation data in a zero-shot setting. (2) Compared with baselines and traditional clean demonstration retrieval methods, selecting instances from both the augmented and mixed demonstration candidate pools can significantly improve the overall performance. This finding is consistent with our conclusion in the section 3.4, proving the effectiveness of incorporating noisy semantic structures in addressing input perturbations. (3) From the perspective of different perturbations, both two types of noisy demonstration strategies show significant improvements in fine-grained perturbations (over 8% improvement in Typos). However, the improvement is not obvious in coarse-grained perturbations, especially in speech errors. This phenomenon indicates that noisy demonstrations are more suitable for fitting the distribution of fine-grained perturbations, while there is still much room for improvement in coarse-grained perturbations that severely disrupt the contextual semantics and slot mentions of the original input. This finding poses further challenges for exploring the robustness of large language models, which will also be the focus of our future research.

4 Conclusion

In this paper, we propose a unified multi-task demonstration-based generative framework for noisy slot filling tasks. Specifically, we introduce three novel noisy auxiliary tasks and a noisy demonstration construction strategy for the generative framework, which aims to learn the semantic structure of perturbations from both explicit and implicit levels. Experiments on two benchmarks show the effectiveness of DemoNSF, Further analysis provides empirical guidance for the practical application of the generative framework.

Limitations

In order to capture semantic structural information of input perturbations at different granularities, we introduce three novel noisy auxiliary tasks in the pre-training stage, which may consume more GPU memory than traditional methods. This drives us to further improve the overall memory efficiency of the framework. Also, our method mainly focuses on the slot filling task. However, we believe it is possible to extend our work to other scenarios, such as few-shot settings, and zero-shot settings. We also reserve them for our future research.

References

  • Bapna et al. (2017) Ankur Bapna, Gokhan Tur, Dilek Hakkani-Tur, and Larry Heck. 2017. Towards zero-shot frame semantic parsing for domain scaling. arXiv preprint arXiv:1707.02363.
  • Brown et al. (2020) Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. 2020. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901.
  • Chen et al. (2022) Xiang Chen, Lei Li, Shumin Deng, Chuanqi Tan, Changliang Xu, Fei Huang, Luo Si, Huajun Chen, and Ningyu Zhang. 2022. Lightner: A lightweight tuning paradigm for low-resource ner via pluggable prompting.
  • Chen et al. (2021) Xiang Chen, Ningyu Zhang, Lei Li, Xin Xie, Shumin Deng, Chuanqi Tan, Fei Huang, Luo Si, and Huajun Chen. 2021. Lightner: A lightweight generative framework with prompt-guided attention for low-resource ner.
  • Coucke et al. (2018) Alice Coucke, Alaa Saade, Adrien Ball, Théodore Bluche, Alexandre Caulier, David Leroy, Clément Doumouro, Thibault Gisselbrecht, Francesco Caltagirone, Thibaut Lavril, Maël Primet, and Joseph Dureau. 2018. Snips voice platform: an embedded spoken language understanding system for private-by-design voice interfaces.
  • Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186.
  • Dong et al. (2022a) Guanting Dong, Daichi Guo, Liwen Wang, Xuefeng Li, Zechen Wang, Chen Zeng, Keqing He, Jinzheng Zhao, Hao Lei, Xinyue Cui, Yi Huang, Junlan Feng, and Weiran Xu. 2022a. PSSAT: A perturbed semantic structure awareness transferring method for perturbation-robust slot filling. In Proceedings of the 29th International Conference on Computational Linguistics, pages 5327–5334, Gyeongju, Republic of Korea. International Committee on Computational Linguistics.
  • Dong et al. (2023a) Guanting Dong, Zechen Wang, Liwen Wang, Daichi Guo, Dayuan Fu, Yuxiang Wu, Chen Zeng, Xuefeng Li, Tingfeng Hui, Keqing He, Xinyue Cui, Qixiang Gao, and Weiran Xu. 2023a. A prototypical semantic decoupling method via joint contrastive learning for few-shot named entity recognition. In ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1–5.
  • Dong et al. (2023b) Guanting Dong, Zechen Wang, Jinxu Zhao, Gang Zhao, Daichi Guo, Dayuan Fu, Tingfeng Hui, Chen Zeng, Keqing He, Xuefeng Li, Liwen Wang, Xinyue Cui, and Weiran Xu. 2023b. A multi-task semantic decomposition framework with task-specific pre-training for few-shot ner.
  • Dong et al. (2023c) Guanting Dong, Jinxu Zhao, Tingfeng Hui, Daichi Guo, Wenlong Wang, Boqi Feng, Yueyan Qiu, Zhuoma Gongque, Keqing He, Zechen Wang, and Weiran Xu. 2023c. Revisit input perturbation problems for llms: A unified robustness evaluation framework for noisy slot filling task. In Natural Language Processing and Chinese Computing, pages 682–694, Cham. Springer Nature Switzerland.
  • Dong et al. (2022b) Qingxiu Dong, Lei Li, Damai Dai, Ce Zheng, Zhiyong Wu, Baobao Chang, Xu Sun, Jingjing Xu, and Zhifang Sui. 2022b. A survey for in-context learning. arXiv preprint arXiv:2301.00234.
  • Fang et al. (2020) Anjie Fang, Simone Filice, Nut Limsopatham, and Oleg Rokhlenko. 2020. Using phoneme representations to build predictive models robust to asr errors. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 699–708.
  • Goo et al. (2018) Chih-Wen Goo, Guang Gao, Yun-Kai Hsu, Chih-Li Huo, Tsung-Chieh Chen, Keng-Wei Hsu, and Yun-Nung Chen. 2018. Slot-gated modeling for joint slot filling and intent prediction. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers), pages 753–757.
  • Gopalakrishnan et al. (2020) Karthik Gopalakrishnan, Behnam Hedayatnia, Longshaokan Wang, Yang Liu, and Dilek Hakkani-Tur. 2020. Are neural open-domain dialog systems robust to speech recognition errors in the dialog history? an empirical study.
  • Gui et al. (2021) Tao Gui, Xiao Wang, Qi Zhang, Qin Liu, Yicheng Zou, Xin Zhou, Rui Zheng, Chong Zhang, Qinzhuo Wu, Jiacheng Ye, Zexiong Pang, Yongxin Zhang, Zhengyan Li, Ruotian Ma, Zichu Fei, Ruijian Cai, Jun Zhao, Xingwu Hu, Zhiheng Yan, Yiding Tan, Yuan Hu, Qiyuan Bian, Zhihua Liu, Bolin Zhu, Shan Qin, Xiaoyu Xing, Jinlan Fu, Yue Zhang, Minlong Peng, Xiaoqing Zheng, Yaqian Zhou, Zhongyu Wei, Xipeng Qiu, and Xuanjing Huang. 2021. Textflint: Unified multilingual robustness evaluation toolkit for natural language processing.
  • Guo et al. (2023) Daichi Guo, Guanting Dong, Dayuan Fu, Yuxiang Wu, Chen Zeng, Tingfeng Hui, Liwen Wang, Xuefeng Li, Zechen Wang, Keqing He, Xinyue Cui, and Weiran Xu. 2023. Revisit out-of-vocabulary problem for slot filling: A unified contrastive frameword with multi-level data augmentations. ArXiv, abs/2302.13584.
  • Hao et al. (2023) Xia Hao, Lu Wang, Hongmei Zhu, and Xuchao Guo. 2023. Joint agricultural intent detection and slot filling based on enhanced heterogeneous attention mechanism. Computers and Electronics in Agriculture, 207:107756.
  • He et al. (2020a) Keqing He, Yuanmeng Yan, and Weiran Xu. 2020a. Learning to tag oov tokens by integrating contextual representation and background knowledge. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 619–624.
  • He et al. (2020b) Keqing He, Jinchao Zhang, Yuanmeng Yan, Weiran Xu, Cheng Niu, and Jie Zhou. 2020b. Contrastive zero-shot learning for cross-domain slot filling with adversarial attack. In Proceedings of the 28th International Conference on Computational Linguistics, pages 1461–1467, Barcelona, Spain (Online). International Committee on Computational Linguistics.
  • He et al. (2020c) Keqing He, Jinchao Zhang, Yuanmeng Yan, Weiran Xu, Cheng Niu, and Jie Zhou. 2020c. Contrastive zero-shot learning for cross-domain slot filling with adversarial attack. In Proceedings of the 28th International Conference on Computational Linguistics, pages 1461–1467.
  • Huang and Chen (2020) Chao-Wei Huang and Yun-Nung Chen. 2020. Learning asr-robust contextualized embeddings for spoken language understanding.
  • Lee et al. (2021) Dong-Ho Lee, Mahak Agarwal, Akshen Kadakia, Jay Pujara, and Xiang Ren. 2021. Good examples make a faster learner: Simple demonstration-based learning for low-resource ner. arXiv preprint arXiv:2110.08454.
  • Lee and Jha (2019) Sungjin Lee and Rahul Jha. 2019. Zero-shot adaptive transfer for conversational language understanding. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 6642–6649.
  • Lewis et al. (2019) Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov, and Luke Zettlemoyer. 2019. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension.
  • Li et al. (2020a) Mingda Li, Xinyue Liu, Weitong Ruan, Luca Soldaini, Wael Hamza, and Chengwei Su. 2020a. Multi-task learning of spoken language understanding by integrating n-best hypotheses with hierarchical attention. In Proceedings of the 28th International Conference on Computational Linguistics: Industry Track, pages 113–123.
  • Li et al. (2020b) Mingda Li, Weitong Ruan, Xinyue Liu, Luca Soldaini, Wael Hamza, and Chengwei Su. 2020b. Improving spoken language understanding by exploiting asr n-best hypotheses.
  • Liang et al. (2022) Xiaozhuan Liang, Ningyu Zhang, Siyuan Cheng, Zhen Bi, Zhenru Zhang, Chuanqi Tan, Songfang Huang, Fei Huang, and Huajun Chen. 2022. Contrastive demonstration tuning for pre-trained language models. arXiv preprint arXiv:2204.04392.
  • Liu and Lane (2015) Bing Liu and Ian Lane. 2015. Recurrent neural network structured output prediction for spoken language understanding. In Proc. NIPS Workshop on Machine Learning for Spoken Language Understanding and Interactions.
  • Liu and Lane (2016) Bing Liu and Ian R. Lane. 2016. Attention-based recurrent neural network models for joint intent detection and slot filling. In INTERSPEECH.
  • Liu et al. (2023) Jiachi Liu, Liwen Wang, Guanting Dong, Xiaoshuai Song, Zechen Wang, Zhengyang Wang, Shanglin Lei, Jinzheng Zhao, Keqing He, Bo Xiao, et al. 2023. Towards robust and generalizable training: An empirical study of noisy slot filling for input perturbations. arXiv preprint arXiv:2310.03518.
  • Liu et al. (2020a) Jiexi Liu, Ryuichi Takanobu, Jiaxin Wen, Dazhen Wan, Hongguang Li, Weiran Nie, Cheng Li, Wei Peng, and Minlie Huang. 2020a. Robustness testing of language understanding in task-oriented dialog. arXiv preprint arXiv:2012.15262.
  • Liu et al. (2020b) Zihan Liu, Genta Indra Winata, Peng Xu, and Pascale Fung. 2020b. Coach: A coarse-to-fine approach for cross-domain slot filling. arXiv preprint arXiv:2004.11727.
  • Lu et al. (2022) Yaojie Lu, Qing Liu, Dai Dai, Xinyan Xiao, Hongyu Lin, Xianpei Han, Le Sun, and Hua Wu. 2022. Unified structure generation for universal information extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 5755–5772, Dublin, Ireland. Association for Computational Linguistics.
  • Ma (2019) Edward Ma. 2019. Nlp augmentation. https://github.com/makcedward/nlpaug.
  • Ma et al. (2022) Tingting Ma, Huiqiang Jiang, Qianhui Wu, Tiejun Zhao, and Chin-Yew Lin. 2022. Decomposed meta-learning for few-shot named entity recognition. arXiv preprint arXiv:2204.05751.
  • Min et al. (2021) Sewon Min, Mike Lewis, Hannaneh Hajishirzi, and Luke Zettlemoyer. 2021. Noisy channel language model prompting for few-shot text classification. arXiv preprint arXiv:2108.04106.
  • Min et al. (2022) Sewon Min, Xinxi Lyu, Ari Holtzman, Mikel Artetxe, Mike Lewis, Hannaneh Hajishirzi, and Luke Zettlemoyer. 2022. Rethinking the role of demonstrations: What makes in-context learning work?
  • Moradi and Samwald (2021a) Milad Moradi and Matthias Samwald. 2021a. Evaluating the robustness of neural language models to input perturbations. arXiv preprint arXiv:2108.12237.
  • Moradi and Samwald (2021b) Milad Moradi and Matthias Samwald. 2021b. Evaluating the robustness of neural language models to input perturbations. Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing.
  • Namysl et al. (2020) Marcin Namysl, Sven Behnke, and Joachim Köhler. 2020. Nat: noise-aware training for robust neural sequence labeling. arXiv preprint arXiv:2005.07162.
  • Namysl et al. (2021) Marcin Namysl, Sven Behnke, and Joachim Köhler. 2021. Empirical error modeling improves robustness of noisy neural sequence labeling. Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021.
  • Niu et al. (2019) Peiqing Niu, Zhongfu Chen, Meina Song, et al. 2019. A novel bi-directional interrelated model for joint intent detection and slot filling. arXiv preprint arXiv:1907.00390.
  • OpenAI (2023) OpenAI. 2023. Gpt-4 technical report.
  • Peng et al. (2020) Baolin Peng, Chunyuan Li, Zhu Zhang, Chenguang Zhu, Jinchao Li, and Jianfeng Gao. 2020. Raddle: An evaluation benchmark and analysis platform for robust task-oriented dialog systems. arXiv preprint arXiv:2012.14666.
  • Peng et al. (2021) Baolin Peng, Chunyuan Li, Zhu Zhang, Chenguang Zhu, Jinchao Li, and Jianfeng Gao. 2021. RADDLE: An evaluation benchmark and analysis platform for robust task-oriented dialog systems. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 4418–4429, Online. Association for Computational Linguistics.
  • Radford et al. (2019) Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, et al. 2019. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9.
  • Raffel et al. (2020) Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer.
  • Reimers and Gurevych (2019) Nils Reimers and Iryna Gurevych. 2019. Sentence-bert: Sentence embeddings using siamese bert-networks.
  • Ruan et al. (2020) Weitong Ruan, Yaroslav Nechaev, Luoxin Chen, Chengwei Su, and Imre Kiss. 2020. Towards an asr error robust spoken language understanding system. In INTERSPEECH, pages 901–905.
  • Shah et al. (2019) Darsh J Shah, Raghav Gupta, Amir A Fayazi, and Dilek Hakkani-Tur. 2019. Robust zero-shot cross-domain slot filling with example values. arXiv preprint arXiv:1906.06870.
  • Touvron et al. (2023) Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, Aurelien Rodriguez, Armand Joulin, Edouard Grave, and Guillaume Lample. 2023. Llama: Open and efficient foundation language models.
  • Wang et al. (2022a) Jianing Wang, Chengyu Wang, Chuanqi Tan, Minghui Qiu, Songfang Huang, Jun Huang, and Ming Gao. 2022a. SpanProto: A two-stage span-based prototypical network for few-shot named entity recognition. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 3466–3476, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
  • Wang et al. (2022b) Liwen Wang, Rumei Li, Yang Yan, Yuanmeng Yan, Sirui Wang, Wei Wu, and Weiran Xu. 2022b. Instructionner: A multi-task instruction-based generative framework for few-shot ner.
  • Wang et al. (2021) Liwen Wang, Xuefeng Li, Jiachi Liu, Keqing He, Yuanmeng Yan, and Weiran Xu. 2021. Bridge to target domain by prototypical contrastive learning and label confusion: Re-explore zero-shot learning for slot filling. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 9474–9480, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
  • Wei et al. (2022) Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Ed Chi, Quoc Le, and Denny Zhou. 2022. Chain of thought prompting elicits reasoning in large language models. arXiv preprint arXiv:2201.11903.
  • Wu et al. (2021) Di Wu, Yiren Chen, Liang Ding, and Dacheng Tao. 2021. Bridging the gap between clean data training and real-world inference for spoken language understanding. arXiv preprint arXiv:2104.06393.
  • Xie et al. (2022a) Sang Michael Xie, Aditi Raghunathan, Percy Liang, and Tengyu Ma. 2022a. An explanation of in-context learning as implicit bayesian inference.
  • Xie et al. (2022b) Tianbao Xie, Chen Henry Wu, Peng Shi, Ruiqi Zhong, Torsten Scholak, Michihiro Yasunaga, Chien-Sheng Wu, Ming Zhong, Pengcheng Yin, Sida I. Wang, Victor Zhong, Bailin Wang, Chengzu Li, Connor Boyle, Ansong Ni, Ziyu Yao, Dragomir Radev, Caiming Xiong, Lingpeng Kong, Rui Zhang, Noah A. Smith, Luke Zettlemoyer, and Tao Yu. 2022b. UnifiedSKG: Unifying and multi-tasking structured knowledge grounding with text-to-text language models. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 602–631, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
  • Yan et al. (2021a) Hang Yan, Tao Gui, Junqi Dai, Qipeng Guo, Zheng Zhang, and Xipeng Qiu. 2021a. A unified generative framework for various NER subtasks. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 5808–5822, Online. Association for Computational Linguistics.
  • Yan et al. (2021b) Hang Yan, Tao Gui, Junqi Dai, Qipeng Guo, Zheng Zhang, and Xipeng Qiu. 2021b. A unified generative framework for various ner subtasks.
  • Zhao et al. (2021) Zihao Zhao, Eric Wallace, Shi Feng, Dan Klein, and Sameer Singh. 2021. Calibrate before use: Improving few-shot performance of language models. In International Conference on Machine Learning, pages 12697–12706. PMLR.

Appendix A Method Details

A.1 Training and Inference

For the multi-level data augmentation, we utilize NLPAug Ma (2019) to construct a noisy candidate data pool from the clean data pool at different levels. During the upstream pre-training stage, we adopt a multi-task training strategy, where the overall loss function is denoted as (13).

In the downstream demonstration-based fine-tuning stage, we directly feed the demonstration-based input [X^;X]^??[\hat{X};X][ over^ start_ARG italic_X end_ARG ; italic_X ] into the model for prediction. During inference, we also employ SBERT Reimers and Gurevych (2019) to retrieve task demonstrations for test data from Daugsubscript????D_{aug}italic_D start_POSTSUBSCRIPT italic_a italic_u italic_g end_POSTSUBSCRIPT, ensuring consistency between the training and inference stages.

Appendix B Experiment Details

B.1 More Details of Datasets

Based on RADDLE Peng et al. (2020) and SNIPS Coucke et al. (2018), we adhere to the evaluation set provided by PSSAT Dong et al. (2022a), which includes two settings: single perturbation and mixed perturbation.

For a single perturbation setting, RADDLE serves as a crowd-sourced diagnostic evaluation dataset that covers a wide range of real-world noisy texts for dialog systems. PSSATDong et al. (2022a) extracts each type of noisy utterance (Typos, Speech, Simplification, Verbose, and Paraphrase) from RADDLE to build the test data pool. Specifically, Typos occur due to non-standard abbreviations or keyboard errors, while Speech arises from recognition and synthesis errors produced by ASR systems. Simplification refers to users expressing their intentions using concise words, while Verbose represents users employing redundant words to convey the same intention. Paraphrase is also prevalent among users who use different words or rephrase the text based on their language habits.

For the multi-perturbations setting, we utilize TextFlint Gui et al. (2021) toolkit to introduce character-level noise (EntTypos), word-level noise (Subword), and sentence-level noise (AppendIrr). We then combine different types of noisy data to construct a multi-perturbations evaluation set. The detailed introduction of these perturbations can be found in the GitHub repository of TextFlint 444https://www.textflint.io/textflint.

Methods Clean Sentence-level Character-level Word-level Perturbed-Avg.
Verbose Paraphrase Simplification Typos Speech
DemoNSF(Backbone) 95.49 81.34 89.13 83.73 62.43 81.13 79.55
+NR??{}_{+NR}start_FLOATSUBSCRIPT + italic_N italic_R end_FLOATSUBSCRIPT 95.17 81.13 88.96 87.98 67.22 82.61 81.58
+RM??{}_{+RM}start_FLOATSUBSCRIPT + italic_R italic_M end_FLOATSUBSCRIPT 95.30 81.04 89.88 86.78 62.31 85.84 81.17
+HD??{}_{+HD}start_FLOATSUBSCRIPT + italic_H italic_D end_FLOATSUBSCRIPT 95.39 81.37 90.36 87.20 64.61 82.66 81.24
+MT??{}_{+MT}start_FLOATSUBSCRIPT + italic_M italic_T end_FLOATSUBSCRIPT 95.82 80.79 89.28 86.89 73.74 84.40 83.02
+CleanDemos??????????{}_{+CleanDemos}start_FLOATSUBSCRIPT + italic_C italic_l italic_e italic_a italic_n italic_D italic_e italic_m italic_o italic_s end_FLOATSUBSCRIPT 95.37 80.28 88.91 85.94 69.00 84.62 81.75
+MixDemos????????{}_{+MixDemos}start_FLOATSUBSCRIPT + italic_M italic_i italic_x italic_D italic_e italic_m italic_o italic_s end_FLOATSUBSCRIPT 95.90 80.04 90.71 87.19 71.23 82.60 82.35
+NoisyDemos??????????{}_{+NoisyDemos}start_FLOATSUBSCRIPT + italic_N italic_o italic_i italic_s italic_y italic_D italic_e italic_m italic_o italic_s end_FLOATSUBSCRIPT 95.08 80.04 90.34 85.39 76.41 85.14 83.46
DemoNSF(Full) 95.72 82.37 89.98 89.49 76.63 87.55 85.20
Table 4: The ablation study results (average F1 score%) on RADDLE. "+" denotes the backbone of DemoNSF with specific module.

B.2 Baselines

In this paper, we focus on comparing DemoNSF with multiple state-of-the-art baselines that use a generative framework, as shown below:
GPT2 Radford et al. (2019) is a decoder-only framework model developed by OpenAI 555https://openai.com/.
It is designed to generate human-like text by predicting the next word in a given sequence. GPT-2 has gained popularity for its impressive ability to generate coherent and contextually relevant text across various domains and has been used for tasks like text completion, translation, and creative writing.
BART is a sequence-to-sequence model architecture introduced by Lewis et al. (2019).
It combines both autoregressive and denoising objectives during training to learn robust representations of input sequences.
T5 is a pre-training model proposed by Raffel et al. (2020).
It utilizes the transformer architecture and transforms various natural language processing tasks into text-to-text transfer tasks.
BARTNER Yan et al. (2021b) is a pointer-based sequence-to-sequence architecture designed for NER tasks.
It converts NER subtasks into a unified sequence generation task by predicting entities and their corresponding type indexes in the input sentence.
LightNER Chen et al. (2022) is a pointer-based sequence-to-sequence model which builds upon the BARTNER.
It introduces a prompt tuning technique that incorporates additional parameters into the attention mechanism.
InstructionNER Wang et al. (2022b) is a multi-task instruction-based generative framework specifically designed for addressing few-shot NER tasks.
It redefines the NER task as a natural language generation problem and introduces descriptive instructions and an optional mechanism to enable the model to understand different tasks and constrain the output space.

B.3 Implementation Details

In the upstream pre-training stage, we set the batch size to 32, and our pre-training process typically takes around 1 hour for 5 epochs. In this paper, we conduct all the experiments without any hyperparameter search. For the multi-task training strategy, we assign equal weights to three noisy auxiliary tasks, i.e., set α?\alphaitalic_α, β?\betaitalic_β, and γ?\gammaitalic_γ to 1313\frac{1}{3}divide start_ARG 1 end_ARG start_ARG 3 end_ARG. The corresponding learning rates are set to 1e-5. For the demonstration-based fine-tuning stage, we also set the batch size to 32 and the training takes an average of 2 hours for 5 epochs, while the learning rates are set to 5e-5. For the selection of demonstrations, we recall the top 2 instances with the highest similarity score from the noisy candidate pool.

In all experiments, we train and test various methods using NVIDIA RTX A6000 GPU. To select the best model, we evaluate the performance on the validation set using the F1 metric every 400 training steps. Experiments in Table 2 and Table 1 use base-version of T5 and BART, while we also adopt the large-version model in Table 5 on single perturbation setting. We retrieve CleanDemos??????????CleanDemositalic_C italic_l italic_e italic_a italic_n italic_D italic_e italic_m italic_o italic_s from Dcleansubscript??????D_{clean}italic_D start_POSTSUBSCRIPT italic_c italic_l italic_e italic_a italic_n end_POSTSUBSCRIPT, NoiseDemos??????????NoiseDemositalic_N italic_o italic_i italic_s italic_e italic_D italic_e italic_m italic_o italic_s from Daugsubscript????D_{aug}italic_D start_POSTSUBSCRIPT italic_a italic_u italic_g end_POSTSUBSCRIPT and MixDemos????????MixDemositalic_M italic_i italic_x italic_D italic_e italic_m italic_o italic_s from Dmixsubscript????D_{mix}italic_D start_POSTSUBSCRIPT italic_m italic_i italic_x end_POSTSUBSCRIPT for the ablation study and the experiments on investigating the impact of the number of demonstrations. We will release our code after a blind review.

Methods Clean Sentence-level Character-level Word-level Perturbed-Avg.
Verbose Paraphrase Simplification Typos Speech
BART 95.21 79.23 87.20 83.81 57.79 75.65 76.74
T5 95.58 82.12 88.36 85.98 68.25 81.31 81.20
BARTNER 95.30 79.96 90.44 87.24 75.32 76.22 81.84
LightNER 96.02 80.32 90.40 87.97 67.28 75.57 80.31
InstructionNER 95.05 82.01 87.82 85.25 69.75 80.09 80.98
DemoNSF(BART) 95.77(±plus-or-minus\pm±1.4) 81.25(±plus-or-minus\pm±0.7) 90.56(±plus-or-minus\pm±0.1) 88.12(±plus-or-minus\pm±0.5) 74.20(±plus-or-minus\pm±0.7) 84.58(±plus-or-minus\pm±0.3) 83.74(±plus-or-minus\pm±0.4)
DemoNSF(T5) 95.81(±plus-or-minus\pm±0.7) 83.77(±plus-or-minus\pm±0.4) 91.58(±plus-or-minus\pm±0.6) 89.78(±plus-or-minus\pm±0.2) 77.70(±plus-or-minus\pm±1.3) 87.96(±plus-or-minus\pm±0.9) 86.16(±plus-or-minus\pm±0.5)
Table 5: F1 scores with standard deviations under 5 different input perturbations on RADDLE. All the models are in large versions

Appendix C More Detailed Experiments

C.1 Ablation Study

We conduct an ablation study to investigate the characteristics of the main components in DemoNSF. As shown in Table 4, we have the following observations: 1) The performance of the model improves when adding any component, which demonstrates that every part of our design is necessary. 2) For the three different granularities of perturbations, we observe significant improvements in auxiliary tasks specifically designed for each. Specifically, the NR task learns the mapping relationship between character-level perturbations and clean data, resulting in a 4.79% improvement in Typos. While the RM task implicitly captures the semantic information of slot entities during the mask-filling procedure. It achieved about 4.71% improvement under word-level perturbations (Speech). As for the HD task, it is able to capture the unique distribution information of perturbed data and significantly improves the performance of the model under coarse-grained perturbations while maintaining generalization, especially in Simplification (3.47%). 3) Adopting joint pre-training tasks (+MT??{}_{+MT}start_FLOATSUBSCRIPT + italic_M italic_T end_FLOATSUBSCRIPT) results in a noticeable improvement compared with adding one of them, which indicates that jointly pre-training objectives have a mutually reinforcing effect (obtain 3.47% improvement on average of perturbed data). 4) We explore the ablation study of three demonstration retrieval strategies. CleanDemos??????????CleanDemositalic_C italic_l italic_e italic_a italic_n italic_D italic_e italic_m italic_o italic_s, MixDemos????????MixDemositalic_M italic_i italic_x italic_D italic_e italic_m italic_o italic_s, and NoisyDemos??????????NoisyDemositalic_N italic_o italic_i italic_s italic_y italic_D italic_e italic_m italic_o italic_s represent retrieve demonstrations from Dcleansubscript??????D_{clean}italic_D start_POSTSUBSCRIPT italic_c italic_l italic_e italic_a italic_n end_POSTSUBSCRIPT, Dmixsubscript????D_{mix}italic_D start_POSTSUBSCRIPT italic_m italic_i italic_x end_POSTSUBSCRIPT and Daugsubscript????D_{aug}italic_D start_POSTSUBSCRIPT italic_a italic_u italic_g end_POSTSUBSCRIPT, respectively. As for MixDemos????????MixDemositalic_M italic_i italic_x italic_D italic_e italic_m italic_o italic_s, We make sure to include both clean and noisy demonstrations. We find that concatenating demonstrations does yield exciting results on perturbed test data. Specifically, while MixDemos????????MixDemositalic_M italic_i italic_x italic_D italic_e italic_m italic_o italic_s is able to absorb more diverse data distributions and performs well on both clean and perturbed data, the NoisyDemos??????????NoisyDemositalic_N italic_o italic_i italic_s italic_y italic_D italic_e italic_m italic_o italic_s used in this paper focuses on introducing the distribution information of the perturbed data, so that the generative model can learn the perturbed sentence and slot entity distribution information to the maximum extent and make it more robust.

Methods Clean Sent Word Char Char+Word Char+Sent Word+Sent Char+Word+Sent
App. Sub. Ent. Ent.+Sub. Ent.+App. App.+Sub. Ent.+Sub.+App.
BART 79.43 65.95 71.20 57.84 58.00 47.27 50.28 38.36
T5 94.12 83.97 85.13 65.90 57.44 56.79 73.47 47.93
BARTNER 86.34 69.33 77.22 62.28 54.83 49.10 58.92 42.25
LightNER 81.39 60.63 70.18 52.59 42.34 35.82 45.44 27.00
InstructionNER 94.69 84.32 84.78 66.93 57.87 58.89 74.45 50.75
DemoNSF (BART) 87.16(±plus-or-minus\pm±0.7) 76.05(±plus-or-minus\pm±1.2) 79.56(±plus-or-minus\pm±1.3) 67.31(±plus-or-minus\pm±0.2) 61.27(±plus-or-minus\pm±0.5) 51.26(±plus-or-minus\pm±0.9) 68.72(±plus-or-minus\pm±0.5) 44.27(±plus-or-minus\pm±0.9)
DemoNSF (T5) 94.75(±plus-or-minus\pm±1.7) 86.83(±plus-or-minus\pm±0.5) 86.81(±plus-or-minus\pm±0.2) 72.79(±plus-or-minus\pm±0.4) 63.59(±plus-or-minus\pm±0.3) 63.94(±plus-or-minus\pm±1.2) 77.69(±plus-or-minus\pm±0.7) 55.12(±plus-or-minus\pm±0.3)
Table 6: F1 scores with standard deviations under 3 kinds of single perturbations and 4 kinds of mixed perturbations on SNIPS.

C.2 Results on Large-version Model

We compare the performance of DemoNSF with other baselines on the large-version model (i.e. T5-large and BART-large). Despite using a model with a larger parameter size, generative models still experience a significant decline in performance when faced with perturbed inputs, especially with fine-grained perturbations. As shown in Table 5, we can find that the model’s performance declined by 37.42% for BART and 27.33% for T5 on Typos and 19.56% and 14.27% for BART and T5, respectively on Speech. Our approach also achieves impressive improvements on both fine-grained and coarse-grained perturbations. To be specific, DemoNSF introduces 5.18% F1 improvements on the average performance of all the perturbations input compared with InstructionNER based on T5 and 1.63% improvements compared with BARTNER based on BART.

C.3 Details of Mixed Perturbations

For the mixed perturbations experiment on SNIPS, we also investigate the performance of DemoNSF on single perturbation (AppendIrr, Sub, and EntTypos). As shown in Table 6, we obtain similar conclusions. Specifically, our approach introduces significant improvements in fine-grained perturbations (e.g. 5.86% on EntTypos). While our approach also maintains exciting performance on coarse-grained perturbations (e.g. 2.51% improvements on AppendIrr).

Appendix D Related Work

D.1 Slot Filling

Sequence Labeling Paradigm. Initially, the slot filling task was commonly defined as a sequence labeling problem. Previous methods can be categorized into two types: one-stage and two-stage approaches. Specifically, one-stage approaches Bapna et al. (2017); Shah et al. (2019); Lee and Jha (2019) conduct slot filling individually for each slot type. It first generates word-level representations and the predictions are based on the concatenated features for each slot type. However, these methods only learn the surface mapping between entities and suffer from multi-prediction problems. To address these limitations, a branch of two-stage methods Liu et al. (2020b); He et al. (2020b); Wang et al. (2021); Ma et al. (2022); Dong et al. (2023a); Wang et al. (2022a); Dong et al. (2023b) are proposed. Firstly, a coarse-grained binary sequence labeling model is used to identify all slot entities in the utterances. Subsequently, the entity value is mapped to the representation of the corresponding slot label in the semantic space in order to classify slot types effectively.

Generative Framework. Recently, some works Wang et al. (2022b) have started to reformulating NER and slot filling tasks to sequence-to-sequence (seq2seq) tasks and integrate generative methods. BARTNER He et al. (2020c) proposes a pointer-based seq2seq architecture, which converts the NER task to a unified sequence generation task and predicts entities from the input sentences and the corresponding type indexes LightNERChen et al. (2021) introduces prompt-tuning to the attention mechanism of BARTNER and achieves promising improvement in low-resource scenarios. Moreover, some prompt-based Generative methods Lu et al. (2022); Xie et al. (2022b) have achieved strong performance in information extraction. Nevertheless, their exploration of generative frameworks on diverse input perturbations remains a blank area, hindering their application in realistic task-oriented dialogue systems.

D.2 Input Perturbation Problem

Recently, there has been a growing interest in enhancing the resilience of NLP systems to input perturbations. Moradi and Samwald (2021b) present empirical evaluations of the robustness of various NLP systems against input perturbations on synthetically generated benchmarks. Namysl et al. (2020, 2021) focus on the robustness of the NER model against Optical Character Recognition (OCR) noise and misspellings. Compared to other NLP systems, dialogue systems would face more diverse input noise due to more frequent interactions with users. Fang et al. (2020); Gopalakrishnan et al. (2020) investigate the robustness of dialogue systems on ASR noise, and Ruan et al. (2020); Li et al. (2020b); Huang and Chen (2020); Li et al. (2020a) mainly focus on the ASR-noise-robustness SLU models in dialogue systems.

Most previous studies (Moradi and Samwald, 2021a; Wu et al., 2021; Liu et al., 2020a) that investigated this robustness problem predominantly focused on rule-based synthetic datasets, which do possess certain limitations. Meanwhile, real-world dialogue systems encounter a wider range of perturbations due to frequent interactions with users. To further explore this direction, RADDLE Peng et al. (2020) offers a crowd-sourced robustness evaluation benchmark for dialog systems, which includes various noisy utterances that existed in real dialogue scenarios. Liu et al. (2020a) introduced Language Understanding Augmentation, a methodology that incorporates four data augmentation techniques to simulate natural perturbations. To cope with more complex noisy scenarios, Dong et al. (2022a); Liu et al. (2023) investigate input perturbation problems on discriminative neural models. In this paper, we mainly focus on the performance of generative frameworks on input perturbation problems.

D.3 Demonstration-based learning

Demonstrations are first introduced by the GPT series Radford et al. (2019); Brown et al. (2020), where a few examples are sampled from training data and transformed with templates into appropriately-filled prompts. Based on the task reformulation and whether the parameters are updated, the existing demonstration-based learning research can be broadly divided into three categories: In-context Learning Brown et al. (2020); Zhao et al. (2021); Min et al. (2021); Wei et al. (2022), Prompt-based Fine-tuning Liang et al. (2022); Dong et al. (2023c), Classifier-based Fine-tuning Lee et al. (2021). However, these approaches mainly adopt demonstration-based learning in the fine-tuning that cannot make full use of the effect of demonstration-based learning.