DiffuD2T: Empowering Data-to-Text Generation with Diffusion
Abstract
:1. Introduction
- We propose to improve prefix tuning for data-to-text generation with the step-by-step refinement diffusion model.
- We propose a look-ahead guiding loss to supervise the refinement process of the diffusion model.
- We propose a planning-then-writing training pipeline to provide the model content planning ability.
- We conducted experiments and analyses on three data-to-text generation datasets and both automatic evaluation criteria (Section 5.2 and Section 5.5) and human evaluation criteria (Section 5.7) show the effectiveness of our model.
2. Literature Review
2.1. Data-to-Text Generation
2.2. Diffusion Model
3. Background
3.1. Task Formulation
3.2. Diffusion Model
4. Methods
4.1. DiffuD2T: Diffusion for Data-to-Text Generation
4.2. Look-Ahead Guiding Loss
4.3. Planning-Then-Writing Pipeline
4.4. Training and Inference
4.4.1. Training
4.4.2. Inference
4.5. Algorithm
Algorithm 1 Our approach DiffuD2T for data-to-text generation |
|
5. Experiments
5.1. Dataset
- The WebNLG dataset focuses on generating textual descriptions for 14 domains. The structured data can be viewed as <entity, type, value> triple. Unlike the other two datasets, it explicitly evaluates the model’s generalization performance on unseen domains; that is, the domain is not covered during training. It consists of 18,025, 870, and 1862 instances for training, validation, and test set. The average input length is 49.6 and the average output length is 30.7. Following the official evaluation script (https://github.com/Yale-LILY/dart, accessed on 27 December 2022), we report BLEU [48], METEOR (MET) [49], and TER [50] scores in this paper.
- The DART dataset focuses on open-domain data-to-text generation based on their structured data. The structured data can be viewed as <entity, type, value> triple. It consists of 62659, 2721, and 4159 instances for training, validation, and test set. The average input length is 38.8 and the average output length is 27.3. Following the official evaluation script (https://github.com/Yale-LILY/dart, accessed on 27 December 2022), we report BLEU [48], METEOR (MET) [49], TER [50], MoverScore [51], BERTScore [52] and BLEURT [53] scores in this paper.
- The E2E dataset focuses on generating restaurant descriptions based on their attributes. The structured data can be viewed as <attribute, value> pairs. It consists of 42,061, 547, and 630 instances for training, validation, and test set. The average input length is 28.5, and the average output length is 27.8. Following the official evaluation script (https://github.com/tuetschek/e2e-metrics, accessed on 27 December 2022), we report BLEU [48], NIST [54], METEOR (MET) [49], ROUGE-L (R-L) [55] and CIDEr [56] scores in this paper.
5.2. Automatic Evaluation Criteria
- BLEU [48]: Based on n-gram matching, BLEU assesses the model’s performance between the generated text and reference text. BLEU is calculated as follows:Equation (13) has three components. The n-gram precision of the generated text compared to the reference text (Equation (11)). The weight for each n-gram is positive. The brevity penalty (Equation (12)) penalizes generated text that is shorter than the reference text. g is the length of the generated text and r is the length of the reference text. Following Papineni et al. [48], we use BLEU with as the criteria.
- NIST [54]: Although it is still an n-gram-based metric similar to BLEU, it introduces information-weighted n-gram precision that favors those more informative n-grams. NIST is calculated as follows:The first fraction of Equation (15) is the information-weighted (Equation (14)) n-gram precision of the generated text compared to the reference text. It gives more weight to those n-grams that occur less frequently, which are considered more informative. is the length of the generated text and is the average length of the reference texts. is used to penalize short text generated by the model. , and is chosen to make this brevity penalty factor as 0.5 when .
- METEOR (MET) [49]: While N-gram matching tends to perform exact string matching between the system-generated text and target text, METEOR uses WordNet to match synonyms because the meanings are the same. In addition, it proposes to organize words into chunks and use this to determine how well-ordered the words are. METEOR is calculated as follows:
- ROUGE [55]: Unlike BLEU, this metric mainly focuses on the models’ performance on recall. Additionally, it uses the Longest Common Subsequence (LCS) to match the system-generated text and reference text.
- CIDEr [56]: Similar to the NIST score, this metric also focus on tokens that are more informative. Differently, it uses Term Frequency Inverse Document Frequency (TF-IDF) to serve the purpose, as it will give more weight to infrequently occurring but informative words in the corpus. CIDEr is calculated as follows:Equation (25) shows that CIDEr uses a TF-IDF (Equation (24)) vector to represent the generated text and reference text, and uses cosine similarity to calculate the score. m is the number of reference texts. The first fraction in Equation (24) is the n-gram ’s term frequency in the text. The second fraction is its inverse document frequency across the documents in the corpus. is the vocabulary of all n-grams and I consists of all texts in the corpus.
- TER [50]: TER is the abbreviation of the Translation Edit Rate. It measures the quality of the system-generated text by calculating the number of edit operations to an exact match between the system-generated text and reference text. TER is calculated as follows:
- MoverScore [51]: Unlike previous metrics that compare system-generated text and reference text only in their surface form, it uses contextualized embedding with Earth Mover’s Distance to evaluate the texts in the semantic level. MoverScore is calculated as follows:and represent the generated text and reference text. C is a matrix of transportation costs to transform into . F consists of n-gram weights. is the n-gram weights for text sequence . is the n-gram weights for text sequence . This criterion measures the semantic distance between and .
- BLEURT [53] and BERTScore [52]: Similar to MoverScore, they are different from those n-gram-based criteria by using a pretrained language model to measure performance in the semantic level. BERTScore is calculated as follows:BERTScore (Equation (30)) is the combination of precision (Equation (29)) and recall (Equation (28)) scores of generated text and reference text x. and maximize the similarity score between the two texts via greedy matching.BLEURT is calculated as follows:
5.3. Comparing Methods
- Diffuseq [26]: While some attempts are made for language modeling with a diffusion model, this is the first and latest diffusion-based sequence-to-sequence model for text generation. We ran their publicly accessible codes and the suggested hyperparameters on WebNLG, E2E, and DART datasets.
- GAS [58]: They propose to model the local and global structural information through a graph neural network and use reinforcement learning to train the model.
- LoRA [39]: LoRA stands for Low-Rank Adaptation, which inserts trainable rank decomposition matrices into each layer of the model and fixes the parameters of the pretrained language model (PLM) to reduce the cost of finetuning.
- Xie et al. [59]: they explore the pretrained language model’s performance on data-to-text generation tasks.
- HierBlock [60]: Based on prefix tuning, this paper integrates hierarchical discourse information into modeling.
- Shen et al. [61]: They draw inspiration from computational pragmatics, which follows the intuitive that the “speaker should generate output text that a listener can use to identify correctly”. They experiment on the E2E dataset, and we report the result in the table.
- An et al. [62]: They propose to finetune both prefixes and input data’s representations so that PLM can better understand unfamiliar input data.
- Hou et al. [63]: They propose to use meta learning to adapt PLM to different tasks better.
- Finetune: This stands for finetuning the corresponding pretrained language model on the corresponding datasets.
- Adapter [64]: It introduces a trainable small-scale neural network module adapter to attach to the pretrained language model. For different tasks, different adapters can be attached to a single PLM to adapt to different tasks.
- Prefix [18]: It freezes the pretrained language model while prepending multiple trainable prefixes to the PLM to adapt it to different tasks. This can be considered our base model.
- +Diff: We combine the diffusion model with prefix tuning. This is the model described in Section 4.1.
- +Diff+LG: Additionally, we apply the look-ahead guiding loss (Section 4.2) to +Diff, in order to better supervise the denoising process of the diffusion model.
- +Diff+LG+CS: This is the full model, which combines the diffusion model (Section 4.1), look-ahead guiding loss (Section 4.2) and the planning-then-writing pipeline (Section 4.3).
5.4. Implementation Details
5.5. Automatic Evaluation Results
- Our models outperform the methods of finetune, adapter, and prefix tuning on the WebNLG dataset’s whole test set (A category), with respect to both GPT-2 Medium or GPT-2 Large as the backbone. For example, our full model +Diff+LG+CS outperforms the prefix tuning by 1.31% on BLEU with GPT-2 Medium. Additionally, our full model outperforms the prefix tuning by 2.19% on BLEU with GPT-2 Large. In addition, our full model (+Diff+LG+CS) on GPT-2 Large outperforms Diffuseq, a Sequence-to-Sequence diffusion-based text generation method, and other competitive baselines as listed in the part of Other Methods. The results above show the effectiveness of our proposed model.
- Our models also outperform finetuning, adapter, and prefix tuning on GPT-2 Medium by 1.69% in terms of BLEU, compared with prefix tuning and 1.66% in terms of BLEU on GPT-2 Large on the DART dataset. The comparison with other methods also indicates the effectiveness of our model on the DART dataset.
- Our models show similar patterns as described above on the E2E dataset. The full model improves prefix tuning by 0.88% in terms of BLEU on GPT-2 Medium and 1.14% on GPT-2 Large. They also outperform other methods on most metrics.
5.6. Ablation Study
5.7. Human Evaluation
6. Case Study
6.1. Diffusion Steps
- The structured data and corresponding reference text are sampled from the test set of WebNLG.
- Given the structured data, we use our proposed full model +Diff+LG+CS model with GPT-2 Large as its pretrained language model backbone to generate text. We explore two settings of diffusion steps: 10 denoising steps and 20 denoising steps (Section 4.1).
6.2. Content Planning
- The data and corresponding reference text are sampled from WebNLG’s test set.
- Content plans are generated by our proposed full model +Diff+LG+CS model with GPT-2 Large as its pretrained language model backbone.
- Extracted plans are automatically extracted from reference text (Section 4.3) and used to train the content planner.
- As shown in all three examples, our model, trained in the content planning stage, obtains good coverage performance, that is, the information in the data is covered in the content plans, by comparing the generated plan and extracted plan.
- However, the overall order of the data in the plans can be different from the one in the reference text, indicating a different style of writing compared to the reference. For instance, in the second example, the generated plan puts “Costa Crociere” and “Genoa” at the end of the plan while the extracted plan from the reference text puts them at the start of it, which means the model tends to first describe “The operator of AIDAstella” before mentioning “Costa Crociere” in the next surface realization stage, as illustrated in Table 9. However, our model may make mistakes during content planning. In the first example, the generated plan puts “Amar Osim” at the end of the plan, which makes it less natural, while the extracted plan and reference text put it in the middle of the text. However, in the planning-then-writing pipeline, models for surface realization can mitigate this, as illustrated in Table 9.
6.3. Surface Realization
- We sample six structured data and reference text from the test set of WebNLG’s unseen categories, that is the domains of these examples are not seen during training, to demonstrate the generalization ability of the models. We compare our proposed full model +Diff+LG+CS with two competitive methods: Prefix and Finetune. All three models use GPT-2 Large as their pretrained language model backbone. The results are reported in Table 9.
- We sample three structured data and reference text from the test set of E2E and three from the test set of DART. We compare our proposed full model +Diff+LG+CS with the baseline model Prefix. Both models use GPT-2 Large as its pretrained language model backbone. Results are reported in Table 10.
- In unseen domains (Table 9), both the prefix-tuning method and finetuning method tend to omit some of the information in the data, as the pattern may not be seen during training. However, our model (+Diff+LG+CS) can sometimes generalize better. For instance, in the sixth example, prefix tuning omitted the genre of the music and finetuning failed to generate fluent text, while our model can include both pieces of information in the structured data.
- Additionally, our model can sometimes generate more faithful content when the domain is unseen. In the second example, both the prefix-tuning and finetuning methods failed to understand what the “owner” in the structured data meant and generated the incorrect owner in this case. Our model can generate it correctly.
- As for generation results on E2E and DART datasets (Table 10), our model can generate more faithful text with higher coverage, that is, cover more information in data, as shown in the second example. In the second example, our model correctly describes “not family-friendly”, while prefix tuning omitted this information. Additionally, the third example shows that our model can generate more fluent text than prefix tuning.
7. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Gatt, A.; Portet, F.; Reiter, E.; Hunter, J.; Mahamood, S.; Moncur, W.; Sripada, S. From data to text in the neonatal intensive care unit: Using NLG technology for decision support and information management. Ai Commun. 2009, 22, 153–186. [Google Scholar] [CrossRef]
- Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 3104–3112. [Google Scholar]
- Luong, T.; Pham, H.; Manning, C.D. Effective Approaches to Attention-based Neural Machine Translation. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 1412–1421. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.u.; Polosukhin, I. Attention is All you Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4 December 2017; pp. 6000–6010. [Google Scholar]
- Novikova, J.; Dušek, O.; Rieser, V. The E2E Dataset: New Challenges For End-to-End Generation. In Proceedings of the 18th Annual SIGdial Meeting on Discourse and Dialogue, Saarbrücken, Germany, 15–17 August 2017; pp. 201–206. [Google Scholar]
- Gardent, C.; Shimorina, A.; Narayan, S.; Perez-Beltrachini, L. The WebNLG Challenge: Generating Text from RDF Data. In Proceedings of the 10th International Conference on Natural Language Generation, Santiago de Compostela, Spain, 4–7 September 2017; pp. 124–133. [Google Scholar]
- Nan, L.; Radev, D.; Zhang, R.; Rau, A.; Sivaprasad, A.; Hsieh, C.; Tang, X.; Vyas, A.; Verma, N.; Krishna, P.; et al. DART: Open-Domain Structured Data Record to Text Generation. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online, 6–11 June 2021; pp. 432–447. [Google Scholar]
- Puduppully, R.; Dong, L.; Lapata, M. Data-to-Text Generation with Content Selection and Planning. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January 2019; pp. 6908–6915. [Google Scholar]
- Puduppully, R.; Dong, L.; Lapata, M. Data-to-text Generation with Entity Modeling. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July 2019; pp. 2023–2035. [Google Scholar]
- Chen, K.; Li, F.; Hu, B.; Peng, W.; Chen, Q.; Yu, H.; Xiang, Y. Neural data-to-text generation with dynamic content planning. Knowl.-Based Syst. 2021, 215, 106610. [Google Scholar] [CrossRef]
- Puduppully, R.; Lapata, M. Data-to-text Generation with Macro Planning. Trans. Assoc. Comput. Linguist. 2021, 9, 510–527. [Google Scholar] [CrossRef]
- Nie, F.; Wang, J.; Yao, J.G.; Pan, R.; Lin, C.Y. Operation-guided Neural Networks for High Fidelity Data-To-Text Generation. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October 2018; pp. 3879–3889. [Google Scholar]
- Zhao, C.; Walker, M.; Chaturvedi, S. Bridging the Structural Gap Between Encoding and Decoding for Data-To-Text Generation. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 2481–2491. [Google Scholar]
- Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language Models are Unsupervised Multitask Learners. OpenAI Blog 2019, 1, 9. [Google Scholar]
- Chen, W.; Su, Y.; Yan, X.; Wang, W.Y. KGPT: Knowledge-Grounded Pre-Training for Data-to-Text Generation. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; pp. 8635–8648. [Google Scholar]
- Chen, Z.; Eavani, H.; Chen, W.; Liu, Y.; Wang, W.Y. Few-Shot NLG with Pre-Trained Language Model. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 183–190. [Google Scholar]
- Chang, E.; Shen, X.; Zhu, D.; Demberg, V.; Su, H. Neural Data-to-Text Generation with LM-based Text Augmentation. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, Online, 19–23 April 2021; pp. 758–768. [Google Scholar]
- Li, X.L.; Liang, P. Prefix-Tuning: Optimizing Continuous Prompts for Generation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Online, 1–6 August 2021; pp. 4582–4597. [Google Scholar]
- Sohl-Dickstein, J.; Weiss, E.; Maheswaranathan, N.; Ganguli, S. Deep unsupervised learning using nonequilibrium thermodynamics. In Proceedings of the International Conference on Machine Learning, Lille, France, 7–9 July 2015; pp. 2256–2265. [Google Scholar]
- Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 7–12 December 2020; pp. 6840–6851. [Google Scholar]
- Nichol, A.Q.; Dhariwal, P.; Ramesh, A.; Shyam, P.; Mishkin, P.; Mcgrew, B.; Sutskever, I.; Chen, M. GLIDE: Towards Photorealistic Image Generation and Editing with Text-Guided Diffusion Models. In Proceedings of the International Conference on Machine Learning, Baltimore, MA, USA, 17–23 July 2022; pp. 16784–16804. [Google Scholar]
- Dhariwal, P.; Nichol, A. Diffusion models beat gans on image synthesis. In Proceedings of the Advances in Neural Information Processing Systems, Online, 6–14 December 2021; pp. 8780–8794. [Google Scholar]
- Kong, Z.; Ping, W.; Huang, J.; Zhao, K.; Catanzaro, B. DiffWave: A Versatile Diffusion Model for Audio Synthesis. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 3–7 May 2021; pp. 1–17. [Google Scholar]
- Yang, D.; Yu, J.; Wang, H.; Wang, W.; Weng, C.; Zou, Y.; Yu, D. Diffsound: Discrete diffusion model for text-to-sound generation. arXiv 2022, arXiv:2207.09983. [Google Scholar] [CrossRef]
- Li, X.; Thickstun, J.; Gulrajani, I.; Liang, P.S.; Hashimoto, T.B. Diffusion-LM Improves Controllable Text Generation. In Proceedings of the Advances in Neural Information Processing Systems, New Orleans, LA, USA, 28 November 2022; pp. 4328–4343. [Google Scholar]
- Gong, S.; Li, M.; Feng, J.; Wu, Z.; Kong, L. DiffuSeq: Sequence to Sequence Text Generation with Diffusion Models. In Proceedings of the International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023; pp. 1–20. [Google Scholar]
- Strudel, R.; Tallec, C.; Altché, F.; Du, Y.; Ganin, Y.; Mensch, A.; Grathwohl, W.; Savinov, N.; Dieleman, S.; Sifre, L.; et al. Self-conditioned Embedding Diffusion for Text Generation. arXiv 2022, arXiv:2211.04236. [Google Scholar]
- Wiseman, S.; Shieber, S.; Rush, A. Challenges in Data-to-Document Generation. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; pp. 2253–2263. [Google Scholar]
- Reiter, E.; Dale, R. Building Natural Language Generation Systems; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar]
- Vinyals, O.; Fortunato, M.; Jaitly, N. Pointer Networks. In Proceedings of the Advances in Neural Information Processing Systems, Palais des Congrès de Montréal, Montréal, QC, Canada, 7 December 2015; pp. 2692–2700. [Google Scholar]
- Schlichtkrull, M.; Kipf, T.N.; Bloem, P.; van den Berg, R.; Titov, I.; Welling, M. Modeling Relational Data with Graph Convolutional Networks. In Proceedings of the The Semantic Web, Cham, Switzerland, 3 June 2018; pp. 593–607. [Google Scholar]
- Chen, W.; Chen, J.; Su, Y.; Chen, Z.; Wang, W.Y. Logical Natural Language Generation from Open-Domain Tables. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 7929–7942. [Google Scholar]
- Chen, Z.; Chen, W.; Zha, H.; Zhou, X.; Zhang, Y.; Sundaresan, S.; Wang, W.Y. Logic2Text: High-Fidelity Natural Language Generation from Logical Forms. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, Online, 16–20 November 2020; pp. 2096–2111. [Google Scholar]
- Zhang, N.; Ye, H.; Yang, J.; Deng, S.; Tan, C.; Chen, M.; Huang, S.; Huang, F.; Chen, H. LOGEN: Few-shot Logical Knowledge-Conditioned Text Generation with Self-training. arXiv 2021, arXiv:2112.01404. [Google Scholar]
- Kasner, Z.; Dusek, O. Neural Pipeline for Zero-Shot Data-to-Text Generation. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Dublin, Ireland, 22–27 May 2022; pp. 3914–3932. [Google Scholar]
- Clive, J.; Cao, K.; Rei, M. Control Prefixes for Parameter-Efficient Text Generation. In Proceedings of the 2nd Workshop on Natural Language Generation, Evaluation, and Metrics (GEM), Abu Dhabi, United Arab Emirates, 7 December 2022; pp. 363–382. [Google Scholar]
- Chen, Y.; Hazarika, D.; Namazifar, M.; Liu, Y.; Jin, D.; Hakkani-Tur, D. Inducer-tuning: Connecting Prefix-tuning and Adapter-tuning. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7–11 December 2022; pp. 793–808. [Google Scholar]
- Lebret, R.; Grangier, D.; Auli, M. Neural Text Generation from Structured Data with Application to the Biography Domain. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, Austin, TX, USA, 1–5 November 2016; pp. 1203–1213. [Google Scholar]
- Hu, E.J.; Wallis, P.; Allen-Zhu, Z.; Li, Y.; Wang, S.; Wang, L.; Chen, W. LoRA: Low-Rank Adaptation of Large Language Models. In Proceedings of the International Conference on Learning Representations, Virtual, 25–29 April 2022; pp. 1–13. [Google Scholar]
- Gao, Z.; Guo, J.; Tan, X.; Zhu, Y.; Zhang, F.; Bian, J.; Xu, L. Difformer: Empowering Diffusion Model on Embedding Space for Text Generation. arXiv 2022, arXiv:2212.09412. [Google Scholar]
- Ito, K.; Johnson, L. The Lj Speech Dataset 2017. Available online: https://keithito.com/LJ-Speech-Dataset (accessed on 1 February 2023).
- Warden, P. Speech commands: A dataset for limited-vocabulary speech recognition. arXiv 2018, arXiv:1804.03209. [Google Scholar]
- Bhagavatula, C.; Le Bras, R.; Malaviya, C.; Sakaguchi, K.; Holtzman, A.; Rashkin, H.; Downey, D.; Yih, W.T.; Choi, Y. Abductive Commonsense Reasoning. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
- Jiang, C.; Maddela, M.; Lan, W.; Zhong, Y.; Xu, W. Neural CRF Model for Sentence Alignment in Text Simplification. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 7943–7960. [Google Scholar]
- Dhingra, B.; Mazaitis, K.; Cohen, W.W. Quasar: Datasets for question answering by search and reading. arXiv 2017, arXiv:1707.03904. [Google Scholar]
- Gu, J.; Bradbury, J.; Xiong, C.; Li, V.O.; Socher, R. Non-Autoregressive Neural Machine Translation. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April 2018. [Google Scholar]
- Rush, A.M.; Chopra, S.; Weston, J. A Neural Attention Model for Abstractive Sentence Summarization. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 379–389. [Google Scholar]
- Papineni, K.; Roukos, S.; Ward, T.; Zhu, W.J. BLEU: A method for automatic evaluation of machine translation. In Proceedings of the Annual meeting on Association for Computational Linguistics, Philadelphia, PA, USA, 6 July 2002; pp. 311–318. [Google Scholar]
- Banerjee, S.; Lavie, A. METEOR: An automatic metric for MT evaluation with improved correlation with human judgments. In Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, Ann Arbor, MI, USA, 23 June 2005; pp. 65–72. [Google Scholar]
- Snover, M.; Dorr, B.; Schwartz, R.; Micciulla, L.; Makhoul, J. A Study of Translation Edit Rate with Targeted Human Annotation. In Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers, Cambridge, MA, USA, 8–12 August 2006; pp. 223–231. [Google Scholar]
- Zhao, W.; Peyrard, M.; Liu, F.; Gao, Y.; Meyer, C.M.; Eger, S. MoverScore: Text Generation Evaluating with Contextualized Embeddings and Earth Mover Distance. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; pp. 563–578. [Google Scholar]
- Zhang, T.; Kishore, V.; Wu, F.; Weinberger, K.Q.; Artzi, Y. BERTScore: Evaluating Text Generation with BERT. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020; pp. 1–43. [Google Scholar]
- Sellam, T.; Das, D.; Parikh, A. BLEURT: Learning Robust Metrics for Text Generation. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 7881–7892. [Google Scholar]
- Doddington, G. Automatic evaluation of machine translation quality using n-gram co-occurrence statistics. In Proceedings of the Second International Conference on Human Language Technology Research, San Francisco, CA, USA, 24 March 2002; pp. 138–145. [Google Scholar]
- Lin, C.Y. Rouge: A package for automatic evaluation of summaries. In Proceedings of the Text Summarization Branches Out, Barcelona, Spain, 25 July 2004; pp. 74–81. [Google Scholar]
- Vedantam, R.; Lawrence Zitnick, C.; Parikh, D. Cider: Consensus-based image description evaluation. In Proceedings of the IEEE conference on computer vision and pattern recognition, Boston, MA, USA, 7–12 June 2015; pp. 4566–4575. [Google Scholar]
- Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
- GAO, H.; WU, L.; ZHANG, H.; WEI, Z.; HU, P.; XU, F.; LONG, B. Triples-to-Text Generation with Reinforcement Learning Based Graph-augmented Structural Neural Networks. arXiv 2021, arXiv:2111.10545. [Google Scholar]
- Xie, T.; Wu, C.H.; Shi, P.; Zhong, R.; Scholak, T.; Yasunaga, M.; Wu, C.S.; Zhong, M.; Yin, P.; Wang, S.I.; et al. UnifiedSKG: Unifying and Multi-Tasking Structured Knowledge Grounding with Text-to-Text Language Models. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7–11 December 2022; pp. 602–631. [Google Scholar]
- Ghazvininejad, M.; Karpukhin, V.; Gor, V.; Celikyilmaz, A. Discourse-Aware Soft Prompting for Text Generation. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7–11 December 2022; pp. 4570–4589. [Google Scholar]
- Shen, S.; Fried, D.; Andreas, J.; Klein, D. Pragmatically Informative Text Generation. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; pp. 4060–4067. [Google Scholar]
- An, S.; Li, Y.; Lin, Z.; Liu, Q.; Chen, B.; Fu, Q.; Chen, W.; Zheng, N.; Lou, J.G. Input-tuning: Adapting unfamiliar inputs to frozen pretrained models. arXiv 2022, arXiv:2203.03131. [Google Scholar]
- Hou, Z.; Salazar, J.; Polovets, G. Meta-learning the difference: Preparing large language models for efficient adaptation. Trans. Assoc. Comput. Linguist. 2022, 10, 1249–1265. [Google Scholar] [CrossRef]
- Lin, Z.; Madotto, A.; Fung, P. Exploring Versatile Generative Language Model Via Parameter-Efficient Transfer Learning. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, Online, 16–20 November 2020; pp. 441–459. [Google Scholar]
- Dušek, O.; Novikova, J.; Rieser, V. Findings of the E2E NLG Challenge. In Proceedings of the 11th International Conference on Natural Language Generation, Tilburg, The Netherlands, 5–8 November 2018; pp. 322–328. [Google Scholar]
- Fleiss, J.L. Measuring nominal scale agreement among many raters. Psychol. Bull. 1971, 76, 378. [Google Scholar] [CrossRef]
Author | Contributions | Dataset | BLEU |
---|---|---|---|
Chen et al. [15] | They propose to pretrain the data-to-text model with large-scale unlabeled text and knowledge graph. | E2E [5], WebNLG [6], WikiBio [38] | 68.05 |
Chang et al. [17] | They propose to use the pretrained language model to generate new text to augment the limited corpus. | E2E [5], WebNLG [6] | 68.88 |
Li and Liang [18] | They propose to finetune lightweight task-specific parameters while freezing the parameters of pre-trained language model during training. | E2E [5], WebNLG [6], DART [7] | 70.30 |
Hu et al. [39] | They propose to insert trainable rank decomposition matrices into each layer of the model and freeze the parameters of pretrained language model (PLM) to reduce the cost of finetuning. | E2E [5], WebNLG [6], DART [7] | 70.40 |
Chen et al. [33] | They propose to manually annotate the logical inference process with logical forms and train the model to do logical inference explicitly. | Logic2Text [33] | 31.44 |
Nie et al. [12] | They propose to execute mathematical operations on structural data to provide more data for generating more accurate text. | RotoWire [28], ESPN [12] | 14.74 |
Puduppully and Lapata [11] | They propose to create paragraph plans for entities. | RotoWire [28], MLB [9] | 15.46 |
Puduppully et al. [8] | They propose a two-stage generation pipeline that selects and plans important information from structured data, and then generates the final text. | RotoWire [28] | 16.50 |
Puduppully et al. [9] | They propose to use memory slots to track the entity’s representation during generation for entity planning. | RotoWire [28], MLB [9] | 16.12 |
Chen et al. [10] | They propose to consider the previously generated text when planning the next structured data. | RotoWire [28], NBAZhn [10] | 16.38 |
Author | Domain | Contributions | Dataset | BLEU |
---|---|---|---|---|
Nichol et al. [21] | Image | They explore the diffusion model for text-conditional image synthesis and found that the classifier-free guidance performs better. | GLIDE (filtered) [21] | N/A |
Kong et al. [23] | Audio | They propose to use the diffusion model for conditional and unconditional audio generation. | LJ [41], Commands [42] | N/A |
Li et al. [25] | Text | They propose a diffusion-based language model for controllable text generation. | -NLG [43] | 7.1 |
Gong et al. [26] | Text | They propose a sequence-to-sequence diffusion-based text generation model for conditional text generation. | Jiang et al. [44], Quasar-T [45] | 36.22 |
Gao et al. [40] | Text | They propose three techniques to mitigate the limitation of the diffusion-based language model [25]. | WMT-14 [46], Gigaword [47] | 27.23 |
BLEU | MET | TER ↓ | |||||||
---|---|---|---|---|---|---|---|---|---|
S | U | A | S | U | A | S | U | A | |
Other Methods | |||||||||
DiffuSeq | 42.74 | 7.80 | 28.07 | 36.52 | 17.29 | 27.36 | 47.62 | 77.15 | 61.16 |
GAS [58] | 57.79 | 26.55 | 44.00 | 41.00 | 26.00 | 34.00 | 41.00 | 66.00 | 53.00 |
LoRA [39] | 64.00 | 48.40 | 57.00 | 45.00 | 39.00 | 42.00 | 32.00 | 45.00 | 38.00 |
GPT-2 | |||||||||
Finetune | 64.13 | 35.34 | 50.57 | 45.67 | 34.51 | 40.33 | 33.19 | 64.73 | 47.65 |
Adapter | 54.71 | 43.66 | 49.72 | 39.64 | 35.38 | 37.67 | 39.85 | 47.33 | 43.28 |
Prefix | 62.29 | 47.23 | 55.49 | 44.38 | 38.22 | 41.51 | 33.71 | 46.04 | 39.36 |
+Diff | 62.48 | 47.30 | 55.63 | 44.22 | 38.23 | 41.43 | 33.84 | 46.00 | 39.41 |
+Diff+LG | 62.87 | 47.43 | 55.91 | 44.49 | 38.09 | 41.51 | 33.18 | 45.41 | 38.78 |
+Diff+LG+CS | 63.06 | 47.86 | 56.22 | 44.77 | 38.69 | 41.93 | 33.74 | 46.29 | 39.50 |
GPT-2 | |||||||||
Finetune | 63.60 | 45.46 | 55.78 | 44.80 | 38.49 | 41.83 | 34.10 | 51.35 | 42.01 |
Adapter | 62.04 | 49.39 | 56.30 | 43.83 | 38.48 | 41.35 | 34.23 | 44.35 | 38.87 |
Prefix | 64.43 | 48.72 | 57.35 | 45.61 | 39.20 | 42.61 | 32.70 | 45.82 | 38.72 |
+Diff | 63.63 | 50.11 | 57.52 | 44.96 | 39.46 | 42.39 | 33.35 | 44.75 | 38.58 |
+Diff+LG | 65.19 | 48.87 | 57.84 | 45.86 | 39.12 | 42.71 | 32.24 | 46.23 | 38.65 |
+Diff+LG+CS | 65.18 | 50.58 | 58.61 | 45.55 | 39.78 | 42.85 | 32.45 | 44.01 | 37.75 |
BLEU | MET | TER ↓ | MoverScore | BERTScore | BLEURT | |
---|---|---|---|---|---|---|
Other Methods | ||||||
DiffuSeq | 12.63 | 31.77 | 60.53 | 62.84 | 89.01 | 27.62 |
Xie et al. [59] | 46.89 | 55.76 | 60.97 | - | 95.00 | 30.00 |
GAS [58] | 39.87 | 32.00 | 57.00 | - | - | - |
HierBlock [60] | 46.60 | 39.00 | 45.00 | 54.00 | 95.00 | - |
LoRA [39] | 47.50 | 39.00 | 45.00 | - | - | - |
GPT-2 | ||||||
Finetune | 45.98 | 38.57 | 45.70 | 67.86 | 94.82 | 39.07 |
Adapter | 42.77 | 36.59 | 47.96 | 66.57 | 94.48 | 34.23 |
Prefix | 45.94 | 38.49 | 45.49 | 67.94 | 94.90 | 39.80 |
+Diff | 46.06 | 38.79 | 46.01 | 68.10 | 94.88 | 40.69 |
+Diff+LG | 46.39 | 38.75 | 45.47 | 68.10 | 94.91 | 40.47 |
+Diff+LG+CS | 46.72 | 38.70 | 45.41 | 68.06 | 94.95 | 40.34 |
GPT-2 | ||||||
Finetune | 46.90 | 39.01 | 45.09 | 68.23 | 94.95 | 40.08 |
Adapter | 46.24 | 38.41 | 45.56 | 67.89 | 94.88 | 39.22 |
Prefix | 46.82 | 38.90 | 45.06 | 68.33 | 94.99 | 41.17 |
+Diff | 47.15 | 38.96 | 44.73 | 68.31 | 95.01 | 41.26 |
+Diff+LG | 47.43 | 38.96 | 44.70 | 68.35 | 95.03 | 41.23 |
+Diff+LG+CS | 47.60 | 39.21 | 44.70 | 68.45 | 95.05 | 41.60 |
BLEU | NIST | MET | ROUGE-L | CIDEr | |
---|---|---|---|---|---|
Other Methods | |||||
DiffuSeq | 48.90 | 7.37 | 39.34 | 59.15 | 1.53 |
Shen et al. [61] | 68.60 | 8.73 | 45.25 | 70.82 | 2.37 |
An et al. [62] | 68.70 | 8.74 | 46.10 | 70.70 | 2.42 |
HierBlock [60] | 67.20 | 8.70 | 45.10 | 69.10 | 2.35 |
LoRA [39] | 70.40 | 8.89 | 46.80 | 72.00 | 2.47 |
Hou et al. [63] | 69.70 | 8.78 | 46.90 | 72.10 | 2.51 |
GPT-2 | |||||
Finetune | 68.37 | 8.71 | 45.93 | 70.89 | 2.41 |
Adapter | 66.24 | 8.53 | 43.59 | 69.21 | 2.26 |
Prefix | 69.75 | 8.79 | 46.39 | 71.54 | 2.51 |
+Diff | 70.03 | 8.81 | 46.57 | 71.64 | 2.51 |
+Diff+LG | 69.83 | 8.79 | 46.21 | 71.38 | 2.50 |
+Diff+LG+CS | 70.37 | 8.86 | 46.42 | 71.77 | 2.52 |
GPT-2 | |||||
Finetune | 68.98 | 8.77 | 45.91 | 71.36 | 2.42 |
Adapter | 68.73 | 8.68 | 46.28 | 71.01 | 2.50 |
Prefix | 70.09 | 8.82 | 46.33 | 72.13 | 2.48 |
+Diff | 70.34 | 8.85 | 46.38 | 71.76 | 2.52 |
+Diff+LG | 70.43 | 8.86 | 46.24 | 71.83 | 2.48 |
+Diff+LG+CS | 70.89 | 8.93 | 46.51 | 72.14 | 2.51 |
Dataset | Model | Quality | Naturalness |
---|---|---|---|
WebNLG | Finetune | −6.11 | −19.22 |
Prefix | 9.67 | −1.22 | |
+Diff+LG+CS | 28.22 | 32.00 | |
DART | Finetune | −0.67 | −18.67 |
Prefix | 15.67 | 1.44 | |
+Diff+LG+CS | 33.00 | 32.11 | |
E2E | Finetune | 1.33 | −17.22 |
Prefix | 7.67 | -3.33 | |
+Diff+LG+CS | 39.44 | 37.67 |
Data | Diffusion Steps: 10 | Diffusion Steps: 20 | Reference Text |
---|---|---|---|
(Apollo 14, was a crew member of, Alan Shepard), (Department of Commerce Gold Medal, higher, Distinguished Service Medal (United States Navy)), (California, deathPlace, Alan Shepard), (New Hampshire, birthPlace, Alan Shepard), (NASA, operator, Apollo 14), (Distinguished Service Medal (United States Navy), awards, Alan Shepard) | Alan Shepard was born in New Hampshire and died in California. He was awarded the Distinguished Service Medal by the United States Navy, which ranks higher than the Department of Commerce Gold Medal. | Alan Shepard was born in New Hampshire and died in California. He was a crew member of Apollo 14 operated by NASA and was awarded the Distinguished Service Medal by the United States Navy, which is higher than the Department of Commerce Gold Medal. | Alan Shepard was born in New Hampshire and died in California. He was a crew member of Apollo 14 which is operated by NASA. He was awarded the Distinguished Service Medal in the US Navy, which is higher than the Dept of Commerce Gold Medal. |
(Switzerland, country, Accademia di Architettura di Mendrisio), (Mario Botta, dean, Accademia di Architettura di Mendrisio), (Mendrisio, city, Accademia di Architettura di Mendrisio), (1996, established, Accademia di Architettura di Mendrisio), (100, academicStaffSize, Accademia di Architettura di Mendrisio), (Ticino, location, Accademia di Architettura di Mendrisio) | The Accademia di Architettura di Mendrisio is located in Ticino, Mendrisio, Switzerland. It was established in 1996 and has 100 academic staff. Its dean is Mario Botta. | The Accademia di Architettura di Mendrisio in Ticino, Switzerland was established in 1996. Its dean is Mario Botta and it has 100 academic staff. | Accademia di Architettura di Mendrisio in Mendrisio, Switzerland has 100 employees, was established in 1996 and is overseen by Dean Mario Botta. The school’s location is Ticino. |
(2006-12-31, epoch, 110 Lydia), (6.7 (kilograms), mass, 110 Lydia), (377,016,000,000.0, periapsis, 110 Lydia) | 110 Lydia has a mass of 6.7 (kilograms) and a periapsis of 377,016,000,000.0. It was published in 2006. | 110 Lydia, which has a mass of 6.7 kg and periapsis of 377,016,000,000.0, was published in 2006. | 110 Lydia’s epoch is 31 December 2006, it has a mass of 6.7 kg and a periapsis of 377,016,000,000.0. |
Data | Generated Plan | Extracted Plan | Reference Text |
---|---|---|---|
(Al Kharaitiyat SC, club, Alaa Abdul-Zahra), (Al Khor, ground, Al Kharaitiyat SC), (Shabab Al-Ordon Club, club, Alaa Abdul-Zahra) (Amar Osim, manager, Al Kharaitiyat SC) | Content plans are: Alaa Abdul Zahra; Al Kharaitiyat SC; Al Khor; Shabab Al Ordon Club; Amar Osim. | Content plans are: Alaa Abdul Zahra; Al Kharaitiyat SC; Al Khor; Amar Osim; Shabab Al Ordon club; Alaa Abdul Zahra. | Alaa Abdul Zahra plays for Al Kharaitiyat SC which is located at Al Khor and managed by Amar Osim. The Shabab Al-Ordon club is associated with Alaa Abdul-Zahra. |
(Genoa, location, Costa Crociere), (AIDA Cruises, operator, AIDAstella), (Costa Crociere, owner, AIDAstella) | Content plans are: AIDAstella; AIDA Cruises; Costa Crociere; Genoa. | Content plans are: Costa Crociere; AIDAstella; Genoa; AIDAstella; AIDA Cruises. | Costa Crociere is the owner of the AIDAstella and are based in Genoa. The operator of AIDAstella is AIDA Cruises. |
(Euro, currency, Netherlands), (Stellendam, birthPlace, Ab Klink), (Netherlands, nationality, Ab Klink) | Content plans are: Ab Klink; Stellendam; Netherlands; Euro. | Content plans are: Ab Klink; Stellendam; Netherlands; euro. | Ab Klink was born in Stellendam in the Netherlands, where the national currency is the euro. |
Data Source or Model Name | Content |
---|---|
Source [Unseen, Athelete] | (Al Kharaitiyat SC, club, Alaa Abdul-Zahra), (Al Khor, ground, Al Kharaitiyat SC), (Shabab Al-Ordon Club, club, Alaa Abdul-Zahra) (Amar Osim, manager, Al Kharaitiyat SC) |
+Diff+LG+CS | Alaa Abdul-Zahra plays for Al Kharaitiyat SC, which is managed by Amar Osim. He also plays for Shabab Al-Ordon Club. |
Prefix | Al Kharaitiyat SC are managed by Amar Osim and play at their ground at Al Khor. Al Kharaitiyat SC are also the club for which Alaa Abdul-Zahra is a player. |
Finetune | Alaa Abdul-Zahra plays for Al-Kharaitiyat SC and Shabab Al-Ordon Club. He also plays for Al-Khor and manages Al-Kharaitiyat SC. |
Reference | Alaa Abdul Zahra plays for Al Kharaitiyat SC which is located at Al Khor and managed by Amar Osim. The Shabab Al-Ordon club is associated with Alaa Abdul-Zahra. |
Source [Unseen, Transportation] | (Genoa, location, Costa Crociere), (AIDA Cruises, operator, AIDAstella), (Costa Crociere, owner, AIDAstella) |
+Diff+LG+CS | AIDAstella is operated by AIDA Cruises and owned by Costa Crociere in Genoa. |
Prefix | AIDAstella is operated by Aida Cruises and is owned by the Costa Rican tourist resort of Genoa. |
Finetune | AIDAstella, operated by AIDA-Cruises, is located in Genoa and is owned by the Costa Rican government. |
Reference | Costa Crociere is the owner of the AIDAstella and are based in Genoa. The operator of AIDAstella is AIDA Cruises. |
Source [Unseen, Politician] | (Euro, currency, Netherlands), (Stellendam, birthPlace, Ab Klink ), (Netherlands, nationality, Ab Klink) |
+Diff+LG+CS | Ab Klink was born in Stellendam, the Netherlands, where the currency is the Euro. |
Prefix | Ab Klink was born in Stellendam and is a national of the Netherlands where the currency is the Euro. |
Finetune | Ab Klink is a national of the Netherlands where the currency is the Euro. He was born in Stellendam. |
Reference | Ab Klink was born in Stellendam in the Netherlands, where the national currency is the euro. |
Source [Unseen, Politician] | (Robert E, Lee, commander, Battle of Salem Church), (American Civil War, isPartOfMilitaryConflict, Battle of Salem Church), (Battle of Salem Church, battles, Aaron S. Daggett) |
+Diff+LG+CS | Robert E. Lee is the commander of the Battle of Salem Church which is part of the American Civil War. The Battle of Salem Church is a military conflict. |
Prefix | Robert E. Lee was the commander of the Battle of Salem Church which was part of the military conflict in the American Civil war. |
Finetune | The Battle of Salem Church is part of the American Civil War and was commanded by Robert E. Lee. |
Reference | Robert E Lee was a commander in the Battle of Salem Church, which was one of the military conflicts in the American Civil War. Aaron S Daggett fought in the same battle. |
Source [Unseen, Artist] | (Christian alternative rock, musicSubgenre, Alternative rock), (Alternative rock, genre, Andrew White (musician)) |
+Diff+LG+CS | Christian alternative rock is the genre of music Andrew White ( musician ) plays. |
Prefix | Andrew White is a Christian alternative rock musician. |
Finetune | Andrew White, a Christian alternative rocker, performs. |
Reference | The musician Andrew White’s genre is alternative rock, the genre which has the sub genre Christian alternative rock. |
Source [Unseen, Artist] | (Hip hop music, genre, Allen Forrest), (solo singer, background, Allen Forrest) |
+Diff+LG+CS | Allen Forrest is a solo singer who was born in the genre of Hip Hop music. |
Prefix | Allen Forrest is a solo singer. |
Finetune | Born in |
Reference | Allen Forrest is a solo singer whose genre is Hip Hop music. |
Data Source or Model Name | Content |
---|---|
Source [E2E] | (name, Cocum), (Type, coffee shop), (customer rating, high), (near, Burger King) |
+Diff+LG+CS | Cocum is a highly rated coffee shop near Burger King. |
Prefix | Cocum is a coffee shop near Burger King with a high customer rating. |
Reference | Near Burger King there is a highly rated coffee shop named Cocum. |
Source [E2E] | (name, The Mill), (Type, restaurant), (food, English), (price, less than £ 20), (customer rating, low), (area, city centre), (family friendly, no), (near, Café Rouge) |
+Diff+LG+CS | The Mill is a restaurant providing English food in the less than £ 20 price range. It is located in the city centre near Café Rouge. It has a low customer rating and is not family - friendly. |
Prefix | The Mill is a restaurant providing English food in the less than £ 20 price range. It is located in the city centre near Café Rouge. Its customer rating is low. |
Reference | The Mill is restaurant in the city centre, near Café Rouge, serving low-priced English food. It is has a low customer rating and is not family-friendly. |
Source [E2E] | (name, The Plough), (Type, pub), (food, Chinese), (price, cheap), (area, riverside), (family friendly, yes), (near, Raja Indian Cuisine) |
+Diff+LG+CS | The Plough is a family friendly Chinese pub in the riverside area near Raja Indian Cuisine with a cheap price range. |
Prefix | The Plough is a pub providing Chinese food in the cheap price range. It is located in the riverside. It is near Raja Indian Cuisine. |
Reference | The Plough is a Chinese pub and pub in the riverside area near Raja Indian Cuisine. It is family friendly and has cheap pricing. |
Source [DART] | (Mark Rutte, leader_name, Netherlands) |
+Diff+LG+CS | The leader of the Netherlands is Mark Rutte. |
Prefix | Mark Rutte is the leader of the Netherlands. |
Reference | The leader of the Netherlands is Mark Rutte. |
Source [DART] | (Michigan, country, United States) |
+Diff+LG+CS | Michigan is in the United States. |
Prefix | Michigan is located within the United States. |
Reference | Michigan is in the United states. |
Source [DART] | (Alfa Romeo 164, assembly, Arese) |
+Diff+LG+CS | Alfa Romeo 164 was assembled by Arese. |
Prefix | Alfa Romeo 164 is assembled arese. |
Reference | The Alfa Romeo 164 was assembled in Arese. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Gong, H.; Feng, X.; Qin, B. DiffuD2T: Empowering Data-to-Text Generation with Diffusion. Electronics 2023, 12, 2136. https://doi.org/10.3390/electronics12092136
Gong H, Feng X, Qin B. DiffuD2T: Empowering Data-to-Text Generation with Diffusion. Electronics. 2023; 12(9):2136. https://doi.org/10.3390/electronics12092136
Chicago/Turabian StyleGong, Heng, Xiaocheng Feng, and Bing Qin. 2023. "DiffuD2T: Empowering Data-to-Text Generation with Diffusion" Electronics 12, no. 9: 2136. https://doi.org/10.3390/electronics12092136