Abstract
The field of Deep Learning (DL) has undergone explosive growth during the last decade, with a substantial impact on Natural Language Processing (NLP) as well. Yet, as with other fields employing DL techniques, there has been a lack of common experimental standards compared to more established disciplines. Starting from fundamental scientific principles, we distill ongoing discussions on experimental standards in DL into a single, widely-applicable methodology. Following these best practices is crucial to strengthening experimental evidence, improve reproducibility and enable scientific progress. These standards are further collected in a public repository to help them transparently adapt to future needs.
| Original language | English |
|---|---|
| Publication date | 29 Apr 2022 |
| Publication status | Published - 29 Apr 2022 |
| Event | ML Evaluation Standards Workshop at ICLR 2022 - Duration: 29 Apr 2022 → … |
Conference
| Conference | ML Evaluation Standards Workshop at ICLR 2022 |
|---|---|
| Period | 29/04/2022 → … |
Keywords
- Deep Learning
- Natural Language Processing
- Experimental Standards
- Reproducibility
- Scientific Methodology
Fingerprint
Dive into the research topics of 'Experimental Standards for Deep Learning Research: A Natural Language Processing Perspective'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver