File:On the Dangers of Stochastic Parrots Can Language Models Be Too Big.pdf

Page contents not supported in other languages.
This is a file from the Wikimedia Commons
Source: Wikipedia, the free encyclopedia.
Go to page
next page →
next page →
next page →

Original file(1,275 × 1,650 pixels, file size: 297 KB, MIME type: application/pdf, 14 pages)

Summary

Description
English: The past 3 years of work in NLP have been characterized by the development and deployment of ever larger language models, especially for English. BERT, its variants, GPT-2/3, and others, most recently Switch-C, have pushed the boundaries of the possible both through architectural innovations and through sheer size. Using these pretrained models and the methodology of fine-tuning them for specific tasks, researchers have extended the state of the art on a wide array of tasks as measured by leaderboards on specific benchmarks for English. In this paper, we take a step back and ask: How big is too big? What are the possible risks associated with this technology and what paths are available for mitigating those risks? We provide recommendations including weighing the environmental and financial costs first, investing resources into curating and carefully documenting datasets rather than ingesting everything on the web, carrying out pre-development exercises evaluating how the planned approach fits into research and development goals and supports stakeholder values, and encouraging research directions beyond ever larger language models.
Date
Source

https://dl.acm.org/doi/abs/10.1145/3442188.3445922

https://doi.org/10.1145/3442188
Author Emily M. Bender, Timnit Gebru, Angelina McMillan-Major, Shmargaret Shmitchell

Licensing

w:en:Creative Commons
attribution
This file is licensed under the Creative Commons Attribution 4.0 International license.
You are free:
  • to share – to copy, distribute and transmit the work
  • to remix – to adapt the work
Under the following conditions:
  • attribution – You must give appropriate credit, provide a link to the license, and indicate if changes were made. You may do so in any reasonable manner, but not in any way that suggests the licensor endorses you or your use.

Captions

On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜

1 March 2021

application/pdf

File history

Click on a date/time to view the file as it appeared at that time.

Date/TimeThumbnailDimensionsUserComment
current12:41, 1 March 2023Thumbnail for version as of 12:41, 1 March 20231,275 × 1,650, 14 pages (297 KB)PamputtUploaded a work by Emily M. Bender, Timnit Gebru, Angelina McMillan-Major, Shmargaret Shmitchell from https://dl.acm.org/doi/abs/10.1145/3442188.3445922 https://doi.org/10.1145/3442188 with UploadWizard
No pages on the English Wikipedia use this file (pages on other projects are not listed).

Global file usage

The following other wikis use this file:

Metadata