diff --git a/brainsteam/content/replies/2022/11/23/1669236617.md b/brainsteam/content/replies/2022/11/23/1669236617.md new file mode 100644 index 0000000..4b132da --- /dev/null +++ b/brainsteam/content/replies/2022/11/23/1669236617.md @@ -0,0 +1,64 @@ +--- +date: '2022-11-23T20:50:17' +hypothesis-meta: + created: '2022-11-23T20:50:17.668925+00:00' + document: + title: + - 2022.naacl-main.167.pdf + flagged: false + group: __world__ + hidden: false + id: b_EbpGtwEe2m8tfhSKM2EQ + links: + html: https://hypothes.is/a/b_EbpGtwEe2m8tfhSKM2EQ + incontext: https://hyp.is/b_EbpGtwEe2m8tfhSKM2EQ/aclanthology.org/2022.naacl-main.167.pdf + json: https://hypothes.is/api/annotations/b_EbpGtwEe2m8tfhSKM2EQ + permissions: + admin: + - acct:ravenscroftj@hypothes.is + delete: + - acct:ravenscroftj@hypothes.is + read: + - group:__world__ + update: + - acct:ravenscroftj@hypothes.is + tags: + - prompt-models + - NLProc + target: + - selector: + - end: 2221 + start: 1677 + type: TextPositionSelector + - exact: "Suppose a human is given two sentences: \u201CNoweapons of mass destruction\ + \ found in Iraq yet.\u201Dand \u201CWeapons of mass destruction found in Iraq.\u201D\ + They are then asked to respond 0 or 1 and receive areward if they are correct.\ + \ In this setup, they wouldlikely need a large number of trials and errors\ + \ be-fore figuring out what they are really being re-warded to do. This setup\ + \ is akin to the pretrain-and-fine-tune setup which has dominated NLP in recentyears,\ + \ in which models are asked to classify a sen-tence representation (e.g.,\ + \ a CLS token) into some" + prefix: task instructions.1 Introduction + suffix: "\u2217Unabridged version available on" + type: TextQuoteSelector + source: https://aclanthology.org/2022.naacl-main.167.pdf + text: This is a really excellent illustration of the difference in paradigm between + "normal" text model fine tuning and prompt-based modelling + updated: '2022-11-23T20:50:17.668925+00:00' + uri: https://aclanthology.org/2022.naacl-main.167.pdf + user: acct:ravenscroftj@hypothes.is + user_info: + display_name: James Ravenscroft +in-reply-to: https://aclanthology.org/2022.naacl-main.167.pdf +tags: +- prompt-models +- NLProc +- hypothesis +type: reply +url: /replies/2022/11/23/1669236617 + +--- + + + +
Suppose a human is given two sentences: “Noweapons of mass destruction found in Iraq yet.”and “Weapons of mass destruction found in Iraq.”They are then asked to respond 0 or 1 and receive areward if they are correct. In this setup, they wouldlikely need a large number of trials and errors be-fore figuring out what they are really being re-warded to do. This setup is akin to the pretrain-and-fine-tune setup which has dominated NLP in recentyears, in which models are asked to classify a sen-tence representation (e.g., a CLS token) into some
This is a really excellent illustration of the difference in paradigm between "normal" text model fine tuning and prompt-based modelling \ No newline at end of file