Large language model

A

Alenoach

Guest
More reference formatting, minor rephrasing

Line 218:Line 218:
Line 361:Line 361:
[td]
← Previous revision
[/td]
[td]
Revision as of 17:13, 30 August 2025
[/td]
[td]=== Emergent abilities ===[/td]
[td]=== Emergent abilities ===[/td]
[td]{{anchor|Emergent abilities}}[[File:LLM emergent benchmarks.png|thumb|At point(s) referred to as [[Broken Neural Scaling Law|breaks]],<ref name="IYm4Q" /> the lines change their slopes, appearing on a linear-log plot as a series of linear segments connected by arcs.]][/td]
[td]{{anchor|Emergent abilities}}[[File:LLM emergent benchmarks.png|thumb|At point(s) referred to as [[Broken Neural Scaling Law|breaks]],<ref name="IYm4Q" /> the lines change their slopes, appearing on a linear-log plot as a series of linear segments connected by arcs.]][/td]
[td]Performance of bigger models on various tasks, when plotted on a log-log scale, appears as a linear extrapolation of performance achieved by smaller models. However, this linearity may be punctuated by "[[Broken Neural Scaling Law|break(s)]]"<ref name="IYm4Q">{{cite arXiv |eprint=2210.14891 |class=cs.LG |first1=Ethan |last1=Caballero |first2=Kshitij |last2=Gupta |title=Broken Neural Scaling Laws |last3=Rish |first3=Irina |last4=Krueger |first4=David |year=2022}}</ref> in the scaling law, where the slope of the line changes abruptly, and where larger models acquire "emergent abilities".<ref name="emergentpaper">{{cite journal |last1=Wei |first1=Jason |last2=Tay |first2=Yi |last3=Bommasani |first3=Rishi |last4=Raffel |first4=Colin |last5=Zoph |first5=Barret |last6=Borgeaud |first6=Sebastian |last7=Yogatama |first7=Dani |last8=Bosma |first8=Maarten |last9=Zhou |first9=Denny |last10=Metzler |first10=Donald |last11=Chi |first11=Ed H. |last12=Hashimoto |first12=Tatsunori |last13=Vinyals |first13=Oriol |last14=Liang |first14=Percy |last15=Dean |first15=Jeff |date=31 August 2022 |title=Emergent Abilities of Large Language Models |url=https://openreview.net/forum?id=yzkSU5zdwD |journal=Transactions on Machine Learning Research |issn=2835-8856 |last16=Fedus |first16=William |access-date=19 March 2023 |archive-date=22 March 2023 |archive-url=https://web.archive.org/web/20230322210052/https://openreview.net/forum?id=yzkSU5zdwD |url-status=live }}</ref><ref name="JM6s1">{{Cite web |title=137 emergent abilities of large language models |url=https://www.jasonwei.net/blog/emergence |access-date=2023-06-24 |website=Jason Wei }}</ref> They arise from the complex interaction of the model's components and are not explicitly programmed or designed.<ref name="Bowman">{{cite arXiv |eprint=2304.00612 |class=cs.CL |first=Samuel R. |last=Bowman |title=Eight Things to Know about Large Language Models |year=2023}}</ref>[/td]
[td]Performance of bigger models on various tasks, when plotted on a log-log scale, appears as a linear extrapolation of performance achieved by smaller models. However, this linearity may be punctuated by "[[Broken Neural Scaling Law|break(s)]]"<ref name="IYm4Q">{{cite arXiv |eprint=2210.14891 |class=cs.LG |first1=Ethan |last1=Caballero |first2=Kshitij |last2=Gupta |title=Broken Neural Scaling Laws |last3=Rish |first3=Irina |last4=Krueger |first4=David |year=2022}}</ref> in the scaling law, where the slope of the line changes abruptly, and where larger models acquire "emergent abilities".<ref name="emergentpaper">{{cite journal |last1=Wei |first1=Jason |last2=Tay |first2=Yi |last3=Bommasani |first3=Rishi |last4=Raffel |first4=Colin |last5=Zoph |first5=Barret |last6=Borgeaud |first6=Sebastian |last7=Yogatama |first7=Dani |last8=Bosma |first8=Maarten |last9=Zhou |first9=Denny |last10=Metzler |first10=Donald |last11=Chi |first11=Ed H. |last12=Hashimoto |first12=Tatsunori |last13=Vinyals |first13=Oriol |last14=Liang |first14=Percy |last15=Dean |first15=Jeff |date=31 August 2022 |title=Emergent Abilities of Large Language Models |url=https://openreview.net/forum?id=yzkSU5zdwD |journal=Transactions on Machine Learning Research |issn=2835-8856 |last16=Fedus |first16=William |access-date=19 March 2023 |archive-date=22 March 2023 |archive-url=https://web.archive.org/web/20230322210052/https://openreview.net/forum?id=yzkSU5zdwD |url-status=live }}</ref><ref name="JM6s1">{{Cite web |title=137 emergent abilities of large language models |url=https://www.jasonwei.net/blog/emergence |access-date=2023-06-24 |website=Jason Wei }}</ref> They arise from the complex interaction of the model's components and are not explicitly programmed or designed.<ref name="Bowman">{{cite journal |journal=Critical AI |url= .1215/2834703X-11556011/400182/Eight-Things-to-Know-about-Large-Language-Models |first=Samuel R. |last=Bowman |title=Eight Things to Know about Large Language Models |year=2023}}</ref>[/td]
[td][/td]
[td][/td]
[td]One of the emergent abilities is [[in-context learning]] from example demonstrations.<ref name="Hahn_20230314">{{cite arXiv |eprint=2303.07971 |class=cs.LG |first1=Michael |last1=Hahn |first2=Navin |last2=Goyal |title=A Theory of Emergent In-Context Learning as Implicit Structure Induction |date=2023-03-14}}</ref> In-context learning is involved in tasks, such as:[/td]
[td]One of the emergent abilities is [[in-context learning]] from example demonstrations.<ref name="Hahn_20230314">{{cite arXiv |eprint=2303.07971 |class=cs.LG |first1=Michael |last1=Hahn |first2=Navin |last2=Goyal |title=A Theory of Emergent In-Context Learning as Implicit Structure Induction |date=2023-03-14}}</ref> In-context learning is involved in tasks, such as:[/td]
[td][/td]
[td][/td]
[td]=== Mental health ===[/td]
[td]=== Mental health ===[/td]
[td]Research and social media posts suggest that some individuals are using LLMs to seek therapy or mental health support.<ref>{{Cite news |last=Zao-Sanders |first=Marc |date=2024-03-19 |title=How People Are Really Using GenAI |url=https://hbr.org/2024/03/how-people-are-really-using-genai |access-date=2025-08-10 |work=Harvard Business Review |language=en |issn=0017-8012}}</ref> In early 2025, a survey by Sentio University found that nearly half (48.7%) of 499 U.S. adults with ongoing mental health conditions who had used LLMs reported turning to them for therapy or emotional support, including help with anxiety, depression, loneliness, and similar concerns.<ref>{{Cite journal |last1=Rousmaniere |first1=Tony |last2=Zhang |first2=Yimeng |last3=Li |first3=Xu |last4=Shah |first4=Siddharth |date=2025-07-21 |title=Large language models as mental health resources: Patterns of use in the United States. |url=https://doi.apa.org/doi/10.1037/pri0000292 |journal=Practice Innovations |language=en |doi=10.1037/pri0000292 |issn=2377-8903|url-access=subscription }}</ref> Studies have found that LLMs can produce hallucinationsβ€”plausible but incorrect statementsβ€”which may mislead users in sensitive mental health contexts.<ref>{{cite arXiv |last1=Ji |first1=Shaoxiong |title=Rethinking Large Language Models in Mental Health Applications |date=2023-12-17 |eprint=2311.11267 |last2=Zhang |first2=Tianlin |last3=Yang |first3=Kailai |last4=Ananiadou |first4=Sophia |last5=Cambria |first5=Erik |class=cs.CL }}</ref> Research also shows that LLMs may express stigma or inappropriate agreement with maladaptive thoughts, reflecting limitations in replicating the judgment and relational skills of human therapists.<ref>{{cite book |last1=Moore |first1=Jared |title=Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency |date=2025-04-25 |arxiv=2504.18412 |last2=Grabb |first2=Declan |last3=Agnew |first3=William |last4=Klyman |first4=Kevin |last5=Chancellor |first5=Stevie |last6=Ong |first6=Desmond C. |last7=Haber |first7=Nick |chapter=Expressing stigma and inappropriate responses prevents LLMS from safely replacing mental health providers |pages=599–627 |doi=10.1145/3715275.3732039 |isbn=979-8-4007-1482-5 }}</ref> Evaluations of crisis scenarios indicate that some LLMs lack effective safety protocols, such as assessing suicide risk or making appropriate referrals.<ref>{{cite arXiv |last1=Grabb |first1=Declan |title=Risks from Language Models for Automated Mental Healthcare: Ethics and Structure for Implementation |date=2024-08-14 |eprint=2406.11852 |last2=Lamparth |first2=Max |last3=Vasan |first3=Nina |class=cs.CY }}</ref><ref>{{Cite journal |last1=McBain |first1=Ryan K. |last2=Cantor |first2=Jonathan H. |last3=Zhang |first3=Li Ang |last4=Baker |first4=Olesya |last5=Zhang |first5=Fang |last6=Halbisen |first6=Alyssa |last7=Kofner |first7=Aaron |last8=Breslau |first8=Joshua |last9=Stein |first9=Bradley |last10=Mehrotra |first10=Ateev |last11=Yu |first11=Hao |date=2025-03-05 |title=Competency of Large Language Models in Evaluating Appropriate Responses to Suicidal Ideation: Comparative Study |journal=Journal of Medical Internet Research |language=EN |volume=27 |issue=1 |pages=e67891 |doi=10.2196/67891 |pmid=40053817 |pmc=11928068 |doi-access=free }}</ref>[/td]
[td]Research and social media posts suggest that some individuals are using LLMs to seek therapy or mental health support.<ref>{{Cite news |last=Zao-Sanders |first=Marc |date=2024-03-19 |title=How People Are Really Using GenAI |url=https://hbr.org/2024/03/how-people-are-really-using-genai |access-date=2025-08-10 |work=Harvard Business Review |language=en |issn=0017-8012}}</ref> In early 2025, a survey by Sentio University found that nearly half (48.7%) of 499 U.S. adults with ongoing mental health conditions who had used LLMs reported turning to them for therapy or emotional support, including help with anxiety, depression, loneliness, and similar concerns.<ref>{{Cite journal |last1=Rousmaniere |first1=Tony |last2=Zhang |first2=Yimeng |last3=Li |first3=Xu |last4=Shah |first4=Siddharth |date=2025-07-21 |title=Large language models as mental health resources: Patterns of use in the United States. |url=https://doi.apa.org/doi/10.1037/pri0000292 |journal=Practice Innovations |language=en |doi=10.1037/pri0000292 |issn=2377-8903|url-access=subscription }}</ref> LLMs can produce hallucinationsβ€”plausible but incorrect statementsβ€”which may mislead users in sensitive mental health contexts.<ref>{{cite arXiv |last1=Ji |first1=Shaoxiong |title=Rethinking Large Language Models in Mental Health Applications |date=2023-12-17 |eprint=2311.11267 |last2=Zhang |first2=Tianlin |last3=Yang |first3=Kailai |last4=Ananiadou |first4=Sophia |last5=Cambria |first5=Erik |class=cs.CL }}</ref> Research also shows that LLMs may express stigma or inappropriate agreement with maladaptive thoughts, reflecting limitations in replicating the judgment and relational skills of human therapists.<ref>{{cite book |last1=Moore |first1=Jared |title=Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency |date=2025-04-25 |arxiv=2504.18412 |last2=Grabb |first2=Declan |last3=Agnew |first3=William |last4=Klyman |first4=Kevin |last5=Chancellor |first5=Stevie |last6=Ong |first6=Desmond C. |last7=Haber |first7=Nick |chapter=Expressing stigma and inappropriate responses prevents LLMS from safely replacing mental health providers |pages=599–627 |doi=10.1145/3715275.3732039 |isbn=979-8-4007-1482-5 }}</ref> Evaluations of crisis scenarios indicate that some LLMs lack effective safety protocols, such as assessing suicide risk or making appropriate referrals.<ref>{{cite journal |journal=COLM |last1=Grabb |first1=Declan |title=Risks from Language Models for Automated Mental Healthcare: Ethics and Structure for Implementation |date=2024-08-14 |url= 2406.11852 |last2=Lamparth |first2=Max |last3=Vasan |first3=Nina }}</ref><ref>{{Cite journal |last1=McBain |first1=Ryan K. |last2=Cantor |first2=Jonathan H. |last3=Zhang |first3=Li Ang |last4=Baker |first4=Olesya |last5=Zhang |first5=Fang |last6=Halbisen |first6=Alyssa |last7=Kofner |first7=Aaron |last8=Breslau |first8=Joshua |last9=Stein |first9=Bradley |last10=Mehrotra |first10=Ateev |last11=Yu |first11=Hao |date=2025-03-05 |title=Competency of Large Language Models in Evaluating Appropriate Responses to Suicidal Ideation: Comparative Study |journal=Journal of Medical Internet Research |language=EN |volume=27 |issue=1 |pages=e67891 |doi=10.2196/67891 |pmid=40053817 |pmc=11928068 |doi-access=free }}</ref>[/td]
[td][/td]
[td][/td]
[td]== See also ==[/td]
[td]== See also ==[/td]

Continue reading...
 


Join 𝕋𝕄𝕋 on Telegram
Channel PREVIEW:
Back
Top