<div><br></div><div><br><div class="gmail_quote"><div dir="ltr" class="gmail_attr">On Fri, Mar 24, 2023 at 2:12 AM Stuart LaForge via extropy-chat <<a href="mailto:extropy-chat@lists.extropy.org">extropy-chat@lists.extropy.org</a>> wrote:<br></div><div dir="ltr" class="gmail_attr"><br></div><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex" dir="auto"><br>But really the meaning of words are quite arbitrary and determined by <br>
the people who use them. Thus the referential meanings of words evolve <br>
and change over time and come to refer to different things</blockquote><div dir="auto"><br></div><div dir="auto">I agree this is a reason for many human miscommunications, but the speaker understands his words to meaning *something* and the hearer understands those words to mean *something*.</div><div dir="auto"><br></div><div dir="auto">As a computational linguist, Bender is on our side. She is obviously very excited about the progress these language models represent, but is reminding that the models do not actually understand words to mean anything whatsoever. </div><div dir="auto"><br></div><div dir="auto">-gts</div></div></div>