[ExI] Language models are like mirrors

Giovanni Santostasi gsantostasi at gmail.com
Mon Apr 3 05:24:47 UTC 2023


Also if you ever want to go into details on how these NLM work here is a
very nice tutorial from the ex- Tesla AI Chief.
https://www.youtube.com/watch?v=kCc8FmEb1nY

Giovanni

On Sun, Apr 2, 2023 at 10:20 PM Giovanni Santostasi <gsantostasi at gmail.com>
wrote:

> And this:
> ChatGPT Writes a Chatbot AI
>
> https://www.youtube.com/watch?v=QumfkMQr47M
>
> On Sun, Apr 2, 2023 at 10:19 PM Giovanni Santostasi <gsantostasi at gmail.com>
> wrote:
>
>>
>> *But hey, here is an optimistic parting shot: let us use ChatGPT as a
>> trainer, ask it to teach us how to set up large language models.  Then we
>> can all try our own hands at it, ja?*It has been done:
>> https://www.youtube.com/watch?v=7QWVJ5rWy2s
>>
>> On Sun, Apr 2, 2023 at 10:02 PM Gordon Swobe via extropy-chat <
>> extropy-chat at lists.extropy.org> wrote:
>>
>>> > The test I proposed of asking ChatGPT about a topic for which you know
>>> the right answers.  GPT fails those, but… in a fun way it comes across as
>>> the talking dog, ja?
>>>
>>> I have an example that is not so fun. For reasons of privacy, I was
>>> debating whether to share it, but I think I can make it sufficiently
>>> abstract. While I am no physician, I am highly knowledgeable about a
>>> particular disease that afflicts someone I know. I have spent more than a
>>> year researching it. No FDA approved meds exist for the treatment of this
>>> disease, but there does exist a class of medications that would seem to
>>> make sense as they are approved for some very similar and related disease.
>>> That class of meds can be divided into two subclasses. In the related
>>> diseases, any med in either subclass is more or less as safe and effective
>>> as any other.
>>>
>>> But in this particular disease that concerns me, meds in one of the
>>> subclasses are contraindicated and strongly so. They are potentially lethal.
>>>
>>> When ChatGPT first went online as version 3.5, I asked what would be
>>> some proper medications for this disease I have in mind, and was appalled
>>> to see it list mostly medications in the contraindicated subclass. I filled
>>> out the feedback form to OpenAI to warn them of the error.
>>>
>>> I'm glad to see now in version 4 that they've got it right. Not only are
>>> the contraindicated meds not listed in the answer, but ChatGPT-4 warns
>>> about them.
>>>
>>> -gts
>>>
>>> On Sun, Apr 2, 2023 at 10:03 PM spike jones via extropy-chat <
>>> extropy-chat at lists.extropy.org> wrote:
>>>
>>>>
>>>>
>>>>
>>>>
>>>> *…*> *On Behalf Of *Adrian Tymes via extropy-chat
>>>> *Subject:* Re: [ExI] Language models are like mirrors
>>>>
>>>>
>>>>
>>>> On Sun, Apr 2, 2023 at 5:32 PM spike jones via extropy-chat <
>>>> extropy-chat at lists.extropy.org> wrote:
>>>>
>>>> So… try it yourself.  Take some area which you know more about that
>>>> anyone, some technical niche perhaps, something on which you are a
>>>> qualified judge, hand it to ChatGPT.  It will take a shot at it and it will
>>>> really sound like it knows from shinola.
>>>>
>>>>
>>>>
>>>> OK then, does it?
>>>>
>>>>
>>>>
>>>> >…HA HA HA no.
>>>>
>>>>
>>>>
>>>> ---
>>>>
>>>>
>>>>
>>>> >…  While this AI is making the same mistakes as humans, it's still
>>>> making mistakes that show it doesn't understand what it's talking about…
>>>>
>>>>
>>>>
>>>> Adrian you said something important there.
>>>>
>>>>
>>>>
>>>> >…I could go on at much further length, but I believe these three
>>>> examples adequately establish the point… Adrian
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> OK cool, now think about Adrian’s point that ChatGPT making the same
>>>> mistake as humans.
>>>>
>>>>
>>>>
>>>> Consider the old joke about the farmer selling a talking dog for ten
>>>> bucks, the dog tells the buyer of his adventures as an undercover FBI
>>>> agent, a stock broker, a Coast Guard officer, and now a farm dog.  The
>>>> astonished owner asks the farmer why he is selling the dog so cheap, at
>>>> which the farmer says “Because he’s a damn liar.  He ain’t done half that
>>>> stuff.”
>>>>
>>>>
>>>>
>>>> OK then, sure, ChatGPT is so dumb it makes the same mistakes humans
>>>> do.  Now consider news stories.  We read those, we assume they are more or
>>>> less correct, but once in a long while we see a news story about something
>>>> we know a lotta lotta about because we were there when it happened.  We
>>>> saw, we heard.  Later we read the news account.  Invariably… we say nooooo
>>>> no no, that’s not what happened, that is a terrible description of the
>>>> event, lousy.
>>>>
>>>>
>>>>
>>>> Then the thought occurs to us: what if… all news stories are this bad?
>>>>
>>>>
>>>>
>>>> The test I proposed of asking ChatGPT about a topic for which you know
>>>> the right answers.  GPT fails those, but… in a fun way it comes across as
>>>> the talking dog, ja?  You don’t take it too seriously in details, and you
>>>> know it certainly isn’t flawless, but… it’s a talking dog fer cryin out
>>>> loud, we don’t expect it to be Mr. Peabody (younger crowd, ask ChatGPT who
>>>> is Mr. Peabody.)
>>>>
>>>>
>>>>
>>>> ChatGPT has its uses, and has demonstrated itself to be a marvelous
>>>> teacher and trainer.  I have used it recently to come up to speed on legal
>>>> terms, and I am convinced it will have enormous impact on society in many
>>>> ways.  I think it produces insight-like comments, but it is clear enough to
>>>> me it found them in a huge database rather than invented them.  Perhaps
>>>> that counts as a kind of legitimate pseudo-insight, and has its uses.  I
>>>> will accept that it is better than humans at many things we do and pay for,
>>>> resulting in some professions going away.  The one that comes to mind first
>>>> is paralegals.  Those guys are adios amigos methinks.
>>>>
>>>>
>>>>
>>>> ChatGPT makes the same mistakes as humans and it is a marvelous novelty
>>>> like a talking dog.  I haven’t been able to convince myself it is going to
>>>> result in the big S Singularity by rewriting itself and becoming a
>>>> Bostrom-style superintelligence.  That is still ahead of us.
>>>>
>>>>
>>>>
>>>> But hey, here is an optimistic parting shot: let us use ChatGPT as a
>>>> trainer, ask it to teach us how to set up large language models.  Then we
>>>> can all try our own hands at it, ja?
>>>>
>>>>
>>>>
>>>> spike
>>>> _______________________________________________
>>>> extropy-chat mailing list
>>>> extropy-chat at lists.extropy.org
>>>> http://lists.extropy.org/mailman/listinfo.cgi/extropy-chat
>>>>
>>> _______________________________________________
>>> extropy-chat mailing list
>>> extropy-chat at lists.extropy.org
>>> http://lists.extropy.org/mailman/listinfo.cgi/extropy-chat
>>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.extropy.org/pipermail/extropy-chat/attachments/20230402/118f5939/attachment-0001.htm>


More information about the extropy-chat mailing list