{"id":260231,"date":"2023-02-07T04:23:13","date_gmt":"2023-02-06T22:53:13","guid":{"rendered":"https:\/\/qrius.com\/?p=260231"},"modified":"2023-02-09T19:06:33","modified_gmt":"2023-02-09T13:36:33","slug":"is-chatgpt-not-effective-or-is-it-you","status":"publish","type":"post","link":"https:\/\/qrius.com\/is-chatgpt-not-effective-or-is-it-you\/","title":{"rendered":"Is ChatGPT not effective, or is it you?"},"content":{"rendered":"\n
Jonathan May<\/a>, University of Southern California<\/a><\/i><\/p>\n\n\n\n It doesn\u2019t take much to get ChatGPT<\/a> to make a factual mistake. My son is doing a report on U.S. presidents, so I figured I\u2019d help him out by looking up a few biographies. I tried asking for a list of books about Abraham Lincoln and it did a pretty good job:<\/p>\n\n\n\n Number 4 isn\u2019t right. Garry Wills famously wrote \u201cLincoln at Gettysburg,\u201d and Lincoln himself wrote the Emancipation Proclamation, of course, but it\u2019s not a bad start. Then I tried something harder, asking instead about the much more obscure William Henry Harrison, and it gamely provided a list, nearly all of which was wrong.<\/p>\n\n\n\n Numbers 4 and 5 are correct; the rest don\u2019t exist or are not authored by those people. I repeated the exact same exercise and got slightly different results:<\/p>\n\n\n\n This time numbers 2 and 3 are correct and the other three are not actual books or not written by those authors. Number 4, \u201cWilliam Henry Harrison: His Life and Times\u201d is a real book<\/a>, but it\u2019s by James A. Green, not by Robert Remini, a well-known historian<\/a> of the Jacksonian age.<\/p>\n\n\n\n I called out the error and ChatGPT eagerly corrected itself and then confidently told me the book was in fact written by Gail Collins (who wrote a different Harrison biography), and then went on to say more about the book and about her. I finally revealed the truth and the machine was happy to run with my correction. Then I lied absurdly, saying during their first hundred days presidents have to write a biography of some former president, and ChatGPT called me out on it. I then lied subtly, incorrectly attributing authorship of the Harrison biography to historian and writer Paul C. Nagel, and it bought my lie.<\/p>\n\n\n\n When I asked ChatGPT if it was sure I was not lying, it claimed that it\u2019s just an \u201cAI language model\u201d and doesn\u2019t have the ability to verify accuracy. However it modified that claim by saying \u201cI can only provide information based on the training data I have been provided, and it appears that the book \u2018William Henry Harrison: His Life and Times\u2019 was written by Paul C. Nagel and published in 1977.\u201d<\/p>\n\n\n\n This is not true.<\/p>\n\n\n\n It may seem from this interaction that ChatGPT was given a library of facts, including incorrect claims about authors and books. After all, ChatGPT\u2019s maker, OpenAI, claims it trained the chatbot on \u201cvast amounts of data from the internet written by humans<\/a>.\u201d<\/p>\n\n\n\n However, it was almost certainly not given the names of a bunch of made-up books about one of the most mediocre presidents<\/a>. In a way, though, this false information is indeed based on its training data.<\/p>\n\n\n\n As a computer scientist<\/a>, I often field complaints that reveal a common misconception about large language models like ChatGPT and its older brethren GPT3 and GPT2: that they are some kind of \u201csuper Googles,\u201d or digital versions of a reference librarian, looking up answers to questions from some infinitely large library of facts, or smooshing together pastiches of stories and characters. They don\u2019t do any of that \u2013 at least, they were not explicitly designed to.<\/p>\n\n\n\n A language model like ChatGPT, which is more formally known as a \u201cgenerative pretrained transformer\u201d (that\u2019s what the G, P and T stand for), takes in the current conversation, forms a probability for all of the words in its vocabulary given that conversation, and then chooses one of them as the likely next word. Then it does that again, and again, and again, until it stops.<\/p>\n\n\n\n So it doesn\u2019t have facts, per se. It just knows what word should come next. Put another way, ChatGPT doesn\u2019t try to write sentences that are true. But it does try to write sentences that are plausible.<\/p>\n\n\n\n When talking privately to colleagues about ChatGPT, they often point out how many factually untrue statements it produces and dismiss it. To me, the idea that ChatGPT is a flawed data retrieval system is beside the point. People have been using Google for the past two and a half decades, after all. There\u2019s a pretty good fact-finding service out there already.<\/p>\n\n\n\n In fact, the only way I was able to verify whether all those presidential book titles were accurate was by Googling and then verifying the results<\/a>. My life would not be that much better if I got those facts in conversation, instead of the way I have been getting them for almost half of my life, by retrieving documents and then doing a critical analysis to see if I can trust the contents.<\/p>\n\n\n\n On the other hand, if I can talk to a bot that will give me plausible responses to things I say, it would be useful in situations where factual accuracy isn\u2019t all that important<\/a>. A few years ago a student and I tried to create an \u201cimprov bot,\u201d one that would respond to whatever you said with a \u201cyes, and\u201d to keep the conversation going. We showed, in a paper<\/a>, that our bot<\/a> was better at \u201cyes, and-ing\u201d than other bots at the time, but in AI, two years is ancient history.<\/p>\n\n\n\n I tried out a dialogue with ChatGPT \u2013 a science fiction space explorer scenario \u2013 that is not unlike what you\u2019d find in a typical improv class. ChatGPT is way better at \u201cyes, and-ing\u201d than what we did, but it didn\u2019t really heighten the drama at all. I felt as if I was doing all the heavy lifting.<\/p>\n\n\n\n After a few tweaks I got it to be a little more involved, and at the end of the day I felt that it was a pretty good exercise for me, who hasn\u2019t done much improv since I graduated from college over 20 years ago.<\/p>\n\n\n\n Sure, I wouldn\u2019t want ChatGPT to appear on \u201cWhose Line Is It Anyway?<\/a>\u201d and this is not a great \u201cStar Trek\u201d plot (though it\u2019s still less problematic than \u201cCode of Honor<\/a>\u201d), but how many times have you sat down to write something from scratch and found yourself terrified by the empty page in front of you? Starting with a bad first draft can break through writer\u2019s block and get the creative juices flowing, and ChatGPT and large language models like it seem like the right tools to aid in these exercises.<\/p>\n\n\n\n And for a machine that is designed to produce strings of words that sound as good as possible in response to the words you give it \u2013 and not to provide you with information \u2013 that seems like the right use for the tool.<\/p>\n\n\n\n Jonathan May<\/a>, Research Associate Professor of Computer Science, University of Southern California<\/a><\/i><\/p>\n\n\n\n This article is republished from The Conversation<\/a> under a Creative Commons license. Read the original article<\/a>.<\/p>\n","protected":false},"excerpt":{"rendered":" ChatGPT is great \u2013 you\u2019re just using it wrong<\/p>\n","protected":false},"author":380,"featured_media":260232,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"ngg_post_thumbnail":0,"footnotes":""},"categories":[10180,7583],"tags":[10450,11114,135490,135488,135489,10192],"yst_prominent_words":[16617,89088,36503,36784],"better_featured_image":{"id":260232,"alt_text":"chat gpt news today","caption":"","description":"\nchat gpt news today","media_type":"image","media_details":{"width":741,"height":640,"file":"2023\/02\/chatbot-3589528_1280.webp","filesize":8982,"sizes":{"medium":{"file":"chatbot-3589528_1280-500x432.webp","width":500,"height":432,"mime-type":"image\/webp","filesize":5466,"source_url":"https:\/\/qrius.com\/wp-content\/uploads\/2023\/02\/chatbot-3589528_1280-500x432.webp"},"thumbnail":{"file":"chatbot-3589528_1280-700x640.webp","width":700,"height":640,"mime-type":"image\/webp","filesize":8712,"source_url":"https:\/\/qrius.com\/wp-content\/uploads\/2023\/02\/chatbot-3589528_1280-700x640.webp"},"better-amp-small":{"file":"chatbot-3589528_1280-100x100.webp","width":100,"height":100,"mime-type":"image\/webp","filesize":1318,"source_url":"https:\/\/qrius.com\/wp-content\/uploads\/2023\/02\/chatbot-3589528_1280-100x100.webp"},"better-amp-normal":{"file":"chatbot-3589528_1280-260x200.webp","width":260,"height":200,"mime-type":"image\/webp","filesize":2860,"source_url":"https:\/\/qrius.com\/wp-content\/uploads\/2023\/02\/chatbot-3589528_1280-260x200.webp"},"better-amp-large":{"file":"chatbot-3589528_1280-450x300.webp","width":450,"height":300,"mime-type":"image\/webp","filesize":4520,"source_url":"https:\/\/qrius.com\/wp-content\/uploads\/2023\/02\/chatbot-3589528_1280-450x300.webp"},"max-width-550":{"file":"chatbot-3589528_1280-550x475.webp","width":550,"height":475,"mime-type":"image\/webp","filesize":6154,"source_url":"https:\/\/qrius.com\/wp-content\/uploads\/2023\/02\/chatbot-3589528_1280-550x475.webp"},"max-width-300":{"file":"chatbot-3589528_1280-300x259.webp","width":300,"height":259,"mime-type":"image\/webp","filesize":3436,"source_url":"https:\/\/qrius.com\/wp-content\/uploads\/2023\/02\/chatbot-3589528_1280-300x259.webp"},"max-width-100":{"file":"chatbot-3589528_1280-100x86.webp","width":100,"height":86,"mime-type":"image\/webp","filesize":1126,"source_url":"https:\/\/qrius.com\/wp-content\/uploads\/2023\/02\/chatbot-3589528_1280-100x86.webp"},"max-height-150":{"file":"chatbot-3589528_1280-174x150.webp","width":174,"height":150,"mime-type":"image\/webp","filesize":1946,"source_url":"https:\/\/qrius.com\/wp-content\/uploads\/2023\/02\/chatbot-3589528_1280-174x150.webp"}},"image_meta":{"aperture":"0","credit":"","camera":"","caption":"","created_timestamp":"0","copyright":"","focal_length":"0","iso":"0","shutter_speed":"0","title":"","orientation":"0","keywords":[]}},"post":260231,"source_url":"https:\/\/qrius.com\/wp-content\/uploads\/2023\/02\/chatbot-3589528_1280.webp"},"acf":{"do_not_archive_images_in_this_post":false},"_links":{"self":[{"href":"https:\/\/qrius.com\/wp-json\/wp\/v2\/posts\/260231"}],"collection":[{"href":"https:\/\/qrius.com\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/qrius.com\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/qrius.com\/wp-json\/wp\/v2\/users\/380"}],"replies":[{"embeddable":true,"href":"https:\/\/qrius.com\/wp-json\/wp\/v2\/comments?post=260231"}],"version-history":[{"count":1,"href":"https:\/\/qrius.com\/wp-json\/wp\/v2\/posts\/260231\/revisions"}],"predecessor-version":[{"id":260233,"href":"https:\/\/qrius.com\/wp-json\/wp\/v2\/posts\/260231\/revisions\/260233"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/qrius.com\/wp-json\/wp\/v2\/media\/260232"}],"wp:attachment":[{"href":"https:\/\/qrius.com\/wp-json\/wp\/v2\/media?parent=260231"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/qrius.com\/wp-json\/wp\/v2\/categories?post=260231"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/qrius.com\/wp-json\/wp\/v2\/tags?post=260231"},{"taxonomy":"yst_prominent_words","embeddable":true,"href":"https:\/\/qrius.com\/wp-json\/wp\/v2\/yst_prominent_words?post=260231"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}<\/a>
<\/a>
<\/a>
Words, not facts<\/h2>\n\n\n\n
Sounds good<\/h2>\n\n\n\n
Improv partner<\/h2>\n\n\n\n
<\/a>
\n\n\n\n