Groups | Search | Server Info | Login | Register
Groups > comp.ai.philosophy > #29815
| Subject | Re: ChatGPT and stack limits (was: Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question |
|---|---|
| Newsgroups | comp.theory, sci.logic, comp.ai.philosophy |
| References | <u6jhqq$1570m$1@dont-email.me> <u6vhv2$2urnr$1@dont-email.me> <u6vkro$30a76$1@dont-email.me> <QiLkM.629$sW_c.553@fx07.iad> <u705ae$323du$1@dont-email.me> |
| From | Richard Damon <Richard@Damon-Family.org> |
| Message-ID | <dDOkM.10063$jlQ4.3709@fx12.iad> (permalink) |
| Organization | Forte - www.forteinc.com |
| Date | 2023-06-21 22:47 -0400 |
Cross-posted to 3 groups.
On 6/21/23 8:40 PM, olcott wrote: > On 6/21/2023 6:01 PM, Richard Damon wrote: >> On 6/21/23 3:59 PM, olcott wrote: >>> On 6/21/2023 2:10 PM, vallor wrote: >>>> On Sat, 17 Jun 2023 00:54:32 -0500, olcott wrote: >>>> >>>>> ChatGPT: >>>>> “Therefore, based on the understanding that self-contradictory >>>>> questions lack a correct answer and are deemed incorrect, one >>>>> could >>>>> argue that the halting problem's pathological input D can be >>>>> categorized as an incorrect question when posed to the halting >>>>> decider H.” >>>>> >>>>> https://chat.openai.com/c/2aae46ef-e7be-444d-a046-b76c1f971c5a It did >>>>> not leap to this conclusion it took a lot of convincing. >>>> >>>> Chatbots are highly unreliable at reasoning. They are designed >>>> to give you the illusion that they know what they're talking about, >>>> but they are the world's best BS artists. >>>> >>>> (Try playing a game of chess with ChatGPT, you'll see what I mean.) >>>> >>> >>> I already know that and much worse than that they simply make up facts >>> on the fly citing purely fictional textbooks that have photos and back >>> stories for the purely fictional authors. The fake textbooks themselves >>> are complete and convincing. >>> >>> In my case ChatGPT was able to be convinced by clearly correct >>> reasoning. >>> >> >> So, you admit that they will lie and tell you want you want to hear, >> you think the fact that it agrees with you means something? >> >>> https://chat.openai.com/c/2aae46ef-e7be-444d-a046-b76c1f971c5a >>> It did not leap to this conclusion it took a lot of convincing. >> >> Which is a good sign that it was learnig what you wanted it to say so >> it finally said it. >> >>> >>> People are not convinced by this same reasoning only because they spend >>> 99.9% of their attention on rebuttal thus there is not enough attention >>> left over for comprehension. >> >> No, people can apply REAL "Correct Reasoning" and see the error in >> what you call "Correct Reasoning". Your problem is that your idea of >> correct isn't. >> >>> >>> The only reason that the halting problem cannot be solved is that the >>> halting question is phrased incorrectly. The way that the halting >>> problem is phrased allows inputs that contradict every Boolean return >>> value from a set of specific deciders. >> >> Nope, it is phrased exactly as needed. Your alterations allow the >> decider to give false answer and still be considered "correct" by your >> faulty logic. >> >>> >>> Each of the halting problems instances is exactly isomorphic to >>> requiring a correct answer to this question: >>> Is this sentence true or false: "This sentence is not true". >>> >> >> Nope. >> >> How is "Does the Machine represented by the input to the decider?" >> isomopric to your statement. >> > > The halting problem instances that ask: > "Does this input halt" > > are isomorphic to asking Jack this question: > "Will Jack's answer to this question be no?" Nope, because Jack is a volitional being, so we CAN'T know the correct answer to the question until after Jack answers the question, thus Jack, in trying to be correct, hits a contradiction. The correct answer to the Halting Problem Question was avaiable as soon as the machine being asked about was defined, so the decider doesn't hit a contradiction in logic, it just is wrong, because it CAN'T "try" to give the other answer, because it just does as it was programmed. All your logic is in designing the machine, and there the contradiction just points out that you can't make a correct machine, which is an acceptable answer. Not all problems are computable, so we can't always make a machine give the answer. > > Which are both isomorphic to asking if this expression > is true or false: "This sentence is not true" Nope. Show how the CAN be. The Halting Problem ALWAYS has a valid yes or no question, since the machine it is being asked on must be defined to ask it, and thus its behavior is FIXED by its code. You just don't seem to understand what a program is, so I guess you faked it when you were working as a programmer. > > That you are unwilling to validate my work merely means that > someone else will get the credit for validating my work. > > I can't "Validate" your work, as it is just incorrect. You think to things of different kind are the same, which is impossible, so your statements are just incorrect. You don't seem to understand that compuations don't have volition, so, you basically don't understand what a computation is at all, and nothing you have done reguarding them has any hope of having a factual basis. You also clearly don't understand how logic works too.
Back to comp.ai.philosophy | Previous | Next — Previous in thread | Next in thread | Find similar
ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-17 00:54 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-17 08:09 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-17 11:59 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-17 13:43 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-17 13:23 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-17 16:27 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Ben Bacarisse <ben.usenet@bsb.me.uk> - 2023-06-17 22:09 +0100
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-17 16:46 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Jeff Barnett <jbb@notatt.com> - 2023-06-17 16:03 -0600
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-17 19:18 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-17 18:44 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-17 21:46 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-17 21:35 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-17 23:03 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-17 19:13 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-17 18:58 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-17 21:31 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-17 21:29 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-17 22:57 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-17 22:10 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-18 08:02 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-18 09:32 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-18 12:31 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-18 11:41 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-18 12:54 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-18 12:09 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-18 13:46 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-18 13:05 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-18 14:20 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-18 13:30 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-18 14:43 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-18 13:47 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-18 15:19 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-18 14:26 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-18 16:10 -0400
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-18 18:43 -0500
Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-18 19:59 -0400
Does input D have semantic property S or is input D [BAD INPUT]? olcott <polcott2@gmail.com> - 2023-06-18 22:31 -0500
Re: Does input D have semantic property S or is input D [BAD INPUT]? Richard Damon <Richard@Damon-Family.org> - 2023-06-19 07:38 -0400
Re: Does input D have semantic property S or is input D [BAD INPUT]? olcott <polcott2@gmail.com> - 2023-06-19 09:30 -0500
Re: Does input D have semantic property S or is input D [BAD INPUT]? Richard Damon <Richard@Damon-Family.org> - 2023-06-19 20:45 -0400
Re: Does input D have semantic property S or is input D [BAD INPUT]? olcott <polcott2@gmail.com> - 2023-06-19 22:57 -0500
Re: Does input D have semantic property S or is input D [BAD INPUT]? Don Stockbauer <donstockbauer@hotmail.com> - 2023-06-20 00:33 -0700
ChatGPT discussion (was: Re: Does input D have semantic property S or is input D [BAD INPUT]? vallor <vallor@vallor.earth> - 2023-06-20 11:16 +0000
Re: Does input D have semantic property S or is input D [BAD INPUT]? Richard Damon <Richard@Damon-Family.org> - 2023-06-20 07:19 -0400
Re: Does input D have semantic property S or is input D [BAD INPUT]? olcott <polcott2@gmail.com> - 2023-06-20 10:09 -0500
Re: Does input D have semantic property S or is input D [BAD INPUT]? Richard Damon <Richard@Damon-Family.org> - 2023-06-20 11:48 -0400
Ben Bacarisse specifically targets my posts to discourage honest dialogue olcott <polcott2@gmail.com> - 2023-06-20 10:06 -0500
Re: Ben Bacarisse specifically targets my posts to discourage honest dialogue Richard Damon <Richard@Damon-Family.org> - 2023-06-20 11:48 -0400
Re: dishonest subject lines Ben Bacarisse <ben.usenet@bsb.me.uk> - 2023-06-20 17:02 +0100
Ben Bacarisse specifically targets my posts to discourage honest dialogue olcott <polcott2@gmail.com> - 2023-06-20 12:25 -0500
Refutation of the Ben Bacarisse Rebuttal [Ben targets my posts to discourage honest dialogue] olcott <polcott2@gmail.com> - 2023-06-20 14:57 -0500
Re: Refutation of the Ben Bacarisse Rebuttal [Ben targets my posts to discourage honest dialogue] Richard Damon <Richard@Damon-Family.org> - 2023-06-20 16:34 -0400
Re: Refutation of the Ben Bacarisse Rebuttal [Ben targets my posts to discourage honest dialogue] olcott <polcott2@gmail.com> - 2023-06-20 15:42 -0500
Re: Refutation of the Ben Bacarisse Rebuttal [Ben targets my posts to discourage honest dialogue] Richard Damon <Richard@Damon-Family.org> - 2023-06-20 16:52 -0400
Re: Refutation of the Ben Bacarisse Rebuttal [Ben targets my posts to discourage honest dialogue] olcott <polcott2@gmail.com> - 2023-06-20 16:39 -0500
Re: Refutation of the Ben Bacarisse Rebuttal [Ben targets my posts to discourage honest dialogue] Richard Damon <Richard@Damon-Family.org> - 2023-06-20 17:53 -0400
Re: Refutation of the Ben Bacarisse Rebuttal [Ben targets my posts to discourage honest dialogue] olcott <polcott2@gmail.com> - 2023-06-20 17:07 -0500
Re: Refutation of the Ben Bacarisse Rebuttal [Ben targets my posts to discourage honest dialogue] Richard Damon <Richard@Damon-Family.org> - 2023-06-20 18:52 -0400
Refutation of the Ben Bacarisse Rebuttal [Ben targets my posts] olcott <polcott2@gmail.com> - 2023-06-20 14:59 -0500
Refutation of the Ben Bacarisse Rebuttal [Ben targets my posts] olcott <polcott2@gmail.com> - 2023-06-20 15:00 -0500
ChatGPT and stack limits (was: Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question vallor <vallor@cultnix.org> - 2023-06-21 19:10 +0000
Re: ChatGPT and stack limits (was: Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question vallor <vallor@vallor.earth> - 2023-06-21 19:23 +0000
Re: ChatGPT and stack limits (was: Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-21 14:59 -0500
Re: ChatGPT and stack limits (was: Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-21 19:01 -0400
Re: ChatGPT and stack limits (was: Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-21 19:40 -0500
Re: ChatGPT and stack limits (was: Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-21 22:47 -0400
Re: ChatGPT and stack limits (was: Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-21 21:58 -0500
Re: ChatGPT and stack limits (was: Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-22 07:26 -0400
Re: ChatGPT and stack limits (was: Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question olcott <polcott2@gmail.com> - 2023-06-22 09:18 -0500
Re: ChatGPT and stack limits (was: Re: ChatGPT agrees that the halting problem input can be construed as an incorrect question Richard Damon <Richard@Damon-Family.org> - 2023-06-22 21:06 -0400
csiph-web