Talk:Artificial intelligence/Archive 13

AI systems are heuristics, not algorithms

It should be noted that AI systems are not algorithms with known results, they are heuristics that approximate the solution. AI is used when complete analysis can be done are rare. AI is used when the input space is large and the decisions hard to make. The neural network or other methods approximate the solution but that solution is approximate as it does not cover all use cases. AI should be treated as a heuristic that gets one closer to the solution but not all the way there. It should not be used to drive cars, in hiring or in healthcare. Those fields are too critical for approximations.

This was posted by 198.103.184.76 North8000 (talk) 15:39, 20 January 2022 (UTC)

"Natural stupidity" listed at Redirects for discussion

  An editor has identified a potential problem with the redirect Natural stupidity and has thus listed it for discussion. This discussion will occur at Wikipedia:Redirects for discussion/Log/2022 January 27#Natural stupidity until a consensus is reached, and readers of this page are welcome to contribute to the discussion. signed, Rosguill talk 20:40, 27 January 2022 (UTC)

So, fuzzy logic is the same as artificial intelligence?

Over at Fuzzy logic#Artificial intelligence, it (currently) says:

AI and fuzzy logic, when analyzed, are the same thing — the underlying logic of neural networks is fuzzy.

Maybe somebody here can improve that section of Fuzzy logic. --R. S. Shaw (talk) 04:21, 16 February 2022 (UTC)

This is a very long article that I really like. Thanks to whoever created this article about AI. Note: There is only AI that controls self-driving-cars like a Tesla. I wonder when AI will control everything. Antiesten (talk) 23:42, 22 March 2022 (UTC)

Where did it go? (On the big copy-edit in the fall of 2021)

This summer and fall, I have copy-edited the entire article for brevity (as well as better organization, citation format, and a non-technical WP:AUDIENCE). The article is down from its peak of 34 text pages down to about 21 or so. Most of this savings was from copy-editing for tighter prose and better organization, but there was a good deal of stuff that was cut. I tried to move as much material as I could down sub-articles like existential risk of AI or machine learning and so on. I've documented exactly where everything I cut has been moved to, and indicated the things I couldn't find a place for (or were otherwise unusable). You can see exactly where this material went here: Talk:Artificial intelligence/Where did it go? 2021. ---- CharlesGillingham (talk) 00:52, 14 October 2021 (UTC)

Thanks for your hard-work. I think many of these topics are related to AI only remotely. AXONOV (talk) 18:51, 16 October 2021 (UTC)

Semi-protected edit request on 29 April 2022

hey i found some extra information i would like to add 12.96.155.31 (talk) 16:58, 29 April 2022 (UTC)

  Not done: it's not clear what changes you want to be made. Please mention the specific changes in a "change X to Y" format and provide a reliable source if appropriate. Cannolis (talk) 17:16, 29 April 2022 (UTC)
You should consider this a request to remove semi-protect. Rklawton (talk) 02:06, 16 May 2022 (UTC)

Google engineers Blaise Agüera y Arcas's and Blake Lemoine's claims about the Google LaMDA chatbot

Not sure if this will gain any traction or get wider spread attention. I believe this Washington Post article and this Economist article are the first mainstream discussions of it. Not saying I personally give it any credibility but it is interesting. If this shows up in any more publications might it be fit for inclusion, or is this just WP:RECENTISM trivia? —DIYeditor (talk) 21:50, 11 June 2022 (UTC)

This got more coverage on the 12th. I guess this would also be relevant to Turing test if it proves enduring. —DIYeditor (talk) 03:27, 13 June 2022 (UTC)

Chatbot, Applications of AI and Turing test are all better places than this article, which is a summary article and is already too long. ---- CharlesTGillingham (talk) 00:47, 12 July 2022 (UTC)

Copyedit

Added a comma to sentence:

' Philosopher Nick Bostrom argues that sufficiently intelligent AI if it chooses actions based on achieving some goal, will exhibit convergent behavior such as acquiring resources or protecting itself from being shut down.'

between "AI" and "if" to improve flow and grammar. Please correct if mistaken, thank you! King keudo (talk) 20:51, 21 September 2022 (UTC)

  Resolved by somebody since this request was made. --Lord Belbury (talk) 09:55, 17 October 2022 (UTC)

Russell definition of AI excludes major fields: CV, transcription, etc.

This article has again been rewritten by someone to once again narrowly define AI as: only autonomous agents are AI. This is based on the Russell definition, which is highly controversial, if not almost generally rejected. This article has been repeatedly been sabotaged by ABM, robotics, killer drones, etc. advocates to narrowly define AI as interactive agents, thereby excluding some of the major key fields of actual AI such as computer vision, speech recognition/transcription, machine translation.

The trick being used to misdefine seem to confuse between AI and AI-systems/AI-based systems/etc.: the former synthesized information, the latter includes an AI component, but also wrongly includes purely procedural steps that have no intelligence to them. A typical misdefinition seems to go like:

AI is difficult to define, AI-based system are things interact with their environment

Google gives this:

the theory and development of computer systems able to perform tasks that normally require human intelligence, such as visual perception, speech recognition, decision-making, and translation between languages.

The explanation of their use of Oxford for all definitions is given here: https://support.google.com/websearch/answer/10106608?hl=en

This article needs a definition that recognizes these major fields like CV and speech recognition as being AI (i.e. not being a part of AI). The productive way is probably to early on state that AI is often encountered in everyday life as part of larger AI-based systems, which can also include procedural components. Bquast (talk) 18:19, 23 October 2022 (UTC)

Further research: the Russel and Norvig "definition" in fact does the same trick as I mentioned above...It states it is difficult to define AI and proceeds that it is easier to work with a definition of AI-based system / agents, etc.
This is NOT a definition of AI, and the definition of another concept should not be used here. This Wikipedia article should define the exact concept of AI, not how AI is used. The "engine" article also doesn't describe how political scientists and lawyers should think about cars.
Similarly, we don't say, human intelligence is hard to define, but a human is a bag of blood and bones that responds to inputs in various ways. Or something along those lines.
Proposed next steps:
  1. Add a link at the top to "Autonomous agents" article
  2. See what references the Oxford dictionary has for its definition
  3. See a definition from the Dartmouth conference on AI
  4. Replace the Russel and Norvig AI-based agents definition with the Oxford definition of Artificial # Intelligence, include in brackets a link to autonomous agents
  5. Revise subsequent paragraphs accordingly where needed
Bquast (talk) 03:35, 27 October 2022 (UTC)

Add a figure with a framework for artificial intelligence in enterprise applications

Being a scientific researcher, I am new to editing Wikipedia. Can you help me, please? I propose to add a unified framework for "Artificial Intelligence in Enterprise Applications". The framework has recently been published in a peer-reviewed, high-quality scientific journal (Scimago Q1), refer to https://www.sciencedirect.com/science/article/pii/S0923474822000467. I am the author of that article and declare a conflict of interest as I am related to the AI article on Wikipedia as a researcher. Specifically, I wanted to contribute my framework's visualization/figure (refer to Figure 6 at the end of the journal article) and an explanatory paragraph for the following reasons: 1) To add further clarity to the current Wikipedia article by depicting the interrelationships of various AI subfields in a visualization/graphic form, and 2) in the proposed explanatory paragraph include cross-links for these subfields to their corresponding areas on Wikipedia. The framework does not contradict anything in the existing Wikipedia article. I published my research article as Open Access and have approval from the publisher to contribute my framework to Wikipedia. Kind regards, Heinzhausw (talk) 06:13, 31 October 2022 (UTC)

References

Copyedit

Under Tools, the first line contains a misplaced modifier. "Many problems in AI can be solved theoretically by intelligently searching through many possible solutions..." The line should probably read: "AI can solve many problems theoretically by intelligently searching through many possible solutions..." LBirdy (talk) 16:02, 5 November 2022 (UTC)

Thank you. I've just changed this to "AI can solve many problems by intelligently searching through many possible solutions." Elspea756 (talk) 16:57, 5 November 2022 (UTC)

implement comment: too many sections, remove intelligent agent section

there has for a long time been an inline comment to remove some sections, there are too many in this article.

I suggest to remove the talk about intelligent agents, it is highly confusing (not in the least because this article was not very accurate with this before), and it does not belong here, there already is an article on intelligent agent. Bquast (talk) 01:41, 17 November 2022 (UTC)

Re: definition of AI

@Bquast: I'm fine with reframing the definition without the term "intelligent agents". This term's popularity peaked back around 2000 or so. A good reworking might even make the underlying philosophical points more clear.

I would be fine with McCarthy's definition, i.e. "Intelligence is the computational part of the ability to achieve goals in the world." (You mentioned above that you would be okay with the definition of AI proposed at The Dartmouth Conference, but I don't believe they made a formal definition -- I assume you had in mind McCarthy's understanding of the term.)

There are several essential elements to the academic definition of AI (as opposed to definitions from popular sources, or dictionaries):

  1. It must be in terms of behavior; it's something it does, not something it is. (That was Turing's main point.)
  2. It must not be in terms of human intelligence. (People like McCarthy have vociferously argued against this.)
  3. It must in terms of goal-directed behavior -- what economists call "rationality". In other words, in terms of well-defined problems with well-defined solutions.

R & N's chapter 2 definition uses a four way categorization: "Thinking humanly", 'acting humanly", "Thinking rationally", "Acting rationally". This is a good way to frame these issues. Two orthogonal dimensions thinking vs. acting, human-like vs goal-directed. ---- CharlesTGillingham (talk) 06:28, 26 November 2022 (UTC)


Oh, and one last thing, which often needs to be said on this page:

There are many, many contradictory sources on AI, whole communities of thinkers who have their own understanding of AI, and many thousands of individual writers who have tried their hand at defining it or re-defining it. The article relies heavily on Russell & Norvig's textbook, in many places, because it is by far the most popular textbook, used in literally thousands of introductory AI courses for almost thirty years now. From Wikipedia's point of view, R & N is the most reliable source we could cite on the topic.

And a parenthetical comment:

By the way, R & N defines "agent" as: "something that perceives and acts", i.e. "something with inputs and outputs". Autonomy or persistence is not a part of their discussion. Any program, any program at all, fits their definition of an "agent". ---- CharlesTGillingham (talk) 06:28, 26 November 2022 (UTC)

Semi-protected edit request on 13 November 2022

I'm requesting to add a section under "risks" of artificial intelligence

Gender Bias in Artificial Intelligence: As artificial intelligence continues to evolve and learn, it’s important to address the fact that the field of AI is extremely male dominated and how that impacts the way AI is learning language and values. In an article written by Susan Leavy from University College Dublin, she talks about the existence of the language used when referencing male and female roles. For example: the term “man-kind” and “man” referring to all of humanity, work roles such as firefighters being seen as a male role, and the words used to describe family such as how a father would be seen as a “family man” and that women don’t have an equal term. If these societal norms aren’t challenged throughout the advancement of AI, then the small ways that language differs between genders will be embedded into the AI’s memory and further reinforce gender inequality for future generations.

Leavy, Susan. “Gender Bias in Artificial Intelligence: Proceedings of the 1st International Workshop on Gender Equality in Software Engineering.” ACM Digital Library, 28 May 2018, https://dl.acm.org/doi/pdf/10.1145/3195570.3195580. Kawahsaki (talk) 19:57, 13 November 2022 (UTC)

  Not done: Hello Kawahsaki, and welcome to Wikipedia! I'm afraid I have to decline to perform this request for a couple of reasons.
When creating edit requests one of the conditions for it being successful is that it be uncontroversial. Gender bias as a topic in whole is certainly controversial in the world today, and so the creation of an entire section based on such a topic would be out of scope here.
Additionally, I have concerns regarding the prose you've written. Wikipedia strives to maintain a neutral point of view when describing topics; our sole goal is to describe what reliable independent sources say on a given topic. This is because we are a tertiary source. Some of your prose seems to fall below this guideline. An example is the phrase it is important to address the fact. Wikipedia may state that a source believes something is important, but Wikipedia would not say something like this in it's own voice.
Now, this page is currently under what we call semi-protection. This means that only editors with accounts that are 3 days old and have a total of 10 edits may edit the page. If you make 9 more editors anywhere on Wikipedia (and there are plenty of eligible pages), and wait until November 16th, you'll be able to edit this page directly.
Feel free to drop by my talk page (Wikipedia's version of direct messages) if you have any questions, or you can ask them at the Teahouse, which is a venue that specializes in answering questions from new editors.
Cheers, and happy editing! —Sirdog (talk) 04:26, 14 November 2022 (UTC)
I think you could add your contribution to the main article on algorithmic bias. This article only has room for a paragraph or so on the topic. ---- CharlesTGillingham (talk) 06:40, 28 November 2022 (UTC)

Why the Oxford Dictionary definition is inadequate

The article currently quotes the Oxford dictionary to define AI: "the theory and development of computer systems able to perform tasks that normally require human intelligence, such as visual perception, speech recognition, decision-making, and translation between languages."

This definition is rejected by the leading AI textbook (see Chpt. 2, Artificial Intelligence: A Modern Approach) and by AI founder John McCarthy (who coined the term "artificial intelligence") (see multiple citations in the article; just search for his name)

A brief introduction to the problems with definition:

The problem is this phrase: "tasks that normally require human intelligence". Consider these two lists:

Tasks that require considerable human intelligence:

  • Multiplying large numbers.
  • Memorizing long lists of information.
  • Doing high school algebra
  • Solving a linear differential equation
  • Playing chess at a beginner's level.

Tasks that do not require human intelligence (i.e. "unintelligent" small children or animals can do it):

  • Facial recognition
  • Visual perception
  • Speech recognition
  • Walking from room to room without bumping into something
  • Picking up an egg without breaking it
  • Noticing who is speaking

The Oxford definition categorizes programs that can do tasks from list 1 as AI, and categorize programs from list 2 as being outside of AI's scope. This is obviously not what is actually happening out in the field -- exactly the opposite, in most cases. All of the problems in list 1 were solved back in the 1960s, with computers far less powerful than the one in your microwave or clock radio. The problems in list 2 have only been solved recently, if at all.

Activities considered "intelligent" when a human does them can sometimes be relatively easy for machines, and sometimes activities that would never appear particularly "intelligent" when a human does them can be incredibly difficult for machines. (See Moravec's paradox) Thus the definition of artificial intelligence can't just be in terms of "human intelligence" -- a more general definition is needed. The Oxford dictionary definition is not adequate.

My recommendation

Scrap the extended definition all together: just stick with the naive common usage definition. Go directly to the examples (i.e. paragraph two of the lede)

Leave the difficult problem of defining "intelligence" (without reference to human intelligence) to the section "Defining AI" deeper in the article. This section considers the major issues, and should settle on "rationality" (i.e. goal-directed behavior) as Russell and Norvig do, and as John McCarthy did.---- CharlesTGillingham (talk) 04:31, 28 November 2022 (UTC)

Actually, I just noticed, it doesn't exist any more! I will restore this very brief philosophical discussion, without any mention of "intelligent agents". And I will leave Google's definition as well. ---- CharlesTGillingham (talk) 04:53, 28 November 2022 (UTC)
please review the history, the Russell definition was moved to the intelligent agent article, it is not adequate for artificial intelligence because it includes all kinds of procedural actions that are of interesting to fields like political science, but are not the essence of AI itself Bquast (talk) 14:29, 30 November 2022 (UTC)
@Bquast: I have not re-added intelligent agents. I agreed with your proposal of eliminating "intelligent agent" from this article entirely. I re-added criticism of the Turing test and of human simulation as a definition of AI. I have restored it, if that's okay with you. ---- CharlesTGillingham (talk) 04:31, 4 December 2022 (UTC)
@Bquast: By the way, I noticed the Google definition doesn't have a working citation, and I can't seem to find it. Would you mind fixing that? ---- CharlesTGillingham (talk) 05:22, 4 December 2022 (UTC)
@CharlesTGillingham ok, sorry then I misunderstood your intention. In general I agree that this current definition is not good. Intelilgence can be human or animal (or plants?). I'm not sure about your list, many of the "dumb" tasks do require intelligence, I would not consider facial recognition _not_ intelilgence.
Regarding the link from Google, I put the direct citation of OED, but you can find it like this: https://www.google.com/search?q=artificial+intelligence+definition I will try to add it soon Bquast (talk) 02:49, 6 December 2022 (UTC)

References, further reading, notes, etc. cleanup

A major cleanup is needed of all these sections. It seems like many authors have inserted their won (maybe) relevant material here. It should contain references of the text used. It should also avoid mentioning the same references in many different places, in particular the confusing Russell and Norvig book. Bquast (talk) 14:32, 30 November 2022 (UTC)

Articles in this area are prone to reference spamming. I've done some work on this at related articles but not on this one. Also keeping a watch on a this and related articles so that it doesn't get worse. North8000 (talk) 21:55, 30 November 2022 (UTC)
Wikipedia requires reliable sources. An article should include only citations to the most reliable sources as possible. These is no reason to include more references to less reliable sources, or to exclude references to the most reliable sources.
There is no more reliable source about AI than Russell and Norvig, the leading textbook, used in thousands of introductory university courses about AI. There is vast body of less reliable sources about AI. There is a lot of dissent, new ideas, outsider perspectives, home brews, sloppy journalism, self-promotion and so on. Wikipedia has to take a NPOV on this huge variety, and we don't have room to cover them all. Thus we, as editors, need to prove that every contribution reflects "mainstream" and "consensus" views on the subject. This is all we have room for. This all that is relevant here. The dozens of citations in this article to the leading text book are a way of showing that each contribution is mainstream and consensus, and a way of weeding out the fringe. ---- CharlesTGillingham (talk) 04:55, 4 December 2022 (UTC)
Please take care that cites you remove are not still in use by referencing. Removing cite that have short form references causes "no target errors". -- LCU ActivelyDisinterested transmissions °co-ords° 09:37, 8 December 2022 (UTC)

A Commons file used on this page or its Wikidata item has been nominated for deletion

The following Wikimedia Commons file used on this page or its Wikidata item has been nominated for deletion:

Participate in the deletion discussion at the nomination page. —Community Tech bot (talk) 22:24, 15 March 2023 (UTC)

Not used here. CharlesTGillingham (talk) 10:10, 23 March 2023 (UTC)

Wiki Education assignment: Research Process and Methodology - SP23 - Sect 201 - Thu

  This article was the subject of a Wiki Education Foundation-supported course assignment, between 25 January 2023 and 5 May 2023. Further details are available on the course page. Student editor(s): Liliability (article contribs).

— Assignment last updated by Liliability (talk) 03:41, 13 April 2023 (UTC)

Future

In the "Future - Technological unemployment" section, would it be appropriate to add a clarifying statement to the quote, "...but they generally agree that it could be a net benefit if productivity gains are redistributed." With how it's presented, there is explicit reasoning that productivity gains would be seen by displaced workers receiving the monetary excess generated by AI's labor. However, this source is a survey of economics professors. Not business leaders speaking on affected industries and not sociologists speaking on affected workers. As a professional writer, presenting a quote like that from experts in a different field feels like an intentional misrepresentation.

Newer and older articles take a different tack, speculating that productivity gains would be seen in industries receiving displaced workers. Elsewhere, it's predicted that productivity gains would be seen from knowledge workers that learn or are able to augment their work with AI as it presents the opportunity to handle repetitive tasks.

Anecdotally, I use AI as an editor and it has tripled my productivity as a writer, which has given me time to edit Wikipedia articles. Software developers with whom I work have announced similar results, without mention of Wikipedia. In that regard, the section on technological unemployment speaks more to the AI boogeyman than it does potential benefit, and I think we should fix that.

NOTE: I am not an AI nor am I employed by an AI or an AI developer. I have no stake in AI and no more interest than ensuring an accurate reporting of the facts. Oleanderyogurt (talk) 00:03, 18 April 2023 (UTC)

I agree that the current sentence is for many reasons problematic. IMO it would be best to simply remove it. North8000 (talk) 21:10, 18 April 2023 (UTC)

Infobox

This article needs an infobox, it could be there general infobox template, or a specific one. Technology standard is a common one, but standard is not correct. Maybe scientific domain or something. What does everyone think? Bquast (talk) 16:24, 21 April 2023 (UTC)

IMHO we're better off without it. I foresee endless problems trying to decide what to put into it for such a broad vaguely defined topic and not much value to what we do put in there. Sincerely, North8000 (talk) 17:20, 21 April 2023 (UTC)

"Tools" section should contain a "machine learning" subsection

I believe machine learning is part of AI and the "Tools" section should contain a subsection named "machine learning methods".

However, currently under the "Tools" section, there is only a subsection named "Classifiers and statistical learning methods". "Classification" is just one task of supervised learning, which is one type of machine learning. Also, not all machine learning methods are statistical.

Changing "classifiers and statistical learning methods" to "machine learning methods" can also make the title simpler and easier to understand.

@CharlesGillingham @CharlesTGillingham

Cooper2222 (talk) 21:40, 16 April 2023 (UTC)

There are many ways to organize this section. The idea was to list the tools without worrying about what they are used for, because in many cases, a particular tool can be used for many different things. This is kind of obvious with Search, Logic and ANNs.
All the things listed there (decision tree, nearest neighbor, Kernel methods, SVM, naive Bayes) are "classifiers" that were developed in the language of the statistics literature (in the 90s) and were mostly applied to machine learning. However, they are also tools for data science and statistical analysis. (Or, at very least, they share a lot in common with other statistical tools.)
Thus, I like the word "statistics" or "statistical" in the title. These are statistical tools. I would be more inclined to strike the "machine learning" part of the title -- we already have a section on machine learning above.
But feel free to be be bold and re-title or reorganize. ---- CharlesTGillingham (talk) 01:48, 27 April 2023 (UTC)
Originally I didn't consider models like k-NN to be statistical, because they are not based on probability. But you said these models all came from statistics. If we consider all these models to be statistical, what is the difference between statistical learning and machine learning? ---- Cooper2222 (talk) 03:30, 28 April 2023 (UTC)
Well, for me, all these tools are all somewhere near the border between AI and statistics, regardless of whether they are generally considered to be inside or outside of AI. It's the shared mathematical language, the way the problems are framed, and the precise way solutions can be judged and measured. All of that comes from statistics, not from previous AI research. ---- CharlesTGillingham (talk) 02:54, 2 May 2023 (UTC)

Sentence cut

I cut this, because it is at the wrong level of detail for the lede (which should primarily be a summary of the contents of the article). Not sure where to move it to, so I put it here for now ---- CharlesTGillingham (talk) 17:42, 1 July 2023 (UTC)

This has changed the purchasing process, being the AI application functions a mediator between the consumer, product, and brand by providing personalized recommendations based on previous consumer purchasing decisions.[1]

References

  1. ^ Curtis, Lee (June 2020). "Trademark Law Playing Catch-up with Artificial Intelligence?". WIPO Magazine.

CharlesTGillingham (talk) 17:42, 1 July 2023 (UTC)

More material temporarily placed here

I cut this from AI § history, for several reasons:

  1. Brevity
  2. I think it reads better if this section is just a social history of AI, and doesn't address technical history or arguable historical interpretation.
  3. The linked article (symbolic AI) has been rewritten to describe a slightly different subject, so links from here are misleading.

It's going to take me some research on how Wikipedia should address the terminological issue of "symbolic AI" vs. "GOFAI" (don't worry about it if you don't know what that is). To keep moving forward, I will just park this stuff here.

By the 1950s, two visions for how to achieve machine intelligence emerged. One vision, known as Symbolic AI or GOFAI, was to use computers to create a symbolic representation of the world and systems that could reason about the world. Proponents included Allen Newell, Herbert A. Simon, and Marvin Minsky. Closely associated with this approach was the "heuristic search" approach, which likened intelligence to a problem of exploring a space of possibilities for answers.

The second vision, known as the connectionist approach, sought to achieve intelligence through learning. Proponents of this approach, most prominently Frank Rosenblatt, sought to connect Perceptron in ways inspired by connections of neurons.[1] James Manyika and others have compared the two approaches to the mind (Symbolic AI) and the brain (connectionist). Manyika argues that symbolic approaches dominated the push for artificial intelligence in this period, due in part to its connection to intellectual traditions of Descartes, Boole, Gottlob Frege, Bertrand Russell, and others. Connectionist approaches based on cybernetics or artificial neural networks were pushed to the background but have gained new prominence in recent decades.[2]

References

  1. ^ Manyika 2022, p. 9.
  2. ^ Manyika 2022, p. 10.

CharlesTGillingham (talk) 18:58, 2 July 2023 (UTC)

Recent changes to intro

@DancingPhilosopher: maybe it is just me, but I can barely understand the new intro. I think the old version was clearer and more accessible. Could you please try to make it more accessible? Vpab15 (talk) 14:43, 18 July 2023 (UTC)

The lede has briefly read:
Artificial intelligence (AI) is firstly an academic discipline with various, often conflicting, views on what constitutes its area of research, as well as goals and approaches used, including logical, knowledge-based approach,[1] on one hand, and machine learning approach, on the other. When probabilistic systems were plagued by theoretical and practical problems of data acquisition and representation,[2] symbolic/knowledge-based approach prevailed and artificial neural networks research had been abandoned by AI and continued outside the AI, as "connectionism", by researchers from other disciplines including Hopfield, Rumelhart, and Hinton.
This is at the wrong level of detail for the lede, a caricature of the conflicts between various approaches, and it's just not true that "artificial neural networks research [has] been abandoned" ---- CharlesTGillingham (talk) 10:18, 19 July 2023 (UTC) CharlesTGillingham (talk) 10:18, 19 July 2023 (UTC)
Only concerning the level of detail: Perhaps we should create an introductory overview article about AI if a comprehensive article is too complex for lay readers. Because I consider the wording of the mentioned lead section to be more accurate than simply stating that AI is intelligence demonstrated by machines, though I do understand that it may be too technical for lay readers. Maxeto0910 (talk) 23:34, 19 July 2023 (UTC)
This is an introductory article. The first line is the most introductory part. The first line should be a definition for someone who has never heard of AI. We have to consider the usage of AI for the whole period 1956-present, primarily non-academic uses such as the popular press or a Google search by a 4th grader. The more specific the definition becomes, the less useful it is for this purpose. The intent here is to use the most general and inarguable definition.
The definition is not intended to describe only the specific state of academic field in the 21st century. (And, I should note, we have struggled for decades to avoid using a definition that raises unsolved philosophical issues.) ---- CharlesTGillingham (talk) 16:24, 26 July 2023 (UTC)

I agree, the version in the box goes off into the weeds in so many directions that it does not communicate the essentials. Thanks for the effort though. Sincerely, North8000 (talk) 19:51, 26 July 2023 (UTC)

Moved some material to the right level of detail

@DancingPhilosopher: I'm moving this contribution to progress in AI, which is the article on this topic

Humans still substantially outperform both GPT-4 and models trained on the ConceptARC benchmark that scored 60% on most, and 77% on one category, while humans 91% on all and 97% on one category.[1]
  1. ^ Biever, Celeste (25 July 2023). "ChatGPT broke the Turing test — the race is on for new ways to assess AI". Nature (journal). Retrieved 26 July 2023.{{cite news}}: CS1 maint: url-status (link)

--- CharlesTGillingham (talk) 21:25, 26 July 2023 (UTC) CharlesTGillingham (talk) 21:25, 26 July 2023 (UTC)

using a non-circular (and more corrent) short description

The current short description is:

Intelligence demonstrated by machines

This is:

  1. circular
  2. machines is too narrow, more to do with machine learning.

Regarding #2, I can "do" AI, e.g. a FF neural network on paper, that is a much AI en silica.

I propose to use (in line with the main body text):

The ability of systems to perceive, synthesize, and infer information

Bquast (talk) Bquast (talk) 15:41, 15 November 2022 (UTC)

To answer your criticism:
  1. Don't think it's circular -- it just assumes the reader already knows what intelligence is. E.g., defining "puppy dog" as "a young dog".
  2. How about replacing "machines" with "machines or software"?
On you definition: "perceive, synthesize, infer" ... hmm ... you left out "learn" ... and "knowledge" ... but, frankly, intelligence is so notoriously difficult to define that we're just opening a can of worms trying to define it here -- you'll have the whole history of philosophy and psychology picking away at you. Better to just leave it out.
My two cents. ---- CharlesTGillingham (talk) 06:24, 28 November 2022 (UTC)
I don't think the definition is circular. Intelligence is a general word not related to artificial intelligence. For "machines", I don't think it's narrow. Currently all AI is done on computers. I don't think you can do a neural network on paper. Cooper2222 (talk) 21:55, 16 April 2023 (UTC)
@Bquast do you have a source for the definition? I'm fine with leaving it in if we can source it, otherwise I put in "to learn and to reason, to generalize, and to infer meaning" because that's close to how the Encyclopedia Britannica characterizes it, and IMHO is better at communicating what sort of things AI researchers (as opposed to other computer science researchers) work on. Rolf H Nelson (talk) 03:35, 21 June 2023 (UTC)
I"m happier with this than the previous. ---- CharlesTGillingham (talk) 16:16, 27 June 2023 (UTC)
I fixed the indents in this conversation, which was driving me crazy. ---- CharlesTGillingham (talk) 22:14, 28 July 2023 (UTC)

Undue weight on AI patents

I am certain that this section doesn't belong here. It's WP:undue weight for a summary article like this with so much to cover. It's very solidly written, accurate and well-sourced, and ideally I would like to find a place for it somewhere else in Wikipedia, but I'm stumped as to where to move it. Does anyone have any ideas? Maybe this should be a short article of its own? Or part of a new stub about ... what? ---- CharlesTGillingham (talk) 05:26, 29 July 2023 (UTC)

 
AI patent families for functional application categories and sub categories. Computer vision represents 49 percent of patent families related to a functional application in 2016.

In 2019, WIPO reported that AI was the most prolific emerging technology in terms of the number of patent applications and granted patents, the Internet of things was estimated to be the largest in terms of market size. It was followed, again in market size, by big data technologies, robotics, AI, 3D printing and the fifth generation of mobile services (5G).[1] Since AI emerged in the 1950s, 340,000 AI-related patent applications were filed by innovators and 1.6 million scientific papers have been published by researchers, with the majority of all AI-related patent filings published since 2013. Companies represent 26 out of the top 30 AI patent applicants, with universities or public research organizations accounting for the remaining four.[2] The ratio of scientific papers to inventions has significantly decreased from 8:1 in 2010 to 3:1 in 2016, which is attributed to be indicative of a shift from theoretical research to the use of AI technologies in commercial products and services. Machine learning is the dominant AI technique disclosed in patents and is included in more than one-third of all identified inventions (134,777 machine learning patents filed for a total of 167,038 AI patents filed in 2016), with computer vision being the most popular functional application. AI-related patents not only disclose AI techniques and applications, they often also refer to an application field or industry. Twenty application fields were identified in 2016 and included, in order of magnitude: telecommunications (15 percent), transportation (15 percent), life and medical sciences (12 percent), and personal devices, computing and human–computer interaction (11 percent). Other sectors included banking, entertainment, security, industry and manufacturing, agriculture, and networks (including social networks, smart cities and the Internet of things). IBM has the largest portfolio of AI patents with 8,290 patent applications, followed by Microsoft with 5,930 patent applications.[2]

  1. ^ "Intellectual Property and Frontier Technologies". WIPO. Archived from the original on 2 April 2022. Retrieved 30 March 2022.
  2. ^ a b "WIPO Technology Trends 2019 – Artificial Intelligence" (PDF). WIPO. 2019. Archived (PDF) from the original on 9 October 2022.

CharlesTGillingham (talk) 05:26, 29 July 2023 (UTC)

I'd recommend keeping it. IP is a big part of this topic, IMO big enough to merit that modest inclusion. Just my 2 cents. North8000 (talk) 14:17, 29 July 2023 (UTC)
One sentence maybe, in history Applications? Surely you don't mean the whole thing, which is as long as the entire section on current applications. ---- CharlesTGillingham (talk) 00:19, 30 July 2023 (UTC)
  Done. I added a sentence covering the boom in patents in History, with the source above. There was already a one-sentence mention of the boom in publications, funding and total jobs. Surely we don't need more coverage of patents than we do of publications, funding or jobs. ---- CharlesTGillingham (talk) 06:46, 30 July 2023 (UTC)

Wiki Education assignment: Research Process and Methodology - SU23 - Sect 200 - Thu

  This article was the subject of a Wiki Education Foundation-supported course assignment, between 24 May 2023 and 10 August 2023. Further details are available on the course page. Student editor(s): NoemieCY, ZhegeID (article contribs).

— Assignment last updated by ZhegeID (talk) 06:23, 7 August 2023 (UTC)

Update to R&N 2021 is complete

I've finished updating the article to be in line with Russell and Norvig 2021 edition. ---- CharlesTGillingham (talk) 00:18, 30 July 2023 (UTC)

Well, not completely. I'm still planning on taking a pass through "ethics", and add any relevant points from R&N. ---- CharlesTGillingham (talk) 18:54, 25 August 2023 (UTC)

Wiki Education assignment: First Year English Composition 1001

  This article was the subject of a Wiki Education Foundation-supported course assignment, between 23 August 2023 and 30 November 2023. Further details are available on the course page. Student editor(s): Cbetters23 (article contribs).

— Assignment last updated by RuthBenander (talk) 14:22, 25 August 2023 (UTC)

I encourage students to add material to these articles, which are at one level below this WP:summary article:
The quality of these articles is up and down: some of them could use more editorial work, and there is plenty of room for more valuable material. ---- CharlesTGillingham (talk) 19:27, 25 August 2023 (UTC)

Almost no mention to computer science and computational complexity of AI models

Currently all the development of AI is concentrated in advancing current algorithms and mathematical models. Several Computer Science departments around the world are pushing this field by researching novel architecture and more complex computational algorithms. It should be clear that AI fulls under computer science, as AI primary goal is to conceive computers/machines the ability to infer information upon unseen data. JoaoL975 (talk) 18:14, 28 August 2023 (UTC)

If you want to add such information, it must be cited to reliable sources (see WP:NOR and WP:V) and it should be added to the body of the article. The opening section summarizes the article body. Adding one's unsourced opinion of the subject to the lead section isn't how Wikipedia is written. MrOllie (talk) 18:36, 28 August 2023 (UTC)

Controversy over the definition of AI (need to add)

Other definitions avoid attributing the quality of intelligence to the computational capacity of machines or software. Jo Adetunji, Editor, The Conversation UK, wrote that the concept of artificial intelligence is being used abusively or, in other words, there is an inflation of the term that harms its realization. ([1])

This is how other definitions arise, such as that of the expert technologist Mauro D. Ríos, who defines AI as the field of information science dedicated to giving software automation, characteristics that simulate the cognitive abilities of the human being, applying these simulations to problem solving and manifesting the results as moving actions, written or spoken language, graphic representations or emerging data.

Ríos, Mauro,Arficial intelligence: When technology is the smallest of the paradigms. (July 26, 2023). Available at SSRN: https://ssrn.com/abstract=4521736 or http://dx.doi.org/10.2139/ssrn.4521736 2800:A4:1782:D300:9488:FEE5:2849:DCDF (talk) 01:18, 25 September 2023 (UTC)

My only problem with this is that there are hundreds, perhaps thousands of such definitions, each by an established academic in a peer reviewed paper. There is no way for this article to cover all of these.(Keep in mind, there are about 1.8 million academic papers about AI.)
In the philosophy section we do Turing, Russell/Norvig & McCarthy because these are the most important and influential. In the lede, we do the most obvious and uncontroversial definition possible, sidestepping all the philosophical problems. ---- CharlesTGillingham (talk) 16:39, 26 September 2023 (UTC)
The current definition in both the lead and the "Defining artificial intelligence" section are fine. Neither of the sources given above contradict what is currently in the article. The first just says the term is being misused, the other is just a rephrasing of what is currently in the lead. That AI is "giving software automation, characteristics that simulate the cognitive abilities of the human being" is just a restating of the article's lead sentence, that "Artificial intelligence (AI) is the intelligence of machines or software, as opposed to the intelligence of humans or animals." I also agree with CharlesTGillingham that we only need to concern ourselves with the most prominent, widely accepted definitions. Elspea756 (talk) 17:00, 26 September 2023 (UTC)
Saying that Artificial intelligence (AI) is the intelligence of machines or software is not the same as saying that it is a simulation of the cognitive abilities of the human being. So yes, the cited sources contradict the assertion that machines have intelligence and think. 2800:A4:E75:2800:4DA8:5840:CB16:28D3 (talk) 00:20, 11 October 2023 (UTC)
Exactly. AI is not a simulation of human cognitive abilities, at least at not according AI's founders and the leading AI textbook. (As discussed in Artificial intelligence § Philosophy) I'm not quite sure how your second sentence follows. ---- CharlesTGillingham (talk) 14:52, 15 October 2023 (UTC)

Wiki Education assignment: Technology and Culture

  This article was the subject of a Wiki Education Foundation-supported course assignment, between 21 August 2023 and 9 December 2023. Further details are available on the course page. Student editor(s): AdvaitPanicker, Ferna235, Boris Zeng (article contribs).

— Assignment last updated by Mbraile (talk) 20:12, 20 October 2023 (UTC)

Wiki Education assignment: Linguistics in the Digital Age

  This article was the subject of a Wiki Education Foundation-supported course assignment, between 21 August 2023 and 6 December 2023. Further details are available on the course page. Student editor(s): Asude Guvener, Ligh1ning (article contribs).

— Assignment last updated by Ligh1ning (talk) 22:23, 29 October 2023 (UTC)

Semi-protection of this talk page

I think that the disruptive editing by uninformed users unfortunately has reached a level where we have to protect this page, like the Talk:ChatGPT#Semi-protection of this talk page page. Or is there a better way? Sjö (talk) 13:59, 16 October 2023 (UTC)

I don't see the issue/problem here. North8000 (talk) 16:56, 16 October 2023 (UTC)
The issue/problem is that volunteer editors are having to spend our limited time removing talk page comments by people who apparently think this page is an AI app that they can type prompts into. Here are some recent examples of reverted pointless edits over the last couple of days: [2], [3], [4], [5], [6], [7], [8], [9], [10] This is a consistent problem across many AI-related article tak pages. Semi-protection of this page (and others) would presumably allow us to spend our time on more productive things. Elspea756 (talk) 23:05, 16 October 2023 (UTC)
I didn't know that that was happening. But it's quite a severe move to forbid (presumably permanently) IP's from posting on the talk page on AI articles. IMO to remove the relatively easy job of removing those inadvertent posts is IMHO not a sufficient reason for such and extreme measure. Sincerely, North8000 (talk) 23:23, 16 October 2023 (UTC)
Protection of a talk page is not something that should be done lightly. But I think that the constant disruption takes time away from more productive editing. Even if this page is protected, IP editors can still request changes at WP:RPP/E, and that link could be added to this page, like it is at the top of Talk:ChatGPT. Sjö (talk) 06:11, 30 October 2023 (UTC)

Suggestion section: Legislation

U.S. President Biden has signed an executive order (admittedly technically not a law or legislation) on AI: https://www.cbsnews.com/news/biden-ai-artificial-intelligence-executive-order/ Kdammers (talk) 21:28, 30 October 2023 (UTC)

I think such a section and also coverage of that particular EO is a good idea. The one downside is that would become a gigantic section. North8000 (talk) 21:36, 30 October 2023 (UTC)
This belongs in the section artificial intelligence § Regulation, and should also be added to the article regulation of artificial intelligence. ---- CharlesTGillingham (talk) 23:44, 30 October 2023 (UTC)

Wiki Education assignment: Research Process and Methodology - FA23 - Sect 202 - Thu

  This article was the subject of a Wiki Education Foundation-supported course assignment, between 6 September 2023 and 14 December 2023. Further details are available on the course page. Student editor(s): Wobuaichifan (article contribs).

— Assignment last updated by Wobuaichifan (talk) 02:03, 10 November 2023 (UTC)

Wiki Education assignment: IFS213-Hacking and Open Source Culture

  This article was the subject of a Wiki Education Foundation-supported course assignment, between 5 September 2023 and 19 December 2023. Further details are available on the course page. Student editor(s): Yaman Shqeirat (article contribs). Peer reviewers: Cvaquera59.

— Assignment last updated by UndercoverSwitch (talk) 03:30, 13 November 2023 (UTC)

No mention to Computer Science in the whole wiki page

How is it possible that a page explaining Artificial Intelligence (computer intelligence) has not one mention about computer science? JoaoL975 (talk) 18:11, 2 December 2023 (UTC)

It is mentioned in the second sentence. ---- CharlesTGillingham (talk) 17:23, 3 December 2023 (UTC)
Wait, no, I lied -- somebody edited it out. I put it back. ---- CharlesTGillingham (talk) 17:25, 3 December 2023 (UTC)

Restructuring the Applications Section

I believe certain parts of the AI Applications section should either be moved (like the section on Chinese facial recognition to the "Bad actors and weaponized AI" subcategory under Ethics) or shortened (like the section on astronomy).

In addition, I think the section is long enough that subheadings should be introduced to make it easier to read through. AdvaitPanicker (talk) 01:03, 5 December 2023 (UTC)

Wiki Education assignment: Linguistics in the Digital Age

  This article was the subject of a Wiki Education Foundation-supported course assignment, between 21 August 2023 and 11 December 2023. Further details are available on the course page. Student editor(s): Asude Guvener, Ligh1ning (article contribs).

— Assignment last updated by Fedfed2 (talk) 00:53, 9 December 2023 (UTC)

Wiki Education assignment: Technology and Culture

  This article was the subject of a Wiki Education Foundation-supported course assignment, between 21 August 2023 and 15 December 2023. Further details are available on the course page. Student editor(s): AdvaitPanicker, Ferna235, Boris Zeng (article contribs). Peer reviewers: Anieukir, Carariney.

— Assignment last updated by Thecanyon (talk) 05:32, 12 December 2023 (UTC)

Section title

It would probably be a bit more accurate if the section "Tools" was named "Techniques" instead. Wiktionary's definition of tool ("A piece of software used to develop software or hardware, or to perform low-level operations") doesn't exactly match what this section is about. Alenoach (talk) 03:43, 1 January 2024 (UTC)

I made the change. You can discuss it here if you disagree. Alenoach (talk) 04:15, 3 January 2024 (UTC)

Simpler animation

I propose replacing the current animation in the subsection "local search" with this one, which is easier to understand. Alenoach (talk) 16:33, 26 December 2023 (UTC)

I made the replacement that I suggested above.
Also, I wonder if this animation in the section "Probabilistic methods for uncertain reasoning" wouldn't be better in another section, like "Classifiers and statistical learning methods". Because what it depicts is mainly how iterative clustering works, not so much how to handle uncertainty. Maybe an image of a simple bayesian network like for example this one would better illustrate the section "Probabilistic methods for uncertain reasoning". It's not very beautiful, but it explains well the concept of bayesian network, which may seem esoteric to a lot of readers even though it's relatively simple.
What do you think? Alenoach (talk) 21:18, 3 January 2024 (UTC)
I support both image changes, the one you've already made replacing the local search image, and the new change you are proposing, moving the existing clustering image and adding the simple Bayesian network image. Elspea756 (talk) 15:29, 4 January 2024 (UTC)

"Age of AI" listed at Redirects for discussion

  The redirect Age of AI has been listed at redirects for discussion to determine whether its use and function meets the redirect guidelines. Readers of this page are welcome to comment on this redirect at Wikipedia:Redirects for discussion/Log/2024 February 8 § Age of AI until a consensus is reached. Duckmather (talk) 23:06, 8 February 2024 (UTC)

"Ai tool" listed at Redirects for discussion

  The redirect Ai tool has been listed at redirects for discussion to determine whether its use and function meets the redirect guidelines. Readers of this page are welcome to comment on this redirect at Wikipedia:Redirects for discussion/Log/2024 February 9 § Ai tool until a consensus is reached. Duckmather (talk) 06:13, 9 February 2024 (UTC)