• Welcome to the Online Discussion Groups, Guest.

    Please introduce yourself here. We'd love to hear from you!

    If you are a CompTIA member you can find your regional community here and get posting.

    This notification is dismissable and will disappear once you've made a couple of posts.
  • We will be shutting down for a brief period of time on 9/24 at around 8 AM CST to perform necessary software updates and maintenance; please plan accordingly!
K

Kathleen Martin

Guest
Artificial intelligence is going mainstream. If you're using Google docs, Ink for All or any number of digital tools, AI is being baked in. AI is already making decisions in the workplace, around hiring, customer service and more. However, a recurring issue with AI is that it can be a bit of a "black box" or mystery as to how it arrived at its decisions. Enter explainable AI.
Explainable Artificial Intelligence, or XAI, is similar to a normal AI application except that the processes and results of an XAI algorithm are able to be explained so that they can be understood by humans. The complex nature of artificial intelligence means that AI is making decisions in real-time based on the insights it has discovered in the data that it has been fed. When we do not fully understand how AI is making these decisions, we are not able to fully optimize the AI application to be all that it is capable of. XAI enables people to understand how AI and Machine Learning (ML) are being used to make decisions, predictions, and insights. Explainable AI allows brands to be transparent in their use of AI applications, which increases user trust and the overall acceptance of AI.
Why and Where We Need Explainable AI
There is a valid need for XAI if AI is going to be used across industries. According to a report by FICO, 65% of surveyed employees could not explain how AI model decisions or predictions are determined. The benefits of XAI are beginning to be well-recognized, and not just by scientists and data engineers. The European Union’s draft AI regulations are specifying XAI as a prerequisite for the eventual normalization of machine learning in society. Standardization organizations including the European Telecommunications Standards Institute (ETSI) and the Institute of Electrical and Electronics Engineers Standards Association (IEEE SA) also recognize the importance of XAI in relation to the acceptance and trust of AI in the future.
Philip Pilgerstorfer, data scientist and XAI specialist at QuantumBlack, has indicated that the benefits of XAI include:
  1. Building trustworthiness: Humans are better able to trust the AI model when the characteristics and rationale of the AI output have been explained.
  2. Satisfying legal requirements: Financial and healthcare industries may be required to incorporate machine learning models into their complex risk assessment strategies in order to fulfill regulatory requirements for effective risk management.
  3. Providing ethics-related justification (and removing unconscious biases): Because XAI is transparent and able to more easily be debugged, unconscious biases can be removed, and ethical decisions explained.
    [/LIST=1]
    Continue reading: https://www.cmswire.com/digital-experience/4-reasons-why-explainable-ai-is-the-future-of-ai/
 

Attachments

  • p0005026.m04694.a2e7d04f0627476d9d02a6ea0bc73c85.png
    p0005026.m04694.a2e7d04f0627476d9d02a6ea0bc73c85.png
    330.6 KB · Views: 45