Page 5 - AI Vol 2: Risks of AI
P. 5

INCORPORATING A "HUMAN IN
              THE LOOP", REQUIRING HUMAN
               OVERSIGHT OF AI-GENERATED
                   DECISIONS, IS A CRITICAL
                 MITIGATION MEASURE FOR
              ADDRESSING POTENTIAL BIASES
                  IN AI SYSTEMS, ESPECIALLY
                  WITHIN PUBLIC AGENCIES.






            When bias is identified after deployment, public
            agencies  should  first  and  foremost  ensure  that   essential, encouraging employees to question and
            the bias does not impact  agency decisions.          challenge AI recommendations when necessary.
            Additionally, agencies can work with developers to   Agencies should also have protocols for reporting
            adjust the model or seek alternative solutions that   potential biases or inaccuracies in AI decisions.
            demonstrate a stronger commitment to unbiased        This feedback can inform the agency's decision
            outcomes.  While the technical  details  of these    on whether to use a specific AI system and may

            corrections  are  managed  by specialists,  agency   also be communicated to the developer to enhance
            officials  should  understand  the  broad  strategies   the model, aiming to mitigate bias.
            used for bias correction,  such as diversifying
            training data or adjusting the algorithm's decision-  Bias may also be minimized  by engaging  in
            making parameters.                                   prompting strategies which promote transparency

                                                                 and are less prone to bias. Instead of asking an AI
            Incorporating  a  “human  in  the  loop,”  requiring   system to make broad assessments or decisions,

            human oversight of  AI-generated  decisions,         a  more  targeted  approach  in  questioning  can
            is a critical  mitigation  measure  for addressing
            potential biases in AI systems, especially within
            public agencies. This approach ensures that AI-
            generated decisions are not made in isolation but
            are instead reviewed and, if necessary, adjusted by         “THE THING THAT’S GOING
            human oversight. For this approach to be effective,            TO MAKE ARTIFICIAL
            agency  personnel involved  in overseeing  AI            INTELLIGENCE SO POWERFUL
            decisions need appropriate training. They should          IS ITS ABILITY TO LEARN, AND
                                                                        THE WAY AI LEARNS IS TO
            be aware of the potential biases AI systems might         LOOK AT HUMAN CULTURE.”
            hold and be equipped  to identify  signs of such
            biases in decision outputs. Additionally, fostering                   - DAN BROWN
            a culture of critical engagement with AI tools is





     RISKS OF AI  |  LOZANOSMITH.COM                                                                       VOLUME  2    |   5
   1   2   3   4   5   6   7   8   9   10