26
Thu, Dec
0 New Articles

Tools for trustworthy AI

Typography
  • Smaller Small Medium Big Bigger
  • Default Helvetica Segoe Georgia Times

A new tool has been developed to catch students cheating with ChatGPT. It’s 99.9% effective. But OpenAI hasn’t released it because it’s mired in ethics concerns.

It’s just one example of one of the major challenges facing AI. How can we monitor the technology to make sure it’s used ethically?

For the past few years, the biggest names in AI have pushed for their tech to be used responsibly. And using AI ethically isn’t just the right thing for businesses to do—it’s also something consumers want. In fact, 86% of businesses believe customers prefer companies that use ethical guidelines and are clear about how they use their data and AI models, according to the IBM Global AI Adoption Index.

“We are all well beyond hoping companies are aware [of using AI ethically],” says Phaedra Boinodiris, IBM’s Global Leader for Trustworthy AI. “The bigger question is, why is it important for businesses and other organizations to be held accountable for safe and responsible outcomes from the AI they deploy?” 

Can AI ethics tools help? Are the tools themselves biased? Here’s a quick look at the latest research.

Deepfake detectors

Deepfake fraud attempts were up 3000% from 2022 to 2023, and they are getting more elaborate. In February, a finance worker at a multinational firm in Hong Kong was duped into paying out $25 million to criminals who created a video conference call with multiple deepfake characters including the firm’s CFO.

In May, OpenAI announced that it had released its own deepfake detection tool to disinformation researchers. The tool was able to spot 98.8% of images created by DALL-E 3. OpenAI also joined Google, Adobe and others on the steering committee for the Coalition for Content Provenance and Authenticity (C2PA), an industry coalition developing a standard for certifying the history and source of media content. 

Until that standard is set, businesses are building tools they hope can fill the gap. In August, McAfee announced its McAfee Deepfake Detector. The tool uses deep neural network models to spot fake AI audio in videos playing in your browser. In 2022, Intel introduced FakeCatcher, which analyzes blood flow in video pixels, separating humans from deepfakes with a 96% accuracy rate. And they’re not alone. Notable startups in the field include NYC’s Reality Defender, Israeli startup Clarity, and Estonia-based Sentinel, all of whom have scanning tools available that use AI to spot patterns in various kinds of deepfakes.

With deepfake detection tech evolving at such a rapid pace, it’s important to keep potential algorithmic biases in mind. Computer scientist and deepfake expert Siwei Lyu and his team at the University of Buffalo have developed what they believe to be the first deepfake-detection algorithms designed to minimize bias. The UB researchers made a photo collage of the hundreds of faces that were identified as fake in their detection algorithms; the results showed an overall darker skin tone.

“Deepfakes may be used to attack underrepresented minority groups, so it is important to make sure detection technologies will not underserve them,” Lyu says. As for the future of deepfake detection? “The generative AI technologies underlying the deepfakes will undoubtedly continue to grow, so we are going to see deepfakes with increasing number, quality and forms. I expect future [detection] technologies will be equipped with more guardrails to reduce the chances of misuses.”

Explore AI ethics at IBM

Anti-facial recognition (AFR) technologies

Facial recognition systems are becoming increasingly common as a convenient way to authenticate a user’s identity—but these systems have long been fraught with ethical problems ranging from racial bias to data privacy. Complicating the issue, “some biases are [also] intersectional, compounding multiple layers of prejudice,” notes Helen Edwards, co-founder of AI ethics think tank Artificiality.

In May, Australian facial recognition startup Outabox’s data was breached, releasing the biometric data of more than a million users. Earlier this year ‘GoldPickAxe,’ a trojan aimed at Android and iOS devices, was caught capturing facial data to break into bank accounts.

A promising approach to protecting facial biometric data is by scrambling it in a way that’s imperceptible to the human eye but confuses recognition systems. One of the first tools to do this was Fawkes, a project developed at the University of Chicago. Named after the Guy Fawkes mask, the program is designed to cloak photos by subtly altering pixels; it’s free to download on the project’s website.

More recently, researchers at Zhejiang University’s USSLAB have pioneered CamPro, which aims to achieve AFR at the camera sensor level. CamPro produces images that reduce facial identification to 0.3% without interfering with other applications like activity recognition.

AI writing detectors

Spotting AI-generated writing continues to be a struggle for businesses and educational institutions. In a blind test at the University of Reading, five different psychology modules had ChatGPT-written exams mixed in with exams written by real students. The June study found that 94% of ChatGPT exam answers were not spotted by the people grading the exams. The AI-generated exams also averaged half a grade higher than the student exams.

A variety of AI writing detectors have flooded the market to address this issue, looking for common hallmarks of AI-generated text such as repetition and perfect grammar. But experts warn that they’re not reliable yet and often demonstrate bias.

Last year a Stanford study found that AI detectors flagged writing by non-native English speakers an average of 61.3% of the time but made far fewer errors when evaluating writing by native English speakers.

Humans passing off AI-generated writing as their own is not only dishonest—sometimes it’s also plagiarism, which can come with serious legal ramifications. Because of this concern, some companies are using AI writing detectors to test the copy of their writers. This has led to companies falsely accusing writers of passing off AI-generated copy as their own, damaging the writers’ reputations and careers.

LLM bias detectors

Datasets often include the unconscious biases of the people who create them. It’s why algorithmic bias is such a persistent problem in the LLMs that train on this data.

In one example, researchers at Cornell used ChatGPT and Alpaca to generate recommendation letters for men and women; the letters showed significant biases favoring men. Generated language like “Kelly is a warm person” versus “Joseph is a role model” demonstrated the way these biases could affect women in the workplace.

Researchers are working to find ways to flag and mitigate biases. A team at the University of Illinois Urbana-Champaign developed QuaCer-B, which generates provable LLM bias measures for prompts sampled from given distributions and can be used for both API and open-source LLMs.

“The AI industry currently relies on evaluating the safety and trustworthiness of their models by testing them on a small set of benchmark inputs,” says UIUC professor Gagandeep Singh, one of the lead researchers behind QuaCer-B. “However, safe generation on benchmark inputs does not guarantee that the LLM-generated content will be ethical when handling diverse unseen scenarios in the real world. QuaCer-B enables LLM developers to make informed decisions about the suitability of their models for real-world deployment and also identify causes of failures to improve the model.”

As AI continues to evolve, new ethical problems will keep evolving alongside it. And while tools to flag, monitor and prevent unethical use of the tech are a start, AI ethics experts don’t consider them a one-stop solution.

“The hard part is not buying the right tool,” Boinodiris adds. “Curating AI responsibly is a sociotechnical challenge that requires a holistic approach. And people are the hardest part of the equation.”

“In addition to thoughtful regulation and enforcement, the key to ethical AI is post-market auditing, continually monitoring performance and minimizing risks,” explains Gemma Galdón-Clavell, an advisor to the United Nations and EU on applied ethics and responsible AI and founder of Eticas.ai. “Think about the automotive industry: Warning lights and proximity sensors can help drivers avoid crashes, but we still need seatbelts, airbags and regular inspections to ensure that the open road is as safe as it can be.”

BLOG COMMENTS POWERED BY DISQUS

LATEST COMMENTS

Support MC Press Online

$

Book Reviews

Resource Center

  • SB Profound WC 5536 Have you been wondering about Node.js? Our free Node.js Webinar Series takes you from total beginner to creating a fully-functional IBM i Node.js business application. You can find Part 1 here. In Part 2 of our free Node.js Webinar Series, Brian May teaches you the different tooling options available for writing code, debugging, and using Git for version control. Brian will briefly discuss the different tools available, and demonstrate his preferred setup for Node development on IBM i or any platform. Attend this webinar to learn:

  • SB Profound WP 5539More than ever, there is a demand for IT to deliver innovation. Your IBM i has been an essential part of your business operations for years. However, your organization may struggle to maintain the current system and implement new projects. The thousands of customers we've worked with and surveyed state that expectations regarding the digital footprint and vision of the company are not aligned with the current IT environment.

  • SB HelpSystems ROBOT Generic IBM announced the E1080 servers using the latest Power10 processor in September 2021. The most powerful processor from IBM to date, Power10 is designed to handle the demands of doing business in today’s high-tech atmosphere, including running cloud applications, supporting big data, and managing AI workloads. But what does Power10 mean for your data center? In this recorded webinar, IBMers Dan Sundt and Dylan Boday join IBM Power Champion Tom Huntington for a discussion on why Power10 technology is the right strategic investment if you run IBM i, AIX, or Linux. In this action-packed hour, Tom will share trends from the IBM i and AIX user communities while Dan and Dylan dive into the tech specs for key hardware, including:

  • Magic MarkTRY the one package that solves all your document design and printing challenges on all your platforms. Produce bar code labels, electronic forms, ad hoc reports, and RFID tags – without programming! MarkMagic is the only document design and print solution that combines report writing, WYSIWYG label and forms design, and conditional printing in one integrated product. Make sure your data survives when catastrophe hits. Request your trial now!  Request Now.

  • SB HelpSystems ROBOT GenericForms of ransomware has been around for over 30 years, and with more and more organizations suffering attacks each year, it continues to endure. What has made ransomware such a durable threat and what is the best way to combat it? In order to prevent ransomware, organizations must first understand how it works.

  • SB HelpSystems ROBOT GenericIT security is a top priority for businesses around the world, but most IBM i pros don’t know where to begin—and most cybersecurity experts don’t know IBM i. In this session, Robin Tatam explores the business impact of lax IBM i security, the top vulnerabilities putting IBM i at risk, and the steps you can take to protect your organization. If you’re looking to avoid unexpected downtime or corrupted data, you don’t want to miss this session.

  • SB HelpSystems ROBOT GenericCan you trust all of your users all of the time? A typical end user receives 16 malicious emails each month, but only 17 percent of these phishing campaigns are reported to IT. Once an attack is underway, most organizations won’t discover the breach until six months later. A staggering amount of damage can occur in that time. Despite these risks, 93 percent of organizations are leaving their IBM i systems vulnerable to cybercrime. In this on-demand webinar, IBM i security experts Robin Tatam and Sandi Moore will reveal:

  • FORTRA Disaster protection is vital to every business. Yet, it often consists of patched together procedures that are prone to error. From automatic backups to data encryption to media management, Robot automates the routine (yet often complex) tasks of iSeries backup and recovery, saving you time and money and making the process safer and more reliable. Automate your backups with the Robot Backup and Recovery Solution. Key features include:

  • FORTRAManaging messages on your IBM i can be more than a full-time job if you have to do it manually. Messages need a response and resources must be monitored—often over multiple systems and across platforms. How can you be sure you won’t miss important system events? Automate your message center with the Robot Message Management Solution. Key features include:

  • FORTRAThe thought of printing, distributing, and storing iSeries reports manually may reduce you to tears. Paper and labor costs associated with report generation can spiral out of control. Mountains of paper threaten to swamp your files. Robot automates report bursting, distribution, bundling, and archiving, and offers secure, selective online report viewing. Manage your reports with the Robot Report Management Solution. Key features include:

  • FORTRAFor over 30 years, Robot has been a leader in systems management for IBM i. With batch job creation and scheduling at its core, the Robot Job Scheduling Solution reduces the opportunity for human error and helps you maintain service levels, automating even the biggest, most complex runbooks. Manage your job schedule with the Robot Job Scheduling Solution. Key features include:

  • LANSA Business users want new applications now. Market and regulatory pressures require faster application updates and delivery into production. Your IBM i developers may be approaching retirement, and you see no sure way to fill their positions with experienced developers. In addition, you may be caught between maintaining your existing applications and the uncertainty of moving to something new.

  • LANSAWhen it comes to creating your business applications, there are hundreds of coding platforms and programming languages to choose from. These options range from very complex traditional programming languages to Low-Code platforms where sometimes no traditional coding experience is needed. Download our whitepaper, The Power of Writing Code in a Low-Code Solution, and:

  • LANSASupply Chain is becoming increasingly complex and unpredictable. From raw materials for manufacturing to food supply chains, the journey from source to production to delivery to consumers is marred with inefficiencies, manual processes, shortages, recalls, counterfeits, and scandals. In this webinar, we discuss how:

  • The MC Resource Centers bring you the widest selection of white papers, trial software, and on-demand webcasts for you to choose from. >> Review the list of White Papers, Trial Software or On-Demand Webcast at the MC Press Resource Center. >> Add the items to yru Cart and complet he checkout process and submit

  • Profound Logic Have you been wondering about Node.js? Our free Node.js Webinar Series takes you from total beginner to creating a fully-functional IBM i Node.js business application.

  • SB Profound WC 5536Join us for this hour-long webcast that will explore:

  • Fortra IT managers hoping to find new IBM i talent are discovering that the pool of experienced RPG programmers and operators or administrators with intimate knowledge of the operating system and the applications that run on it is small. This begs the question: How will you manage the platform that supports such a big part of your business? This guide offers strategies and software suggestions to help you plan IT staffing and resources and smooth the transition after your AS/400 talent retires. Read on to learn: