Close Menu
ceofeature.com

    Subscribe to Updates

    Subscribe to our newsletter for the latest leadership tips, exclusive interviews, and expert advice from top CEOs. Simply enter your email below and stay ahead of the curve!.

    What's Hot

    U.S. Inserts Itself Into Israel-Iran War and Strikes 3 Iranian Nuclear Sites

    June 22, 2025

    Bitcoin falls 4% to $99,237

    June 22, 2025

    How a ‘brag doc’ can help you ace interviews and land a job, says recruiter

    June 22, 2025
    Facebook X (Twitter) Instagram
    ceofeature.com
    ceofeature.com
    ceofeature.com
    • Home
    • Business
    • Lifestyle
    • CEO News
    • Investing
    • Opinion
    • Market
    • Magazine
    Facebook X (Twitter) Instagram YouTube
    Subscribe
    ceofeature.com
    Home»Business»AI models need more standards and tests, say researchers
    Business

    AI models need more standards and tests, say researchers

    Daniel snowBy Daniel snowJune 22, 20254 Mins Read
    Share
    Facebook Twitter LinkedIn Pinterest Email Copy Link


    As the usage of artificial intelligence — benign and adversarial — increases at breakneck speed, more cases of potentially harmful responses are being uncovered.

    Pixdeluxe | E+ | Getty Images

    As the usage of artificial intelligence — benign and adversarial — increases at breakneck speed, more cases of potentially harmful responses are being uncovered. These include hate speech, copyright infringements or sexual content.

    The emergence of these undesirable behaviors is compounded by a lack of regulations and insufficient testing of AI models, researchers told CNBC.

    Getting machine learning models to behave the way it was intended to do so is also a tall order, said Javier Rando, a researcher in AI.

    “The answer, after almost 15 years of research, is, no, we don’t know how to do this, and it doesn’t look like we are getting better,” Rando, who focuses on adversarial machine learning, told CNBC.

    However, there are some ways to evaluate risks in AI, such as red teaming. The practice involves individuals testing and probing artificial intelligence systems to uncover and identify any potential harm — a modus operandi common in cybersecurity circles.

    Shayne Longpre, a researcher in AI and policy and lead of the Data Provenance Initiative, noted that there are currently insufficient people working in red teams.

    While AI startups are now using first-party evaluators or contracted second parties to test their models, opening the testing to third parties such as normal users, journalists, researchers, and ethical hackers would lead to a more robust evaluation, according to a paper published by Longpre and researchers.

    “Some of the flaws in the systems that people were finding required lawyers, medical doctors to actually vet, actual scientists who are specialized subject matter experts to figure out if this was a flaw or not, because the common person probably couldn’t or wouldn’t have sufficient expertise,” Longpre said.

    Adopting standardized ‘AI flaw’ reports, incentives and ways to disseminate information on these ‘flaws’ in AI systems are some of the recommendations put forth in the paper.

    With this practice having been successfully adopted in other sectors such as software security, “we need that in AI now,” Longpre added.

    Marrying this user-centred practice with governance, policy and other tools would ensure a better understanding of the risks posed by AI tools and users, said Rando.

    We're pursing a path of AI development that's extremely harmful to a lot of people, says Karen Hao

    No longer a moonshot

    Project Moonshot is one such approach, combining technical solutions with policy mechanisms. Launched by Singapore’s Infocomm Media Development Authority, Project Moonshot is a large language model evaluation toolkit developed with industry players such as IBM and Boston-based DataRobot.

    The toolkit integrates benchmarking, red teaming and testing baselines. There is also an evaluation mechanism which allows AI startups to ensure that their models can be trusted and do no harm to users, Anup Kumar, head of client engineering for data and AI at IBM Asia Pacific, told CNBC.

    Evaluation is a continuous process that should be done both prior to and following the deployment of models, said Kumar, who noted that the response to the toolkit has been mixed.

    “A lot of startups took this as a platform because it was open source, and they started leveraging that. But I think, you know, we can do a lot more.”

    Moving forward, Project Moonshot aims to include customization for specific industry use cases and enable multilingual and multicultural red teaming.

    Higher standards

    Pierre Alquier, Professor of Statistics at the ESSEC Business School, Asia-Pacific, said that tech companies are currently rushing to release their latest AI models without proper evaluation.

    “When a pharmaceutical company designs a new drug, they need months of tests and very serious proof that it is useful and not harmful before they get approved by the government,” he noted, adding that a similar process is in place in the aviation sector.

    AI models need to meet a strict set of conditions before they are approved, Alquier added. A shift away from broad AI tools to developing ones that are designed for more specific tasks would make it easier to anticipate and control their misuse, said Alquier.

    “LLMs can do too many things, but they are not targeted at tasks that are specific enough,” he said. As a result, “the number of possible misuses is too big for the developers to anticipate all of them.”

    Such broad models make defining what counts as safe and secure difficult, according to a research that Rando was involved in.

    Tech companies should therefore avoid overclaiming that “their defenses are better than they are,” said Rando.



    Source link

    Follow on Google News Follow on Flipboard
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Daniel snow
    • Website

    Related Posts

    How a ‘brag doc’ can help you ace interviews and land a job, says recruiter

    June 22, 2025

    Gulf markets end higher Iran strike

    June 22, 2025

    U.S. is at war with Iran’s nuclear program

    June 22, 2025
    Leave A Reply Cancel Reply

    Top Posts

    Steven E. Orr Redefines FinTech and Financial Media with Quasar Markets

    June 4, 2025

    Redefining leadership and unlocking human potential, Meet Janice Elsley

    June 4, 2025

    Queen of the North: How Ravinna Raveenthiran is Redefining Real Estate with Resilience and Compassion

    October 22, 2024

    Mervina Nyampulu with “I Am” charts a success Story and A Legacy of her own

    May 22, 2025
    Don't Miss

    U.S. Inserts Itself Into Israel-Iran War and Strikes 3 Iranian Nuclear Sites

    By Daniel snowJune 22, 2025

    A maritime center warns American Navy ships are at “high” risk of being targeted by…

    Bitcoin falls 4% to $99,237

    June 22, 2025

    How a ‘brag doc’ can help you ace interviews and land a job, says recruiter

    June 22, 2025

    Gulf markets end higher Iran strike

    June 22, 2025
    Stay In Touch
    • Facebook
    • Twitter

    Subscribe to Updates

    Subscribe to our newsletter for the latest leadership tips, exclusive interviews, and expert advice from top CEOs. Simply enter your email below and stay ahead of the curve!.

    About Us
    About Us

    Welcome to CEO Feature, where we dive deep into the exhilarating world of entrepreneurs and CEOs from across the globe! Brace yourself for captivating stories that will blow your mind and leave you inspired.

    Facebook X (Twitter)
    Featured Posts

    5 Simple Tips to Take Care of Larger Air Balloons

    January 4, 2020

    5 Ways Your Passport Can Ruin Your Cool Holiday Trip

    January 5, 2020

    Tokyo Officials Plan For a Safe Olympic Games Without Quarantines

    January 6, 2020
    Worldwide News

    5 Ways Your Passport Can Ruin Your Cool Holiday Trip

    January 5, 20200

    Fun Games: Kill The Boredom And Enjoy Your Family Time

    January 7, 20200

    A Diverse Collection of Museum Quality Artifacts Sculptures

    January 8, 20200
    • www.ceofeature.com
    @2025 copyright by ceofeature

    Type above and press Enter to search. Press Esc to cancel.