Close Menu
Ztoog
    What's Hot
    Crypto

    Ethereum Will Skyrocket Due To AI DAO Revolution: Arthur Hayes

    Crypto

    Will $0.055 Launch a Recovery Phase?

    Technology

    Kagan: Florida social media law seems like “classic First Amendment violation”

    Important Pages:
    • About Us
    • Contact us
    • Privacy Policy
    • Terms & Conditions
    Facebook X (Twitter) Instagram Pinterest
    Facebook X (Twitter) Instagram Pinterest
    Ztoog
    • Home
    • The Future

      How I Turn Unstructured PDFs into Revenue-Ready Spreadsheets

      Is it the best tool for 2025?

      The clocks that helped define time from London’s Royal Observatory

      Summer Movies Are Here, and So Are the New Popcorn Buckets

      India-Pak conflict: Pak appoints ISI chief, appointment comes in backdrop of the Pahalgam attack

    • Technology

      Ensure Hard Work Is Recognized With These 3 Steps

      Cicada map 2025: Where will Brood XIV cicadas emerge this spring?

      Is Duolingo the face of an AI jobs crisis?

      The US DOD transfers its AI-based Open Price Exploration for National Security program to nonprofit Critical Minerals Forum to boost Western supply deals (Ernest Scheyder/Reuters)

      The more Google kills Fitbit, the more I want a Fitbit Sense 3

    • Gadgets

      Apple plans to split iPhone 18 launch into two phases in 2026

      Upgrade your desk to Starfleet status with this $95 USB-C hub

      37 Best Graduation Gift Ideas (2025): For College Grads

      Backblaze responds to claims of “sham accounting,” customer backups at risk

      Snapdragon X Plus Could Bring Faster, More Powerful Chromebooks

    • Mobile

      What are people doing with those free T-Mobile lines? Way more than you’d expect

      Samsung doesn’t want budget Galaxy phones to use exclusive AI features

      COROS’s charging adapter is a neat solution to the smartwatch charging cable problem

      Fortnite said to return to the US iOS App Store next week following court verdict

      Chinese tech icon is about to raise the stakes in a battle with US chipmaker over AI processors

    • Science

      Trump administration cuts off all future federal funding to Harvard

      Does kissing spread gluten? New research offers a clue.

      Why Balcony Solar Panels Haven’t Taken Off in the US

      ‘Dark photon’ theory of light aims to tear up a century of physics

      Signs of alien life on exoplanet K2-18b may just be statistical noise

    • AI

      How to build a better AI benchmark

      Q&A: A roadmap for revolutionizing health care through data-driven innovation | Ztoog

      This data set helps researchers spot harmful stereotypes in LLMs

      Making AI models more trustworthy for high-stakes settings | Ztoog

      The AI Hype Index: AI agent cyberattacks, racing robots, and musical models

    • Crypto

      ‘The Big Short’ Coming For Bitcoin? Why BTC Will Clear $110,000

      Bitcoin Holds Above $95K Despite Weak Blockchain Activity — Analytics Firm Explains Why

      eToro eyes US IPO launch as early as next week amid easing concerns over Trump’s tariffs

      Cardano ‘Looks Dope,’ Analyst Predicts Big Move Soon

      Speak at Ztoog Disrupt 2025: Applications now open

    Ztoog
    Home » Supporting benchmarks for AI safety with MLCommons – Google Research Blog
    AI

    Supporting benchmarks for AI safety with MLCommons – Google Research Blog

    Facebook Twitter Pinterest WhatsApp
    Supporting benchmarks for AI safety with MLCommons – Google Research Blog
    Share
    Facebook Twitter LinkedIn Pinterest WhatsApp

    Posted by Anoop Sinha, Technology and Society, and Marian Croak, Google Research, Responsible AI and Human Centered Technology group

    Standard benchmarks are agreed upon methods of measuring necessary product qualities, and so they exist in lots of fields. Some commonplace benchmarks measure safety: for instance, when a automobile producer touts a “five-star overall safety rating,” they’re citing a benchmark. Standard benchmarks exist already in machine studying (ML) and AI applied sciences: for occasion, the MLCommons Association operates the MLPerf benchmarks that measure the velocity of innovative AI {hardware} resembling Google’s TPUs. However, although there was important work finished on AI safety, there are as but no related commonplace benchmarks for AI safety.

    We are excited to assist a brand new effort by the non-profit MLCommons Association to develop commonplace AI safety benchmarks. Developing benchmarks which can be efficient and trusted goes to require advancing AI safety testing know-how and incorporating a broad vary of views. The MLCommons effort goals to convey collectively professional researchers throughout academia and business to develop commonplace benchmarks for measuring the safety of AI techniques into scores that everybody can perceive. We encourage the entire neighborhood, from AI researchers to coverage specialists, to affix us in contributing to the trouble.

    Why AI safety benchmarks?

    Like most superior applied sciences, AI has the potential for great advantages however might additionally result in detrimental outcomes with out applicable care. For instance, AI know-how can increase human productiveness in a variety of actions (e.g., enhance well being diagnostics and analysis into ailments, analyze power utilization, and extra). However, with out enough precautions, AI is also used to assist dangerous or malicious actions and reply in biased or offensive methods.

    By offering commonplace measures of safety throughout classes resembling dangerous use, out-of-scope responses, AI-control dangers, and so on., commonplace AI safety benchmarks might assist society reap the advantages of AI whereas making certain that enough precautions are being taken to mitigate these dangers. Initially, nascent safety benchmarks might assist drive AI safety analysis and inform accountable AI growth. With time and maturity, they might assist inform customers and purchasers of AI techniques. Eventually, they could possibly be a useful instrument for coverage makers.

    In pc {hardware}, benchmarks (e.g., SPEC, TPC) have proven a tremendous potential to align analysis, engineering, and even advertising and marketing throughout a complete business in pursuit of progress, and we imagine commonplace AI safety benchmarks might assist do the identical on this important space.

    What are commonplace AI safety benchmarks?

    Academic and company analysis efforts have experimented with a spread of AI safety assessments (e.g., RealToxicityPrompts, Stanford HELM equity, bias, toxicity measurements, and Google’s guardrails for generative AI). However, most of those assessments deal with offering a immediate to an AI system and algorithmically scoring the output, which is a helpful begin however restricted to the scope of the check prompts. Further, they normally use open datasets for the prompts and responses, which can have already got been (typically inadvertently) included into coaching information.

    MLCommons proposes a multi-stakeholder course of for deciding on assessments and grouping them into subsets to measure safety for explicit AI use-cases, and translating the extremely technical outcomes of these assessments into scores that everybody can perceive. MLCommons is proposing to create a platform that brings these present assessments collectively in a single place and encourages the creation of extra rigorous assessments that transfer the cutting-edge ahead. Users will be capable of entry these assessments each by way of on-line testing the place they will generate and overview scores and offline testing with an engine for non-public testing.

    AI safety benchmarks must be a collective effort

    Responsible AI builders use a various vary of safety measures, together with computerized testing, handbook testing, pink teaming (through which human testers try to supply adversarial outcomes), software-imposed restrictions, information and mannequin best-practices, and auditing. However, figuring out that enough precautions have been taken may be difficult, particularly because the neighborhood of corporations offering AI techniques grows and diversifies. Standard AI benchmarks might present a strong instrument for serving to the neighborhood develop responsibly, each by serving to distributors and customers measure AI safety and by encouraging an ecosystem of sources and specialist suppliers centered on enhancing AI safety.

    At the identical time, growth of mature AI safety benchmarks which can be each efficient and trusted shouldn’t be attainable with out the involvement of the neighborhood. This effort will want researchers and engineers to come back collectively and supply modern but sensible enhancements to safety testing know-how that make testing each extra rigorous and extra environment friendly. Similarly, corporations might want to come collectively and supply check information, engineering assist, and monetary assist. Some points of AI safety may be subjective, and constructing trusted benchmarks supported by a broad consensus would require incorporating a number of views, together with these of public advocates, coverage makers, teachers, engineers, information employees, enterprise leaders, and entrepreneurs.

    Google’s assist for MLCommons

    Grounded in our AI Principles that have been introduced in 2018, Google is dedicated to particular practices for the secure, safe, and reliable growth and use of AI (see our 2019, 2020, 2021, 2022 updates). We’ve additionally made important progress on key commitments, which is able to assist guarantee AI is developed boldly and responsibly, for the advantage of everybody.

    Google is supporting the MLCommons Association’s efforts to develop AI safety benchmarks in a variety of methods.

    1. Testing platform: We are becoming a member of with different corporations in offering funding to assist the event of a testing platform.
    2. Technical experience and sources: We are offering technical experience and sources, such because the Monk Skin Tone Examples Dataset, to assist be certain that the benchmarks are well-designed and efficient.
    3. Datasets: We are contributing an inner dataset for multilingual representational bias, in addition to already externalized assessments for stereotyping harms, resembling SeeGULL and SPICE. Moreover, we’re sharing our datasets that target gathering human annotations responsibly and inclusively, like DICES and SRP.

    Future course

    We imagine that these benchmarks shall be very helpful for advancing analysis in AI safety and making certain that AI techniques are developed and deployed in a accountable method. AI safety is a collective-action downside. Groups just like the Frontier Model Forum and Partnership on AI are additionally main necessary standardization initiatives. We’re happy to have been a part of these teams and MLCommons since their starting. We stay up for further collective efforts to advertise the accountable growth of latest generative AI instruments.

    Acknowledgements

    Many due to the Google group that contributed to this work: Peter Mattson, Lora Aroyo, Chris Welty, Kathy Meier-Hellstern, Parker Barnes, Tulsee Doshi, Manvinder Singh, Brian Goldman, Nitesh Goyal, Alice Friend, Nicole Delange, Kerry Barker, Madeleine Elish, Shruti Sheth, Dawn Bloxwich, William Isaac, Christina Butterfield.

    Share. Facebook Twitter Pinterest LinkedIn WhatsApp

    Related Posts

    AI

    How to build a better AI benchmark

    AI

    Q&A: A roadmap for revolutionizing health care through data-driven innovation | Ztoog

    AI

    This data set helps researchers spot harmful stereotypes in LLMs

    AI

    Making AI models more trustworthy for high-stakes settings | Ztoog

    AI

    The AI Hype Index: AI agent cyberattacks, racing robots, and musical models

    AI

    Novel method detects microbial contamination in cell cultures | Ztoog

    AI

    Seeing AI as a collaborator, not a creator

    AI

    “Periodic table of machine learning” could fuel AI discovery | Ztoog

    Leave A Reply Cancel Reply

    Follow Us
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    Top Posts
    Gadgets

    Pebble’s founder wants to relaunch the e-paper smartwatch for its fans

    “We’re making new Pebble watches,” writes unique Pebble founder Eric Migicovsky on the “rePebble” launch…

    AI

    How machine learning models can amplify inequities in medical diagnosis and treatment | Ztoog

    Prior to receiving a PhD in laptop science from MIT in 2017, Marzyeh Ghassemi had…

    The Future

    Motorola launches its cheapest 5G device in Australia with the Moto G54

    Motorola, a sub model of Lenovo, has launched its cheapest 5G device which the firm…

    AI

    MIT in the media: 2023 in review | Ztoog

    It was an eventful journey round the solar for MIT this 12 months, from President…

    AI

    Building supply chain resilience with AI

    The Canadian fertilizer firm Nutrien, for instance, operates two dozen manufacturing and processing amenities unfold…

    Our Picks
    Science

    Caves seen on the surface of a comet for the first time

    Mobile

    Try Galaxy app now allows iPhone users to see what foldables are like

    Crypto

    How Crypto Mining Stocks Have Performed In Comparison To Bitcoin

    Categories
    • AI (1,482)
    • Crypto (1,744)
    • Gadgets (1,795)
    • Mobile (1,838)
    • Science (1,852)
    • Technology (1,789)
    • The Future (1,635)
    Most Popular
    Mobile

    ASUSTOR AS3302T v2 review: A good 2-bay starter NAS with 2.5GbE connectivity

    AI

    Meet FinGPT: An Open-Source Financial Large Language Model (LLMs)

    The Future

    Image-generating AI creates uncanny optical illusions

    Ztoog
    Facebook X (Twitter) Instagram Pinterest
    • Home
    • About Us
    • Contact us
    • Privacy Policy
    • Terms & Conditions
    © 2025 Ztoog.

    Type above and press Enter to search. Press Esc to cancel.