Close Menu
Ztoog
    What's Hot
    Science

    High altitude balloons spy on dark matter

    Mobile

    Get $50 for reserving the Galaxy Ring right now ahead of Samsung Unpacked

    Crypto

    Bitcoin Price Falls Below Short-Term Holders’ Realized Price Of $66,200

    Important Pages:
    • About Us
    • Contact us
    • Privacy Policy
    • Terms & Conditions
    Facebook X (Twitter) Instagram Pinterest
    Facebook X (Twitter) Instagram Pinterest
    Ztoog
    • Home
    • The Future

      Any wall can be turned into a camera to see around corners

      JD Vance and President Trump’s Sons Hype Bitcoin at Las Vegas Conference

      AI may already be shrinking entry-level jobs in tech, new research suggests

      Today’s NYT Strands Hints, Answer and Help for May 26 #449

      LiberNovo Omni: The World’s First Dynamic Ergonomic Chair

    • Technology

      A Replit employee details a critical security flaw in web apps created using AI-powered app builder Lovable that exposes API keys and personal info of app users (Reed Albergotti/Semafor)

      Gemini in Google Drive can now help you skip watching that painfully long Zoom meeting

      Apple iPhone exports from China to the US fall 76% as India output surges

      Today’s NYT Wordle Hints, Answer and Help for May 26, #1437

      5 Skills Kids (and Adults) Need in an AI World – O’Reilly

    • Gadgets

      Future-proof your career by mastering AI skills for just $20

      8 Best Vegan Meal Delivery Services and Kits (2025), Tested and Reviewed

      Google Home is getting deeper Gemini integration and a new widget

      Google Announces AI Ultra Subscription Plan With Premium Features

      Google shows off Android XR-based glasses, announces Warby Parker team-up

    • Mobile

      Microsoft is done being subtle – this new tool screams “upgrade now”

      Wallpaper Wednesday: Android wallpapers 2025-05-28

      Google can make smart glasses accessible with Warby Parker, Gentle Monster deals

      vivo T4 Ultra specs leak

      Forget screens: more details emerge on the mysterious Jony Ive + OpenAI device

    • Science

      Analysts Say Trump Trade Wars Would Harm the Entire US Energy Sector, From Oil to Solar

      Do we have free will? Quantum experiments may soon reveal the answer

      Was Planet Nine exiled from the solar system as a baby?

      How farmers can help rescue water-loving birds

      A trip to the farm where loofahs grow on vines

    • AI

      Rationale engineering generates a compact new tool for gene therapy | Ztoog

      The AI Hype Index: College students are hooked on ChatGPT

      Learning how to predict rare kinds of failures | Ztoog

      Anthropic’s new hybrid AI model can work on tasks autonomously for hours at a time

      AI learns how vision and sound are connected, without human intervention | Ztoog

    • Crypto

      GameStop bought $500 million of bitcoin

      CoinW Teams Up with Superteam Europe to Conclude Solana Hackathon and Accelerate Web3 Innovation in Europe

      Ethereum Net Flows Turn Negative As Bulls Push For $3,500

      Bitcoin’s Power Compared To Nuclear Reactor By Brazilian Business Leader

      Senate advances GENIUS Act after cloture vote passes

    Ztoog
    Home » Automated system teaches users when to collaborate with an AI assistant | Ztoog
    AI

    Automated system teaches users when to collaborate with an AI assistant | Ztoog

    Facebook Twitter Pinterest WhatsApp
    Automated system teaches users when to collaborate with an AI assistant | Ztoog
    Share
    Facebook Twitter LinkedIn Pinterest WhatsApp

    Artificial intelligence fashions that pick patterns in pictures can usually accomplish that higher than human eyes — however not all the time. If a radiologist is utilizing an AI mannequin to assist her decide whether or not a affected person’s X-rays present indicators of pneumonia, when ought to she belief the mannequin’s recommendation and when ought to she ignore it?

    A personalized onboarding course of may assist this radiologist reply that query, in accordance to researchers at MIT and the MIT-IBM Watson AI Lab. They designed a system that teaches a person when to collaborate with an AI assistant.

    In this case, the coaching methodology would possibly discover conditions the place the radiologist trusts the mannequin’s recommendation — besides she shouldn’t as a result of the mannequin is incorrect. The system mechanically learns guidelines for the way she ought to collaborate with the AI, and describes them with pure language.

    During onboarding, the radiologist practices collaborating with the AI utilizing coaching workouts primarily based on these guidelines, receiving suggestions about her efficiency and the AI’s efficiency.

    The researchers discovered that this onboarding process led to a few 5 % enchancment in accuracy when people and AI collaborated on an picture prediction activity. Their outcomes additionally present that simply telling the person when to belief the AI, with out coaching, led to worse efficiency.

    Importantly, the researchers’ system is totally automated, so it learns to create the onboarding course of primarily based on information from the human and AI performing a selected activity. It may also adapt to totally different duties, so it may be scaled up and utilized in many conditions the place people and AI fashions work collectively, akin to in social media content material moderation, writing, and programming.

    “So often, people are given these AI tools to use without any training to help them figure out when it is going to be helpful. That’s not what we do with nearly every other tool that people use — there is almost always some kind of tutorial that comes with it. But for AI, this seems to be missing. We are trying to tackle this problem from a methodological and behavioral perspective,” says Hussein Mozannar, a graduate pupil within the Social and Engineering Systems doctoral program inside the Institute for Data, Systems, and Society (IDSS) and lead creator of a paper about this coaching course of.

    The researchers envision that such onboarding might be an important a part of coaching for medical professionals.

    “One could imagine, for example, that doctors making treatment decisions with the help of AI will first have to do training similar to what we propose. We may need to rethink everything from continuing medical education to the way clinical trials are designed,” says senior creator David Sontag, a professor of EECS, a member of the MIT-IBM Watson AI Lab and the MIT Jameel Clinic, and the chief of the Clinical Machine Learning Group of the Computer Science and Artificial Intelligence Laboratory (CSAIL).

    Mozannar, who can also be a researcher with the Clinical Machine Learning Group, is joined on the paper by Jimin J. Lee, an undergraduate in electrical engineering and pc science; Dennis Wei, a senior analysis scientist at IBM Research; and Prasanna Sattigeri and Subhro Das, analysis employees members on the MIT-IBM Watson AI Lab. The paper might be introduced on the Conference on Neural Information Processing Systems.

    Training that evolves

    Existing onboarding strategies for human-AI collaboration are sometimes composed of coaching supplies produced by human specialists for particular use circumstances, making them troublesome to scale up. Some associated strategies depend on explanations, the place the AI tells the person its confidence in every choice, however analysis has proven that explanations are not often useful, Mozannar says.

    “The AI model’s capabilities are constantly evolving, so the use cases where the human could potentially benefit from it are growing over time. At the same time, the user’s perception of the model continues changing. So, we need a training procedure that also evolves over time,” he provides.

    To accomplish this, their onboarding methodology is mechanically realized from information. It is constructed from a dataset that incorporates many cases of a activity, akin to detecting the presence of a visitors gentle from a blurry picture.

    The system’s first step is to gather information on the human and AI performing this activity. In this case, the human would attempt to predict, with the assistance of AI, whether or not blurry pictures include visitors lights.

    The system embeds these information factors onto a latent area, which is a illustration of information wherein comparable information factors are nearer collectively. It makes use of an algorithm to uncover areas of this area the place the human collaborates incorrectly with the AI. These areas seize cases the place the human trusted the AI’s prediction however the prediction was incorrect, and vice versa.

    Perhaps the human mistakenly trusts the AI when pictures present a freeway at night time.

    After discovering the areas, a second algorithm makes use of a big language mannequin to describe every area as a rule, utilizing pure language. The algorithm iteratively fine-tunes that rule by discovering contrasting examples. It would possibly describe this area as “ignore AI when it is a highway during the night.”

    These guidelines are used to construct coaching workouts. The onboarding system reveals an instance to the human, on this case a blurry freeway scene at night time, in addition to the AI’s prediction, and asks the person if the picture reveals visitors lights. The person can reply sure, no, or use the AI’s prediction.

    If the human is incorrect, they’re proven the proper reply and efficiency statistics for the human and AI on these cases of the duty. The system does this for every area, and on the finish of the coaching course of, repeats the workouts the human acquired incorrect.

    “After that, the human has learned something about these regions that we hope they will take away in the future to make more accurate predictions,” Mozannar says.

    Onboarding boosts accuracy

    The researchers examined this system with users on two duties — detecting visitors lights in blurry pictures and answering a number of alternative questions from many domains (akin to biology, philosophy, pc science, and many others.).

    They first confirmed users a card with details about the AI mannequin, the way it was educated, and a breakdown of its efficiency on broad classes. Users had been cut up into 5 teams: Some had been solely proven the cardboard, some went via the researchers’ onboarding process, some went via a baseline onboarding process, some went via the researchers’ onboarding process and got suggestions of when they need to or mustn’t belief the AI, and others had been solely given the suggestions.

    Only the researchers’ onboarding process with out suggestions improved users’ accuracy considerably, boosting their efficiency on the visitors gentle prediction activity by about 5 % with out slowing them down. However, onboarding was not as efficient for the question-answering activity. The researchers imagine it’s because the AI mannequin, ChatGPT, supplied explanations with every reply that convey whether or not it ought to be trusted.

    But offering suggestions with out onboarding had the alternative impact — users not solely carried out worse, they took extra time to make predictions.

    “When you only give someone recommendations, it seems like they get confused and don’t know what to do. It derails their process. People also don’t like being told what to do, so that is a factor as well,” Mozannar says.

    Providing suggestions alone may hurt the person if these suggestions are incorrect, he provides. With onboarding, alternatively, the most important limitation is the quantity of accessible information. If there aren’t sufficient information, the onboarding stage gained’t be as efficient, he says.

    In the long run, he and his collaborators need to conduct bigger research to consider the short- and long-term results of onboarding. They additionally need to leverage unlabeled information for the onboarding course of, and discover strategies to successfully cut back the variety of areas with out omitting vital examples.

    “People are adopting AI systems willy-nilly, and indeed AI offers great potential, but these AI agents still sometimes makes mistakes. Thus, it’s crucial for AI developers to devise methods that help humans know when it’s safe to rely on the AI’s suggestions,” says Dan Weld, professor emeritus on the Paul G. Allen School of Computer Science and Engineering on the University of Washington, who was not concerned with this analysis. “Mozannar et al. have created an innovative method for identifying situations where the AI is trustworthy, and (importantly) to describe them to people in a way that leads to better human-AI team interactions.”

    This work is funded, partly, by the MIT-IBM Watson AI Lab.

    Share. Facebook Twitter Pinterest LinkedIn WhatsApp

    Related Posts

    AI

    Rationale engineering generates a compact new tool for gene therapy | Ztoog

    AI

    The AI Hype Index: College students are hooked on ChatGPT

    AI

    Learning how to predict rare kinds of failures | Ztoog

    AI

    Anthropic’s new hybrid AI model can work on tasks autonomously for hours at a time

    AI

    AI learns how vision and sound are connected, without human intervention | Ztoog

    AI

    How AI is introducing errors into courtrooms

    AI

    With AI, researchers predict the location of virtually any protein within a human cell | Ztoog

    AI

    Google DeepMind’s new AI agent cracks real-world problems better than humans can

    Leave A Reply Cancel Reply

    Follow Us
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    Top Posts
    Mobile

    Samsung Galaxy Z Fold5 vs. Xiaomi Mix Fold 3

    The Samsung Galaxy Z fold collection has changed into an echelon within the foldable smartphone…

    Mobile

    Google Keep may soon use AI to help you make lists, and this is how

    What you want to knowGoogle plans to use synthetic intelligence in Google Keep to help…

    Technology

    Nvidia announces a range of AI products and updates, including the DGX GH200 supercomputer platform, data center networking tools, and a robotics platform (Bloomberg)

    Bloomberg: Nvidia announces a range of AI products and updates, including the DGX GH200 supercomputer…

    Crypto

    Hut 8 Secures $50 Million Credit Facility from Coinbase Credit

    Canadian prime crypto mining agency Hut 8 Mining Corp has obtained a million-dollar credit score…

    Crypto

    Andalusia Labs raises $48M Series A to improve digital asset risk infrastructure

    Digital asset risk infrastructure-focused Andalusia Labs, previously often known as RiskHarbor, has raised $48 million…

    Our Picks
    Gadgets

    Alexa and Google Assistant fall on hard times, agree to be speaker roommates

    Mobile

    Overplay turns your videos into mobile video games even without any coding knowledge

    The Future

    Time Doctor vs. TimeCamp: A head-to-head comparison for 2023

    Categories
    • AI (1,493)
    • Crypto (1,753)
    • Gadgets (1,805)
    • Mobile (1,850)
    • Science (1,866)
    • Technology (1,802)
    • The Future (1,648)
    Most Popular
    Gadgets

    Lava Teases ProWatch XN: The Next Generation Smartwatch With Gorilla Glass 3

    The Future

    Elon Musk rants about work from home folks again on Tesla earnings call

    AI

    A computer scientist pushes the boundaries of geometry | Ztoog

    Ztoog
    Facebook X (Twitter) Instagram Pinterest
    • Home
    • About Us
    • Contact us
    • Privacy Policy
    • Terms & Conditions
    © 2025 Ztoog.

    Type above and press Enter to search. Press Esc to cancel.