Close Menu
Ztoog
    What's Hot
    The Future

    Companion’s Director and Cast on Its Gruesome Finish and Future

    AI

    Using societal context knowledge to foster the responsible application of AI – Google Research Blog

    Crypto

    Ethereum Defies Expectations With Lower Volatility Than Bitcoin

    Important Pages:
    • About Us
    • Contact us
    • Privacy Policy
    • Terms & Conditions
    Facebook X (Twitter) Instagram Pinterest
    Facebook X (Twitter) Instagram Pinterest
    Ztoog
    • Home
    • The Future

      How I Turn Unstructured PDFs into Revenue-Ready Spreadsheets

      Is it the best tool for 2025?

      The clocks that helped define time from London’s Royal Observatory

      Summer Movies Are Here, and So Are the New Popcorn Buckets

      India-Pak conflict: Pak appoints ISI chief, appointment comes in backdrop of the Pahalgam attack

    • Technology

      Ensure Hard Work Is Recognized With These 3 Steps

      Cicada map 2025: Where will Brood XIV cicadas emerge this spring?

      Is Duolingo the face of an AI jobs crisis?

      The US DOD transfers its AI-based Open Price Exploration for National Security program to nonprofit Critical Minerals Forum to boost Western supply deals (Ernest Scheyder/Reuters)

      The more Google kills Fitbit, the more I want a Fitbit Sense 3

    • Gadgets

      Maono Caster G1 Neo & PD200X Review: Budget Streaming Gear for Aspiring Creators

      Apple plans to split iPhone 18 launch into two phases in 2026

      Upgrade your desk to Starfleet status with this $95 USB-C hub

      37 Best Graduation Gift Ideas (2025): For College Grads

      Backblaze responds to claims of “sham accounting,” customer backups at risk

    • Mobile

      Samsung Galaxy S25 Edge promo materials leak

      What are people doing with those free T-Mobile lines? Way more than you’d expect

      Samsung doesn’t want budget Galaxy phones to use exclusive AI features

      COROS’s charging adapter is a neat solution to the smartwatch charging cable problem

      Fortnite said to return to the US iOS App Store next week following court verdict

    • Science

      Failed Soviet probe will soon crash to Earth – and we don’t know where

      Trump administration cuts off all future federal funding to Harvard

      Does kissing spread gluten? New research offers a clue.

      Why Balcony Solar Panels Haven’t Taken Off in the US

      ‘Dark photon’ theory of light aims to tear up a century of physics

    • AI

      How to build a better AI benchmark

      Q&A: A roadmap for revolutionizing health care through data-driven innovation | Ztoog

      This data set helps researchers spot harmful stereotypes in LLMs

      Making AI models more trustworthy for high-stakes settings | Ztoog

      The AI Hype Index: AI agent cyberattacks, racing robots, and musical models

    • Crypto

      ‘The Big Short’ Coming For Bitcoin? Why BTC Will Clear $110,000

      Bitcoin Holds Above $95K Despite Weak Blockchain Activity — Analytics Firm Explains Why

      eToro eyes US IPO launch as early as next week amid easing concerns over Trump’s tariffs

      Cardano ‘Looks Dope,’ Analyst Predicts Big Move Soon

      Speak at Ztoog Disrupt 2025: Applications now open

    Ztoog
    Home » Researchers from UC Berkeley Propose RingAttention: A Memory-Efficient Artificial Intelligence Approach to Reduce the Memory Requirements of Transformers
    AI

    Researchers from UC Berkeley Propose RingAttention: A Memory-Efficient Artificial Intelligence Approach to Reduce the Memory Requirements of Transformers

    Facebook Twitter Pinterest WhatsApp
    Researchers from UC Berkeley Propose RingAttention: A Memory-Efficient Artificial Intelligence Approach to Reduce the Memory Requirements of Transformers
    Share
    Facebook Twitter LinkedIn Pinterest WhatsApp

    A kind of deep studying mannequin structure known as Transformers in the context of many state-of-the-art AI fashions. They have revolutionized the subject of synthetic intelligence, significantly in pure language processing and numerous different duties in machine studying. It relies on a self-attention mechanism the place the mannequin weighs the significance of totally different components of the enter sequence when making predictions. They consist of an encoder and a decoder to course of the inputs.  

    However, scaling up the context size of Transformers takes lots of work. It is due to the inherited self-attention. Self-attention has reminiscence price quadratic in the enter sequence size, which makes it difficult to scale to the longer enter sequences. Researchers at UC Berkley developed a technique known as Ring Attention to deal with this based mostly on a easy remark. They noticed that when self-attention and feedforward community computations are carried out blockwise, the sequences will be distributed throughout a number of gadgets and simply analyzed.

    They distribute the outer loop of computing blockwise consideration amongst hosts, every machine managing its respective enter block. For the inside loop, they compute blockwise consideration and feedforward operations particular to its designated enter block for all gadgets. Their host gadgets kind a conceptual ring and ship a replica of its key-value blocks getting used for blockwise computation to the subsequent machine in the ring. They additionally concurrently obtain key-value blocks from the earlier one.

    The block computations take longer than block transfers. The group overlapped these processes, leading to no added overhead in contrast to customary transformers. By doing so, every machine requires solely reminiscence proportional to the block measurement, impartial of the unique enter sequence size. This successfully eliminates the reminiscence constraints imposed by particular person gadgets. 

    Their experiments present that Ring Attention can scale back the reminiscence necessities of Transformers by enabling them to practice greater than 500 instances longer sequences than prior reminiscence environment friendly state-of-the-arts. This technique additionally permits coaching sequences that exceed 100 million in size with out making approximations to consideration. As Ring Attention eliminates the reminiscence constraints imposed by particular person gadgets, one also can obtain near-infinite context sizes. However, one would require many quantity of gadgets as sequence size is proportional to the quantity of gadgets.

    The analysis solely includes an analysis of the effectiveness of the technique with out the large-scale coaching fashions. As the scale context size depends upon the quantity of gadgets, the mannequin’s effectivity depends upon the optimization; they’ve solely labored on the low-level operations required for reaching optimum laptop efficiency. The researchers say that they want to work on each most sequence size and most laptop efficiency in the future. The chance of near-infinite context introduces many thrilling alternatives, comparable to massive video-audio-language fashions, studying from prolonged suggestions and trial-and-errors, understanding and producing codebase, and adapting AI fashions to perceive scientific information comparable to gene sequences.


    Check out the Paper. All Credit For This Research Goes To the Researchers on This Project. Also, don’t neglect to be a part of our 31k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, the place we share the newest AI analysis information, cool AI tasks, and extra.

    If you want our work, you’ll love our e-newsletter..

    We are additionally on WhatsApp. Join our AI Channel on Whatsapp..


    Arshad is an intern at MarktechPost. He is at present pursuing his Int. MSc Physics from the Indian Institute of Technology Kharagpur. Understanding issues to the elementary stage leads to new discoveries which lead to development in expertise. He is keen about understanding the nature basically with the assist of instruments like mathematical fashions, ML fashions and AI.


    ▶️ Now Watch AI Research Updates On Our Youtube Channel [Watch Now]

    Share. Facebook Twitter Pinterest LinkedIn WhatsApp

    Related Posts

    AI

    How to build a better AI benchmark

    AI

    Q&A: A roadmap for revolutionizing health care through data-driven innovation | Ztoog

    AI

    This data set helps researchers spot harmful stereotypes in LLMs

    AI

    Making AI models more trustworthy for high-stakes settings | Ztoog

    AI

    The AI Hype Index: AI agent cyberattacks, racing robots, and musical models

    AI

    Novel method detects microbial contamination in cell cultures | Ztoog

    AI

    Seeing AI as a collaborator, not a creator

    AI

    “Periodic table of machine learning” could fuel AI discovery | Ztoog

    Leave A Reply Cancel Reply

    Follow Us
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    Top Posts
    Mobile

    AT&T is offering the Pixel Fold for half the price

    After many months of leaks and hypothesis, Google lastly made the Pixel Fold official final…

    Crypto

    SBF’s trial is coming to a close — here’s what you missed

    Welcome again to Chain Reaction. To get a roundup of Ztoog’s greatest and most essential…

    AI

    Finding value in generative AI for financial services

    According to a McKinsey report, generative AI may add $2.6 trillion to $4.4 trillion yearly in…

    Gadgets

    The best boxing gloves in 2023, according to experts

    We might earn income from the merchandise out there on this web page and take…

    Gadgets

    The best camping chairs of 2023

    We might earn income from the merchandise obtainable on this web page and take part…

    Our Picks
    Mobile

    The problem with Passkeys | Android Central

    The Future

    Watch a robot with living muscles walk through water

    Technology

    Sources: Appin co-founder Rajat Khare used law firms to threaten outlets in the US, UK, and other countries to kill stories about the Indian hack-for-hire firm (Lachlan Cartwright/The Daily Beast)

    Categories
    • AI (1,482)
    • Crypto (1,744)
    • Gadgets (1,796)
    • Mobile (1,839)
    • Science (1,853)
    • Technology (1,789)
    • The Future (1,635)
    Most Popular
    Crypto

    SBF trial: Everything to know from the FTX courtroom ahead of his testimony

    AI

    Four things you need to know about China’s AI talent pool 

    Crypto

    What’s Happened In The Past?

    Ztoog
    Facebook X (Twitter) Instagram Pinterest
    • Home
    • About Us
    • Contact us
    • Privacy Policy
    • Terms & Conditions
    © 2025 Ztoog.

    Type above and press Enter to search. Press Esc to cancel.