Close Menu
Ztoog
    What's Hot
    Mobile

    Google removes decade-old feature from the camera app on Pixel 8, Pixel 8 Pro

    Gadgets

    Discord starts down the dangerous road of ads this week

    Crypto

    Why Is Bitcoin Price Up Today?

    Important Pages:
    • About Us
    • Contact us
    • Privacy Policy
    • Terms & Conditions
    Facebook X (Twitter) Instagram Pinterest
    Facebook X (Twitter) Instagram Pinterest
    Ztoog
    • Home
    • The Future

      OPPO launches A5 Pro 5G: Premium features at a budget price

      How I Turn Unstructured PDFs into Revenue-Ready Spreadsheets

      Is it the best tool for 2025?

      The clocks that helped define time from London’s Royal Observatory

      Summer Movies Are Here, and So Are the New Popcorn Buckets

    • Technology

      What It Is and Why It Matters—Part 1 – O’Reilly

      Ensure Hard Work Is Recognized With These 3 Steps

      Cicada map 2025: Where will Brood XIV cicadas emerge this spring?

      Is Duolingo the face of an AI jobs crisis?

      The US DOD transfers its AI-based Open Price Exploration for National Security program to nonprofit Critical Minerals Forum to boost Western supply deals (Ernest Scheyder/Reuters)

    • Gadgets

      Maono Caster G1 Neo & PD200X Review: Budget Streaming Gear for Aspiring Creators

      Apple plans to split iPhone 18 launch into two phases in 2026

      Upgrade your desk to Starfleet status with this $95 USB-C hub

      37 Best Graduation Gift Ideas (2025): For College Grads

      Backblaze responds to claims of “sham accounting,” customer backups at risk

    • Mobile

      Samsung Galaxy S25 Edge promo materials leak

      What are people doing with those free T-Mobile lines? Way more than you’d expect

      Samsung doesn’t want budget Galaxy phones to use exclusive AI features

      COROS’s charging adapter is a neat solution to the smartwatch charging cable problem

      Fortnite said to return to the US iOS App Store next week following court verdict

    • Science

      Nothing is stronger than quantum connections – and now we know why

      Failed Soviet probe will soon crash to Earth – and we don’t know where

      Trump administration cuts off all future federal funding to Harvard

      Does kissing spread gluten? New research offers a clue.

      Why Balcony Solar Panels Haven’t Taken Off in the US

    • AI

      Hybrid AI model crafts smooth, high-quality videos in seconds | Ztoog

      How to build a better AI benchmark

      Q&A: A roadmap for revolutionizing health care through data-driven innovation | Ztoog

      This data set helps researchers spot harmful stereotypes in LLMs

      Making AI models more trustworthy for high-stakes settings | Ztoog

    • Crypto

      Ethereum Breaks Key Resistance In One Massive Move – Higher High Confirms Momentum

      ‘The Big Short’ Coming For Bitcoin? Why BTC Will Clear $110,000

      Bitcoin Holds Above $95K Despite Weak Blockchain Activity — Analytics Firm Explains Why

      eToro eyes US IPO launch as early as next week amid easing concerns over Trump’s tariffs

      Cardano ‘Looks Dope,’ Analyst Predicts Big Move Soon

    Ztoog
    Home » Generative Logic – O’Reilly
    Technology

    Generative Logic – O’Reilly

    Facebook Twitter Pinterest WhatsApp
    Generative Logic – O’Reilly
    Share
    Facebook Twitter LinkedIn Pinterest WhatsApp

    Alibaba’s newest mannequin, QwQ-32B-Preview, has gained some spectacular opinions for its reasoning skills. Like OpenAI’s GPT-4 o1,1 its coaching has emphasised reasoning slightly than simply reproducing language. That appeared like one thing value testing out—or not less than enjoying round with—so after I heard that it in a short time grew to become accessible in Ollama and wasn’t too massive to run on a reasonably well-equipped laptop computer, I downloaded QwQ and tried it out. I additionally tried just a few competing fashions: GPT-4 o1 and Gemma-2-27B. GPT-4 o1 was the primary mannequin to say that it had been educated particularly for reasoning. Gemma, so far as I do know, makes no such declare.

    How do you check a reasoning mannequin? It’s not as simple as simply checking information. I believed, “Maybe I could ask it to prove Fermat’s Last Theorem,” however I shortly realized that the one response I’d be capable of consider was “I can’t prove this.” (Which was, in reality, the response I acquired from all three fashions.) I can’t show Fermat’s Last Theorem both, nor do I declare to grasp any of the large proof that mathematicians have accepted. If QwQ sputtered out just a few thousand tokens of “proof,” I might pretty confidently say that its reasoning was flawed—simply because I do know the precise proof is for much longer, 100 pages of dense mathematical reasoning. But that’s hardly a legitimate check.



    Learn sooner. Dig deeper. See farther.

    What about Fermat’s Little Theorem? That’s rather a lot simpler, but it surely’s a textbook instance, so regardless of the consequence, I’d by no means know whether or not QwQ reasoned appropriately or if it simply managed to parrot one thing from its coaching set. There are various math textbooks on-line, and it’s honest to imagine that every one of them are within the coaching knowledge.

    So I attempted a barely extra obscure instance: Prove that if a quantity consists solely of ones (like 11,111,111) and is prime, the variety of digits (all 1s) should even be prime. I’ve managed to show this, but it surely’s not an issue that’s well-known sufficient to draw textbook consideration. So let’s go! Here’s QwQ’s proof:

    >>> Can you show that if all of the digits of a chief quantity are 1,
    then the variety of digits should be prime?

    Let’s deal with this drawback step-by-step. So, the assertion is: If all of the digits of a chief quantity are 1, then the variety of digits should be prime. Hmm, fascinating. Let’s break this down.

    First, let’s perceive what we’re coping with. We have a chief quantity the place all its digits are 1. For instance, numbers like 11, 111, 1111, and so forth., however solely contemplating these which can be prime. We must show that the variety of digits in such a chief quantity should itself be a chief quantity.

    Let me take into consideration this. Let’s denote the variety of digits by ( n ). So, the quantity consists of ( n ) ones. Such numbers are referred to as repunit numbers. A repunit is a quantity consisting solely of ones in its decimal illustration. For instance, ( R_2 = 11 ), ( R_3 = 111 ), ( R_4 = 1111 ), and so forth, the place ( R_n ) denotes a repunit of size ( n ).

    Our aim is to indicate that if ( R_n ) is prime, then ( n ) should be prime.

    Let’s think about the contrapositive of this assertion. The contrapositive of “If ( R_n ) is prime, then ( n ) is prime” is “If ( n ) is not prime, then ( R_n ) is not prime.” If we are able to show the contrapositive, that can set up the unique assertion.

    …and so forth, till…

    **Final Answer**
    [ boxed{text{If } R_n text{ is prime,
    then } n text{ is prime.}} ]

    I’ve spared you the (very) chatty proof—if you wish to see it, put QwQ to the textual content your self. Suffice it to say that it was right. Both GPT-4 o1 and Gemma-2-27B gave me the identical proof however with much less chatter. Gemma skipped some steps {that a} human may think about apparent.

    Next, can we go a step farther and generalize?

    >>> does this proof maintain for
    bases apart from base 10?

    That’s an fascinating query! Let’s discover whether or not this proof could be prolonged to bases apart from base 10.

    In base ( b ), a repunit of size ( n ) is outlined as:
    [ R_n^{(b)} = frac{b^n - 1}{b - 1} ]

    …and so forth…

    Again, attempt it your self—and once more, QwQ proved it appropriately. This second proposition isn’t as prone to present up in textbooks or different sources; it’s much less of a proof than an remark that the unique proof didn’t make any assumptions in regards to the base.

    When I requested GPT to show the identical theorem, I acquired a really related (and proper) proof, said extra formally and with much less shade commentary. That isn’t significantly stunning, since GPT has additionally been educated to motive. I used to be extra stunned to see that Gemma-2-27B additionally gave me an accurate proof. Gemma has been educated on mathematical texts however not particularly on “reasoning.” (Perhaps Google’s advertising and marketing by no means thought to name this coaching “reasoning.”) Gemma omitted a few of the steps—steps a daily human would most likely omit as apparent however {that a} mathematician would write out for completeness. (Just to verify, I requested GPT to substantiate that Gemma’s proof was right. It agreed.)

    Have we confirmed that coaching fashions to motive “works”? Well, we are able to’t declare to have confirmed something on the premise of 1 profitable trial—or, for that matter, on the premise of an especially massive variety of trials. (In case you’re questioning, Gemma-2-7B, a good smaller mannequin, failed.) But now we have discovered one thing essential. Think in regards to the measurement of the fashions: OpenAI has mentioned nothing in regards to the measurement of GPT-4 o1, however it’s rumored to have over a trillion parameters. QwQ weighs in at 32 billion parameters, and Gemma-2-27B at 27 billion. So QwQ and Gemma2 are between practically two orders of magnitude smaller than GPT. Furthermore, GPT runs on what should be thought-about one of many world’s largest supercomputers. We don’t know the dimensions, however we do know that OpenAI’s infrastructure is huge and contains a big proportion of the world’s high-end GPUs. QwQ and Gemma ran fortunately on my MacBook Pro. They made the fan spin and sucked down the battery however nothing extraordinary. Granted, GPT is serving hundreds of customers concurrently, so it isn’t actually a good comparability. But it’s vital to appreciate that GPT isn’t the one sport on the town and that fashions working regionally can equal GPT on nontrivial duties. Most individuals who have experimented with working fashions regionally have come to related conclusions, however take into consideration what this implies. If you’re constructing an AI utility, you don’t should tie your self to OpenAI. Smaller open fashions can do the job—and so they’ll protect you from OpenAI’s payments (and inevitable worth will increase), they’ll allow you to preserve your knowledge native, and so they’ll go away you answerable for your future.

    What else can we be taught? I’ve questioned how a language mannequin could be educated for logic; my instinct mentioned that might be a tougher and extra advanced drawback than coaching it for language. My instinct was mistaken. I don’t understand how these fashions had been educated, however I now assume that producing logic efficiently is, in some ways, easier than producing language. Why? QwQ’s verbosity provides us an enormous trace: “Let’s consider the contrapositive of this statement.” A contrapositive is solely a logical sample: If A implies B, then not B implies not A. What different logical patterns can we consider? Syllogisms: If A implies B and B implies C, then A implies C. Proof by contradiction: To show that A implies B, assume that A implies not B and present that assumption is fake. Induction: Show that if A(n) implies B(n), then A(n+1) implies B(n+1); then present that A(0) implies B(0).

    It could be simple to develop a for much longer listing of patterns. There are higher notations to symbolize these patterns, however an extended listing and higher representations aren’t vital right here. What’s vital is to appreciate that these are patterns—and that composing logical patterns into logical statements or proofs isn’t basically totally different from composing phrases (or tokens) into sentences. Is pushing patterns across the essence of logic? That’s not a good query: It’s logic when you do it appropriately, illogic when you don’t. The logic isn’t within the patterns however in understanding how one can assemble the patterns to unravel issues—and the method of assembling patterns needs to be the main target of coaching, tens of millions of examples of logical reasoning to mannequin the best way patterns are assembled into wholes. Any of those logical patterns can lead you astray when you’re not cautious; it’s simple to assemble false syllogisms by beginning with premises which can be incorrect. I don’t count on logic to treatment the issue of hallucination. But I believe that coaching a mannequin in logical patterns is a greater means for the mannequin to “learn” logic than merely coaching it on phrases (human utterances). That’s the wager that OpenAI, Alibaba, and presumably Google are making—and so they appear to be profitable.

    Can we go additional? Are there other forms of patterns that language fashions may very well be educated on? Yes. Generative AI has confirmed helpful for producing code however hasn’t (but) made important inroads into software program design. Could coaching fashions particularly on design patterns be a breakthrough?2 I don’t know, however I’d prefer to see somebody attempt. A mannequin specialised for software program design could be value having.

    Could we do higher with generative music if we educated fashions on the patterns analyzed in music principle, along with audio? Applications like Suno are a variety of enjoyable, however whenever you get all the way down to it, they’re simply repeating the clichés of widespread musical types. Would it assist to present Suno some data of music principle, data of the patterns behind music along with the music itself? Would language fashions write higher poetry in the event that they had been educated on the patterns present in poetic language (rhetorical gadgets, figurative speech) slightly than simply phrases? One of my first experiments with generative AI was to ask GPT-3 to put in writing a Petrarchan sonnet, which has a distinct construction from the extra widespread Shakespearian sonnet. GPT-3 and its contemporaries failed. It was a very long time earlier than I discovered a mannequin that would do this efficiently; though most fashions might outline a Petrarchan sonnet, they might solely generate Shakespearean sonnets. That era of fashions was educated solely on the phrases, not the bigger patterns.

    Is this a means ahead for AI? I don’t know, however I’d prefer to see AI researchers attempt. In the meantime, although, it’s sufficient to appreciate that, highly effective because the GPT fashions are, you’ll be able to run small open fashions on a laptop computer or a telephone that carry out equally nicely.


    Footnotes

    1. I examined on the Preview, which has now been promoted to GPT-4 o1. I didn’t retest with the ultimate o1, which presumably has had additional coaching and optimization.
    2. Design patterns are usually related to object-oriented design, however the idea is basically extra normal. Design patterns try to call for options to issues that you simply see each day; naming the answer means that you can speak about it. That definition is relevant to any self-discipline, together with useful programming and (after all) structure.

    Share. Facebook Twitter Pinterest LinkedIn WhatsApp

    Related Posts

    Technology

    What It Is and Why It Matters—Part 1 – O’Reilly

    Technology

    Ensure Hard Work Is Recognized With These 3 Steps

    Technology

    Cicada map 2025: Where will Brood XIV cicadas emerge this spring?

    Technology

    Is Duolingo the face of an AI jobs crisis?

    Technology

    The US DOD transfers its AI-based Open Price Exploration for National Security program to nonprofit Critical Minerals Forum to boost Western supply deals (Ernest Scheyder/Reuters)

    Technology

    The more Google kills Fitbit, the more I want a Fitbit Sense 3

    Technology

    Sorry Shoppers, Amazon Says Tariff Cost Feature ‘Is Not Going to Happen’

    Technology

    Vibe Coding, Vibe Checking, and Vibe Blogging – O’Reilly

    Leave A Reply Cancel Reply

    Follow Us
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    Top Posts
    AI

    Enhancing Underwater Image Segmentation with Deep Learning: A Novel Approach to Dataset Expansion and Preprocessing Techniques

    Underwater picture processing mixed with machine studying provides important potential for enhancing the capabilities of…

    Gadgets

    Get an iPad Air, accessories, and Beats headphones for $99.97 with this Labor Day Savings

    We might earn income from the merchandise accessible on this web page and take part…

    Crypto

    Pudgy Penguins’ approach may be the answer to fixing NFTs’ revenue problems

    Pudgy Penguins was as soon as solely identified for its 8,888 NFT assortment. But ever…

    Technology

    KeeperFX keeps Dungeon Keeper alive by making it actually playable

    Enlarge / If it had been me, I might merely not burrow my means on…

    The Future

    OpenAI founder Sam Altman is ‘seeking $7 trillion investment’

    The CEO of OpenAI, Sam Altman, is looking for a $7 trillion funding to overtake…

    Our Picks
    The Future

    AI costs too much to automate vision-related jobs – for now

    The Future

    Agree With the Judges? Rewatch the Mike Tyson vs. Jake Paul Fight on Netflix

    Gadgets

    Green Chef’s Meal Kit Makes Dinner Delicious—and Organic

    Categories
    • AI (1,483)
    • Crypto (1,745)
    • Gadgets (1,796)
    • Mobile (1,839)
    • Science (1,854)
    • Technology (1,790)
    • The Future (1,636)
    Most Popular
    The Future

    Today’s NYT Strands Hints, Answer and Help for August 3 #153

    The Future

    How to Enhance Energy Efficiency in the Industrial Sector

    Crypto

    0G Labs launches with whopping $35M pre-seed to build a modular AI blockchain

    Ztoog
    Facebook X (Twitter) Instagram Pinterest
    • Home
    • About Us
    • Contact us
    • Privacy Policy
    • Terms & Conditions
    © 2025 Ztoog.

    Type above and press Enter to search. Press Esc to cancel.