Close Menu
    Trending
    • Revisiting Benchmarking of Tabular Reinforcement Learning Methods
    • Is Your AI Whispering Secrets? How Scientists Are Teaching Chatbots to Forget Dangerous Tricks | by Andreas Maier | Jul, 2025
    • Qantas data breach to impact 6 million airline customers
    • He Went From $471K in Debt to Teaching Others How to Succeed
    • An Introduction to Remote Model Context Protocol Servers
    • Blazing-Fast ML Model Serving with FastAPI + Redis (Boost 10x Speed!) | by Sarayavalasaravikiran | AI Simplified in Plain English | Jul, 2025
    • AI Knowledge Bases vs. Traditional Support: Who Wins in 2025?
    • Why Your Finance Team Needs an AI Strategy, Now
    AIBS News
    • Home
    • Artificial Intelligence
    • Machine Learning
    • AI Technology
    • Data Science
    • More
      • Technology
      • Business
    AIBS News
    Home»Machine Learning»#1minPapers Francois Chollet: use LLMs for tree-search instead of next token prediction | by Gwen Cheni | Jan, 2025
    Machine Learning

    #1minPapers Francois Chollet: use LLMs for tree-search instead of next token prediction | by Gwen Cheni | Jan, 2025

    Team_AIBS NewsBy Team_AIBS NewsJanuary 10, 2025No Comments3 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    Not a paper, however 90min of Chollet is all the time value watching! The ARC Problem is fascinating as a result of it’s a fast adaptation, evolution, emergence of latest mannequin species.

    1. Check-time compute elevated efficiency from 10% accuracy to 50–60% accuracy in ARC, however test-time compute is just attainable for quantifiable input-output pairs like duties in ARC.
    2. Finetuning will be autonomous when utilizing demonstration pairs.
    3. Within the 2020 ARC Kaggle competitors, highest rating was 20% by way of brute drive. However combining the entire submissions bought to 49% (people would get to 99% accuracy), as a result of 1/2 of the non-public take a look at set was brute-force-able, which implies the benchmark was flawed (inadequate process variety and complexity). Must co-evolve the issue with the answer.
    4. If an enter is steady, neural networks (discrete symbolic applications) will not be a superb construction to strategy these kinds of sample recog issues. Vector-based applications could also be higher at sure issues.
    5. Induction is formally verifiable. Transduction is guessing what the reply is likely to be, with out a solution to confirm if it’s the fitting guess — all of the incorrect solutions are incorrect for various causes, however the fitting reply is correct for a similar purpose. Transduction requires extra sampling. Higher to begin with induction, but when induction doesn’t work, fall again to transduction.
    6. In the event you have a look at the issue from totally different angles, you usually tend to provide you with the true form of the issue. Very true for NN as a result of NN tends to latch onto noise and irregularities. Completely different angles additionally act as a regularization mechanism, the place noises from diff angles counter one another.
    7. Utilizing a VAE learns a way more structured, smoother latent areas, which is essential to creating test-time gradient descent work.
    8. Chollet would clear up ARC by way of deep-learning program synthesis, not utilizing LLMs for subsequent token technology, however as a graph of operators. Program synthesis is a tree-search course of. Use LLMs to information this tree-search course of.
    9. People clear up ARC by first describing the objects, contents, properties, causal relationships, then use this to constrain the search house, doubtlessly even eliminating the necessity for search.
    10. Turing-complete language (Python) vs DSL? The language should be capable of be taught, such that upon seeing an analogous drawback, it will possibly save compute. It additionally wants to put in writing higher-level capabilities.
    11. The basic cognitive unit in our mind is fuzzy sample recognition. System2 planning is making use of our instinct in a structured kind — which is deep-learning program synthesis. Iteratively guessing with guardrails, to assemble a symbolic artifact. With out guardrails, it’s dreaming- constantly intuiting with out consistency to the previous. Consistency requires back-and-forth loops, bringing the previous to the current.
    12. Some recombination patterns of the constructing blocks will happen extra typically in sure contexts, extract this as a reservoir kind (higher-level abstraction fitted to the issue), add it again to the constructing blocks, such that subsequent time you clear up it in fewer steps.
    13. Speculate how o1 works: search course of within the house of attainable chains-of-thought. By backtracking and enhancing which branches work higher, it finally ends up with a pure language program describing what the mannequin needs to be doing, adapting to novelty. It’s clearly doing search in chain-of-thought house at test-time: telltale signal is compute and latency rising.
    14. Full interview right here: https://www.youtube.com/watch?v=w9WE1aOPjHc



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleSupreme Court to hear TikTok’s final plea against US ban
    Next Article The Best Way to Prepare for Data Science and Machine Learning Interviews | by Marina Wyss – Gratitude Driven
    Team_AIBS News
    • Website

    Related Posts

    Machine Learning

    Is Your AI Whispering Secrets? How Scientists Are Teaching Chatbots to Forget Dangerous Tricks | by Andreas Maier | Jul, 2025

    July 2, 2025
    Machine Learning

    Blazing-Fast ML Model Serving with FastAPI + Redis (Boost 10x Speed!) | by Sarayavalasaravikiran | AI Simplified in Plain English | Jul, 2025

    July 2, 2025
    Machine Learning

    From Training to Drift Monitoring: End-to-End Fraud Detection in Python | by Aakash Chavan Ravindranath, Ph.D | Jul, 2025

    July 1, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Revisiting Benchmarking of Tabular Reinforcement Learning Methods

    July 2, 2025

    I Tried Buying a Car Through Amazon: Here Are the Pros, Cons

    December 10, 2024

    Amazon and eBay to pay ‘fair share’ for e-waste recycling

    December 10, 2024

    Artificial Intelligence Concerns & Predictions For 2025

    December 10, 2024

    Barbara Corcoran: Entrepreneurs Must ‘Embrace Change’

    December 10, 2024
    Categories
    • AI Technology
    • Artificial Intelligence
    • Business
    • Data Science
    • Machine Learning
    • Technology
    Most Popular

    Transformer Crisis: Deepak Divan Sounds the Alarm

    February 10, 2025

    WhatsApp backs Apple in its legal row with the UK over user data

    June 11, 2025

    5 Ways to Spend Less and Sell More

    February 25, 2025
    Our Picks

    Revisiting Benchmarking of Tabular Reinforcement Learning Methods

    July 2, 2025

    Is Your AI Whispering Secrets? How Scientists Are Teaching Chatbots to Forget Dangerous Tricks | by Andreas Maier | Jul, 2025

    July 2, 2025

    Qantas data breach to impact 6 million airline customers

    July 2, 2025
    Categories
    • AI Technology
    • Artificial Intelligence
    • Business
    • Data Science
    • Machine Learning
    • Technology
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2024 Aibsnews.comAll Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.