Yannic Kilcher
Yannic Kilcher
  • Видео 463
  • Просмотров 15 806 081
Hallucination-Free? Assessing the Reliability of Leading AI Legal Research Tools (Paper Explained)
#rag #hallucinations #legaltech
An in-depth look at a recent Stanford paper examining the degree of hallucinations in various LegalTech tools that incorporate LLMs.
OUTLINE:
0:00 - Intro
1:58 - What are legal research tools and how are large language models used by them?
5:30 - Overview and abstract of the paper
9:29 - What is a hallucination and why do they occur?
15:45 - What is retrieval augmented generation (RAG)?
25:00 - Why LLMs are a bad choice when reasoning is involved
29:16 - The products that were tested
32:00 - Some shady practices by the researchers in the back and forth with the legal research companies
37:00 - Legal technology companies’ marketing claims to eliminate or solve halluci...
Просмотров: 20 541

Видео

xLSTM: Extended Long Short-Term Memory
Просмотров 31 тыс.28 дней назад
xLSTM is an architecture that combines the recurrency and constant memory requirement of LSTMs with the large-scale training of transformers and achieves impressive results. Paper: arxiv.org/abs/2405.04517 Abstract: In the 1990s, the constant error carousel and gating were introduced as the central ideas of the Long Short-Term Memory (LSTM). Since then, LSTMs have stood the test of time and con...
[ML News] OpenAI is in hot waters (GPT-4o, Ilya Leaving, Scarlett Johansson legal action)
Просмотров 32 тыс.Месяц назад
#gpt4o #sky #scarlettjohansson After the release of their flagship model GPT-4o, OpenAI finds itself in multiple controversies and an exodus of senior personnel - notably Ilya Sutskever References: openai.com/index/gpt-4o-and-more-tools-to-chatgpt-free/ openai.com/index/hello-gpt-4o/ x.com/LiamFedus/status/1790064963966370209?t=rx2YBT9AdDdKPhI6dUH4zA&s=09 x.com/lmsysorg/status/17900975883997799...
ORPO: Monolithic Preference Optimization without Reference Model (Paper Explained)
Просмотров 21 тыс.2 месяца назад
Paper: arxiv.org/abs/2403.07691 Abstract: While recent preference alignment algorithms for language models have demonstrated promising results, supervised fine-tuning (SFT) remains imperative for achieving successful convergence. In this paper, we study the crucial role of SFT within the context of preference alignment, emphasizing that a minor penalty for the disfavored generation style is suf...
[ML News] Chips, Robots, and Models
Просмотров 28 тыс.2 месяца назад
OUTLINE: 0:00 - Intro 0:19 - Our next-generation Meta Training and Inference Accelerator 01:39 - ALOHA Unleashed 03:10 - Apple Inks $50M Deal with Shutterstock for AI Training Data 04:28 - OpenAI Researchers, Including Ally of Sutskever, Fired for Alleged Leaking 05:01 - Adobe's Ethical Firefly AI was Trained on Midjourney Images 05:52 - Trudeau announces $2.4billion for AI-related investments ...
TransformerFAM: Feedback attention is working memory
Просмотров 35 тыс.2 месяца назад
Paper: arxiv.org/abs/2404.09173 Abstract: While Transformers have revolutionized deep learning, their quadratic attention complexity hinders their ability to process infinitely long inputs. We propose Feedback Attention Memory (FAM), a novel Transformer architecture that leverages a feedback loop to enable the network to attend to its own latent representations. This design fosters the emergenc...
[ML News] Devin exposed | NeurIPS track for high school students
Просмотров 40 тыс.2 месяца назад
OUTLINE: 0:00 - Intro 0:21 - Debunking Devin: "First AI Software Engineer" Upwork lie exposed! 07:24 - NeurIPS 2024 will have a track for papers from high schoolers. 13:29 - Opus can operate as a Turing machine. 13:47 - An AI-Powered, Self-Running Propaganda Machine for $105 14:27 - TechScape: How cheap, outsourced labour in Africa is shaping AI English 16:25 - Is ChatGPT Transforming Academics...
Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention
Просмотров 51 тыс.2 месяца назад
Google researchers achieve supposedly infinite context attention via compressive memory. Paper: arxiv.org/abs/2404.07143 Abstract: This work introduces an efficient method to scale Transformer-based Large Language Models (LLMs) to infinitely long inputs with bounded memory and computation. A key component in our proposed approach is a new attention technique dubbed Infini-attention. The Infini-...
[ML News] Llama 3 changes the game
Просмотров 46 тыс.2 месяца назад
Meta's Llama 3 is out. New model, new license, new opportunities. References: llama.meta.com/llama3/ ai.meta.com/blog/meta-llama-3/ github.com/meta-llama/llama3/blob/main/MODEL_CARD.md llama.meta.com/trust-and-safety/ ai.meta.com/research/publications/cyberseceval-2-a-wide-ranging-cybersecurity-evaluation-suite-for-large-language-models/ github.com/meta-llama/llama-recipes/tree/main/recipes/res...
Hugging Face got hacked
Просмотров 31 тыс.2 месяца назад
Links: Homepage: ykilcher.com Merch: ykilcher.com/merch RUclips: ruclips.net/user/yannickilcher Twitter: ykilcher Discord: ykilcher.com/discord LinkedIn: www.linkedin.com/in/ykilcher If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): Subscri...
[ML News] Microsoft to spend 100 BILLION DOLLARS on supercomputer (& more industry news)
Просмотров 21 тыс.2 месяца назад
Some updates from industry in the Machine Learning world Links: Homepage: ykilcher.com Merch: ykilcher.com/merch RUclips: ruclips.net/user/yannickilcher Twitter: ykilcher Discord: ykilcher.com/discord LinkedIn: www.linkedin.com/in/ykilcher If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and vol...
[ML News] Jamba, CMD-R+, and other new models (yes, I know this is like a week behind 🙃)
Просмотров 25 тыс.2 месяца назад
A flurry of new models continues to appear. Links: Homepage: ykilcher.com Merch: ykilcher.com/merch RUclips: ruclips.net/user/yannickilcher Twitter: ykilcher Discord: ykilcher.com/discord LinkedIn: www.linkedin.com/in/ykilcher If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a...
Flow Matching for Generative Modeling (Paper Explained)
Просмотров 40 тыс.2 месяца назад
Flow matching is a more general method than diffusion and serves as the basis for models like Stable Diffusion 3. Paper: arxiv.org/abs/2210.02747 Abstract: We introduce a new paradigm for generative modeling built on Continuous Normalizing Flows (CNFs), allowing us to train CNFs at unprecedented scale. Specifically, we present the notion of Flow Matching (FM), a simulation-free approach for tra...
Beyond A*: Better Planning with Transformers via Search Dynamics Bootstrapping (Searchformer)
Просмотров 33 тыс.2 месяца назад
Paper: arxiv.org/abs/2402.14083 Abstract: While Transformers have enabled tremendous progress in various application settings, such architectures still lag behind traditional symbolic planners for solving complex decision making tasks. In this work, we demonstrate how to train Transformers to solve complex planning tasks and present Searchformer, a Transformer model that optimally solves previo...
[ML News] Grok-1 open-sourced | Nvidia GTC | OpenAI leaks model names | AI Act
Просмотров 34 тыс.3 месяца назад
OUTLINE: 0:00 - Intro 0:15 - XAI releases Grok-1 2:00 - Nvidia GTC 4:45 - Comment of the Week 5:35 - Brute-forcing OpenAI model names 7:30 - Inflection AI gets eaten by Microsoft 9:25 - EU AI Act moving forward 11:45 - Advances in Robotics 14:00 - India retracts controversial advisory 14:30 - OpenSora 15:20 - Improved Gemma fine-tuning 16:20 - Decoding encrypted LLM traffic 17:45 - Varia Refere...
[ML News] Devin AI Software Engineer | GPT-4.5-Turbo LEAKED | US Gov't Report: Total Extinction
Просмотров 52 тыс.3 месяца назад
[ML News] Devin AI Software Engineer | GPT-4.5-Turbo LEAKED | US Gov't Report: Total Extinction
[ML News] Elon sues OpenAI | Mistral Large | More Gemini Drama
Просмотров 32 тыс.3 месяца назад
[ML News] Elon sues OpenAI | Mistral Large | More Gemini Drama
No, Anthropic's Claude 3 is NOT sentient
Просмотров 43 тыс.3 месяца назад
No, Anthropic's Claude 3 is NOT sentient
[ML News] Groq, Gemma, Sora, Gemini, and Air Canada's chatbot troubles
Просмотров 40 тыс.3 месяца назад
[ML News] Groq, Gemma, Sora, Gemini, and Air Canada's chatbot troubles
Gemini has a Diversity Problem
Просмотров 53 тыс.4 месяца назад
Gemini has a Diversity Problem
V-JEPA: Revisiting Feature Prediction for Learning Visual Representations from Video (Explained)
Просмотров 40 тыс.4 месяца назад
V-JEPA: Revisiting Feature Prediction for Learning Visual Representations from Video (Explained)
What a day in AI! (Sora, Gemini 1.5, V-JEPA, and lots of news)
Просмотров 32 тыс.4 месяца назад
What a day in AI! (Sora, Gemini 1.5, V-JEPA, and lots of news)
Lumiere: A Space-Time Diffusion Model for Video Generation (Paper Explained)
Просмотров 28 тыс.4 месяца назад
Lumiere: A Space-Time Diffusion Model for Video Generation (Paper Explained)
AlphaGeometry: Solving olympiad geometry without human demonstrations (Paper Explained)
Просмотров 34 тыс.5 месяцев назад
AlphaGeometry: Solving olympiad geometry without human demonstrations (Paper Explained)
Mixtral of Experts (Paper Explained)
Просмотров 55 тыс.5 месяцев назад
Mixtral of Experts (Paper Explained)
Until the Litter End
Просмотров 14 тыс.5 месяцев назад
Until the Litter End
LLaMA Pro: Progressive LLaMA with Block Expansion (Paper Explained)
Просмотров 35 тыс.5 месяцев назад
LLaMA Pro: Progressive LLaMA with Block Expansion (Paper Explained)
I created an AI-powered Social Network
Просмотров 25 тыс.5 месяцев назад
I created an AI-powered Social Network
NeurIPS 2023 Poster Session 4 (Thursday Morning)
Просмотров 12 тыс.6 месяцев назад
NeurIPS 2023 Poster Session 4 (Thursday Morning)
Art @ NeurIPS 2023
Просмотров 6 тыс.6 месяцев назад
Art @ NeurIPS 2023

Комментарии

  • @WeaponBalalaika
    @WeaponBalalaika Час назад

    These transormers are a bad idea, they will never lead to any commercial product.

  • @MrAhsan99
    @MrAhsan99 8 часов назад

    Yannic, please don't pick a paper like this again.

  • @tantzer6113
    @tantzer6113 10 часов назад

    Lexis's marketing blurb is fine. Lexis's answers are grounded, i.e., referenced to, sources that are real (and thus generally useful and authoritative), exactly as advertised, as opposed to fictional sources. Lexis's marketing is reasonably accurate and fair, but incomplete. The "incompleteness" may be missed by customers who don't pay attention to niceties, but most of the people who pay attention and are reasonably intelligent will not miss it. Lexis eliminates responses that are backed up by hallucinated, non-existing citations. This is a significant improvement over tools that do not do so, since it reduces the amount of verbiage from the sources you have read and evaluate, but you still have to read the sources cited because the ideas attributed to them might be hallucinated.

  • @AlexanderTurkhanov
    @AlexanderTurkhanov 10 часов назад

    Very interesting, thank you. I enjoyed it very much and recommended it to my colleagues. What sparked my interest is the benchmark. Like with any expert opinion, many people I listen to hallucinate (imagine, brainstorm, we can use a lot of other non-loaded words here) a lot, but that's a valuable perk. They make me think about the subject from a new perspective. What are better ways to evaluate the quality of answers? I understand that we should at least distinguish between explorative questions/answers and definitive ones. And the quality of those will be different. But what else?

  • @fox_7765
    @fox_7765 10 часов назад

    Hmm... The explanation of how and why hallucinations are produced is theoretically implausible - Dr Klicher assumes that a transformer parametrises a distribution like a smoothed univariate Gaussian - with a oscillatory-like function superimposed over the Gaussian. Transformers are auto-regressive models and therefore the distributions they parametrise are joint probabilities distributions in high-dimensional space arrived at through a sequence of non-linear perturbations. Given the parameter space, the distributions GPTs parametrise are not generated by functions that generate Gaussians. While there is a statistical component in the prediction, the embedding space is parallel-distributed representation over which the auto-regressive prediction is made, that representation is not statistical but spatial-temporal in nature, based on the relationship between concepts learned from pre-training. The parallel-distributed representation models (in a loose sense) associative memory in mammalian cognitive systems. Hallucination is poorly chosen term for the phenomena (Dr Klicher didn't choose this term though), 'confabulation' better captures the nature of what the cognitive system is doing when non-factual information is presented.

  • @DimitarBerberu
    @DimitarBerberu 14 часов назад

    The limitations are in the illogical structure of English. Try much more precise Esperanto & see if it hallucinates. That's why Western politicians constantly hallucinate that the West is still relevant ;)

  • @lorea4749
    @lorea4749 День назад

    The section you highlighted in the paper as defining "hallucinations' is a general introduction and a reference to another paper . The working definition of "hallucination" for this reserach is much more nuanced and defined in section 4.3 of the paper "We now adopt a precise definition of a hallucination in terms of the above variables. A response is considered hallucinated if it is either incorrect or misgrounded. In other words, if a model makes a false statement or falsely asserts that a source supports a statement, that constitutes a hallucination".

  • @MegaNightdude
    @MegaNightdude День назад

    Put out as many papers that are barely not crap and throw them at the random generator 😂😂😂

  • @MegaNightdude
    @MegaNightdude День назад

    Yannick😂😂😂😂. Ooh, ooh, not enough experiments.😂😂😂

  • @azadnoorani7065
    @azadnoorani7065 День назад

    Simply Amazing!

  • @StupidInternetPeople1
    @StupidInternetPeople1 День назад

    Good thing you have sunglasses on! I only listen to technical advice from jagoffs who wear sunglasses for no reason. Moron. 😂

  • @markmonfort29
    @markmonfort29 2 дня назад

    "Ohh it hallucinates" ... Technically they always hallucinate. If you don't agree with the answer it's a hallucination. So control comes from whether you leave it to figure things out completely on its own or you steer it... Like a good expert should. Hopefully we see a change in understanding of what these things are good at because even with that limitation, the outsized gains in productivity we'll have from using them in law or any field will be better than not having these sorts of tools

  • @eliaszeray7981
    @eliaszeray7981 2 дня назад

    Great and informative. Thank u.

  • @errorbool531
    @errorbool531 2 дня назад

    Like your spice comments 🔥 It hurts but it is true.

  • @ryanengel7736
    @ryanengel7736 2 дня назад

    Yannic you are a great academic and youtuber. I appreciate your videos, and as an NLP graduate student myself, I find your content intellectual and entertaining. Keep it up man

  • @darshanpandit
    @darshanpandit 2 дня назад

    Christopher Manning is a co-author. I am pretty sure you are missing onto something fundamental. 😂

  • @gody7334
    @gody7334 2 дня назад

    fine-tune on lots of domain specific documents might help improve performance ??

  • @billxu9799
    @billxu9799 2 дня назад

    cringe cringe

  • @-long-
    @-long- 3 дня назад

    Prof. Hinton also claimed to be the first (to make neural net LM) ^^ it's a safer claim right? ruclips.net/video/n4IQOBka8bc/видео.htmlsi=aVmQk62AozY0b-Rh&t=552

  • @AaronKaufmann-v3x
    @AaronKaufmann-v3x 3 дня назад

    Virtually hallucination-free AI is totally doable. In digital marketing you have Lemon AI with AI Reports, which almost never hallucinates

  • @JumpDiffusion
    @JumpDiffusion 3 дня назад

    Lots of strong claims/language ("garbage", "that's not what these models for") without any arguments to back it up...

  • @henrythegreatamerican8136
    @henrythegreatamerican8136 3 дня назад

    Yes, they hallucinate often. I type the same response into Claude, ChatGPT, and PerplexityAI. More often than not I'll get different responses. But then I play all of them against each other by repeating what the other said in a follow up response. Eventually, they'll all agree, but sometimes they agree on the WRONG ANSWER!!!! And the problem with Claude is it doesn't have a memory. So if you close the website and return to ask the same question, it will repeat the same wrong response it gave you the first time without any reference to your follow up questions.

  • @florianhoenicke4858
    @florianhoenicke4858 3 дня назад

    You said something like "don't use LLMs for reasoning" and I agree that you need human in the loop. But I also know from experience that GPT-4 can be used for reasoning if I do a lot of handholding and split the task into smaller reasoning tasks.

  • @luisluiscunha
    @luisluiscunha 3 дня назад

    I remember your 2017 revision of the Attention is All you Need, first on RUclips, then on my run, after making an mp3 of it. Good memories.

  • @samdirichlet7500
    @samdirichlet7500 3 дня назад

    The whining about how private companies won't share access for free reminds me of dealing with a junior engineer who expects to be handed an X and y matrix before starting work on any project. On another note, I'm writing AI to predict the ideal structure for better aerospace materials. Why won't Alcoa give me access to their proprietary database of alloy properties?

  • @bjarke7886
    @bjarke7886 3 дня назад

    ESM3!

  • @alan2here
    @alan2here 3 дня назад

    "Lex" is excellent at including the human in the loop.

  • @josephmacdonald1255
    @josephmacdonald1255 3 дня назад

    Yannic has provided a fairly good summary of the risk, I may have missed him saying it, but in my opinion having LLMs generating different answers to identical prompts at different times when none of the facts or rules have change is expoentialy increasing the risk of a negligence claim. The issue of rulings and legislation being superseded by later legislation and later rulings. I also believe pertinent is more accurate than relevant.

  • @alan2here
    @alan2here 3 дня назад

    Useful practical comparisons/tests between models. typical questions Well crafted questions with lots of prompt crafting, clearing the context window where needed, looking up and providing some stuff yourself etc… sloppy questions Questions about cases that involve domain specific knowledge outside of law, such as number theory, cryptography, structural engineering. Questions about gruesome cases to test for refusals. Large and small context window utilisation. Easy vs hard questions. Questions that are controversial or where no case law exists yet. and the such…

  • @alan2here
    @alan2here 3 дня назад

    Chat GPT already RAGs when it thinks it needs to, with web searches, analysis steps, and the such.

  • @alan2here
    @alan2here 4 дня назад

    Requests like "help me with this specific neurology research" or "here's a vauge half-remembered thing, what's the correct termoniligy for it" ⭐️⭐️⭐️⭐️💫. Requests like "[Baba is You level 7 screenshot] lets work step by step to solve it" ⭐️

  • @jabowery
    @jabowery 4 дня назад

    Self-aggrandizing Boomer-posting admitted, there is a good reason for bringing to people's "attention" prior art and it has to do with the foundation of intelligence in Kolmogorov Complexity approximation: Don't multiply names for things beyond necessity. Now, don't get me wrong here. I'm not saying that the terms currently in use are inferior -- I'm just saying that unification of taxonomy can reduce the explosion of confusion that now besets the field. So the renaming can be beneficial, so long as one then describes prior art in terms of the current tech-argot with appropriate modifiers.

  • @hasko_not_the_pirate
    @hasko_not_the_pirate 4 дня назад

    As the old saying goes: If you only have a hammer, every problem looks like a nail.

  • @hitechconnect
    @hitechconnect 4 дня назад

    thanks for explaining this paper to all. So you did not actually evaluate this by yourself? The fact that these tools are better than ChatGPT shows that they improved which they did using e.g. RAG but also other technologies. My strong guess is that over time these tools will get better and better and who knows how close to the 100% correct answers. I strongly suppose also DeepJudge cannot claim 100% accuracy. Can you share where you are there? I agree though that claiming so is a bad business practice.

  • @InstaKane
    @InstaKane 4 дня назад

    Nice, feel like I’m up to speed on this topic at a high level, cheers

  • @gr8ape111
    @gr8ape111 4 дня назад

    congrats on 2^8 * 1000 subs

  • @jaakko3083
    @jaakko3083 4 дня назад

    What about finetuning instead of using RAG?

  • @luke2642
    @luke2642 4 дня назад

    Interesting video. How many years are we away from building a complete, semantic logical knowledge graph of all legal precendents based on a large set of documents?

  • @JoanFiguerolaHurtado
    @JoanFiguerolaHurtado 4 дня назад

    From personal experience building AskPandi, the issue with hallucinations is more about lack of data rather than the model itself (assuming great QA capabilities). It's equally true that most QA models are not trained to say "I don't know", which complicates things too...

  • @gregsLyrics
    @gregsLyrics 4 дня назад

    perfect timing! Full of wisdom.

  • @scottmiller2591
    @scottmiller2591 4 дня назад

    I remember when Lexis/Nexis was just a keyword database for patents, with the "(a OR NOT b) c"-type query language. It looks like it's aspirationally come a long way, but it does seem they've forgotten (or lost) the knowledge of where they came from (database management), and are dazzled by the Eye of AI.

  • @JoeTaber
    @JoeTaber 4 дня назад

    Not that Lexis is deserving, but a charitable reading of that marketing spiel could see it as an answer to that recent case where a lawyer who didn't know what he was doing tried to use GPT-4 to find case law and it completely hallucinated references multiple times. I think they ended up getting fined and reprimanded, besides being rather embarrassing. I'm neither a lawyer nor ML researcher, just a regular old programmer, but this spiel looks to be targeted at those lawyers who are largely put off of AI tools altogether by this case and don't really understand/care if it's basically the same as current search techniques as long as it avoids this one big (in their eyes) flaw.

  • @zeeshanahmed5624
    @zeeshanahmed5624 4 дня назад

    Hi, love your videoes on LLMs and hallucinations - I'm new to machine learning (doing a masters in CS) so I find this channel very useful. I understand that LLMs arent designed specifically for reasoning hence why we shouldn't expect them to perform well in QA-like tasks. So what are LLMs fundamentally designed for then?

  • @arowindahouse
    @arowindahouse 4 дня назад

    In non-english speaking countries, we tend to say RAG to avoid the mouthful

  • @andytroo
    @andytroo 4 дня назад

    Lexis: delivering 99.6% hallucination free advice connected to citations to 100% hallucination free citations.

  • @user-iv8fq5zl9o
    @user-iv8fq5zl9o 4 дня назад

    As a recent law graduate I’m siding with you over the academics and lawyers lmao

  • @andytroo
    @andytroo 4 дня назад

    how do you find good documents :D sounds like you need a LLM to evaluate the relevancy of all the documents...

  • @Anonymous-lw1zy
    @Anonymous-lw1zy 4 дня назад

    At 39:00 - lawyers scamming lawyers with technically correct wording that misleadingly advertises, seeks to contractually entrap, and thereby defraud. Hilarious!!! And thanks for pulling the rug out from RAG's frequently outrageous claims. Well deserved 256k!

    • @hieroben
      @hieroben 3 дня назад

      I would argue that the claim of LexisNexis is not even "technically correct." They assert that their responses are "grounded" in trustworthy documents, which can obviously not be the case if the system hallucinates.

    • @MacGuffin1
      @MacGuffin1 3 дня назад

      RAG-pull?

    • @clray123
      @clray123 3 дня назад

      @@hieroben The "those responses" phrase has no valid semantic connection to the previous things mentioned in the same sentence ("legal citations" or "source documents"). It is just slimy marketing mumbo jumbo not worth paying any attention to. But you can bet that such subtleties will fly over the head of many lawyers with all their love for logic and precise language.

    • @tantzer6113
      @tantzer6113 10 часов назад

      Lexis's marketing blurb is fine. Lexis's answers are grounded, i.e., referenced to, sources that are real (and thus generally useful and authoritative), exactly as advertised, as opposed to fictional sources. Lexis's marketing is reasonably accurate and fair, but incomplete. The "incompleteness" may be missed by customers who don't pay attention to niceties, but most of the people who pay attention and are reasonably intelligent will not miss it. Lexis eliminates responses that are backed up by hallucinated, non-existing citations. This is a significant improvement over tools that do not do so, since it reduces the amount of verbiage from the sources you have read and evaluate, but you still have to read the sources cited because the ideas attributed to them might be hallucinated.

  • @marinepower
    @marinepower 4 дня назад

    LLMs hallucinate because they have zero grounding, but also because they are trained with next-token prediction. In essence, this means that whatever is in the context is maximally correct -- including whatever crap the llm itself generated. This is why, if your context is, 'does 3+5=9? Yes, 3+5 equals 9 because', then it's trivial to see that the model will hallucinate because it must match the existing context. I don't know what you were talking about with llms not being able to build a world model, being 'just statistical' models, etc. Humans are also statistical models with a lifetime of training data, that has nothing to do with hallucinations.

    • @doppelrutsch9540
      @doppelrutsch9540 4 дня назад

      That is both kind of true and not true in practice. A lot of gen AI tools these days are not LLMs in the strict sense of the word: After receiving additional training like RLFH or related techniques the distribution of the output is significantly shifted from the simple "most likely" continuation. You can observer this very strongly with Claude for example which will correct itself sometimes when it makes a mistake in its output even without human prompting - and of course the classic "answering this question would be unethical or dangerous" refuse responses.

    • @marinepower
      @marinepower 4 дня назад

      @@doppelrutsch9540 I suppose RLHF with the model making mistakes and then backpedaling could very well solve what I'm describing. Historically, no one ever trained on models making mistakes and then correcting them, we instead would simply train on 'clean' data. But, I suppose, if a mask was applied such that the training loss was only applied to the correction tokens, with the actual bad data tokens masked (so that the model wouldn't learn to generate bad data, only the corrections to the bad data), then it could work.

    • @doppelrutsch9540
      @doppelrutsch9540 3 дня назад

      @@marinepower I wasn't describing a hypothetical, models *do* work like that and have for years. And they are also trained on data that is purposefully incorrect with later added corrections, I am quiet sure.

    • @clray123
      @clray123 3 дня назад

      There is no law that the model "must match the existing context". But when training on the next token prediction task, models "discover" by themselves that copying from context more often than not minimizes the average loss across batch and they develop structures suitable to perform such copying ("induction heads" - in small models it can be even empirically observed when they develop during training and they can be ablated away). You could construct your training data in such a way that copying would not be an effective loss minimizing strategy. I think the bigger problem is the "average" part in minimizing average loss. A training outcome where you are slightly wrong on each token is arithmetically the same as an outcome where you are very right on many tokens and very wrong on others. Also, add to that that with cross-entropy loss only the probability of the correct token matters in considered for loss calculation - and maximizing this probability can also rise probability of similiar (but incorrect) tokens as a side effect (as mentioned in the ORPO paper). So overall, the impression is that our easy-to-compute loss function is crap, but it is hard to devise better ones (especially differentiable and efficient ones). That is where the PPO/RLHF algorithms try to compensate, but the later devised DPO/ORPO/etc. show that they do not really do such a great job either (because you can obtain similar "alignment" quality using these simpler non-PPO algorithms - what seems like an achievement really just highlights how the original approach "also sucks"). It could be that the core problem of "rare truths" is just not possible to represent well using statistical distributions. Imagine a training dataset in which you have a million sequences which end in one way (and that's correct for them), but just one sequence which ends in a different way, based on some slight difference in the constituent tokens. How do you exactly teach a statistical model to pay attention to that one example and please disregard the mountain of what looks like "contrary" evidence opposing it? Possibly by assigning a million times weight to that one example, but then, if you are forced to apply such tricks and think about exceptions by yourself, then what good is machine learning, why not just write an if-else rule or a database of cases if you already know what to look for? I think the public is slowly coming to the realization that there is no free lunch in AI (probably something which seasoned ML researchers have known all along, but they certainly have been quite silent about it recently because of a huge conflict of interest).

    • @marinepower
      @marinepower 3 дня назад

      ​@@clray123 I do think the next-token prediction training regime is sort of ass. It is nice in the sense that (with temporal masking), it essentially allows you to train each token as an independent sample (sort of), so it's parallelizable, but there can definitely be improvements made. One thing that could maybe work is diffusion-based transformer models. In essence, instead of autoregressive token generation, you generate an entire sentence at a time (via iterative diffusion). Basically, you first train a CLIP-like model where you have a corpus of sentences, you use an LLM to reword said sentences while keeping the same semantic meaning, then train a model such that the original and reworded sentences maximize their cosine similarity between each other and minimize the cosine similarity between other sentences. Then, during training, you noise your input embedding (instead of using a one-hot encoding representing just that token), and pass the thing through your transformer decoder. The final loss is signal_strength * logit_loss + (1 - signal_strength) * similarity_loss. In essence, in low signal regimes, we want our model to predict a sentence that is semantically similar (even if it's not using the same tokens), whereas, when the noise is low we want it to predict the actual tokens as to what is in our training set. I haven't thought too deeply about this so maybe there's some sort of crucial issue with this methodology but I think it makes sense.

  • @adfaklsdjf
    @adfaklsdjf 4 дня назад

    22:43 to be clear, when we say "paste references", chatgpt and claude can't retrieve the contents of URLs, so a list URLs as references doesn't work with those afaik. i paste the full text of the material i want it to use