InternLM - A Strong Agentic Model?

SdĂ­let
VloĆŸit
  • čas pƙidĂĄn 10. 07. 2024
  • In this video I look at InternLM an LLM which focus on math, reasoning and being able to support function calling.
    Colab: drp.li/mxJrX
    Github: github.com/InternLM/InternLM
    LM Deploy: github.com/InternLM/InternLM/...
    HF: huggingface.co/internlm/inter...
    đŸ•”ïž Interested in building LLM Agents? Fill out the form below
    Building LLM Agents Form: drp.li/dIMes
    đŸ‘šâ€đŸ’»Github:
    github.com/samwit/langchain-t... (updated)
    github.com/samwit/llm-tutorials
    ⏱Time Stamps:
    00:00 Intro
    01:33 Hugging Face Leaderboard
    01:57 InternLM Github
    03:02 InternLM: LMDeploy
    04:29 InternLM: Lagent
    06:36 InternLM Paper
    08:29 InternLM Hugging Face Models and Datasets
    08:39 InternLM on Ollama
    08:54 Code Time
    09:15 InternLM Hugging Face Implementation (Colab)
    13:12 InternLM Chat Format
    13:39 InternLM Function Calling
    15:01 InternLM Running Locally through Ollama
  • Věda a technologie

Komentáƙe • 24

  • @toadlguy
    @toadlguy Pƙed 5 dny +2

    Thank you, Sam, for once again highlighting the most interesting new models/techniques in this fascinating field. I note InternLM 2.5 explicitly notes that it "supports gathering information from over 100 websites" with an implementation using Lagent. I'm sure a LangChain implementation could be easily created as well. Actually fine tuning models with Sources for information not in the model (like current weather or news) with function calling and JSON support and using LangChain for finer control would be a great method for using smaller local models. (I feel more comfortable using LangChain than a model specific framework, if possible.) I would love to see other models add this approach. I wonder how much this is done in pretraining vs the base model. (guess I'll have to look at the paper 😉).

  • @LaHoraMaker
    @LaHoraMaker Pƙed 5 dny +3

    LMDeploy is a quite interesting framework to deploy and quantize most of the Chinese models. It also works in Kaggle fairly well given it supports also older GPUs.

  • @keithmatthews2707
    @keithmatthews2707 Pƙed 3 dny

    Very useful content thank you Sam for your valuable insights into these topic areas

  • @mickelodiansurname9578
    @mickelodiansurname9578 Pƙed 5 dny +1

    thats a nice SMALL model for function calling alright... appreciate you bringing it to my attention.

  • @omarelfaqir3627
    @omarelfaqir3627 Pƙed 4 dny

    Hello Sam, Thanks to bring this wonderful model to our attention. There is just a confusion in the video between commercial usage and commercial licence: commercial usage is allowed without submitting any form, but with the Open Source licence you might need to Open Source any derivative work (ie finetuning you make for example). If you want to make non open source stuff with it (why would you😊?) you will need to submit the form to obtain a commercial licence, allowing you to do that.
    It is a quite classic business model in Open Source software

  • @waneyvin
    @waneyvin Pƙed 5 dny

    great job mate! And this is a bit like glm4, not sure about the comparison of benchmark. Both are agentic designed, and could be trained with agentic instructions.

  • @tlfmcooper
    @tlfmcooper Pƙed 5 dny

    Thanks

  • @kenchang3456
    @kenchang3456 Pƙed 5 dny

    Kind of interesting that if one of the stronger points of InternLM 2.5 is being able to support agents, I wonder what part of the training data makes it more capable of supporting agents if function calling data only accounts for 16%. Thanks for the video, I'll have to find a way to make time to try it out.

  • @choiswimmer
    @choiswimmer Pƙed 5 dny

    Nice

  • @SonGoku-pc7jl
    @SonGoku-pc7jl Pƙed 5 dny

    thanks! in spanish is regular but good that all evolution :)

  • @lapozzunk
    @lapozzunk Pƙed 5 dny +2

    If each model gets a higher rating than its predecessors, when will we reach 100? Also, if I don't watch such videos, will this happen later?

  • @ManjaroBlack
    @ManjaroBlack Pƙed 5 dny

    I couldn’t get InternLM to work well with RAG or any embedding. It gives ok answers to simple prompting.

  • @attilavass6935
    @attilavass6935 Pƙed 4 dny

    Am I the only one who misses a memory module from Lagent? I'm gonna test this though ASAP

  • @aa-xn5hc
    @aa-xn5hc Pƙed 3 dny

    Please try lmagent with 2.5

  • @WillJohnston-wg9ew
    @WillJohnston-wg9ew Pƙed 5 dny

    What is the agentic aspect? Maybe I don't understand something or missed something?

    • @Schaelpy
      @Schaelpy Pƙed 3 dny

      He talks about it at 4:45

  • @wickjohn3854
    @wickjohn3854 Pƙed 5 dny +4

    ask him what happen in 1989 LOL

  • @TheGuillotineKing
    @TheGuillotineKing Pƙed 5 dny

    Fun fact these Chinese models are banned in the USA and can’t be used for a commercial product

    • @ringpolitiet
      @ringpolitiet Pƙed 4 dny

      Quite an enigma how you combine an interest in rather techy stuff like tool calling LLMs with a straight off the turnip truck view of other things that seems as easy or easier to get informed about.

    • @dinoscheidt
      @dinoscheidt Pƙed 3 dny

      Fun fact: A source helps. @TheGuillotineKing seems cognitively challenged holding apart the current talks to maybe restrict the EXPORT of OSS Models vs the other way around.

    • @TheGuillotineKing
      @TheGuillotineKing Pƙed 3 dny

      @@dinoscheidt Fun Fact your mother swallowed a gallon of đŸ„œđŸ„œđŸ„œđŸ„œđŸ„œđŸżïžđŸżïžđŸżïž juice and that's how she had you

    • @toadlguy
      @toadlguy Pƙed 2 dny

      @@dinoscheidt Well, he is right that they can’t be used for commercial projects due to the license. 😉