Analyzing the Costs of Large Language Models in Production

Sdílet
Vložit
  • čas přidán 10. 09. 2024

Komentáře • 15

  • @CresGallego
    @CresGallego Před 8 měsíci +2

    Really great insights. Economics is well explained.

  • @mohamedfouad1309
    @mohamedfouad1309 Před 9 měsíci +2

    😊

  • @billykotsos4642
    @billykotsos4642 Před 8 měsíci +1

    Being handed a bill based on tokens generated by a model is preposterous...
    These LLM apps cost so much right now that you need to have a solid use case in mind....
    Else you just wait for a couple more years when inferencing these LLMs wont be as expensive... the only reason these LLMs are so expensive to run is that they are SOTA and Nvidia is the only player right now.

  • @billykotsos4642
    @billykotsos4642 Před 8 měsíci +1

    the economics are broken because the hardware setup just isnt there...
    instead of paying by the hour you pay by the token/call which is insane..... Cloud has been build on the idea that you fire up the instance and you know what you pay.... but these days you need huge cloud instances to run these huge models...
    The costs will go down significantly to run these models in about 3 years.... you wont have to think about these things...

  • @lionhuang9209
    @lionhuang9209 Před 8 měsíci +1

    Where can we download the slides?

  • @loopaal
    @loopaal Před 7 měsíci

    fantastic

  • @balainblue
    @balainblue Před 7 měsíci +1

    Can you explain the math of 5 requests per minute translating it to 9,000$ per month?

    • @tensorops
      @tensorops  Před 7 měsíci +1

      We recommend looking here
      gptforwork.com/tools/openai-chatgpt-api-pricing-calculator
      Assuming 220K requests, with proper prompts that are usually 1000-2000 tokens you can get to these costs.
      Additionally we want to remind that often a single request to an LLM application triggers more than one API call to an LLM

    • @balainblue
      @balainblue Před 7 měsíci

      @@tensorops Thank you so much.

    • @balainblue
      @balainblue Před 6 měsíci

      @@tensorops Can you please elaborate on that? "A single request to an LLM application triggers more than one API call to an LLM"

    • @tensorops
      @tensorops  Před 6 měsíci +1

      @@balainblue We give an example on the next webinar where you have one query that triggers many LLM calls. Sometimes even simple chains like Map-Reduce or Refine can cause many LLM calls to OpenAI for a simple action as "summarization"

    • @balainblue
      @balainblue Před 6 měsíci

      @@tensorops Thank you. I look forward to it.