First Impressions of Gemini Flash 1.5 - The Fastest 1 Million Token Model

Sdílet
Vložit
  • čas přidán 30. 05. 2024
  • Just checked out Google's new Gemini Flash at Google I/O. It's a super-fast AI model designed for handling big tasks - think processing videos, audios, or huge codebases, all while keeping costs low. I put it through its paces against giants like GPT 3.5 and GPT 4.0, looking at performance, costs, and how it handles real-world tasks. I even tried confusing it with tricky questions and coding challenges in Google AI Studio. Spoiler: it's not perfect, but for speed and efficiency on a budget? Pretty cool stuff. Stick around to see how Gemini Flash holds up in the AI arena!
    🦾 Discord: / discord
    ☕ Buy me a Coffee: ko-fi.com/promptengineering
    |🔴 Patreon: / promptengineering
    💼Consulting: calendly.com/engineerprompt/c...
    📧 Business Contact: engineerprompt@gmail.com
    Become Member: tinyurl.com/y5h28s6h
    💻 Pre-configured localGPT VM: bit.ly/localGPT (use Code: PromptEngineering for 50% off).
    Signup for Advanced RAG:
    tally.so/r/3y9bb0
    LINKS:
    aistudio.google.com/
    TIMESTAMPS:
    00:00 Introducing Gemini Flash: Google's Answer to GPT-4
    00:53 Why Choose Gemini Flash? Performance and Cost Analysis
    02:20 Hands-On Testing: Unveiling Gemini Flash's Capabilities
    03:05 Exploring Safety Features and Customization
    03:44 Prompt-Based Testing: Analyzing Model Responses
    12:11 Advanced Testing: Contextual Understanding and Mathematics
    15:30 Programming Challenges: Assessing Gemini Flash's Coding Prowess
    All Interesting Videos:
    Everything LangChain: • LangChain
    Everything LLM: • Large Language Models
    Everything Midjourney: • MidJourney Tutorials
    AI Image Generation: • AI Image Generation Tu...
  • Věda a technologie

Komentáře • 25

  • @Omesh7
    @Omesh7 Před dnem

    What software did you use to create this video recording, looks interesting

  • @ksanjaykumar4412
    @ksanjaykumar4412 Před 15 dny +1

    thank you for your video! Quick question:
    I'm trying to use the Gemini 1.5 Pro API and to host it on a Streamlit app. is it possible if we could host the Gemini 1.5 Pro API on streamlit to receive video file as inputs from users and perform the desired output (to get summary or whatever). Basically just like what we have in google studio but using the API in the streamlit app. Is it possible?

    • @engineerprompt
      @engineerprompt  Před 15 dny

      I think its not possible yet but you can have a work around. You can generate frames and then use those in sequence to get response. Here is one way of doing it: tinyurl.com/5n8ywwpt

  • @caseyhoward8261
    @caseyhoward8261 Před 16 dny

    Could the Gemini Flash 1.5 model handle most of the heavy lifting, allowing 4.0 to manage more complex, value-oriented tasks, such as image recognition? For instance, could Gemini handle routine tasks and delegate more complex ones to 4.0 when necessary?In an agent-based problem-solving chatbot with a RAG implementation, wouldn't this approach be more cost-effective?I'm working on a grow buddy chatbot for plant health and environmental monitoring, which includes image recognition. The backend is being developed in Python, running on a cloud server, with communication to an Android front-end via RESTful APIs.I'm relatively new to coding, having started just six months ago. Any advice or insights tha5 anyone can provide would be greatly appreciated!
    P.s. It's the implementation of vision recognition is where I'm completely stuck at. 😔

  • @LTBLTBLTBLTB
    @LTBLTBLTBLTB Před 16 dny +4

    Is it possible to use GEMINI API in visual studio code?

    • @engineerprompt
      @engineerprompt  Před 16 dny

      Yes, the code segment you get from ai studio shows how to set that up.

  • @TheReferrer72
    @TheReferrer72 Před 16 dny

    nice model.

  • @unclecode
    @unclecode Před 15 dny

    .Interesting, it seems both OpenAI & Google are stepping up their game, improving performance, speed, and benchmarks a bit. But no groundbreaking intelligence yet. Is this a plateau for transformer models? Or are these companies slowing down for steady competition? Either way, it's cool that the tech behind these models isn't a secret anymore. KV Cache, RoPE and ... Watching your video got me thinking, how cool would it be to invest in something like Llama3-Flash with a 200k content window length! 🚀

    • @engineerprompt
      @engineerprompt  Před 15 dny +1

      That's the beauty of open source :) I still think the main limiting factor would be the availability of compute! but the way I think about it is the for most use cases, you don't even need SOTA models. A specialized Llama3-Flash-200k will be more than enough :)

    • @unclecode
      @unclecode Před 14 dny

      That is exactly what I believe. Giant companies aim to dominate the AGI market for future valuation, but we need to focus on “topic-specific language models” and orchestrate dozens of them, from 100M to 1B. This is the way to build a distributed “intelligent” system, similar to blockchain where each block is a small model. I think this is the way to AGI, making it more consumer-device friendly and energy-efficient. As Mandalorians say, this is the way.@@engineerprompt

  • @NLPprompter
    @NLPprompter Před 15 dny

    it is amaze me when transformers does that... I remember when stability AI never be able to logic the human fingers even it's already "seen" them... I wonders does it think human hands have so many fingers because everytime prompt come in it's instant whole tokens. and since "it" limited by cpu bottle neck which makes "it" think all being have 1 hands to hold a thing and only able to out put something one at a time
    (please forget this talk I'm hallucinating... too much time with them made me thinks to hallucinate too, it... it... some what fun to hallucinate,.... words )

  • @ssoundtrack
    @ssoundtrack Před 3 dny

    geimini 1.5 pro vs gemini 1.5 pro flash what is better ?

  • @bastabey2652
    @bastabey2652 Před 15 dny

    looking for a better chatGPT 3.5 in the same price range, I m impressed with Gemini Flash..

  • @contentfreeGPT5-py6uv
    @contentfreeGPT5-py6uv Před 16 dny

    yes, in my project i use is so fast, really like this future, like your video

  • @bgNinjaart
    @bgNinjaart Před 15 dny

    You forgot to hide your email

  • @sofianeben2490
    @sofianeben2490 Před 15 dny

    Mistral Large is died 😅

  • @Priming-AI
    @Priming-AI Před 16 dny

    Gemini is shit.. GPT4o the best !

    • @caseyhoward8261
      @caseyhoward8261 Před 16 dny +2

      In production, wouldn't using 4.0 be super expensive, especially for a personal assistant with vision recognition capabilities?Could the Gemini Flash 1.5 model handle most of the heavy lifting, allowing 4.0 to manage more complex, value-oriented tasks, such as image recognition? For instance, could Gemini handle routine tasks and delegate more complex ones to 4.0 when necessary?In an agent-based problem-solving chatbot with a RAG implementation, wouldn't this approach be more cost-effective?I'm working on a grow buddy chatbot for plant health and environmental monitoring, which includes image recognition. The backend is being developed in Python, running on a cloud server, with communication to an Android front-end via RESTful APIs.I'm relatively new to coding, having started just six months ago. Any advice or insights you can provide would be greatly appreciated!Thank you for your time!

    • @jakeboardman5212
      @jakeboardman5212 Před 15 dny

      @caseyhoward8261 Yes, it would be more expensive, and absolutely, it would be. Ironically, Gemini has a larger token capacity than the GPT4o having an overall capacity of 128,000, Gemini, exceeds this with a token capacity of 1,000,000 making it far more superior.

    • @caseyhoward8261
      @caseyhoward8261 Před 15 dny

      ​@@jakeboardman5212Thank you.

    • @sofianeben2490
      @sofianeben2490 Před 15 dny

      Not the same price. Flash seem better than Haiku and better than mistral