Matt Williams
Matt Williams
  • 257
  • 1 257 253
My favorite way to run Ollama: Gollama
Gollama is pretty amazing...
smcleod.net/2024/06/gollama-ollama-model-manager/
github.com/sammcj/gollama
Be sure to sign up to my monthly newsletter at technovangelist.substack.com/subscribe
zhlédnutí: 12 062

Video

The Matt and Ryan Chat on June 4 - Cleaned Up
zhlédnutí 258Před dnem
Trying out a new platform and there were some hiccups, but overall went pretty well. 00:00:00 - Start 00:00:12 - Trying out Riverside 00:00:42 - Why use Streamyard? 00:06:19 - Datadog DASH 00:06:46 - Has Generative AI Peaked 00:11:05 - Dependencies You Can't Control 00:11:44 - When Google Deleted a Company 00:14:46 - Don't Judge based on Mistakes, but on the Recovery 00:24:04 - Content Creation...
how to reliably get json out of ollama. Just a rough demo for a discord user
zhlédnutí 2,8KPřed dnem
Be sure to sign up to my monthly newsletter at technovangelist.substack.com/subscribe
I fumbled my chance
zhlédnutí 735Před dnem
On the live stream somebody asked me if I had any memories of specific conference demos that just did not go well. And it reminded me of the time I had a moment with Bill Gates. You can find the code for every video I make at github.com/technovangelist/videoprojects. Then find the folder name that starts with the date this video was published and a title that makes sense for what the video cove...
Have You Picked the Wrong AI Agent Framework?
zhlédnutí 36KPřed 14 dny
Are you diving into the world of agent-based AI workflows and finding it more complicated than it needs to be? You’ve probably chosen the wrong approach, burdened with verbose, time-consuming frameworks. In this video, I’ll show you a far easier method to build your AI workflow, cutting through the unnecessary boilerplate and complexity. Using CrewAI as an example, inspired by Code with Brandon...
Better Searches With Local AI
zhlédnutí 23KPřed 21 dnem
Better Searches With Local AI
Popularity doesn't always mean Great, But Pretty Good is Possible
zhlédnutí 6KPřed 28 dny
Popularity doesn't always mean Great, But Pretty Good is Possible
This Chrome Extension Surprised Me
zhlédnutí 15KPřed měsícem
This Chrome Extension Surprised Me
A video essay about AI...where are we now
zhlédnutí 3,1KPřed měsícem
A video essay about AI...where are we now
Does parallel embedding work in Ollama yet?
zhlédnutí 5KPřed měsícem
Does parallel embedding work in Ollama yet?
Ask Ollama Many Questions at the SAME TIME!
zhlédnutí 11KPřed měsícem
Ask Ollama Many Questions at the SAME TIME!
This may be my favorite simple Ollama GUI
zhlédnutí 21KPřed měsícem
This may be my favorite simple Ollama GUI
Is Open Webui The Ultimate Ollama Frontend Choice?
zhlédnutí 57KPřed měsícem
Is Open Webui The Ultimate Ollama Frontend Choice?
How I make the sausage
zhlédnutí 1,9KPřed měsícem
How I make the sausage
Design Your Own Ollama Model Now!
zhlédnutí 9KPřed 2 měsíci
Design Your Own Ollama Model Now!
Discover The Secrets Of Your Chromadb
zhlédnutí 4,6KPřed 2 měsíci
Discover The Secrets Of Your Chromadb
Supercharge Your Typescript Projects With Retrieval Augmented Generation
zhlédnutí 4,6KPřed 2 měsíci
Supercharge Your Typescript Projects With Retrieval Augmented Generation
Supercharge your Python App with RAG and Ollama in Minutes
zhlédnutí 28KPřed 2 měsíci
Supercharge your Python App with RAG and Ollama in Minutes
Unlocking The Power Of AI: Creating Python Apps With Ollama!
zhlédnutí 19KPřed 2 měsíci
Unlocking The Power Of AI: Creating Python Apps With Ollama!
Level Up Your Typescript Skills: Adding Ollama To Your Apps!
zhlédnutí 23KPřed 2 měsíci
Level Up Your Typescript Skills: Adding Ollama To Your Apps!
Getting Started on Ollama
zhlédnutí 34KPřed 2 měsíci
Getting Started on Ollama
Choosing the right Chunk Size for RAG
zhlédnutí 4,8KPřed 2 měsíci
Choosing the right Chunk Size for RAG
Whats the best Chunk Size for LLM Embeddings
zhlédnutí 10KPřed 3 měsíci
Whats the best Chunk Size for LLM Embeddings
Let's use Ollama's Embeddings to Build an App
zhlédnutí 16KPřed 3 měsíci
Let's use Ollama's Embeddings to Build an App
Installing Ollama is EASY Everywhere #mac #windows #linux #brevdev #paperspace
zhlédnutí 6KPřed 3 měsíci
Installing Ollama is EASY Everywhere #mac #windows #linux #brevdev #paperspace
Starcoder 2 on Ollama with 0.1.28
zhlédnutí 8KPřed 3 měsíci
Starcoder 2 on Ollama with 0.1.28
Unlocking The Power Of GPUs For Ollama Made Simple!
zhlédnutí 22KPřed 3 měsíci
Unlocking The Power Of GPUs For Ollama Made Simple!
How to run Ollama on Docker
zhlédnutí 20KPřed 3 měsíci
How to run Ollama on Docker
Ollama 0.1.26 Makes Embedding 100x Better
zhlédnutí 43KPřed 3 měsíci
Ollama 0.1.26 Makes Embedding 100x Better
The Secret Behind Ollama's Magic: Revealed!
zhlédnutí 28KPřed 3 měsíci
The Secret Behind Ollama's Magic: Revealed!

Komentáře

  • @johngoad
    @johngoad Před hodinou

    Thanks ... I appreciate you POV... I have used CrewAI, AutoGen and more.. I have wasted tons of time learning stuff that I don't use because at some point (usually 3/4 of the way in) I realize the limitations. I mostly just learn why it is not for me. Not sure if I can put a finger exactly on what it is, but it seems to me like most stuff is built these days with training wheels.

  • @continuouslearner
    @continuouslearner Před 15 hodinami

    Would have been good to cover what ollama is and what problems does it solve, for about 30sec-1min, before going into hardwarre requirements etc.

  • @Alex29196
    @Alex29196 Před 21 hodinou

    Interesting Matt, but not enough to draw me in yet, also no windows friendly. Thanks !

    • @technovangelist
      @technovangelist Před 20 hodinami

      What makes you say not windows friendly?

    • @Alex29196
      @Alex29196 Před 18 hodinami

      @@technovangelist Matt, I meant that the tool requires Windows Subsystem for Linux (WSL) in order to follow the installation instructions found in most repositories on GitHub.

    • @technovangelist
      @technovangelist Před 14 hodinami

      I don't think it does. just build it on windows. one simple comand. If you are comfy using a dev tool like Ollama, you can run a build tool.

  • @kameshsingh7867
    @kameshsingh7867 Před 23 hodinami

    Thank you

  • @AlfredNutile
    @AlfredNutile Před dnem

    Great points about Kubernetes being overkill and just the over focus on Python etc. I am building an open source tool inspired by Laravel PHP just to help me see how it is not the language but so much more. And good points about automation, to me I want normal non coding users to see the power and ease of this as they see the power of prompts. Anyways great one!

  • @alx8439
    @alx8439 Před dnem

    Not sure if anyone told you, but these hours long streams will be probably skipped by the vast majority of your audience, just because most of the ppl with brains in their heads don't have that much free time to watch all that :)

    • @technovangelist
      @technovangelist Před dnem

      They aren’t for everyone. But I am always amazed how many thousand watch them. They are the best thing for my channel considering the effort and watch time. But I will keep building short videos too. Most 10 minute videos take about 10 to 20 hours to make while a 2 hour stream takes 2 hours to make from start to finish. They are both equally enjoyable to make which is probably the most important factor.

  • @RamiAwar
    @RamiAwar Před dnem

    I love this. Exactly. Spot on. This is why I'm building my new project from scratch.

  • @michaelthompson8251

    curious. maybe a bench marking of war and peace using the the various data bases based on size and or based on speed

    • @technovangelist
      @technovangelist Před dnem

      it really should be a more recent long document. War and Peace is probably a part of every model already. But I need to find something written in the last year or so that is long and not part of the training data for every model.

  • @RupeeSalzburg
    @RupeeSalzburg Před dnem

    Hey, I have an 'ask' alias too. Pretty cool! I like piping the output to 'say'.

    • @technovangelist
      @technovangelist Před dnem

      Nice. But I don't really like listening to the voices of 'say'

  • @YotamGuttman
    @YotamGuttman Před 2 dny

    fascinating. thank you for these videos!

  • @michaelgaga
    @michaelgaga Před 2 dny

    Now, it's 4'16 in the video, but I still can't get any practical tips. I must go now to cut the loss...

    • @technovangelist
      @technovangelist Před 2 dny

      Weren’t paying attention? The main tip is don’t use a framework. It’s easier to build without.

  • @henrythomas7112
    @henrythomas7112 Před 2 dny

    Thank you so much for this video. Super helpful and well-presented. Your time and effort is most appreciated!

  • @tharun2003
    @tharun2003 Před 2 dny

    You saved my day. Thank you Matt.

  • @gavinknight8560
    @gavinknight8560 Před 2 dny

    If the cost of abstraction is always complexity. All these frameworks are transitional. As the new layer becomes standardised, the boilerplate tends to disappear, but so does transparency.

    • @technovangelist
      @technovangelist Před 2 dny

      But these frameworks offer little for transparency and a lot of boilerplate and far more complexity...so far. I am sure its just growing pains and they will become amazing after a bit more work.

  • @hiltonwong5419
    @hiltonwong5419 Před 2 dny

    I have watched a few of your videos. I love the way you explain things simply and clearly. Keep it up. Thank you for your work to all of us.

  • @robcz3926
    @robcz3926 Před 2 dny

    with these frameworks it’s the usual ‘sell shovels during a gold rush’ attitude and I’m getting sick of it tbh. especially with those langchain cowboys on yt…

  • @stinkymccheese8010
    @stinkymccheese8010 Před 2 dny

    It would be interesting to see what AI and machine learning could come up with as far as Neuro-training and neuro-therapy protocols.

    • @technovangelist
      @technovangelist Před 2 dny

      Can you tell me more about that? What does that mean?

    • @stinkymccheese8010
      @stinkymccheese8010 Před 2 dny

      @@technovangelist it’s a behaviorist approach to mental health, basically a psychoanalyst does an initial assessment then a technician uses neurofeedback equipment to generate a baseline on the client’s brain function, then the psychoanalyst uses the baseline combined with the initial consult to identify issues and develop a course of treatment which involves using the neurofeedback to interact with a program that will help restore the problem areas of the brain to some degree of normality. All this is based on the theory that mental illness is a function of neurological deregulation, which is just a fancy way of saying that for any number of reasons the sundry Brodmann Areas of the client’s brain are out of sync and not communicating properly and need help getting back into sync.

    • @technovangelist
      @technovangelist Před 2 dny

      Interesting. That's a world I know nothing about. Thanks so much for filling me in.

    • @stinkymccheese8010
      @stinkymccheese8010 Před 2 dny

      @@technovangelist yea after I asked the question I realized it was probably a little too niche I sometimes forget not everyone has read the same books as me.

  • @TokyoNeko8
    @TokyoNeko8 Před 2 dny

    lol... casual day at the household with regular conversations.

    • @technovangelist
      @technovangelist Před 2 dny

      A reference to my wife asking about going to the pub?

    • @TokyoNeko8
      @TokyoNeko8 Před 2 dny

      Yes. I caught the steam the end and saw she came in and you gotta go. It was pretty funny like a reality show stream.

  • @MarcRene
    @MarcRene Před 3 dny

    Hi Matt, thanks for this video. But what would you recommend for the following? I want to develop software with agents. The target is not a specific workflow but a actual software that they need to create.

    • @technovangelist
      @technovangelist Před 2 dny

      Not sure....all i can say is that I wouldn't recommend most of the existing frameworks

    • @MarcRene
      @MarcRene Před 2 dny

      @@technovangelist oki, thanks for getting back at me.

  • @fintech1378
    @fintech1378 Před 3 dny

    is it possible for long range task with non deterministic input and output

  • @NLPprompter
    @NLPprompter Před 3 dny

    thank you i didn't know iTerm2 exists with ollama copilot plugin we can also chat with vault... i put johnntdt contact info on one of machine learning notes... then ask llama3 who is jhonny, and llama3 give me johnydt it said it's misplaced in out of context note.

    • @technovangelist
      @technovangelist Před 2 dny

      its still called iterm2 even though its v3. but that update happened a bunch of years ago... warp.dev is another popular terminal client.

    • @NLPprompter
      @NLPprompter Před 2 dny

      @@technovangelist wow thank you.

  • @Ammarsays
    @Ammarsays Před 3 dny

    Shell-GPT can integrate AI in any terminal. Local models are supported.

  • @samferrer
    @samferrer Před 3 dny

    No body knows anymore

  • @SouthbayCreations
    @SouthbayCreations Před 3 dny

    Darn I missed the live! I'll catch the replay.

    • @technovangelist
      @technovangelist Před 2 dny

      Someday I will plan and schedule a specific time for these

  • @t.genedavis2047
    @t.genedavis2047 Před 3 dny

    Great video. I see now how to use Ollama to get back a specific scheme form the LLM when requesting it. However, how do I provide Ollama with a list of schemes with descriptions of their use, and have the LLM response give a normal message, with a potentially empty list of schemes with their tool names for my code to handle for the LLM to then convert into natural language response? The killer feature of "function calling" is to allow the LLM to assert what functions (if any) need calling. Proper scheme is just icing on the cake.

    • @technovangelist
      @technovangelist Před 3 dny

      the way i have seen others do it, the model just returns the one schema for the output. and the proper scheme is critical if you want to actually call the function, otherwise you need to do a lot of error handling. But generally if you need the model to decide which functions to run you have probably done something wrong in your code.

  • @patriot0971
    @patriot0971 Před 4 dny

    I am just sick of all the AI applications explicitly using CUDA drivers. Anyone with AMD and Intel GPUs are out of luck.

    • @technovangelist
      @technovangelist Před 4 dny

      So you aren’t using ollama? It has support for amd and intel arc is trickling in now

    • @patriot0971
      @patriot0971 Před 3 dny

      @@technovangelist I am using ollama however I am using it on my M2 mac mini. I would love to use it on my gaming rig that has 64GB RAM and a dedicated Intel ARC GPU

    • @technovangelist
      @technovangelist Před 3 dny

      Intel Arc is being added... not sure if you need to do anything special to enable, but the last couple versions have had it. Nothing in the release notes yet.

    • @technovangelist
      @technovangelist Před 3 dny

      I was really surprised to see this coming because Intel has never made it easy to build for their gpus, compared to nvidia or Apple.

    • @patriot0971
      @patriot0971 Před 3 dny

      @@technovangelist Thank you for the update. I just updated Ollama and retested on my gaming rig and same result, it just uses the AMD CPU (slow) vs the Intel ARC GPU . I will keep an eye on the release notes.

  • @chrisogonas
    @chrisogonas Před 4 dny

    Very well illustrated! Thanks Matt

  • @NeeleshKamkolkar
    @NeeleshKamkolkar Před 4 dny

    I am just getting started on learning about agents. So glad to find this video and hear your perspective.

  • @riok4523
    @riok4523 Před 4 dny

    Hey Matt - love your videos. Not really too relavant to Ollama specifically but can you make a video on model types? Im having a hard time understanding what fill-mask models are. Thanks!

    • @technovangelist
      @technovangelist Před 2 dny

      do you mean all the ways to filter for models on HF?

    • @riok4523
      @riok4523 Před 2 dny

      @@technovangelist moreso a description on what some of the model types are, and how to use them. sorry this might be a super amateur question. i primarily use mistral or nous-hermes2 by CLI on my mac but wanted to try using the bert-uncased model

  • @NLPprompter
    @NLPprompter Před 4 dny

    what is kubernetes? me: my nightmare

    • @technovangelist
      @technovangelist Před 2 dny

      used to be mine for my last year or so at datadog and then my first year with what became known as Ollama..

    • @NLPprompter
      @NLPprompter Před 2 dny

      @@technovangelist might a roller coaster ride doing such, 🎢

  • @MeinDeutschkurs
    @MeinDeutschkurs Před 4 dny

    What a great thing. Temporary embeddings for web research. ChromaDB + python + selenium. Should do it. Thx for inspiring me!

  • @MeinDeutschkurs
    @MeinDeutschkurs Před 4 dny

    Really? Multiple models? I know a way to do this on different ports. But on the same port? How?

    • @technovangelist
      @technovangelist Před 4 dny

      That’s been in there for a month or more

    • @MeinDeutschkurs
      @MeinDeutschkurs Před 4 dny

      @@technovangelist , automatically? Do I have to change settings? I have the feeling that a model gets unload during a request on another model. I have to check this out.

    • @technovangelist
      @technovangelist Před 4 dny

      There’s an env var. look at the last few releases and it’s in the notes

  • @MeinDeutschkurs
    @MeinDeutschkurs Před 4 dny

    Funny, today I criticized on a channel that I think that everything is blown up for all of the use cases I’ve seen so far. And I do agree on your opinion! I use python and even for complex or more open actions, I avoid libs like CrewAI or LangChain.

    • @technovangelist
      @technovangelist Před 2 dny

      perfect. my opinion is pretty awesome...at least in my opinion it is.

  • @threepe0
    @threepe0 Před 4 dny

    The model effectively decides when to call the function, and which function is appropriate at the time by formatting the output for the particular function. If I give it a query where taking a headless web screenshot is appropriate, then the output is formatted for that particular function. And the output does NOT particularly need to be in json either, depending on how the function is described to the openai API and how you're choosing to detect that the model "wants" to call a particular function. Respectfully, I think "function calling" is pretty appropriate, and it really seems like there's a misunderstanding of what function calling does here. Not trying to be negative, just pointing out that "just formatting" is missing the point. Of course the model doesn't have access to actually execute code on your local machine. But the "formatting" decided upon (by the model deciding which function to use) is essentially doing the exact same thing. Based on scrolling through the comments here, disappointingly I don't expect you to be very receptive to this feedback. Just my $0.02, and I stand by it. Function calling has a lot of really great uses, and I think "just formatting," while technically true, sorely misses the point. You can (and I do, not just guessing here,) effectively give the LLM a set of tools, trigger those tools based on output from the LLM, and provide the output from those tools back to the LLM if needed. That's pretty impressive I think. All that being said, I really do enjoy your videos. Thank you for sharing, and I look forward to the next one.

    • @technovangelist
      @technovangelist Před 4 dny

      Everything you said here describes what ollama does. Function calling is just formatting. That is the entire point of function calling. People try to say it’s more but even open ai disagrees. I am very receptive to anyone who has anything backing up the opinion that it’s more than just formatting. Please. Show something. Anything. Ollama can be given a list of tools and output the appropriate json to allow the user to call whichever function was chosen. But it is just formatting.

  • @brusdaar
    @brusdaar Před 4 dny

    While I agree with the sentiment that simpler is better. Sometimes you don't know that you could benefit from the ability to let the LLM talk back and forth between other LLMs until you've tried it on several workflows. If your plan is super well defined then it makes sense to simply the design, but for now I'll continue using crewai to scaffold out AI based workflows till I can determine the true limits of the models that are available. Good video though! Thank you for putting it together. It made me think!

    • @technovangelist
      @technovangelist Před 4 dny

      but you don't need crew to add that. crew adds nothing and just complicates

  • @curtkeisler7623
    @curtkeisler7623 Před 4 dny

    I remember that Charles Petzold book . . . learned a lot from that. He's a cool guy. You should check out his book on Turing or his book on Code. Neat guy.

  • @karlroberts5138
    @karlroberts5138 Před 4 dny

    NLU and LLM the way to go, NLU still king as is assured "CERTAINTY", intersplice with tactical LLM agents, LLM with Rag Educate an Agent GUI. This is the model that is forging ahead. If you want a real service there's about another 50 things you need to consider, and most tools, especially codey tools cover <10%. Multi-Customer/Sass, Analytics/Timezones, Chat Client, Skinnable, Configurable, Plug and Go, Voice/Speech/Multi-Modal, AI Agents (LLMS), NLU /LLM Hybrids. Escalations and Triggers. Communication API's, IPAAS back office Services. Multi-Messaging Agents Groups / Messaging Cockpit, or Call centre. Multi- CC Connectors, Pro-Active Tuning, Auto LLM Tuning, and Live Spech Rec tuning. and more.

  • @albugsp
    @albugsp Před 5 dny

    I think for external you go to settings-> API Keys -> Add New Key -> Open AI Compatible, but it didn't work for me 😕

    • @technovangelist
      @technovangelist Před 2 dny

      the key was to set the url to end with /v1 i think.

  • @knoopx
    @knoopx Před 5 dny

    function __fish_complete_ollama_list ollama list 2>/dev/null | tail -n +2 | string replace --regex "\s.*" "" end complete -c ollama -f -a "(__fish_complete_ollama_list)" --condition '__fish_seen_subcommand_from run'

    • @technovangelist
      @technovangelist Před 5 dny

      That kinda of ridiculousness I what I did before. Gollama does a whole lot more too

  • @jlgabrielv
    @jlgabrielv Před 5 dny

    Thanks Matt for the video! great introduction to Msty

    • @technovangelist
      @technovangelist Před 2 dny

      there are a lot of updates in the last few versions and I will be putting out another video when a few more things are added.

  • @vanderleigoncalves5184

    Very good

  • @PankajDoharey
    @PankajDoharey Před 6 dny

    You mentioned Charles Petzold, when i first read it i was 18. A lot of time has passed.

  • @oomraden
    @oomraden Před 6 dny

    i did not see any manager agent in this video. what did i miss?

    • @technovangelist
      @technovangelist Před 6 dny

      If you have a workflow that requires one add it. You probably don’t.

  • @banzie74
    @banzie74 Před 6 dny

    Subscribed and linked. Definitely an underrated video. Yet to check out your other videos, and will check those out too. You mentioned that you were using a m1 based laptop (assuming, due to the reference to battery life). Not sure if you sped up the inference part of the video, but if that is the native inference speed, it really is great. Is there any video where you've covered the specs of this laptop ?

    • @technovangelist
      @technovangelist Před 6 dny

      I have a M1 Max MacBook Pro with 64gb ram and 4tb disk. It’s great. I’ve had it for about 2.5 years. It was expensive at the time but I see it pop up on eBay for 1600 (usd) or so every now and then.

  • @attaboyabhi
    @attaboyabhi Před 6 dny

    having a RAG will be cool

    • @technovangelist
      @technovangelist Před 2 dny

      Its pretty nice now, but i am looking forawrd to some improvements in the next version or two

  • @kwokallsafe5642
    @kwokallsafe5642 Před 6 dny

    VID SUGGESTION ~ (Resolve Error Response: Invalid Volume Specification) - Thanks test@xz97:~$ docker run -d --gpus=all -v /home/test/models/:root/.ollama -p 11434:11434 --name ollama ollama/ollama docker: Error response from daemon: invalid volume specification: '/home/test/_models/:root/.ollama': invalid mount config for type "bind": invalid mount path: 'root/.ollama' mount path must be absolute.

    • @technovangelist
      @technovangelist Před 2 dny

      that’s more of a support request.... the error message is all you need. You specified a relative path rather than a absolute one. refer to the docs on docker hub for the image

    • @kwokallsafe5642
      @kwokallsafe5642 Před 2 dny

      @@technovangelist - Thanks Matt for your reply - Discovered the is a "/" SLASH missing before root - (Problem solved) Thanks again.

  • @mcpduk
    @mcpduk Před 6 dny

    old skool....loads of good frameworks make embedding VERY VERY simple

    • @technovangelist
      @technovangelist Před 6 dny

      It’s pretty hard to beat the simplicity without a framework. And most frameworks just complicate without benefit. Like langchain and llamaindex.

  • @winkler1b
    @winkler1b Před 6 dny

    You can close a Dialog with the ESC key. I had the exact same response.... especially because the window border is so faint. Took me a while to realize was in a modal.

    • @technovangelist
      @technovangelist Před 2 dny

      I'll have to review it again to remember what I did

  • @xspydazx
    @xspydazx Před 7 dny

    very good !!! qell understood.... (quick advice)... Tempreture is related to the training also (as things which were not trained deeply will need higher tempreture ... and things deeply embedded will be ok with lowest tempreture: how do people train thier odels and what are thier acceptable levels? as some are .=0.5 and under whilst other dont care and let the model complete an epoch on large dataset and assume the data took .. as long as thier final output was preferable : when in fact all the data which did not go in at the loss below .0.5 did not take and is not retrivable perhaps its there ephemeallly ... as it is like a pretraining ... its just used for next word prediction... but we are doing tasks ! which is whole sequece prediction/recall so when we train for a task we expect the whole of the data set to be fit in range .... so low temptrture 1 should be acceptable losses ... Some say tha this effect the soft max of possiblisty chosen byu the topk sample as well as the topP percentage of cutt of... but this is when there are many sample chosen... but this also depicts the values that were trained at thatr rate of loss .... so it will be collecting sample from the level under the temptrture rate of 1 ( a lot ) so this will need constraining with topP (selecting the highest of probablitys ... but the softmax will also spread them alowing for more random also, when the model has been over trained.) .. so an over trained model can be loosened by raising the temptretue and a wild model tamed ! lol...

    • @xspydazx
      @xspydazx Před 7 dny

      i would like to see a vdieo on publishing a modl really !

  • @miscodices
    @miscodices Před 7 dny

    i hope add option for change folder or c: to d: for download models or install olama

    • @technovangelist
      @technovangelist Před 7 dny

      That’s something you set in environment variables before you start Ollama, so not really something this would do