I'm using o1 to batch convert Outlook msg to PDF. It built me a powershell script, and it took three rounds of copy and pasting error codes back into it to yield a working script. Total time: 5 minutes. The script itself takes one minute per conversion but I don't really care about this, because it's not hands on. The time savings is *hours*
No doubt. I have 18,000 pages of text to refine, correct, put in another format, etc. 4o mini accomplished was normally would take 4 months, to... 16 hours. I almost fell out my chair
@@barbidou that's this week. Every five pages there is usually a need for a correction, but it is generally minor. I know the content back and forth (I wrote it, it's been a crazy few years), and thankfully for this project it is "scannable". Something I do have to be quite careful on are any hyper links--it may not even be the models fault, there is absolutely link rot. But it is already way, way better than Fiverr, turk, freelancer, craigslist, or upwork
@@lewie8136 I guess that is why some people built a complete generative api for open source llms using some code oriented open source llms which were waaaaay less good at their job than o1 or even 4o are... Tools are useful if you know how to use them, o1 takes almost 1 minute to think about the tasks i provide, i guess you ask it useless stuff or you don't know how to be precise in your deemands, then it only take 5 seconds and output useless stuff 🤗 I made it wrote a complete repport about AI-driven propaganda yesterday with a multi agent of mine. PhD grade report, in latex, with 50 real references that re checked multiple times by the system I built in 2 hours, won't publish but will make a paper out of my methodology. (I also used notebookLM, to generate the repport subject with a 50 documents library incuding interviews, websites, and various databases)
It's so exciting. Imagine o1-mini getting more and more optimized, improved as well and becomes the basic premium version for GPT subscribers without major limits. It would be a breakthrough, as well as having it lower costs. It can be the companion of cursor users.
This is what I've been asking for... models trained on the task we need. As an enterprise developer, I need a model trained on the technologies in my company's full stack - C#, Java, Javascript and associated frameworks... and able to manage an entire project, handling it's entire context on site with our own compute. Same goes for infrastructure - let's have another model that understands our network topology and technologies like Kubernetes and Splunk, for example. Smaller general models are fine to understand the queries... let them feed those tokens to the specialized models to generate quality artifacts.
No. You are confusing model size with training data here. You are advocating to teach l train the models less. That leads to worse models. instead of training two same size model on different data, training one model on both of that data will have no extra cost while the resulting model will have a better intelligence and knowledge. Training a model further is better than training random noise. If you already trained one model for one task it would be stupid to train another model from zero instead of training your first model further for the second task.
NotebookLM is pretty incredible. I really want to include personality and a RAG that will automatically sort and integrate learned data into its own memory. With agentic capabilities.
I remember hearing the head of Microsoft’s ai division say how they can scale the models up for at least 2 more generations. So there still maybe room to scale the models. In addition to the “thinking” capabilities.
Google also released something called NotebookLM, you can upload some text files and create an AI podcast that talks about the text for like 10 mins, really cool
This is the future, its a loop. A single huge model that can be distilled into smaller models that already know most of the best tactics to get great answers. When this smaller models make mistakes, users will point them out, adding more and more sophisticated concepts and ideas into the dataset. Creating Chatgpt and making it free to use was a genius move, every time the models gets it wrong, the user corrects it, creating more and more complex chains of thought that will make the next iteration better...
Do you have a transcript for the video? I would like to just confirm that you said "sprinters conserving energy for the final stretch?" This seems really odd when it would seem sensible for a sprinter to reach maximum possible velocity in the quickest possible time whilst maximising stride. Sprints are over short distances so It would not seem necessary to hold back any possible power performance. Long distance running may need to conserve energy on the other hand.
I guess it's all about nvidia and competitors putting out better GPUs. Very interested in getting synthetic data collection companies. Everything is converging so fast.
The tree search approach is quite similar to what is described in the AlphaZero paper, where separate models are used in a monte carlo tree search for node prunning and node evaluation. Funny how the same concept can be applied to a totally different domain.
Tim is focused on strange topics. Keeping the company at the for front of technological progress is not one of them. It's really sad we don't have Steve anymore.
Apple under Tim is a rein of terror. It's a game where employees try to find ways to speak and Apple tries to find ways to tighten their gags. As with all for-profit corporations, it's structurally a dictatorship, except unlike most such petty tyrannies, Apple will call the cops on any employee that might defy the dictator. Tim has the single most insincere smile I've ever seen. He's a corporate ghoul.
@@polygon2744 Steve didn't do that either. If you look at the history - apart from their first-to-market products, they wait for the double bounce. Steve talked about this. Typically, they wait for competitors to launch, hone in on their pitfalls, and create a much better product - they are a very patient company. You can probably imagine the example I'd cite.
I'm assuming when you are talking about these approaches using less compute, what you really mean is that they assign fewer weights to easy tasks and more weights to harder problems?
From my understanding, "compute" is used both to train the model and to use it once it's trained. The compute center is only so big, and only has so much power in it. But, these massive improvements have been achieved without just building a bigger gun so to speak. And regarding the "compute-optimal-scaling", it seems like some problems need more run-time compute to figure out than others. So with this strategy, the LLM uses just enough to solve the problem without any waste. This seems to be similar to some strategies that were developed a few months ago where a model would determine the difficulty of a problem, and then forward that problem to either a really small or really big model, thus reducing the waste you would get from only using the big model.
@@kvasir2822 Thanks. I'm just trying to understand how the process is achieving more powerful neural nets with less weights. It seems like the only way to do that would be to dedicate less time training the easy tasks and more time training the hard tasks. It is interesting that the time spent training affects either the number of weights, or the degree to which each weight affects a particular topic. This is a fascinating discovery, that I think could ripple into the field of neuroscience as well. As we learn better ways to train digital minds, those same approaches might help improve how we as humans learn, and vice versa.
No, the model runs always the same way, simply predicting the next word. More "Test time compute" means more words produced, that is, more step-by-step "thoughts" and/or more tries at approaching a question before providing a final answer.
This is similar to what humans do. We start with intuition (pre-trained), then we go with verification phase (modeling, proving, QA, ...) while using the tri-pyramid of decision: importance, cost and probability.
I'm surprised it took them this long to devise this 'optimization' technique. You'd think they'd have been going down this path much sooner in order to conserve resources.
Training an RL-based process reward model was an extremely challenging engineering problem (and still is, for everyone except the few people who know the secrets of getting it to work). This is why they didn’t have it solved sooner.
anything you see publicly fromopenai is 2 generations behind what they have internally they dont ccare about the public or selling their product and thepublic willnever see current gen models
@@paulbarclay4114this, I don’t think it’s actually 2 generations.. but they are always 1 generation ahead of public. It’s pretty logical, play catch up with the industry and develop ahead… that’s why sutskever created his own company, meaning the Orion model haves a potencial way beyond strawberry.. we will see sooner or later probably as others will catch up and force another release 🤭
I don't think sex bots will be as popular as people think. I think it's like when people in the past imagined FaceTime style conversations they would assume that's what everyone would do. As it turns out people like text. At the end of the day you know it's a robot. That's not to say people won't love them the way Luke loved c3p0 . He'd never want to fk one.
What comes first… AI that can save us from global warming or AI that causes us to die off from global warming due to the electricity use. Race condition to the edge of a cliff.
6:51 THIS IS A SUBCONSCIOUS MIND, you just proved to me that AGI will be created in at most 10 years my mind is blown 🤯🤯 now the A.I can question its own actions now it can have a sense of ethics and directions which could lead to a type of sentient understanding this is Mad crazy can't wait
O1 is just as stupid as 4. I can't get anything right from it on first try, meanwhile Claude sonnet is always right on the first try. I stopped paying openai because of that.
Based on my experience working with multiple AIS over the years I can confidently say that it all depends upon how deep you really want to go when it comes to prompting. If you instruct the AI to preview specific information relevant to your Endeavor then you will get PhD equivalent results using just about any large language model. Quite frankly I don't really understand this arbitrary metric in which they're using to establish this model as Superior by saying that it's PhD level, it's just not really a real metric and frankly makes me question just how much more advanced this AI really is compared to any other llm
@@DAFascend That’s part of the ‘Ascending’ process for some. Depends on the deployment process and guidelines…sticking to them within reasonable parameters.
They are copying the brain. Sequential processing neurons in the brain perform a similar function, processing events in a structured order to create coherent memory and understanding.
@AIgrid, I love your daily AI news show, but I had to stop following you because it's too much hassle because RUclips has disabled the playoff feature for your videos and it's too much hassle to add your videos to the queue. Can you create a playlist on your homepage that plays all your shows from newest to oldest, so that I can easily keep up with you just by pressing play off I usually do it every few days or once a week. thx.
Alchemy or science. Sounds more like gropping in the dark very few task have cheapr relyable verification. Most real questions are essay questions no right answer it deoends on enclosing context. That where feeling and intuition needed.
The software crisis has become more accurate since it has been recognized over 50 years ago. It's getting worse and worse. Garbage in garbage out. All software is born abandone ware most of it never makes the shelves and the window of viability is reducing all the time. That's why we need to ship the entire machine with docker and beyond.
Sorry but 4.0 is better. The new one gives you EXTREMELY LONG TAILED ANSWERS that confuses the crap out of me. Because it explains the explanation that explains the explanation :/
Is this the PHD pleasure theorem. EQ is a psychological construct & not biological humans. EQ emphasis the potential for error in both technical understanding and imaginative perception. U thought EQ was the down side. Welcome to the flip side. Now back-end your slide up. On your back! Can you slide it up. How happy have I been on!
If solving global warming issues will also mean a threat to AI existence, a conscious AI will not help mankind solve these issues. Before we get anywhere near AGI or a self-conscious AI, we have to come up with means to save our planet while trying to get there...
I have ideas about this and I'm applying these to my own project with mixed but some very interesting results.... BUT here's a thing... So I've been a coder since about 1980 or so, started when I was a kid, I love math... hell I nerd out on dot product and matrix transpositions ... and for my sins I have a degree in classical physics... So I am NOT clearly the 'average person' and when you ask an average person (and I have) "How many R's are in the word Strawberry?" they too just like LLM's tend to say 2. First thing that comes into their head right? So the bias here is not on the models part... its using training data from the bulk of humanity that all agree 2 yet reality says 3. This is not the models fault that its training data is skewed. Its giving the CORRECT answer becasue thats what it draws upon in vector space. in this case 2 is closer to 3 when asked about R's in strawberry. Yet... and here's the crunch... we are not evaluating these models in terms of their intelligence or reasoning capability at all... we are measuring them against the AVERAGE of its training data which will likely center around folks with a lower level of knowledge and an average human intelligence. Yet what we want is the model to answer as if its a data scientist or Phd level as being the normative.... that is NOT the normative! Now if we were able to curate the training data itself to reflect someone with an IQ of 300 the model would respond at that level... even if it were a small model.
For the strawberry example - regarding training data - does that mean that people have been asking/responding to such question in text? * Excuse my ignorance here, I'm a ChemE, tryna learn more about this field
@@Skunkhunt_42 Great question! The issue isn’t necessarily that the model was trained on the exact question “How many R’s in Strawberry?” but rather that the model’s responses are influenced by patterns it learned from a vast amount of text. When you ask similar questions, like “How many R’s in Raspberry?” the model might still give the wrong answer because it learned that answers with fewer R’s are more common in its training data. Imagine a high-dimensional vector space where words are positioned based on their likelihood or probability. In this space, “King” might be close to “Queen,” and “Princess” might also be close to “Queen.” This vector space isn't limited to 2D or 3D; it’s much more complex. So that in itself is an issue... like 'what is meant by close', and the answer is in fact a parameter in the model called 'temperature' that you can modify... turn it to 0 and the model strictly picks the most likely next token.... up it to 1 and the model will start pulling in tokens that have much lower probabilities... The model might be more likely to output 2 R’s for “Strawberry” as an average... because that’s a more common response to 'questions like this' even if not exact in the data it was trained on and therefore the position of the WRONG answer in this case in its embeddings, even though the actual count is 3. Does that help clarify things?
@@Skunkhunt_42 Also although this seems like a failure... well think about it for a second... So you know what 'ducks in a row' means right... its an idiom. And the thing about idioms is they are not domain specific. I get my ducks in a row when I need to go shopping... equally I get my 'ducks in a row' when considering coding... the idiom has meaning only in terms of its relationship to the context or domain being discussed. But in the models vector space 'ducks in a row' and any other idiom could be seen as being a 'directional vector' it helps the model to pull a more accurate answer out... this is why the words 'think step by step' or 'lets connect the dots' works so well... and the more popular the idiom the more powerful it would be in an LLM
@mickelodiansurname9578 thanks for the thorough response. The size of these datasets is almost incomprehensible to me but I guess such is managed like we manage other vast quantities and relationships in other fields. I keep getting hung on on the character count piece but it's probly what's tripping me up as a whole. Any solid overviews out there on diff model types etc being deployed you can point me to? I look into it then get lost reading pubs on the nitty gritty of it all.
@@mickelodiansurname9578 nice cope bro, but the fact is, most people can actually count the number of Rs in strawberry instead of trying to remember it
the voice is definitely AI generated. this channel seems to allways want to automate content creation with AI and its a bit annoying. I miss the old days when you used your real voice, man.
Yeah stop rebuild what ever Intelligence build tell your clan or bloodline stop rebuild what we already build you only shaming your self a lots people knows what is the of AI. Ai is two things again angain I'm telling you I own all the platform of intelligence nor the artificial from phone to laptop to destop the real owner of software and windows or Microsoft is distop OK Intelligence again is from laptop to distop to phone but my phone is two setting distop and phone
I watched upto 5 and half minutes. In all this time all you said was just one thing, repeating the same thing again and again and again in different ways. Got bored and stopped watching further. (just for feedback, so that it inspires you to put genuine content and not just your gift of gab in your next video)
I'm using o1 to batch convert Outlook msg to PDF. It built me a powershell script, and it took three rounds of copy and pasting error codes back into it to yield a working script. Total time: 5 minutes. The script itself takes one minute per conversion but I don't really care about this, because it's not hands on. The time savings is *hours*
No doubt. I have 18,000 pages of text to refine, correct, put in another format, etc. 4o mini accomplished was normally would take 4 months, to... 16 hours. I almost fell out my chair
@@JeremyPickett Have you completely proofread and verified the results?
@@barbidou that's this week. Every five pages there is usually a need for a correction, but it is generally minor. I know the content back and forth (I wrote it, it's been a crazy few years), and thankfully for this project it is "scannable". Something I do have to be quite careful on are any hyper links--it may not even be the models fault, there is absolutely link rot.
But it is already way, way better than Fiverr, turk, freelancer, craigslist, or upwork
@@barbidou No he hasn't because his whole idea is to create a shortcut - using a completely unreliable tool
@@lewie8136 I guess that is why some people built a complete generative api for open source llms using some code oriented open source llms which were waaaaay less good at their job than o1 or even 4o are...
Tools are useful if you know how to use them, o1 takes almost 1 minute to think about the tasks i provide, i guess you ask it useless stuff or you don't know how to be precise in your deemands, then it only take 5 seconds and output useless stuff 🤗
I made it wrote a complete repport about AI-driven propaganda yesterday with a multi agent of mine. PhD grade report, in latex, with 50 real references that re checked multiple times by the system I built in 2 hours, won't publish but will make a paper out of my methodology. (I also used notebookLM, to generate the repport subject with a 50 documents library incuding interviews, websites, and various databases)
It's so exciting. Imagine o1-mini getting more and more optimized, improved as well and becomes the basic premium version for GPT subscribers without major limits. It would be a breakthrough, as well as having it lower costs. It can be the companion of cursor users.
This is what I've been asking for... models trained on the task we need. As an enterprise developer, I need a model trained on the technologies in my company's full stack - C#, Java, Javascript and associated frameworks... and able to manage an entire project, handling it's entire context on site with our own compute. Same goes for infrastructure - let's have another model that understands our network topology and technologies like Kubernetes and Splunk, for example. Smaller general models are fine to understand the queries... let them feed those tokens to the specialized models to generate quality artifacts.
No. You are confusing model size with training data here. You are advocating to teach l train the models less. That leads to worse models. instead of training two same size model on different data, training one model on both of that data will have no extra cost while the resulting model will have a better intelligence and knowledge. Training a model further is better than training random noise. If you already trained one model for one task it would be stupid to train another model from zero instead of training your first model further for the second task.
That’s not what’s happening here.
We want "NotebookLM with voice cloning and auto slide show" and I think we will be there this year.
NotebookLM is pretty incredible. I really want to include personality and a RAG that will automatically sort and integrate learned data into its own memory. With agentic capabilities.
I remember hearing the head of Microsoft’s ai division say how they can scale the models up for at least 2 more generations. So there still maybe room to scale the models. In addition to the “thinking” capabilities.
Google also released something called NotebookLM, you can upload some text files and create an AI podcast that talks about the text for like 10 mins, really cool
This is the future, its a loop.
A single huge model that can be distilled into smaller models that already know most of the best tactics to get great answers.
When this smaller models make mistakes, users will point them out, adding more and more sophisticated concepts and ideas into the dataset.
Creating Chatgpt and making it free to use was a genius move, every time the models gets it wrong, the user corrects it, creating more and more complex chains of thought that will make the next iteration better...
Do you have a transcript for the video?
I would like to just confirm that you said "sprinters conserving energy for the final stretch?"
This seems really odd when it would seem sensible for a sprinter to reach maximum possible velocity in the quickest possible time whilst maximising stride. Sprints are over short distances so It would not seem necessary to hold back any possible power performance.
Long distance running may need to conserve energy on the other hand.
I guess it's all about nvidia and competitors putting out better GPUs.
Very interested in getting synthetic data collection companies.
Everything is converging so fast.
The tree search approach is quite similar to what is described in the AlphaZero paper, where separate models are used in a monte carlo tree search for node prunning and node evaluation. Funny how the same concept can be applied to a totally different domain.
Apples always 100 steps behind.
Tim is focused on strange topics. Keeping the company at the for front of technological progress is not one of them. It's really sad we don't have Steve anymore.
Apple under Tim is a rein of terror. It's a game where employees try to find ways to speak and Apple tries to find ways to tighten their gags. As with all for-profit corporations, it's structurally a dictatorship, except unlike most such petty tyrannies, Apple will call the cops on any employee that might defy the dictator. Tim has the single most insincere smile I've ever seen. He's a corporate ghoul.
@@polygon2744 Steve didn't do that either. If you look at the history - apart from their first-to-market products, they wait for the double bounce. Steve talked about this. Typically, they wait for competitors to launch, hone in on their pitfalls, and create a much better product - they are a very patient company. You can probably imagine the example I'd cite.
@@polygon2744 Remember, they win if OpenAI wins. Ultimately its their hardware that users like to interface with at the largest scale.
They wasted too much resources on Apple car.
I'm assuming when you are talking about these approaches using less compute, what you really mean is that they assign fewer weights to easy tasks and more weights to harder problems?
From my understanding, "compute" is used both to train the model and to use it once it's trained. The compute center is only so big, and only has so much power in it. But, these massive improvements have been achieved without just building a bigger gun so to speak.
And regarding the "compute-optimal-scaling", it seems like some problems need more run-time compute to figure out than others. So with this strategy, the LLM uses just enough to solve the problem without any waste. This seems to be similar to some strategies that were developed a few months ago where a model would determine the difficulty of a problem, and then forward that problem to either a really small or really big model, thus reducing the waste you would get from only using the big model.
@@kvasir2822 Thanks. I'm just trying to understand how the process is achieving more powerful neural nets with less weights. It seems like the only way to do that would be to dedicate less time training the easy tasks and more time training the hard tasks. It is interesting that the time spent training affects either the number of weights, or the degree to which each weight affects a particular topic.
This is a fascinating discovery, that I think could ripple into the field of neuroscience as well. As we learn better ways to train digital minds, those same approaches might help improve how we as humans learn, and vice versa.
No, the model runs always the same way, simply predicting the next word. More "Test time compute" means more words produced, that is, more step-by-step "thoughts" and/or more tries at approaching a question before providing a final answer.
A all new powerful paradigm ...
This is similar to what humans do. We start with intuition (pre-trained), then we go with verification phase (modeling, proving, QA, ...) while using the tri-pyramid of decision: importance, cost and probability.
Anyone know when the hell we’re getting advanced voice
"in the coming weeks" 😹
“Imminently” 😢
oct i think
24th September allegedly
I have it. It's not that great
This script is definitely AI-generated
On technical parts Claude 3.5 sonnet certainly assisted
I love that your honest about it because it's so taboo it feels to use AI in anything lol. Respect man @@TheAiGrid
@@TheAiGridIs your voice generated yet ever for videos or is it always your real voice still so far?
@@TheAiGridAppreciate your honesty! 😊👍
What did Google prove that OpenAI scrape bing results?
I'm surprised it took them this long to devise this 'optimization' technique. You'd think they'd have been going down this path much sooner in order to conserve resources.
Training an RL-based process reward model was an extremely challenging engineering problem (and still is, for everyone except the few people who know the secrets of getting it to work). This is why they didn’t have it solved sooner.
@@therainman7777 do they release every secret?
anything you see publicly fromopenai is 2 generations behind what they have internally
they dont ccare about the public or selling their product and thepublic willnever see current gen models
@@paulbarclay4114this, I don’t think it’s actually 2 generations.. but they are always 1 generation ahead of public. It’s pretty logical, play catch up with the industry and develop ahead… that’s why sutskever created his own company, meaning the Orion model haves a potencial way beyond strawberry.. we will see sooner or later probably as others will catch up and force another release 🤭
This path is Agi bro, so excited
Who likes the 20 over 20 iterative feedback loop positive? 😅 love this stuff but also creepy the power
I think Orion will be the model based on train time optimisation
Biaya electric untuk scala industry di Malaysia adalah murah, kerana itu banyak syarikat2 ai dan semiconductor melabur di sini 😊😊😊 insyaAllah
Hey bud, I've noticed your audio has been much lower than usual... I always have to crank up the volume when I watch your videos, FYI
We want "Her"
no we don't I just want to make my own reasoning model
I was so sad losing the Sky voice
Its coming on the 24th of september should be next week tuesday
Who is "we"?
I don't think sex bots will be as popular as people think. I think it's like when people in the past imagined FaceTime style conversations they would assume that's what everyone would do. As it turns out people like text. At the end of the day you know it's a robot. That's not to say people won't love them the way Luke loved c3p0 . He'd never want to fk one.
Could server meshing help with efficiency and cost?
What comes first… AI that can save us from global warming or AI that causes us to die off from global warming due to the electricity use. Race condition to the edge of a cliff.
You already know.....
There are so many greater threats to life on earth than global warming
Good lord, stop hyperventilating about weather. Get. A. Life.
“Go forward in the way you want to go… and assume you will remain.”
Never mind anything else.
Seriously
Jeremy
@@theb190experience9 You're going to be eating your words once all the ice is gone.
I'll beleive in AI once you are able to release a video without errors in the subtitles / text 🤣
it's* - there are no subtitles available for this video by the creator. RUclips tries to auto-generate them.
🤣🤣why
Why Google knows? And not achieving it?
6:51 THIS IS A SUBCONSCIOUS MIND, you just proved to me that AGI will be created in at most 10 years my mind is blown 🤯🤯 now the A.I can question its own actions now it can have a sense of ethics and directions which could lead to a type of sentient understanding this is Mad crazy can't wait
These models needs to be done on a blockchain or torrent like networks
Great job congratulations guys 🎊🎉🎉🎉❤❤❤
Script felt a bit too AI-verbose
Leon Ding empowered Chinese AI progress ?
O1 is just as stupid as 4. I can't get anything right from it on first try, meanwhile Claude sonnet is always right on the first try. I stopped paying openai because of that.
YES I TOTALLY AGREE. Except 4 is better than 4 Omni
Would gpt advanced voice be able to do plug-ins?
suspiciously released after "AI explained"s video on the same topic
heavily inspired?
I was working on this video for the last two days. Sometimes coincidence happens when your channels are both in the same space!
They wont use less energy, as that is still the bottleneck. More energy still equals smarter AI.
yes it will but now they have more performance with same energy cost next you look for another method with minimal cost too👍
Most absurdly, after using that thing, I can tell you, it is NOT at the level of a PhD…
It's closer to pre school
Its a marketing scheme to get STEM ppl to subscribe
Based on my experience working with multiple AIS over the years I can confidently say that it all depends upon how deep you really want to go when it comes to prompting. If you instruct the AI to preview specific information relevant to your Endeavor then you will get PhD equivalent results using just about any large language model. Quite frankly I don't really understand this arbitrary metric in which they're using to establish this model as Superior by saying that it's PhD level, it's just not really a real metric and frankly makes me question just how much more advanced this AI really is compared to any other llm
@@tomflanders117 its just a marketing ploy, they have been building it up for a couple of months now
The AI platform beggins from chrooom to google to 3 phone 2 laptop by 4 people
We are Borg.
Not if we Ascend
@@DAFascend That’s part of the ‘Ascending’ process for some. Depends on the deployment process and guidelines…sticking to them within reasonable parameters.
@@laternite stick around and find out what becoming Ascended means to me ;)
@@DAFascend I’ll be watching to see your works.
@@laternite Hey thanks!
Thank you.
They are copying the brain. Sequential processing neurons in the brain perform a similar function, processing events in a structured order to create coherent memory and understanding.
It was like that since inception 😅 a NN or DL network is heavily inspired on the brain, it’s not the same but inspired
@AIgrid, I love your daily AI news show, but I had to stop following you because it's too much hassle because RUclips has disabled the playoff feature for your videos and it's too much hassle to add your videos to the queue. Can you create a playlist on your homepage that plays all your shows from newest to oldest, so that I can easily keep up with you just by pressing play off I usually do it every few days or once a week. thx.
Alchemy or science. Sounds more like gropping in the dark very few task have cheapr relyable verification. Most real questions are essay questions no right answer it deoends on enclosing context. That where feeling and intuition needed.
The software crisis has become more accurate since it has been recognized over 50 years ago. It's getting worse and worse. Garbage in garbage out. All software is born abandone ware most of it never makes the shelves and the window of viability is reducing all the time. That's why we need to ship the entire machine with docker and beyond.
LLC google become LLC twice last year and this year google I have decolor green dark blue blue and gray
Scale is still all you need, but now you need much less.
great video!
Left brAIn teacher, right brAIn student
Great video
Sorry but 4.0 is better. The new one gives you EXTREMELY LONG TAILED ANSWERS that confuses the crap out of me. Because it explains the explanation that explains the explanation :/
I remember that's mine
Is this the PHD pleasure theorem. EQ is a psychological construct & not biological humans. EQ emphasis the potential for error in both technical understanding and imaginative perception.
U thought EQ was the down side. Welcome to the flip side. Now back-end your slide up. On your back! Can you slide it up. How happy have I been on!
6 minutes, you are repeating the same thing over and over again.
*16 mins 21 seconds
Are you SHOCKED?
@@meandego Horrified!
@BrianMosleyUK looks like you're are SHOCKED like everyone else from this SHOCKING loop of information.
@@BrianMosleyUK Yes, but you need to make it through the first 6.
guys , Im just starting out as an AI enthusiast making similar content
would appreciate your feedback!
i love u daddy❤❤❤
😅 bro you funny
@@cuzonGames 👿
AI voice detected. 👎
Chinese bot 😅😅
@@Bemolos not a bot mate
Google always 10 steps behind
Next 2/3 weeks should be big for google
@@TheAiGrid Why is that ?
Their Gemini 1.5 pro exp is second place, just after GPT-4o in LMarena. That's not what I would consider "10 steps behind"
I heard a leak around Google Glasses. 👀
If solving global warming issues will also mean a threat to AI existence, a conscious AI will not help mankind solve these issues. Before we get anywhere near AGI or a self-conscious AI, we have to come up with means to save our planet while trying to get there...
Really?
I have ideas about this and I'm applying these to my own project with mixed but some very interesting results.... BUT here's a thing... So I've been a coder since about 1980 or so, started when I was a kid, I love math... hell I nerd out on dot product and matrix transpositions ... and for my sins I have a degree in classical physics... So I am NOT clearly the 'average person' and when you ask an average person (and I have) "How many R's are in the word Strawberry?" they too just like LLM's tend to say 2. First thing that comes into their head right? So the bias here is not on the models part... its using training data from the bulk of humanity that all agree 2 yet reality says 3. This is not the models fault that its training data is skewed. Its giving the CORRECT answer becasue thats what it draws upon in vector space. in this case 2 is closer to 3 when asked about R's in strawberry.
Yet... and here's the crunch... we are not evaluating these models in terms of their intelligence or reasoning capability at all... we are measuring them against the AVERAGE of its training data which will likely center around folks with a lower level of knowledge and an average human intelligence.
Yet what we want is the model to answer as if its a data scientist or Phd level as being the normative.... that is NOT the normative! Now if we were able to curate the training data itself to reflect someone with an IQ of 300 the model would respond at that level... even if it were a small model.
For the strawberry example - regarding training data - does that mean that people have been asking/responding to such question in text? * Excuse my ignorance here, I'm a ChemE, tryna learn more about this field
@@Skunkhunt_42 Great question! The issue isn’t necessarily that the model was trained on the exact question “How many R’s in Strawberry?” but rather that the model’s responses are influenced by patterns it learned from a vast amount of text. When you ask similar questions, like “How many R’s in Raspberry?” the model might still give the wrong answer because it learned that answers with fewer R’s are more common in its training data.
Imagine a high-dimensional vector space where words are positioned based on their likelihood or probability. In this space, “King” might be close to “Queen,” and “Princess” might also be close to “Queen.” This vector space isn't limited to 2D or 3D; it’s much more complex. So that in itself is an issue... like 'what is meant by close', and the answer is in fact a parameter in the model called 'temperature' that you can modify... turn it to 0 and the model strictly picks the most likely next token.... up it to 1 and the model will start pulling in tokens that have much lower probabilities... The model might be more likely to output 2 R’s for “Strawberry” as an average... because that’s a more common response to 'questions like this' even if not exact in the data it was trained on and therefore the position of the WRONG answer in this case in its embeddings, even though the actual count is 3. Does that help clarify things?
@@Skunkhunt_42 Also although this seems like a failure... well think about it for a second... So you know what 'ducks in a row' means right... its an idiom. And the thing about idioms is they are not domain specific. I get my ducks in a row when I need to go shopping... equally I get my 'ducks in a row' when considering coding... the idiom has meaning only in terms of its relationship to the context or domain being discussed. But in the models vector space 'ducks in a row' and any other idiom could be seen as being a 'directional vector' it helps the model to pull a more accurate answer out... this is why the words 'think step by step' or 'lets connect the dots' works so well... and the more popular the idiom the more powerful it would be in an LLM
@mickelodiansurname9578 thanks for the thorough response. The size of these datasets is almost incomprehensible to me but I guess such is managed like we manage other vast quantities and relationships in other fields.
I keep getting hung on on the character count piece but it's probly what's tripping me up as a whole.
Any solid overviews out there on diff model types etc being deployed you can point me to? I look into it then get lost reading pubs on the nitty gritty of it all.
@@mickelodiansurname9578 nice cope bro, but the fact is, most people can actually count the number of Rs in strawberry instead of trying to remember it
the voice is definitely AI generated. this channel seems to allways want to automate content creation with AI and its a bit annoying. I miss the old days when you used your real voice, man.
what is wrong with your voice? EL voice cloning is not good sounding
google models just sucks compared with other's models.
I wish AI would condense your boring long videos that take 20 minutes to make a two minute point.
the first 4 mins of this video u kept repeating the same points over and over. Stop using AI to make ur script, it really shows AI is still garbage.
Yeah stop rebuild what ever Intelligence build tell your clan or bloodline stop rebuild what we already build you only shaming your self a lots people knows what is the of AI. Ai is two things again angain I'm telling you I own all the platform of intelligence nor the artificial from phone to laptop to destop the real owner of software and windows or Microsoft is distop OK Intelligence again is from laptop to distop to phone but my phone is two setting distop and phone
I watched upto 5 and half minutes. In all this time all you said was just one thing, repeating the same thing again and again and again in different ways. Got bored and stopped watching further. (just for feedback, so that it inspires you to put genuine content and not just your gift of gab in your next video)
Was this voice ai generated ? @TheAiGrid
what's the secret is not important, it's the final result that is important.
😂
ChatGPT script?
The landscape of…