- Видео 548
- Просмотров 105 485
Richard Aragon
США
Добавлен 28 сен 2011
I am the CEO of Applicable Grounded Innovations If you are interested in AI Services, feel free to visit our website: www.applicablegroundedinnovations.com!
Also feel free to browse my products Gumroad Page if you like: aragonized.gumroad.com/
Also feel free to browse my products Gumroad Page if you like: aragonized.gumroad.com/
Revealing the Platonic Forms of Data: AI’s Ultimate Breakthrough!
Cross Entropy Is All You Need Research Paper: arxiv.org/abs/2410.21869?fbclid=IwY2xjawHiXVhleHRuA2FlbQIxMAABHa8MUw-nBQnBCs7auXNz0GVaXwXWXNUl_mEAZOC6Jt8_x0m0B2t6QAhPmQ_aem_BOUiAkGAN9t8MV-yY0AfFQ
Colab Notebook: colab.research.google.com/drive/1mm7TDgg8lqjUn7tNZ6rBEWRdcikmAloR?usp=sharing
The video discusses a research paper titled "Cross Entropy is All You Need to Invert the Data Generating Process". The paper explores how AI models learn and understand data through shapes and topology. It argues that AI models create a "platonic form" of representation for concepts, which is a simplified and idealized version of the real-world concept. The video uses the example of the book "Moby Dick" to i...
Colab Notebook: colab.research.google.com/drive/1mm7TDgg8lqjUn7tNZ6rBEWRdcikmAloR?usp=sharing
The video discusses a research paper titled "Cross Entropy is All You Need to Invert the Data Generating Process". The paper explores how AI models learn and understand data through shapes and topology. It argues that AI models create a "platonic form" of representation for concepts, which is a simplified and idealized version of the real-world concept. The video uses the example of the book "Moby Dick" to i...
Просмотров: 36
Видео
SmolAgents Overview and Tutorial
Просмотров 4954 часа назад
Link to Colab Notebook: colab.research.google.com/drive/1sfFObHxhYqRXONpFnGG2szHegBx-5MiO?usp=sharing This video is about the small agents library, a new and relatively undocumented tool. The video starts by introducing the small agents library and its purpose. It explains that the library is designed to be simple and easy to use, allowing developers to quickly define and deploy AI agents. It a...
Training Neural Networks On Ghost Memories: Better Than RAG Tuning
Просмотров 58912 часов назад
Link to Colab Notebook: colab.research.google.com/drive/12HD_OY_XIRPSEhMCw9iaoFAOiPgyOty4?usp=sharing Link to Github Repository: github.com/RichardAragon/GhostMemory The video is about a phenomenon called "ghost memory" in neural networks, which is a trace of a memory that exists even after the network's weights have been reset. The speaker, Richard Aragon, starts off by discussing the concept ...
Encoding Data To The Intersection of Lines: Not Clickbait Showcasing The Actual Process
Просмотров 48014 часов назад
Link to Colab Notebook: colab.research.google.com/drive/13XzQSmxTA4WSLM2jTveEOAWEAqyokrGl?usp=sharing This video demonstrates a novel method of encoding data within the intersections of lines in Euclidean space. The creator begins by explaining the theoretical foundation of this concept, drawing on computational theory and the idea of the universe as a computational system. They then delve into...
A Neural Network Based On 98% Chaos Theory: A Beautiful Mind
Просмотров 19416 часов назад
Link to Colab Notebook: colab.research.google.com/drive/1RcKhesZtoBjC5196v7Asqu719SvU0KXX?usp=sharing This video is about a neural network built on 98% chaos theory. The creator of the video, Richard Aragon, explains that neural networks always involve a linear process and a predictive process. He decided to take the opposite approach and see what happens if he creates a model that is 98% proba...
Training A Self Supervised Convolutional Neural Network (CNN)
Просмотров 25719 часов назад
Link to Colab Notebook: colab.research.google.com/drive/1PkUlUqHlsRiA0DCJ3ZFIxEqIInH0MKuX?usp=sharing This video is about training a self-supervised convolutional neural network (CNN). The speaker, Richard Aragon, starts by showcasing the results of a CNN that he trained for 2 hours on the MNIST fashion dataset. The model is able to reconstruct images with only minor errors. Richard then discus...
I Truly Think The Universe Is Hyperbolic And Not Euclidean
Просмотров 342День назад
The math checks out, you can double check it yourself: colab.research.google.com/drive/1t_ZUB1r2h5Q0ToU1S9X3Ca7gHjFt3L08?usp=sharing Richard Aragon starts the video by going back in time to 1595 when Galileo was working and experimenting with Euclidean math to make calculations. Through his calculations and geometry, Galileo discovered that it was mathematically impossible for the Earth to be t...
You Can Control How Well AI Models Learn By Controlling Physics
Просмотров 567День назад
Link to Arxiv Research Paper: arxiv.org/abs/2411.00660 Link to Colab Notebook: colab.research.google.com/drive/14PB_g-yiJXYia4aIXYxNXw03W6Xmh3DP?usp=sharing
Experimenting With LCM Models (Meta's Alternative To LLM Models)
Просмотров 6 тыс.14 дней назад
Link to Colab Notebook: colab.research.google.com/drive/1BxPSbAzp724qaeNducQ_tgWUasH7yBip?usp=sharing In this video, Richard Aragon explores Large Concept Models (LCMs), a new architecture developed by Meta. He explains that LCMs are not Transformer models, but rather utilize a small portion of Transformers. Aragon discusses the architecture of LCMs, highlighting their unique approach to combin...
The First Map Of Digital Space
Просмотров 9421 день назад
Link to Google Doc: docs.google.com/document/d/14yRABlAbsXXVqm_iuZ5a3IF9O6WCc3EGsZXfatyK9bM/edit?usp=sharing This video is about the first map of digital space. The speaker talks about the different types of digital spaces and how they can be used to model real-world phenomena. The speaker also discusses the key relationships that they have discovered within digital space. The speaker starts by...
How To Conduct Science Experiments With AI
Просмотров 6721 день назад
The video is about Richard Aragon, the creator of the channel, who talks about his philosophy and approach for utilizing AI models in scientific experiments. Richard is a strong believer in AI as a knowledge tool, and he emphasizes the importance of understanding how these models work to effectively conduct experiments. Richard also stresses the importance of having base knowledge in the scient...
The Geometric Perpetual Energy Machine: Unlimited Energy
Просмотров 10521 день назад
Link to Github Repository: github.com/RichardAragon/Digital-Perpetual-Energy-Machine Richard Aragon demonstrates a digital perpetual energy machine, a closed-loop system that generates energy over time. The machine is built in a 3D hyperbolic space and consists of three main components: - A sphere, the base element, containing agents with random energy levels. - A pyramid that compresses the ag...
The Physics of Different Geometric Dimensions
Просмотров 6621 день назад
This video is a part of my 100% free course on Udemy, 'AI Geometry: Understanding How Shape Impacts AI Learning' Check out the full course here, it's free: www.udemy.com/course/ai-geometry-understanding-how-shape-impacts-ai-learning
Why Did I Decide To Get Into AI Research?
Просмотров 13521 день назад
I woke up feeling introspective and decided to create a video about my journey into AI research. I've been seeing a lot of videos from people in the tech industry who are feeling chaotic about the current state of affairs, but I've been expecting this for a while now. About a year ago, I decided to go all-in on AI research and diversify my income. I've been making videos about this for two year...
Real World Use Cases Of AI: Building A Stock Trader Optimization App
Просмотров 11828 дней назад
Link to Colab Notebook: colab.research.google.com/drive/1tJ9PrLcGu4-5ZyNQnYkjOfTZbEkBGVGe?usp=sharing Richard Aragon explores the real-world applications of AI by delving into the scientific concept of redox reactions - chemical reactions involving electron transfer. The video outlines the concept of redox reactions, demonstrates how AI can be used to build a simplified model of this process, a...
Natural Language Reinforcement Learning Implementation and Walkthrough
Просмотров 66328 дней назад
Natural Language Reinforcement Learning Implementation and Walkthrough
Hyper Dimensional Computing (HDC) Decoders vs Transformers Decoders
Просмотров 15028 дней назад
Hyper Dimensional Computing (HDC) Decoders vs Transformers Decoders
Hyper Dimensional Computing (HDC) Encoders vs Transformers Encoders
Просмотров 19528 дней назад
Hyper Dimensional Computing (HDC) Encoders vs Transformers Encoders
I Ran An Experiment That Proves LLM Models Can Reason Beyond Their Training Data
Просмотров 550Месяц назад
I Ran An Experiment That Proves LLM Models Can Reason Beyond Their Training Data
Super Basic Intro To Hyperdimensional Computing
Просмотров 143Месяц назад
Super Basic Intro To Hyperdimensional Computing
I Solved Quantum Wave Collapse And Can Prove Why The Observer Effect Happens
Просмотров 167Месяц назад
I Solved Quantum Wave Collapse And Can Prove Why The Observer Effect Happens
Formal Presentation of the Argument: The Laws of Physics Must Arise External of Euclidean Space
Просмотров 40Месяц назад
Formal Presentation of the Argument: The Laws of Physics Must Arise External of Euclidean Space
The Geometry Awareness Function For Neural Networks
Просмотров 104Месяц назад
The Geometry Awareness Function For Neural Networks
I Created An Auto Trainer To Teach LLM Models To Play Chess
Просмотров 112Месяц назад
I Created An Auto Trainer To Teach LLM Models To Play Chess
A Straight Line Is Not Actually Straight In Curved Space: This Changes The Universe For AI
Просмотров 198Месяц назад
A Straight Line Is Not Actually Straight In Curved Space: This Changes The Universe For AI
Euclidean Vs Hyperbolic Spaces For Businesses
Просмотров 364Месяц назад
Euclidean Vs Hyperbolic Spaces For Businesses
How LLM Models Create Shapes (LLM Models And Graph Theory)
Просмотров 177Месяц назад
How LLM Models Create Shapes (LLM Models And Graph Theory)
Philosophically Understanding How Reinforcement Learning Works For Business
Просмотров 153Месяц назад
Philosophically Understanding How Reinforcement Learning Works For Business
love the idea of noisy representation
Just recently found your channel and am enjoying the content. This video especially resonates. Since learning about the concept of embeddings, I have wondered what a visual space using the embeddings of our greatest writers would look like. Shakespeare's writings, for example, had beauty in both meaning and structure, it would be interesting to see if there is structure or beauty in an embedded space. Thanks for the great content.
Someone actually explaining the research paper code ❤
my brother from another mother. I agree completely with literally everything. And yes, fuck OpenAI.
Can LCM be combined with the Byte Latent Transformer and the Continuous Train of Thought?
uv > pip : pip install uv ; uv pip install [modules]
Thank you i have reviewed and subbed great content !!!!
I wouldn't per se it is quantum probabilities. There is some wild stuff, going on. But let's not confuse people. Good video otherwise, though. the concept space utilizing sentence atoms , is a awesome. Note, they may not be full sentences think more of an atomized sentence, without the further reduction into tokenezation. this is really cook actually, and surprised it took so long to even conisder it. the diffuser usage is kind of interesting also. Currently I'm working on a cross attention implementation to map visual atoms, (using florence2 segmentaiton) of conceptual vision. Now what's cool is the diffuser usage allows for some hackery that can be forward or reverse through diffusion for both generative and visual encoding. I've had some success with a few attempts, but it's kind of beefy...too beefy, so I am going to have to shave it off a bit and try to get it to use a lot less vram.
I really enjoyed this video and your explanations within. Subscribed and will be checking out your other content Keep it up !!
Wait what
The idea of encoding data with line intersections is interesting, however after looking at the code of the linked colab I don't think the code does what you think it does. To keep things short I'll only refer to the last iteration, the EuclideanBraidEncoder. The biggest problem is that the message is never recovered from the crossings so there is no demonstration, that the encoding actually works. Whenever the message is printed it refers to the original message variable. The next issue is, that the encode_message function just takes each crossing and assigns it a bit of the message as a label. The crossing is not altered in any way, so all the information is in the label. It is not possible to get the message back from the crossings. It is possible to get the message from the labels, but there is no need for any lines or crossings at all. A bit array of the original message would do the same. The next issue is with the find_all_crossings function. It does not find the actual intersections of lines, but just points along the lines that are closer than a threshold. when using a different resolution and/or threshold this will give completely different crossings. It may also find more than one crossing per pair of lines which should be impossible in euclidean geometry. This is in fact what happens. Each line can intersect every other line at most once. For 7 lines that is (7*(7-1))/2 = 21 crossings at most so 88 is way too many. It is however much worse, because every one of the lines goes from (-x, -y) to (x, y) for different values for x and y. It is easy to see, that the midpoint between those two points is (0,0) for all lines. That means that all seven lines actually only intersect in a single point: the origin. It would be interesting to see what a working version of this idea might actually look like
Following closely
What’s the accuracy metric measuring in this case? Whether the model can successfully predict the next (target) embedding (implying the output of interest is an entire embedding vector)? And are these numbers with respect to the training or test set?
All of these wonderful questions are exactly why I provide you the code! Do you have any questions about the code?
Thanks, it actually helped me with my engeneering progect.
Well don't leave us hanging, share the deets?
I dont understand anything, but it seem wonderful work
Very Interesting! I wonder what would happen if we just deactivate 50% of the neurons during inference. And see how much of model's performance degrades.
maybe the linear layers need to be trained to "aim" or "shape" the inputs to the fractal intuition space, and similarly to collect the outputs?
I will experiment with this! There are a lot of instances where I intuitively think training like this would not work, then it works. I think this could work lol.
@richardaragon8471 🍻 challenging our intuition is both a useful and unintuitive tool!
what the fk are u yapping about
Would this architecture still support GPU splitting? In other words would a 70b model fit onto two cards with 24 gb each If total size of model was 30-40 gb after quantitization … and could you even quantize it?
I added one more cell to the code for you. Yes, you could do this in super cool ways actually. You could put the hidden dimensions on a separate GPU than the rest of the layers for example. It is all reliant on PyTorch. Thanks for inspiring me to think about this! It hadn't crossed my mind.
@@richardaragon8471 This is a huge weakness for image generation tools. They typically don't support multiple cards, but most current LLMs do support splitting; so I was worried since this sounds like it has a similar design to image generation. Glad you found a way to do it! You're awesome. Can't wait to see this model grow in popularity.
Nice vid
dude saying it passionately but i failed to understand... i wish i have bigger brain
to get a similar graph effect: what if every weight was embedded in a N-d (ex: 2 or 3) graph layout space (entirely separate from the weight's value). then use a graph layout algorithm like force directed or UMAP and compute the edges and clusters according to N-nearest neighbors. ?
My hypothesis is more and more that there is never a single way to accomplish anything. Any concept imaginable, the universe lays out a multitude of ways to accomplish functionally the same thing. Some ways are more optimal than others in different conditions. Your method seems like another optimal way to achieve this result. I love it, thank you for sharing!
That light mode made me blind at middle of the night : (
I don't get how to make it conversational. I send it "hello" and it continues with some gibberish about being a new user to some tool. It seems more like an autocomplete than some assistant. Do I have the wrong expectancies here? Ultimately I was interested in learning to talk to an AI who could call my functions.
To be honest this sounds like what I was trying to approach in terms of encrypting interactions with LLM's like ChatGPT-- I was trying to figure out how to steal the processing of their "logic clouds" without collapsing it into definite outputs-- Basically means I turn everything I am talking about into a shape, I tell chatgpt what to do with that shape, it processes the change in the shape, and that shape is returned, and then the key points of the shape is turned into symbols which then collapses into logic and reason at the level of words-- The whole idea here was to keep the central processing completely unaware of what I was doing, while giving it the fundamentals to process-- However, I was thinking about how to do this purely from a client side; where as this is doing the same thing from an architectural side-- Do I sound entirely off base?
You sound entirely on base. I can put you in touch with a team of people literally building exactly this. That's incredible lol.
aaaa so excited to start doing this myself. you inspire everyday!
Good job you proved something that was proved in 1919 when a solar eclipse was observed and light was bent by the suns gravity on its way to earth. Einstein predicted it was the case, anything consisting of matter distorts the space and time around it. But anti matter not being fully understood that could be a different story but it is also believed it has the same effect on space time if not an even greater effect, resulting in the un even pattern you see in our universe. Its believed to be more antimatter than matter.
Unfortunately, people are generally no better than monkeys in every way. Can you show me a physics book that shows my observations are already documented physics? You know enough about physics to know why your response is trolling, pedantic, and literally just here to incite violence. Why?
Good Work
Man, love your content. Would be very exited to work with you/for you. I have intermediate level skills with building LLM agents, full stack development, and deep learning models development. Don't have to pay me, I fact I will pay you to work under you.
I am happy to teach you anything you would like to learn! If you're serious, the best email address to get ahold of me is richardaragon186atgmail.com. We could for sure build some things together and I am more than happy to teach.
Interesting. There is only one reason to postulate dark energy, to explain the exponential expansion rate of the universe. The expansion was discovered in 1929. In 1998 it was discovered that the expansion is accelerating, this is when the existence of galaxies was confirmed. Electricity is drawn towards potential and the universe as a whole behaves the same way. Electricity comes into our homes because the neutral circuit provides the potential. Electricity is drawn towards grounding rods for the same reason. If something accelerates at a constant rate it will get faster and faster. If a spaceship travels at a constant 1g acceleration rate it would achieve about 95% light speed in 1 year. Physicists in the last century did not postulate dark energy because they understood that the expansion is a fundamental property of the universe. The fact that the expansion is accelerating does not invalidate their reasoning, it's what the known laws of physics would predict provided gravitational forces are not strong enough to counteract the process. All studies to find dark energy have been fruitless because it doesn't exist. To say there is dark energy is to say there is 5 fundamental forces, there is 4. All the great physicists of the last century did not postulate dark energy for these reasons.
ok i have never read into this much so i cant really say anything about this, im commenting anyway so algorithm sends you some more people who might be able to talk and think about this more
You know a lot about many different fields. What is your educational background? I would guess Comp Sci or theoretical physics?
Subscribed. Exactly the type of material I love.
It's rather fitting to say LCMs are an upgrade to LLMs. Meta has been killing it lately with its Byte Latent Transformer and Continuous Chain of Thought paper. Btw, great explanation. Learning a lot from you! Much love from India :)
What would be the accuracy with best performing/architecture (Falcon3 or Lllama3)?
you are so cool! i wanna learn more about everything. can i be an apprentice?
i can help w whatever you’d want me to help with. i learn fast, am good w research, design, sales and marketing. i’d love to help as much as i can if you teach me.
daddy's about to come in from creative mode to get in on this breakthrough 🤩
I don’t know what you are talking about but I am going to subscribe anyway.
you really have to explain things in detail. i didn't understand most of what you said.
You're expecting one source to bring everyone that visits it to the same level of understanding? Feels like a very passive way to approach learning imo.
Great Video, really one point. Thanks for sharing your code, very interesting
Hey Thanks a lot! I have a question, Let say i want to prepare embeddings specifically for Financial statements(10K), I can simply provide 10 bunch of FS and then index the data? will that work?
Yup, that would work and that's exactly the theory behind Contextual Document Embeddings! Indexing the data properly is a bit tedious but it is the only actual hard step.
Great session, keep up the good work ❤
Thank you. This is the first time I've seen this channel, so I just subscribed. Thanks for the detailed explanation!
Keep on the good work
I just finished having a discussion with Grok about the differences between Latent Consistency Models (LCMs) and Large Concept Models (LCMs) as well as what would be involved in combining the two architectures.
Love it thanks
Wow
Did you try 1bitLLM/bitnet_b1_58-3B huggingface or 1bitLLM/bitnet_b1_58-large?
As I'm sure you realize (especially since you did bring at least one of the laws of thermodynamics), building a physical version would require discarding some well-established laws of physics, and replacing them with new ones - which is a tough sell. If you really believe that you may have somehow uncovered a perpetual energy machine, and you think that your theoretical framework could be experimentally verified in the real world, the absolute first thing you should be doing is assuming that you probably made an error somewhere, and then either publish to something like Arxiv, and then be asking those in the field to falsify your claims. Unless I'm incorrect about what I think you're implying - that whatever you seem to have uncovered in the digital/mathematical world could be built/manifested in the physical world - then the scientific method would dictate that you provide extraordinary (and in the case of unlimited energy - overwhelming) evidence before you even seriously consider that what you have here is something. The epistemologically diligent thing to do here is to assume that you are mistaken. And if the laws of thermodynamics don't deter you from exploring this path further on their own, then you would need to be rigorous about trying to figure out why/where you're wrong. Only if/when you and others have exhausted reasonable efforts to falsify your claim would it make sense to even consider that you may have something here. I think you're doing some excellent work, and I very much support and commend you on your motivations and aspirations, and your ethical reasons for doing your work. I also want to be transparent and say that I am not, and can not falsify what I'm assuming are your implications here - I do not have the advanced enough mathematics or physics education to do that. I do, however, have a solid understanding of the scientific method and its applications. I'm communicating my $0.02 to you because I think I would want that feedback if things were reversed, and even though I don't know you personally or beyond the few videos that I've watched here, I care, and I respect what you're doing in general, so I hope this feedback doesn't come across as shootdown in any way. Keep doing the amazing work and sharing your ideas my dude - and thanks for all you do!
Perpetual motion existing in the physical world would in fact 100% break the laws of relativistic physics. You are not wrong and the math in that is not wrong! There exists two, as yet not unified frameworks to physics though. If perpetual motion were to exist in the physical world, I would imagine it would exist in the same way that I am able to experiment with it in the digital realm. I rely on quantum mechanics to do it in the digital realm, not relativistic physics. To do this in the physical realm would involve an understanding of being able to control dimensions. Therefore, it is only physically impossible as long as you do not know how to control dimensions. Sadly, I do not know how to control dimensions lol.
Thanks for this we should really try building a concept embedding using high dimensional vectors
I am working on exactly that!