- Видео 123
- Просмотров 690 649
Learn Data with Mark
Великобритания
Добавлен 30 май 2022
Weekly 5-7 minute videos on data and LLMs.
Agentic Analytics with PhiData and DuckDB
Dive into the world of agentic analytics with the Phidata Agent library and DuckDB! In this video, we'll build a Streamlit app designed for a tennis commentator, showcasing how to leverage Phidata tools for data analysis and uncover meaningful insights. Whether you're new to Phidata or experienced in data engineering, this walkthrough offers practical examples to enhance your skills.
Resources:
🟡 Code - github.com/mneedham/LearnDataWithMark/tree/main/phidata-agentic-analytics
🟡 DuckDB tool - docs.phidata.com/tools/duckdb
🟡 Streamlit - docs.phidata.com/tools/duckdb
Timestamps:
0:00 - Introduction to ad hoc analysis with Phidata and DuckDB.
0:07 - Exploring the concept of agentic analytics.
0:20 - ...
Resources:
🟡 Code - github.com/mneedham/LearnDataWithMark/tree/main/phidata-agentic-analytics
🟡 DuckDB tool - docs.phidata.com/tools/duckdb
🟡 Streamlit - docs.phidata.com/tools/duckdb
Timestamps:
0:00 - Introduction to ad hoc analysis with Phidata and DuckDB.
0:07 - Exploring the concept of agentic analytics.
0:20 - ...
Просмотров: 1 206
Видео
Building an AI agent with PhiData and Streamlit
Просмотров 1,6 тыс.14 дней назад
In this video, we'll learn how to create a sports research agent using PhiData with Streamlit acting as our front-end. 🟡 Streamlit - streamlit.io 🟡 PhiData - github.com/phidatahq/phidata 🟡 Code - github.com/mneedham/LearnDataWithMark/tree/main/phidata-streamlit
How to add memory to a PhiData agent
Просмотров 59821 день назад
Learn how to enhance your PhiData agent with memory capabilities! In this tutorial, we'll explore how to implement both chat history memory and user memories to make your agent more efficient and personalized. 🛠️ Resources 🟡 Memory docs - docs.phidata.com/agents/memory 🟡 Code - github.com/mneedham/LearnDataWithMark/blob/main/phidata-memory/code.ipynb ⌚ Timestamps: 0:00 - Introduction 0:05 - Qui...
An intro to the PhiData agent library
Просмотров 2,8 тыс.Месяц назад
In this video, we dive into PhiData, an exciting agent library for AI development. Join us as we explore its features and capabilities, demonstrating how it can be used with different models and tools. Learn how PhiData integrates various AI models, tools, and knowledge bases to create powerful and flexible agents. Whether you're using OpenAI's GPT-4 or Ollama's models, PhiData offers a compreh...
Do LLMs understand markdown tables?
Просмотров 739Месяц назад
In this video, I explore using the Docling library to convert Wikipedia pages into Markdown format, focusing on Novak Djokovic's career statistics page. I demonstrate how Large Language Models (LLMs) can better understand and interpret complex tables when presented in Markdown format. Key points covered: Introduction to Dockling for converting web pages to Markdown Extracting and parsing Markdo...
Intro to burr: A State Machine for LLM apps
Просмотров 1,1 тыс.2 месяца назад
This video is an introduction to burr, a state machine for your LLM applications. We're gonna use to help us come up with titles for my last RUclips videos. We'll learn how state works, how to create transitions, and how to put a human in the loop. 🟡 burr - burr.dagworks.io/getting_started/simple-example/# 🟡 Code: Part 1 - github.com/mneedham/LearnDataWithMark/blob/main/burr-playground/app_part...
Llama 3.2-vision: The best open vision model?
Просмотров 7 тыс.2 месяца назад
Llama 3.2-vision is a multimodal large language model by Meta, and it's finally available on Ollama! We're going to see how well it performs on a few different tasks: extracting code from an image, reading my terrible handwriting, and critiquing one of my RUclips thumbnails. 🟡 Model on Ollama - ollama.com/library/llama3.2-vision
Moonshine: Real-Time Speech-To-Text on your laptop
Просмотров 1,4 тыс.2 месяца назад
In this video, we're going to learn about Moonshine, a series of models optimized for real-time ASR (Automatic Speech Recognition). We learn why it's better than OpenAI's Whisper models if low latency if what we need, and then I read a TechCrunch article and we see how well it works 🟡 TechCrunch article - techcrunch.com/2024/11/06/the-other-election-night-winner-perplexity/ 🟡 moonshine - github...
NuExtract: An LLM that extracts information
Просмотров 1,7 тыс.2 месяца назад
In this video, we're going to learn about NuExtract, an LLM designed for structured extraction of information from documents. It's fine tuned on top of Phi3.5 and we're going to use it via Ollama. Timestamps 00:00 - Intro to NuExtract 00:58 - Functions to call NuExtract with Ollama 01:40 - The Hello World example 02:08 - Social media post 02:39 - Weather information 03:13 - Chocolate cake recip...
Using LLMs on the command line
Просмотров 1,7 тыс.3 месяца назад
In this video, we're going to learn how to use Simon Willison's library, which lets us run almost any LLM model from the command line. OpenAI, Claude, Grok, Ollama, they're all covered! We'll see how to run basic commands, how to have the LLM summarise the output of the tree command, and finally I'll share some shell scripts that use the tool. 🟡 llm documentation - llm.datasette.io/en/stable/us...
Ollama: Running Hugging Face GGUF models just got easier!
Просмотров 4 тыс.3 месяца назад
In this video, we're going to learn the new and improved way to running Hugging Face GGUF models on Ollama. 🟡 Hugging Face/Ollama docs - huggingface.co/docs/hub/en/ollama 🟡 Ministral model - huggingface.co/bartowski/Ministral-8B-Instruct-2410-HF-GGUF-TEST
The fastest way to run OpenAI Whisper Turbo on a Mac
Просмотров 4,3 тыс.3 месяца назад
mlx-whisper is the fastest way to do automatic speech recognition on a Mac with OpenAI's Whisper models. In this video, we'll learn how to use it to transcribe an episode of the Tech Meme Ride Home podcast. We'll then see how well an LLM can answer questions about the episode based on the transcript and we'll see how fast mlx-whisper is compared to insanely-fast-whisper. #openai #asr #mlxwhispe...
Ollama: How to send multiple prompts to vision models
Просмотров 1,4 тыс.3 месяца назад
In this video we're going to learn how to have a conversation or send multiple prompts to vision models when using Ollama. 🟡 Code - github.com/mneedham/LearnDataWithMark/blob/main/ollama-images-multiprompt/multi-prompt.ipynb 🟡 Ollama - ollama.com/
Running OpenAI Whisper Turbo on a Mac
Просмотров 4,4 тыс.3 месяца назад
In this video, we're going to learn how to run OpenAI's new Whisper Turbo model on a Mac, which is used for transcribing audio to text. We'll try out the openai-whisper CLI, before moving onto insanely-fast-whisper. We'll also use Ollama and the Llama3.1 8B model to compare the transcripts produced by Turbo and Large-V3. 🟡 Blog - www.markhneedham.com/blog/2024/10/02/insanely-fast-whisper-runnin...
An intro to rerankers: A uniform API for reranking models
Просмотров 2,3 тыс.3 месяца назад
In this video, we'll learn how to use rerankers, a new library by Jeremy Howard and his team at Answer AI. rerankers provides a uniform API over various re-ranking models, making it super easy to switch/change/test the different models. Re-ranking is a technique where you re-sort a set of documents against a provided query to make sure that the most relevant documents are at the top of the list...
DuckDB dynamic column selection gets even better
Просмотров 1,1 тыс.4 месяца назад
DuckDB dynamic column selection gets even better
Ollama and LanceDB: The best combination for Local RAG?
Просмотров 6 тыс.4 месяца назад
Ollama and LanceDB: The best combination for Local RAG?
Searching images on my laptop with LanceDB
Просмотров 1,1 тыс.4 месяца назад
Searching images on my laptop with LanceDB
Rewriting RAG Queries with OpenAI Structured Outputs
Просмотров 1,6 тыс.4 месяца назад
Rewriting RAG Queries with OpenAI Structured Outputs
DuckDB function chaining: The simpler SQL you didn't know you needed
Просмотров 1,1 тыс.5 месяцев назад
DuckDB function chaining: The simpler SQL you didn't know you needed
Why OpenAI's new Structured Outputs feature is awesome!
Просмотров 2,7 тыс.5 месяцев назад
Why OpenAI's new Structured Outputs feature is awesome!
How to evaluate retrieval in RAG pipelines
Просмотров 2 тыс.5 месяцев назад
How to evaluate retrieval in RAG pipelines
Hybrid Search for RAG in DuckDB (Reciprocal Rank Fusion)
Просмотров 1,7 тыс.5 месяцев назад
Hybrid Search for RAG in DuckDB (Reciprocal Rank Fusion)
Full-Text Search vs Vector Search (RAG with DuckDB)
Просмотров 3,5 тыс.6 месяцев назад
Full-Text Search vs Vector Search (RAG with DuckDB)
Search-Based RAG with DuckDB and GLiNER
Просмотров 1,8 тыс.6 месяцев назад
Search-Based RAG with DuckDB and GLiNER
A UI to quantize Hugging Face LLMs
Просмотров 1,1 тыс.6 месяцев назад
A UI to quantize Hugging Face LLMs
Mistral 7B Function Calling with llama.cpp
Просмотров 2,6 тыс.7 месяцев назад
Mistral 7B Function Calling with llama.cpp
Does Mistral 7B function calling ACTUALLY work?
Просмотров 1,3 тыс.7 месяцев назад
Does Mistral 7B function calling ACTUALLY work?
Now how to use this streaming data (logs) from kafka with llm ?
What fart with the auto translation
why you bald?
Replied on the other one!
bro, how you lose all your hairs? is there a secret?
I have alopecia - dunno if it's secret, but that's the reason
hi excellent demo, can you also demo the chat history functionality of chainlit without using Literal ai?
arrow was pointing to the right... not left
Just for fun, I guess there were a few print statements in check_connectivity and toggle_wifi("off")
I need something more barney style
What's barney style?!
closed source is so 2020
Do you mean with the model? I tried using some of the Ollama models but the tool calling wasn't working well with them so that's why I used gpt-4o. If you mean the code for the video itself, there's a link in the description.
@@learndatawithmark yes the model, supporting scammers who switch from NGO to for-profit should really be in the title.
he has given u the code just use another model man
Great Tutorial. How does Phidata compare to crewai when building ai agents?
It truly sucks for OCR for LaTeX, it feels embecile to give it instructions and always gets them wrong...Qwen2-VL (72B) Instruct is better, but still far from perfect... sadly
Oh I haven't tried anything LaTeX - are the frontier models able to OCR LaTeX?
This is exactly what I was looking for! Thanks a lot. Could you also provide a Github repository that can be clone to play with it?
Yup. Code is here - github.com/mneedham/LearnDataWithMark/tree/main/phidata-streamlit
GPT-4o recommend this video and it was 100% correct.
hah, did it? Might be the first time an LLM brought somebody here!
i'm getting this error from a llama3.2:1b-Instruct, ollama create -f Modelfile myllama transferring model data 100% converting model Error: json: cannot unmarshal array into Go struct field .model.merges of type string
Thanks for the video! I had success running this with "mlx-community/whisper-tiny" and "mlx-community/whisper-turbo" but anytime I tried any of the other models on the huggingface page I ran into this error: "TypeError: audio_features has an incorrect dtype: mlx.core.float32" Has anyone encountered this issue?
Thanks for the great video. I have Ollama version 0.5.5. So theoretically these parameters should already be set and I also should be able to change them before launching Ollama server. However no matter what I set for these parameters I see no difference in GPU usage. Is there a way to actually check how many models are loaded in GPU?
This is very hard to follow. You don't explain the intent and purpose of what you're doing or show what files you'll be working with, let alone *why* you're doing what you're doing. Only someone who is either very familiar with these scripts and coding for AI or very patient and determined could follow these instructions. Please, go slower and explain more.
Thanks for the feedback. I did another video introducing PhiData over here - ruclips.net/video/V9rB3OkyanA/видео.html - and I wasn't sure whether to explain again some of the stuff I did in that video. Or maybe I should have referenced that video earlier and suggested people look at that one if they hadn't done anything with agents before? Or perhaps I should have done a brief recap of what was covered in that video as a middle ground.
can i know which technique is used for orchestration of the agents : sequential or hierarchical
Hi Mark, 'style_pandas' is missing in repo, could you upload it please? BTW I like your channel - great topics in short videos and on top code examples - the best suits my needs. THANKS!
Thanks - have uploaded that file now!
My favorite youtuber and favorite library, what can go wron here? hahaha
Mark - you are the best.
I just came across your content on RUclips about re-ranking. Great work, and I’ll be following you from now on. Wishing you all the best!
Amazing one!! Could you also share the commands and sample data files. Would be good to try these 😃
So happy to see Phidata here! Thanks a bunch for this! 😊
My favourite agent library!
Wow! Thats amazing! Are there similar performance gains with normal llm models in mlx format as well? And what about stable diffusion models?
what human isn't using windows....
Great 👍 (after playing with transformers and cpu offloading, searching for ideas i found ctransformers and realized that ollama is no magic at all, its just gui 😅)
Can phidata have enough to memory to talk to me months later of chatting and reflect and advise future ideas? Like giving it enough info of my teams progress and what new AI tools I may need to speed up the process to fulfill the deadline? Or maybe personal bot conversations? I figured ai had this already with GPT having so much data to reference but learned that it’s behind babyAGI by a year just now introducing chain of thought and doesn’t have good memory within the chat … is phidata + GPT + maybe lang chain or something (for non coders and coders alike?) have a way to give the AI very long memories? Really need to know lol.
I think you can do it using their memory summarisation feature. I haven't tried it out yet but it's described in the docs - docs.phidata.com/agents/memory
The llm does a correct job in this case, and I'm happy that the model works well with the prompt. You get a different answer when you ask the question differently, because for each question you retrieve different chunks from the vector-store. The chunks for the first question don't have the answer. By default, the similarity_search returns k=4 documents. In your second question, the third chunk shows the answer: "disrupted by rain.The tournament......" This text-chunk is in the top 4 chunks similar to the input question. For the first question this chunk is not chosen, probably there were some other more common words of other chunks similar to the question. Idk yet what is an optimal choice for k, or how many clusters to choose. But you would get a better result if you increase k. Still keep in mind, if you make k too large or the size of the text chunks is large, then the model loses context and "does not remember" everything you pass to it. Or maybe it's a good idea to search differently, and decompose the input question into tokens that we can best align with a retriever.
I think models don't have people face
I would have thought they'd mostly seen famous faces though given they're usually trained on internet datasets which have many famous people!
Phidata vs pydanticAI?
I haven't tried PydanticAI properly yet, but skimming the documentation it seems more verbose than PhiData
Can we use any database other than these? Postgres Sqlite Singlestore DynamoDB JSON YAML And if yes how, I would appreciate any help.
Your Formula for code programming the LLM is shit, I'm a master computer engineer and what you're doing is completes bullshit. let's see if you can handle what im about to say to you. Mathematical Inconsistencies: The provided code snippets and parameter reductions do not add up, indicating a potential fundamental flaw. Deviation from Established Best Practices: The approach contradicts Hugging Face's documented guidelines for parameter optimization, which emphasize balanced trade-offs between complexity and performance. Lack of Empirical Evidence: The presentation fails to provide conclusive experimental data supporting the claimed efficiency gains. The Elephant in the Room: Running LLMs on Laptops without GPUs Warning: Attempting to run Hugging Face LLMs on laptops without dedicated GPUs is ill-advised and likely the root cause of: Code inconsistencies (e.g., errors in the code modifier) System performance issues experienced by those trying to replicate the approach 10 Compelling Reasons to Run LLMs on Computers with Dedicated GPUs: Computational Intensity: LLMs require massive parallel processing, which GPUs are designed to handle. Memory Constraints: Laptops' integrated graphics often lack sufficient VRAM, leading to memory overflow errors. Thermal Throttling: Laptops' cooling systems may struggle to keep up with intense computations, throttling performance. Power Consumption: Running LLMs on laptops can lead to excessive power drain and reduced battery life. Optimization Limitations: Without a GPU, optimizations are severely limited, compromising model performance. Scalability Issues: Laptops lack the scalability to handle large LLMs or multiple model instances. Dependence on CPU: Relying solely on CPUs for computations can lead to significant performance bottlenecks. Increased Risk of Errors: Insufficient computational resources increase the likelihood of errors and inconsistencies. Incompatibility with Future Models: As LLMs evolve, laptops without GPUs will become increasingly incompatible. Wasted Resources: Attempting to run LLMs on under-equipped laptops can result in wasted time, energy, and computational resources. Mathematical Demonstrations: To illustrate the importance of proper computational resources, let's examine two examples: Example 1: Matrix Multiplication (a fundamental LLM operation) CPU (Laptop): O(n^3) time complexity, where n is the matrix dimension GPU (Dedicated): O(n^2) time complexity, leveraging massively parallel processing Example 2: Gradient Descent (a crucial LLM training algorithm) CPU (Laptop): O(m \* n) time complexity, where m is the number of samples and n is the feature dimension GPU (Dedicated): O(m + n) time complexity, exploiting GPU-accelerated linear algebra libraries. Conclusion: Running Hugging Face LLMs on laptops without dedicated GPUs is not recommended. The presented approach is flawed, and the lack of empirical evidence undermines its credibility. For optimal performance, scalability, and reliability, always deploy LLMs on computers equipped with dedicated GPUs. Rating: -5/5 (due to the identified technical inaccuracies and disregard for best practices). In literal terms You're computer skills are garbage and dogshit, How about you go back to a normal desk boy job or pushing mail cart's or whatever it is you do and leave the engineering to me and my team of more advanced engineers then you.
ValueError: could not broadcast input array from shape (384,) into shape (1024,)
Sounds like the embedding model is using a dimension of 384 but the database expects 1024
@@learndatawithmark I got the error here: operation_info = client.upsert( collection_name="podcast", wait=True, points=points ) how can i fix it
Very nice tutorial, thank you for curating and creating these content mark,
Phidata is by far, the easiest way to create agents, thanks for sharing this amazing tool
I’ve tried setting up everything correctly but still can’t achieve parallel responses. I’m testing this by hitting the same cURL request for the same model from two different places (Postman or directly via terminal). Here are the steps I followed: Stopped the service: systemctl stop ollama.service Set the environment variables: export OLLAMA_NUM_PARALLEL=4 export OLLAMA_MAX_LOADED_MODELS=4 export OLLAMA_MAX_QUEUE=2048 Verified the variables: env | grep OLLAMA OLLAMA_MAX_LOADED_MODELS=4 OLLAMA_MAX_QUEUE=2048 OLLAMA_HOST=My-localhost:11434 OLLAMA_NUM_PARALLEL=4 Started the server: ollama serve Despite this, I’m still unable to achieve parallel responses. Could you please help me identify what might be going wrong?
Thanks a ton man... outstanding video and tutorials
Great video! I'm also exploring multi agentic frameworks, on my top list there are ag2ai and PydanticAI, which IMO have great motivations and potential to grow. BTW, the LanceDB part of your git code requires OPENAI_API_KEY, I don't have it so I skipped this part.
Would it be possible (In SQL) to have a function that combines GROUP BY and PIVOT into a single CROSSTAB() function?
How can we setup a local RAG which can redact PII information when extracting the data from the PD before creating the chunk, enrich the chunks by using a LLM to create a title, summary, keywords, embeddings and then use PhiData to chat with the document?
Yes please, need at least 10+ videos series for phidata, so much to explore. Also swarms (not openai) have great multi agents architecture. For web3, I recommend eliza by ai16z.
Nice video, thanks! Have you tried swarm (OpenAI) or pydantic agent before? if yes which one do you prefer compared to PhiData? I am looking for a library to easily create agents, but that it's simple to use, light weight and that is not so cluttered and doesn't obfuscate or over abstract everything like CrewAI and Langchain do.
Nice to see the LLM lord is turning into the agentic side! Good stuff!
Finally took the advice of the Ice Lord
Absolutely fantastic for people who use mac!
Nice
Damn you computers fast mine took 5 minutes to respond at 0.07 tokens/s
why you look like mahatma gandhi?
Just found your site. I’m learning so much. You do a great job explaining things. Thanks for all you do to help people.
I really love to see your way to explain this, Thanks a lot!