Really glad I stumbled upon your videos. Super high quality, informative, and helpful. I'm sure your subscriber count will be ballooning in the coming weeks and months. Great stuff!
The most powerful, and by far the best of all these, today are Alpaca 7B and Vicuna 13B. Yes, these days innovation runs fast and frenetic in this sector.
Alpaca 30B - 4bit quantized: > What is the opposite of up? Down. > Answer the following question by reasoning step by step. The cafeteria had 23 apples. If they used 20 for lunch, and bought 6 more, how many apple do they have? They now have 18 apples (the original 23 minus the 5 that were eaten at lunch). > Can Geoffrey Hinton have a conversation with George Washington? Give the rationale before answering. No, because they are not in the same time period and thus cannot communicate directly. However, if we had access to some sort of portal or device that could allow them both to be present at once then it would theoretically be possible for Geoffrey Hinton to have a conversation with George Washington. The others were'nt good enough.
Hi Sam, Thanks so much for the video. It is quite cool and funny as well. You mention a paper in there, is it a research paper? Could we get ahold of it? Thanks so much again and all the best!
Yeah all papers should be on Arxiv if not in the description of the video. That video is from quite a while back, so not sure which paper you are talking about, but take a look on Arxiv.
The creative task should have been done with much higher Temp than 0.1 - I think. I see it's not easy to setup params by task (compare), but I might try this with higher Temps later.
Sam thank you these LLM reviews could become Yelp for AI 😆 I have a question can you post a video about computational LLM? For example to provide an actual table of data and then have a conversation or derive insights w the LLM. Langchain have a csv agent but do others?
The CSV idea is certainly interesting, I am planning to show fine tuning a custom model for natural language to SQL. Maybe I will try to incorporate the csv or tables analysis in there.
@@samwitteveenai this GPT plugin announcement could make so many computational use cases possible. There are langchain text to sql agents but csv is such a simple use case. Imagine uploading an entire budget integrated with a custom LLM corpus. Or an ML plugin with label predictions on-the-fly.
I do think there is a lot of discreet applications currently , but I am also very interested in neural plugins eg separate models within a plugin for ChatGPT etc
Very nice information. Giving a LLM a kind of memory is a little bit challenging. Currently I´m trying to fine tune LLaMA for text-to-sql. Then it could access the DB for previous information. The DB schema is not so easy to design, maybe some (artificial) intelligence could help :D
@@samwitteveenai Thanks for asking. No, I generate the dataset by myself. To be honest, I let GPT4 do all the heavy lifting. It is also more a spare time project and maybe a simple and naive approach. To give you an example, using your example sentence, a little bit extended: 1. Let the AI create a semantic interpretation of sentence Q: Find the subject an it´s attributes in the following sentence and classify the attributes. The sentence is "Hi, I´m Sam and 42 years old". A: "Sam" is an attribute of the subject that specifies the name of the person being referred to, and "42 years old" is another attribute that specifies the age of the person. 2. Let the AI create a class object, e.g. in C# Q: Could you give me a c# class that models the subject. A: public class Person { public string Name { get; set; } public int Age { get; set; } } 3. Let the AI create the SQL to create the table and store the concrete object CREATE TABLE Persons ( Id INT IDENTITY(1,1) PRIMARY KEY, Name NVARCHAR(50) NOT NULL, Age INT NOT NULL ); INSERT INTO Persons (Name, Age) VALUES ('Sam', 42); The training data point consists then of: {'instruction': 'Convert text to sql', 'input': 'Hi, I´m Sam and 42 years old', 'output': 'CREATE TABLE Persons ( Id INT IDENTITY(1,1) PRIMARY KEY, Name NVARCHAR(50) NOT NULL, Age INT NOT NULL ); INSERT INTO Persons (Name, Age) VALUES ('Sam', 42);' Not sure if and how it will work. There is maybe more context in the instruction and/or input necessary.
Really glad I stumbled upon your videos. Super high quality, informative, and helpful. I'm sure your subscriber count will be ballooning in the coming weeks and months. Great stuff!
At the moment because I am a beginner I focus only on the GPT model, but maybe in the future, I will use others. Thank you for the video.
The most powerful, and by far the best of all these, today are Alpaca 7B and Vicuna 13B. Yes, these days innovation runs fast and frenetic in this sector.
Checkout Koala 13B it s like Vicuna but trained on more data.
@@samwitteveenai Thank you very much!
Alpaca 30B - 4bit quantized:
> What is the opposite of up?
Down.
> Answer the following question by reasoning step by step. The cafeteria had 23 apples. If they used 20 for lunch, and bought 6 more, how many apple do they have?
They now have 18 apples (the original 23 minus the 5 that were eaten at lunch).
> Can Geoffrey Hinton have a conversation with George Washington? Give the rationale before answering.
No, because they are not in the same time period and thus cannot communicate directly. However, if we had access to some sort of portal or device that could allow them both to be present at once then it would theoretically be possible for Geoffrey Hinton to have a conversation with George Washington.
The others were'nt good enough.
I haven't tried the 4bit version but the 30B model should have better results than that if finetuned properly.
What are my options If i want to compare really large LLMs?
Thank you for your wonderful video 👌how can I compare the performance of different finetuned LLMs based on my dataset?
Hi Sam,
Thanks so much for the video. It is quite cool and funny as well. You mention a paper in there, is it a research paper? Could we get ahold of it?
Thanks so much again and all the best!
Yeah all papers should be on Arxiv if not in the description of the video. That video is from quite a while back, so not sure which paper you are talking about, but take a look on Arxiv.
Great video, do you think we can do the same evaluation for qa retrieval from a doc?
yeah you should be able to, but it might have to be a bit more manual about how you do it.
@@samwitteveenai , I have been working on one, got mpnet embedding and flan as retrieval to work, would be great to collaborate with you on it!
It will be great if you compared the alpaca one here as well
The creative task should have been done with much higher Temp than 0.1 - I think.
I see it's not easy to setup params by task (compare), but I might try this with higher Temps later.
yes totally the more randomness (creativity) you want in the higher the temp.
Sam thank you these LLM reviews could become Yelp for AI 😆 I have a question can you post a video about computational LLM? For example to provide an actual table of data and then have a conversation or derive insights w the LLM. Langchain have a csv agent but do others?
The CSV idea is certainly interesting, I am planning to show fine tuning a custom model for natural language to SQL. Maybe I will try to incorporate the csv or tables analysis in there.
@@samwitteveenai this GPT plugin announcement could make so many computational use cases possible. There are langchain text to sql agents but csv is such a simple use case. Imagine uploading an entire budget integrated with a custom LLM corpus. Or an ML plugin with label predictions on-the-fly.
I do think there is a lot of discreet applications currently , but I am also very interested in neural plugins eg separate models within a plugin for ChatGPT etc
What a funny video you made! I love it.
Thanks! 😃
Very nice information. Giving a LLM a kind of memory is a little bit challenging. Currently I´m trying to fine tune LLaMA for text-to-sql. Then it could access the DB for previous information. The DB schema is not so easy to design, maybe some (artificial) intelligence could help :D
Are you using a public dataset?
@@samwitteveenai Thanks for asking. No, I generate the dataset by myself. To be honest, I let GPT4 do all the heavy lifting. It is also more a spare time project and maybe a simple and naive approach. To give you an example, using your example sentence, a little bit extended:
1. Let the AI create a semantic interpretation of sentence
Q: Find the subject an it´s attributes in the following sentence and classify the attributes. The sentence is "Hi, I´m Sam and 42 years old".
A: "Sam" is an attribute of the subject that specifies the name of the person being referred to, and "42 years old" is another attribute that specifies the age of the person.
2. Let the AI create a class object, e.g. in C#
Q: Could you give me a c# class that models the subject.
A: public class Person
{
public string Name { get; set; }
public int Age { get; set; }
}
3. Let the AI create the SQL to create the table and store the concrete object
CREATE TABLE Persons (
Id INT IDENTITY(1,1) PRIMARY KEY,
Name NVARCHAR(50) NOT NULL,
Age INT NOT NULL
);
INSERT INTO Persons (Name, Age)
VALUES ('Sam', 42);
The training data point consists then of:
{'instruction': 'Convert text to sql',
'input': 'Hi, I´m Sam and 42 years old',
'output': 'CREATE TABLE Persons (
Id INT IDENTITY(1,1) PRIMARY KEY,
Name NVARCHAR(50) NOT NULL,
Age INT NOT NULL
);
INSERT INTO Persons (Name, Age)
VALUES ('Sam', 42);'
Not sure if and how it will work. There is maybe more context in the instruction and/or input necessary.
@@toddnedd2138 Checkout WikiSQL www.kaggle.com/datasets/shahrukhkhan/wikisql you may find this useful.
@@samwitteveenai Thank you for the link, appreciate it very much. 🤩