- Видео 14
- Просмотров 17 614
AI Architect
Германия
Добавлен 25 окт 2013
Rust is not Complicated - An architects perspective
Link to the article - rust-on-nails.com/blog/rust-complicated/
Просмотров: 3 212
Видео
Kubernetes is not THAT complicated.
Просмотров 36821 день назад
A look at the DHH decision to self host and the way they went about it.
How Coding Agents Work - A look at the Cline system prompt.
Просмотров 191Месяц назад
Taking a look at a MASSIVE prompt.
Building Web Application in Rust - Rust on Nails Tutorial
Просмотров 6847 месяцев назад
Building Web Application in Rust - Rust on Nails Tutorial
Integrating Airbyte with Bionic-GPT - Enterprise grade RAG pipelines.
Просмотров 34510 месяцев назад
Here we integrate Airbyte with Bionic-GPT. bionic-gpt.com
Create a RAG pipeline in less than 5 minutes
Просмотров 1,9 тыс.Год назад
An overview of using Bionic-GPT to create a no code Retrieval Augmented Generation (RAG) pipeline. bionic-gpt.com
Install Bionic-GPT on Kubernetes with the Bionic-GPT Operator
Просмотров 833Год назад
Using a Kubernetes operator we install Bionic-GPT. bionic-gpt.com/
Bionic GPT in 39 Seconds
Просмотров 893Год назад
bionic-gpt.com We provide 4 main services in 1 neat package. 1. Text and Code Generation 2. No Code - RAG 3. API Key Proxy 4. A full audit trail
Installing Generative AI with Bionic GPT on Kubernetes
Просмотров 532Год назад
Walk through of a possible setup for K8's installation of Bionic GPT. bionic-gpt.com/
My first surfing 360 at the Jochen Schweitzer Arena. Munich.
Просмотров 277Год назад
My first surfing 360 at the Jochen Schweitzer Arena. Munich.
Bionic GPT - Run Generative AI On Premise
Просмотров 6 тыс.Год назад
A demonstration of BionicGPT which allows enterprises who can't use Chat GPT to run generative AI on premise.
I learnt Rust and became proficient at it over the course of two years and I totally agree with you. It's software development/engineering that is inherently complex, not Rust. Rust however, unlike other languages, doesn't hide that complexity from us. I see the compiler like a strict but helpful computer science teacher that doesn't accept blatantly bad code. I learnt so much more about software by learning Rust, especially as I asked ChatGPT to explain me the tricky concepts I didn't instantly grasp in Rust.
Yep, this is where the genAIs shine. In fact, they're often better at correctly explaining concepts than getting the generated code exactly right!
@@kevinmcfarlane2752 I agree 100%. Nowadays that's how I use genAIs. Instead of asking for a code snippet, I ask for a conceptual description of some tricky concepts, or for some guidance on how to structure my code from a high level perspective. Sometimes I might ask for a code snippet to illustrate some coding pattern for instance, but not a direct solution for my problem. For Rust in particular, it helped me understanding asynchronous code / futures or lifetimes way better than reading some documentation.
Rust has its own unnecessary complexity. Not engineering. For engineering is C.
@@hahavx512 In my opinion for simpler projects Rust's complexity may indeed be too distracting and unnecessary. In those cases I agree C, or even better Zig could be a great option. However, after a certain problem domain complexity you are either an extremely proficient C programmer and don't ever make mistakes, or you use something better like Rust, which will help you catch those mistakes, even if at the cost of a much more complex language. As always, it's about the best tool for the job and most importantly, using whatever you understand or feel better with.
i'm a hobbyist programmer and i love peeking into different languages. i admire the determination of Rust devs to combine C with Haskell, but i think that the result is underwhelming both in comparison to C and to ML style languages. like C is so good for low-level precisely bc it doesn't encourage RAII (which can be very slow for a large number of small structs), and because it doesn't have phantom types and traits and all that bs -- it's raw pointers and offsets. functional languages on the other hand have a lot of stuff to manage traits and other type frankensteins (which only offer some basic safety checks, btw). C's greatest weakness is lack of proper compile-time programming. though Rust has a powerful macro system, it's very unpleasant, and i can't see why it couldn't be more like in Zig or Julia.
Like he said either you learn c and other languages for different domain, or you learn rust for all domain and you master the language, or you use different languages for each domain and have a shallow knowledge about the languages
@danielt8880 i'm not against anybody else programming Rust, i just think it's ass. imo to be effective for high-level programming, language needs something like jit-compiler/interactive interpreter and a format similar to jupyter notebooks. Rust definitely does not fit there and never will. To be effective for low-level programming, language needs good templates, macros, high control over memory allocation and layout. and here i can see that Rust is not very good with macros, and not ergonomic with allocation & layout, even compared to C (which is a fairly low bar imo).
@@danielt8880 isn't it a bit strange that we even have a debate, what PL is better, one made 50y ago when bytes could have 12 bits, or the one that was made ~10y ago?
brilliant
I so so so much agree.
Rust is complicated
Skill issue
Naa Rust is complex , I can justify, developer want to leave C bcz there you have to manage memory by your won , and same goes to rust too , it's you have to think about owner ship model , so in simple words you are doing same thing but in diffrent language and different way , Rust is safe but it's comes with a cost of high complexity non-readable
Every program has to manage memory, but in most languages you just don't know what it's doing.
@Bruno-cb5gk if you are not developing OS , you don't need to know
@@codeking4585 people like you are the reason so many modern apps and websites are painfully slow and buggy.
@Bruno-cb5gk naaa man naa they aren't slow bcz language, they are slow bcz of bad resource management, and no any language can solve this , it's a developer responsibility.
@ you said you don't need to know about memory management, and then you said programs are slow because of bad resource management. How are you meant to manage resources well if you don't know about it?
"It's software engineering that is complicated. Rust doesn't allow you to avoid its complexities." ...I think I just added something to my vocabulary.
I'm impressed how fast you scrolled through the article unlike some other youtubers who'd shout, scream and regurgigate every other sentence just to add nothing.
Yeah I agree with you
@@jostasizzi818cough prima... cough
Oh I know which other RUclipsrs you are talking about and I complete agree haha
Rust definitely is not for people who are not willing to learn difficult stuff. I don't see a point in lowering expectations here. Is Rust worth it anyway? It was for me, but that doesn't mean it will be for everyone. What I have seen in my career is, that many employers and superisors think, that when it looks like it works, then they could have gotten away with less work and want to ensure not to repeat that mistake in the future, then Rust is rather like a land mine. If your goal is to create high quality software and your superiors have your back, Rust is the main contender.
I would argue that if you are not willing to learn difficult stuff you have no place in this industry.
@artxiom Ideally it would be like you say, but I have seen enough developer jobs where doing anything but keeping the office chair warm, was discouraged or even punished. There are even more dev jobs where nobody ever deals with anything nearly as complex as Rust.
I cant agree more, this is such a modern and amazing take. Exactly why I code websites in axum
@@Filipcorobivblenderi Thanks.
Actix better
@ why. I went for axum cos no unsafe used and was "simpler" to use. But i really was hesitant about using axum or actix
@@Filipcorobivblenderi Idk it's probably the same nowadays but actix is the OG. I was there when it got popular and when everyone bullied the author lol
I was wondering what is happening in the background and cannot find correct information. This video exactly what i was searching for.Thanks !
Indeed. They are just recreating the k8s. I’ve tried interacting with them on X about various ego issues they have. Shockingly I never get a response lol
Is k3s good for production environment?
I've used it in production and had no issues so far.
thank you for the video , how do we limit the knowledge base only limited more to the pdf uploaded. .
@@macx75 if you create a dataset and only upload 1 pdf. That should do it.
When other videos?
Can I use local models in bionic-gpt?
@@asparagess Yes. Any inference engine that supports the open ai API.
Hi, great work on Bionic GPT. Tried installing it locally in. my macbook but I am encountering errors when using local Llama model via Oomla. It gives me an error 111 refused connection error. Can you please point me to the resources which may help me sort this out? Thank you.
Seems like it cut before the end. Any plan on adding the rest ? Thanks !
does it work with ollama?
Yes
How do I connect it to ollama? Do you have a video for that?
Thanks for this, the system looks amazing !!! The questions I have are - how to set this up on a mac or windows? docker? - can it run on a single machine? I'm not overly familiar with kubernetes but can a cluster just be one machine ? Many thanks in advance !
Is it possible to use bionic GPT with Azure OpenAI Service?
Great!
So how do we add models? 🤔
Hello, I am new to this but I was wondering can this chat run on 16GB Ram/ i5-10300H CPU @ 4.5 Hz/ GTX 1650 4GB ? thank you in advance
Well done, you helped me a lot! I have a question about 0:58 - How to get more models?
Is it possible to do Multi-document summarization with BionicGPT? I would imagine an agent on top of each document and one agent on top comparing the outputs of all the agents and feeding it to the LLM model. Or is BionicGPT primarily set up to find the most relevant source of information based on the prompt you give and then gives a specific answer together with the LLM?
It could be something we could add.
Haven't been able to get the embeddings-api service to run with GPU support. Any advice? I've installed the Nvidia toolkit for docker, and GPU works for the llm-api service.
Can you raise an issue on Github?
for windows ??
Let's say that i've got a bladecenter that has 16x 16-core 2.10GHz processors just laying around... How would/could I cluster the hardware together to run something like this?
Thank you very much, Ian, for the amazing work you guys are doing. I can’t wait to try it out.
is there an installation video, step by step guide on hot to download on each operating system,. such as windows, or mac.
Hi @Ian I have just installed as a docker container and opened the UI at localhost. When I submit a prompt it shows the following error on the console. I created new prompts and tried but no luck. Please help object has no attribute 'request' oauth2-proxy-1 | 2024/01/05 18:27:42 httputil: ReverseProxy read error during body copy: unexpected EOF oauth2-proxy-1 | 2024/01/05 18:27:42 httputil: ReverseProxy read error during body copy: unexpected EOF oauth2-proxy-1 | 2024/01/05 18:27:42 httputil: ReverseProxy read error during body copy: unexpected EOF
Can you raise a ticket on our GitHub?
Hi@@enterprise-architecture Even after installing ollama and running llama2 locally, I get the same errors. Please help
Thank you been playing around with it.
I am incredibly excited about your work. Great job, so far!
Does this work on a Macbook M2 Pro? I tried every way to install it but failed. Getting this error Error response from daemon: could not select device driver "nvidia" with capabilities: [[gpu]]. The only way that works to install and run is via the "Installation (CPU)" route but then no responses are generated by the LLM, I get an error with every prompt. I'd appreciate any help. I'm not very technical but can do some basics. It would be so great to use this.
This is the error in the console when I prompt the engine using the CPU installation process: response="status = 422 Unprocessable Entity, message = error trying to connect: tcp connect error: Connection refused (os error 111)" Either way, I'm not sure where the error is...
Was in touch with your co-founder. Found the issue and will be fixed. Awesome tool! Looking forward to digging in.
Hi Ian, I tried it out and it runs fine with 16gb RAM. However I am having trouble setting up a new LLM. I tried adding Mistral 7B but it doesn't work. I went through the documentation on your website but still facing the same issue. Can you please assist for the same??
Mistral 7b you could try with our ollama setup. bionic-gpt.com/docs/running-locally/gpu-setup-ollama/
how many GB of RAM is needed? I tried to install it on a 8GB RAM notebook, could get into the UI but then the models did not respond Thanks in advance and thank you for making it possible
16GB works I know that. 8GB is too small as the model itself with tage 5GB. It's great you tried it out. Thanks.
@@enterprise-architecture i am not too deep into technology topics... thats why i am wondering if i want to deploy bionicgpt for a company on german servers i will need to use kubernetes or can it be a dockercontainer on a server? I am just wondering how the process is
Spectacular .. Genius Mate!
Looking forward to capability extending to MacOS and M hardware.
We're working on it.
if my notebook is 8GB of RAM.. doesn't this work? Because when i start a chat it says "Processing prompt" and the loading icon occurs.
8GB is probably not enough. You could try running the model on it's own to see if that works.
Great work guys! Id love to connect with with autogen. Is this possible?
Yes, it's possible. We export an Open AI compatible rest API which applications can connect to.
This project will go big. Great work.
That's the plan!
can you pls show/explain how to integrate with other top open source models like OpenChat3.5 i get Error occurred while generating
This might help. bionic-gpt.com/docs/running-locally/osxwindows/
getting error Error occurred while generating. in docker logs getting 2023-12-02T14:29:51.838799Z ERROR axum_server::errors: response="status = 422 Unprocessable Entity, message = query returned an unexpected number of rows" help please
We're tracking that here github.com/bionic-gpt/bionic-gpt/issues/134
It works great! Would be nice if i can switch to gpu instead of cpu
There's some documentation for that here. bionic-gpt.com/docs/administration/external-api/
Great! esp for small teams to deploy. thanks
Thanks for watching.
wow. I shall learn to do that too! Lots of fun.
You should!
This looks awesome. Thanks for making it open source, I am going to try it out with my organization.
Thanks. If you need help just let us know.
love it