Scrape ANY Website with AI!!! 💥 AI powered Web Scraping 💥

Поделиться
HTML-код
  • Опубликовано: 28 янв 2025

Комментарии • 25

  • @vaibhavgeek
    @vaibhavgeek 3 месяца назад +6

    I wonder whats' untitled 1 to 34? do you have access to AGI???

  • @romanemul1
    @romanemul1 2 месяца назад +2

    The problem at the end was that it was array of objects and it appeared as multiple objects without enclosing square brackets

  • @ChristianReyes-r9k
    @ChristianReyes-r9k 17 часов назад

    Can HasData handle web scraping as efficiently as AI models do in Crawl4AI?

  • @aadeshabhang2657
    @aadeshabhang2657 2 месяца назад

    man i cant figure this out always some IndentationError or SyntaxError: incomplete input

  • @softmangaming4251
    @softmangaming4251 3 месяца назад

    Can it scrape all pages from the website?

  • @vivekchandan6255
    @vivekchandan6255 3 месяца назад +2

    Hi bro could you please upload it using local models or if you can provide any links also helpful, Thanks.

    • @1littlecoder
      @1littlecoder  3 месяца назад

      I'll try to put together something bro!

    • @vivekchandan6255
      @vivekchandan6255 3 месяца назад +1

      @1littlecoder thanks bro once uploaded that would be helpful for many students 😁

  • @aditya_sahani
    @aditya_sahani 3 месяца назад

    Is llama 3.2 eb model running locally enough to do the scrapping? How to do it?

    • @unclecode
      @unclecode 3 месяца назад

      In my experience it does a pretty job, specially if you pick up some good fine-tuned model to generate structured data, like what Groq fine tuned and its available in HF,

  • @abubakkarsiddique13
    @abubakkarsiddique13 3 месяца назад

    Nice, for ollama llm, how can we setup on craw4ai, any tutorial?

    • @1littlecoder
      @1littlecoder  3 месяца назад +1

      Will work on it soon!

    • @abubakkarsiddique13
      @abubakkarsiddique13 3 месяца назад

      @@1littlecoder Thanks, looking forward

    • @unclecode
      @unclecode 3 месяца назад

      It simply support ollama, in the `provider` property you just pass `ollama/MODEL_NAME`

  • @d.d.z.
    @d.d.z. 3 месяца назад

    Nice and easy.

  • @minglv
    @minglv 3 месяца назад +1

    Can this method crawl all the information on the website?

    • @d.d.z.
      @d.d.z. 3 месяца назад +1

      I was thinking the same but it seems is mainly to build datasets.

    • @Tanvir1337x
      @Tanvir1337x 3 месяца назад

      @@d.d.z. yup

    • @harisjaved1379
      @harisjaved1379 3 месяца назад

      Yes you can do it! Just pass in the field names and how you want them back

  • @st-hs2ve
    @st-hs2ve 3 месяца назад +1

    Is it legal to crawl?

    • @unclecode
      @unclecode 3 месяца назад +1

      In many cases, scraping data that is publicly accessible on the internet is considered legal. However, the purpose of using a crawler is not just to scrape other people's websites. Now, many website owners, online magazines, bloggers, RUclipsrs, and even individuals with an Instagram account want to crawl their own information. Enterprise companies, having many private or public websites, forum or similar sources that they use the help of crawlers that can extract all the data. For instance, they use their own fine-tuned large language models internally, and they allow other companies to use their trawlers. As a result, they need an open-source engine to use for data enrichment and running their own servers. I believe that everybody should be able to have their own language model fine-tuned by their own personal data to democratize AI. That means everybody should be able to extract the information they have in a proper way. In the not-so-distant future, I envision everyone being able to crawl through their own messages from social media accounts and emails, and have control over them. What they do with that information is up to their own units at their own stages. This is one of the most interesting uses of trawlers.

  • @MichealScott24
    @MichealScott24 3 месяца назад +1