Web Scraping with GPT-4 Vision AI + Puppeteer is Mind-Blowingly EASY!

preview_player
Показать описание


👉 Discord: all my courses have a private Discord where I actively participate

⏱️ Timestamps:
00:00 Intro
00:48 ChatGPT for HTML
02:01 OpenAI API
04:08 Puppeteer
04:28 Avoid restrictions (Bright Data)
05:11 Get HTML with Puppeteer + Proxy
08:35 Processing HTML
10:00 Give HTML to OpenAI
12:19 ChatGPT for scraper code
14:41 Vision API
19:40 Vision API pricing
21:12 Vision vs Text pricing
22:52 Future of web scraping (AI + Bright Data)

#webdevelopment #programming #coding #reactjs #nextjs
Рекомендации по теме
Комментарии
Автор

🎯 Key Takeaways for quick navigation:

00:00 *🌐 Web scraping has been revolutionized by AI, particularly with the latest Vision AI model, making data extraction more efficient.*
01:07 *💻 Manually copying HTML and using Chat GPT for extraction is one method, but OpenAI's API offers programmable solutions for scalability.*
02:16 *🔄 Using Puppeteer with Bright Data's scraping browser helps circumvent website restrictions and rate limiting during scraping.*
05:33 *🖥️ Puppeteer allows for easy scraping of HTML content, but there's a need to manage and clean up the extracted data before analysis.*
08:35 *💡 Extracting only necessary data from HTML can optimize costs when using OpenAI's models for analysis.*
12:17 *💰 Text-based scraping methods can be cost-effective, but they require ongoing maintenance due to HTML structure changes.*
14:49 *📸 Utilizing OpenAI's GPT-4 Vision API enables data extraction from screenshots, potentially offering a more robust solution for complex web scraping tasks.*
17:52 *🖼️ Using base64 encoding allows passing images to models, enhancing data processing capabilities.*
18:49 *💸 Consider cost-effectiveness when choosing between complex HTML-based or text-based approaches for web scraping.*
19:58 *🎚️ Adjusting image resolution can significantly decrease token usage in web scraping, but it may increase the likelihood of errors.*
20:53 *🖼️🔄 Balance image resolution and price when utilizing Vision API for web scraping, as higher resolution images incur higher costs.*
21:19 *🧹 Clean up HTML before web scraping to reduce token usage and ensure accuracy in results.*
22:57 *🤖 Explore advanced features of AI tools, such as identifying clickable elements, to enhance web scraping automation.*

Made with HARPA AI

hxxzxtf
Автор

This is such a timely video — i'm doing something similar to resurrect a website from the wayback machine.

zeeeeeman
Автор

Wow. this video provides GREAT value. Just in time for what I´m doing now. Thanks mate!

beemerrox
Автор

Thank you infinitely for sharing this masterclass lesson with the universe for free. Subbed

SupCortez
Автор

what an amazing video - like its so niche but so useful

reidevanson
Автор

Hey man, mind if I ask what programming languages you know other than Javascript/TS ?

hellokevin_
Автор

It's interesting, but what if I want pagination?
I will still need to select next button in old way.
Is there any other way of doing the pagination?

imranhrafi
Автор

Great video. Some question though. What about hallucinating? How can be sure is not doing it?

felipeblin
Автор

Can you create a video how to deploy puppeteer and next js to vercel?

juliushernandez
Автор

Have you thought about or tried using a local model to scrape, it would save all the costs

RobShocks
Автор

Where can I learn basic coding from scratch to be able to do that?

dupatrio
Автор

and how do you get to the next page to scrape?

gregsLyrics
Автор

How to do this using Braina AI? Braina can run GPT-4 Vision.

LifeTrekchannel
Автор

This is a great video. But the problem with scraping has hardly ever been parsing the HTML or maintaining the parsers.

The biggest problem is efficiently accessing websites that actively try to block you by gating their content being a login or captchas. Then comes IP blocking (or worse data obfuscation) if you Scrape their website in a large volume.

Lars
Автор

I am interested in creating a price comparison website featuring approximately 10-20 shops, each offering around 10, 000 similar products. Unfortunately, these shops do not provide APIs for direct access to their data. What would be the most efficient approach to setting up such a website while keeping maintenance costs reasonable?

amadeuszg
Автор

I am scrapping (dropping html) with python code with selenium (aprrox 60, 000 articles) and later creating vector embeddings for Llama 3 and asking it to write article for me.

amitjangra