- 88
- 91 244
Adrian | The Web Scraping Guy
United States
Registrace 21. 10. 2016
Hi! My name is Adrian and I build custom scraping solutions for your business needs. I scrape the internet using Node.js (JavaScript) and make videos about how to scrape.
I am taking on scraping projects now so hit me up if you need some data scraped from a website adrian@thewebscrapingguy.com 🤘
adrian_horning_
I am taking on scraping projects now so hit me up if you need some data scraped from a website adrian@thewebscrapingguy.com 🤘
adrian_horning_
My Social Media Scraper API's Demo
API docs: quick-dungeon-673.notion.site/Social-Media-Scraper-API-s-f18de4220c1848b3a7a6c6be6fdeaef3
If you need anything scraped, email me: adrian@thewebscrapingguy.com
Cheap Proxies that I use: stormproxies.com/clients/aff/go/michael_horning
dashboard.smartproxy.com/register?referral_code=67ce4005cc20c309ebfca78aa10c91621e305e3d
If you need anything scraped, email me: adrian@thewebscrapingguy.com
Cheap Proxies that I use: stormproxies.com/clients/aff/go/michael_horning
dashboard.smartproxy.com/register?referral_code=67ce4005cc20c309ebfca78aa10c91621e305e3d
zhlédnutí: 332
Video
How To Scrape Zillow with Node.js
zhlédnutí 438Před měsícem
If you need anything scraped, email me: adrian@thewebscrapingguy.com Cheap Proxies that I use: stormproxies.com/clients/aff/go/michael_horning dashboard.smartproxy.com/register?referral_code=67ce4005cc20c309ebfca78aa10c91621e305e3d
How to use Proxies in Node.js
zhlédnutí 236Před 2 měsíci
If you need anything scraped, email me: adrian@thewebscrapingguy.com Cheap Proxies that I use: stormproxies.com/clients/aff/go/michael_horning dashboard.smartproxy.com/register?referral_code=67ce4005cc20c309ebfca78aa10c91621e305e3d
How to upload Images to Supabase
zhlédnutí 108Před 2 měsíci
If you need anything scraped, email me: adrian@thewebscrapingguy.com Cheap Proxies that I use: stormproxies.com/clients/aff/go/michael_horning
How to take a screenshot with puppeteer (node.js)
zhlédnutí 223Před 2 měsíci
If you need anything scraped, email me: adrian@thewebscrapingguy.com Cheap Proxies that I use: stormproxies.com/clients/aff/go/michael_horning
Baddie Finder VSL
zhlédnutí 1,2KPřed 2 měsíci
Can be found at: baddiefinder.com/ Would love your feedback! adrian@thewebscrapingguy.com
I scraped 90k+ tiktok creators + emails 👀
zhlédnutí 610Před 3 měsíci
Sign up now for lifetime access: tokfinder.co/ If you need anything scraped, email me: adrian@thewebscrapingguy.com Cheap Proxies that I use: stormproxies.com/clients/aff/go/michael_horning
I wrote a free script that scrape Facebook groups
zhlédnutí 3,1KPřed 4 měsíci
Script: thewebscrapingguy.com/articles/I-made-a-free-script-that-scrapes-facebook-group-posts If you need anything scraped, email me: adrian@thewebscrapingguy.com Cheap Proxies that I use: stormproxies.com/clients/aff/go/michael_horning
How I scraped the 130,000 chrome extensions in the chrome store
zhlédnutí 331Před 4 měsíci
If you need anything scraped, email me: adrian@thewebscrapingguy.com Cheap Proxies that I use: stormproxies.com/clients/aff/go/michael_horning
I built a bot that notifies me when Stanley Tumblers Restock
zhlédnutí 305Před 4 měsíci
Email me and I'll add you to it: adrian@thewebscrapingguy.com If you need anything scraped, email me: adrian@thewebscrapingguy.com Cheap Proxies that I use: stormproxies.com/clients/aff/go/michael_horning
How I scraped 150k leads from Angi list! 🤯
zhlédnutí 693Před 5 měsíci
If you want all the 150k leads, email me: adrian@thewebscrapingguy.com If you need anything scraped, email me: adrian@thewebscrapingguy.com Cheap Proxies that I use: stormproxies.com/clients/aff/go/michael_horning
My Process for Scraping Any Site
zhlédnutí 1,3KPřed 6 měsíci
If you need anything scraped, email me: adrian@thewebscrapingguy.com Proxies that I use: stormproxies.com/clients/aff/go/michael_horning
How Web Apps Get Their Data
zhlédnutí 243Před 6 měsíci
If you need anything scraped, email me: adrian@thewebscrapingguy.com This is a video geared towards beginners. Knowing how web apps get their data is crucial to figuring out the best way to scrape them. Cheap Proxies that I use: stormproxies.com/clients/aff/go/michael_horning
How to recreate scraping bee 🐝
zhlédnutí 808Před 7 měsíci
If you need anything scraped, email me: adrian@thewebscrapingguy.com Some great, cheap proxies that I use: stormproxies.com/clients/aff/go/michael_horning
How to scrape any website with AI 😱
zhlédnutí 3,4KPřed 7 měsíci
If you need anything scraped, email me: adrian@thewebscrapingguy.com Using Node.js (javascript) Langchain documentation: js.langchain.com/docs/use_cases/question_answering/ Cheap Proxies that I use: stormproxies.com/clients/aff/go/michael_horning
How to get cookies from requests and responses
zhlédnutí 2KPřed 7 měsíci
How to get cookies from requests and responses
Scraping TikTok Ads from the Creative Center
zhlédnutí 1KPřed 7 měsíci
Scraping TikTok Ads from the Creative Center
How to scrape ALL Google Search Results with Node.js
zhlédnutí 1,3KPřed 9 měsíci
How to scrape ALL Google Search Results with Node.js
How to get around "Please enable JavaScript"
zhlédnutí 340Před 9 měsíci
How to get around "Please enable JavaScript"
How to scrape LinkedIn Jobs with Node js 🕸️
zhlédnutí 1,9KPřed 9 měsíci
How to scrape LinkedIn Jobs with Node js 🕸️
How to scrape Google Jobs with Puppeteer
zhlédnutí 573Před 9 měsíci
How to scrape Google Jobs with Puppeteer
How to Scrape Google Maps with Node.js Puppeteer 🌍
zhlédnutí 7KPřed 10 měsíci
How to Scrape Google Maps with Node.js Puppeteer 🌍
🚀 How to Get Immediate Notifications for New Jobs in Upwork
zhlédnutí 2,1KPřed 10 měsíci
🚀 How to Get Immediate Notifications for New Jobs in Upwork
How to proxy requests with Node Fetch
zhlédnutí 1,1KPřed 10 měsíci
How to proxy requests with Node Fetch
Is Batchleads Pre Foreclosure List Accurate?? 🤔
zhlédnutí 95Před 10 měsíci
Is Batchleads Pre Foreclosure List Accurate?? 🤔
How to Find Expired Listings for FREE Using Realtor.com!
zhlédnutí 1,3KPřed 10 měsíci
How to Find Expired Listings for FREE Using Realtor.com!
Batch Leads Only Updates Once A Month!? 😱🤯
zhlédnutí 131Před 10 měsíci
Batch Leads Only Updates Once A Month!? 😱🤯
Are Batch Leads Expired Listings Accurate??? 🤔
zhlédnutí 190Před 10 měsíci
Are Batch Leads Expired Listings Accurate??? 🤔
Hello! I am tasked with scraping the entire website for all the properties, that's half a million I believe. I'm a junior dev and I stumbled upon your video after having no luck for a few days, and this is the first time I have to scrape a website. First of all, is this even possible? If so, can it be done in pure vanilla JS? Or are external libraries needed for this many properties? Ideally, I need to write the script in vanilla JS without using any libraries because I need to deploy it as a Worker on Cloudflare that will fetch the data as JSON and then send it off somewhere else. I'm feeling like this is a lot of data to scrape. Thank you and have a nice day :)
grande!
Hey mate, when I try to paste it I get a warning that avoids me pasting the code into the console, this is the warning: Warning: Don’t paste code into the DevTools Console that you don’t understand or haven’t reviewed yourself. This could allow attackers to steal your identity or take control of your computer. Please type ‘allow pasting’ below and hit Enter to allow pasting. Is there any step I messed up?
I need bypass bot 🤔
possible to scrap high frequency data like sports odds?
Which site would you scrape it from?
@@the-web-scraping-guyFanDuel
How do you automated it?
do you have a codepen for the tiktok.js file?
Nope
I think if you made a guide of the 10 most popular requests you have gotten and why there is value for clients (maybe it's lead generation or it saves 100 hours of time), I would watch and share that video
Good idea!
Hey, I'm really interested in this. I'm trying to sign up on the website but the sign up links are just looping back to the top of the home page. Are you still working on this? I'm part of a few PUA groups and can get you a lot of sign ups if this works well
If you can share anything that can be used for free to get notifications, it would be really helpful
Can you share the script you use for notifications on your mail
Thank you, Have you succeeded in creating a solution that extracts both the images and all the reviews?
Not yet
@@the-web-scraping-guy Okay, I'm creating. When I finish I will inform you.
I deployed my express app with pupeteer on vercel. It cannot find the chromium, will using spartacuz work ? or this is a completely different thing for AWS only ?
Yeah its only for AWS
hi , the link of the script is corrupted, can you share a github repo ? thanks
It’s not corrupted at all
thank you sir very useful
Great tutorial, Do we have a CDK implementation of this, looking for something similar
Does this still work?
Hey Adrian Thats helpful. Thank you very much. Im just wondering if its possible to scrape all the posts in a subreddit (not just the initial 1000).
Don’t think so. Not to my knowledge
Great video! I was wondering if you have a bot that scrapes the pre foreclosure listings?
Needs to be custom created
wow!! this is so awesome!
Thanks!
my fear its that google detects thats a bot. what do you think?
Yeah this way isn’t the best. I have an IG reel going over how to do it a better way
thank you @the-web-scraping-guy . it works very good .
Awesome!
I use the same method scrap chrome extension, but only scraped 1524 extension at most, why?
🤷♂️
"Definitely with Big Miss there" lmaooo. Great vid!
Thanks 🤙
out of curiosity, is this legal to get someone's data like that?
Grey area. Ultimately, did you get the data from a page that needed a login? If the answer is no, then do what you want. If the answer is yes, then the social media site will have a problem with you breaking the terms of service. However, the possession of the data isn't illegal since it's still available whether you used a bot to get it or not. You are open to legal issues from the platform, but until you're a big fish Captcha is your only foe.
Nope, sitting in jail right now, plz send help
@@the-web-scraping-guylmao
I've seen several people say that this doesn't work. My first attempt was the same. The csv file had no data in it. I then tried it on a different Facebook group and it worked. I don't know why it works on some but not others.
Hi Adrian, how does the chrome launch for you, it doesn't launch for me after I run the code. It might be a silly question, I just started learning node. Thanks in advance!!
I would ask chat gpt
I am trying to do but getting CSRF failed. Please suggest
what do you use to record your beautiful videos
Screen studio
Thank you so much on that. Finally get my puppeteer to work on Lambda
Awesome!
Thank you very much on this. Finally get the puppeteer to work on Lambda
Thank you so much, really helpful!
😎🙏
the best web scraper
🙇♂
I am trying to generate PDF from HTML Using puppeteer with Lambda but not working can you share any solution
Mmmm, I dunno if I’ve ever done that before. Try chat gpt
Adrian, the video link seems to be private. Can you share it? Thank you!
Hola, muchas gracias! Great, great script. Could you adapt it to get the number of reactions for each group post? Please
hello thanks you for sharing. the columns email, post_text, postText, postAuthor, postAuthorId are empty
How much is this?
There are actually 270k leads now. $500
@@the-web-scraping-guy are you selling the program or just the date
@@the-web-scraping-guy data
How to find sales data for a specific product through this method?
Can you still do this
Yeah
cool tutorial
Thank!
very useful both in headless:ture/false mode here is the code help to demo "puppeteer": "~22.6.3", "puppeteer-extra": "^3.3.6", "puppeteer-extra-plugin-user-preferences": "^2.4.1", puppeteerExtra.use(puppeteerExtraPluginUserPreferences({ userPrefs: { download:{ prompt_for_download:false, open_pdf_in_system_reader:true, }, plugins:{ always_open_pdf_externally:true } } })) let url = "somehost.pdf" const browser = await puppeteerExtra.launch({ headless: false, }) const page = await browser.newPage(); const downloadPath = path.resolve('./downloads') const client = await page.createCDPSession() await client.send('Page.setDownloadBehavior', { behavior: 'allow', downloadPath: downloadPath }); await page.goto(url);
straight to the point, keep it simple, love it
Thank!
Any API I use to request depop via postman is a 403...
I think they updated auth for endpoints, same here
will watch it later
Your tabs are giving me anxiety! Have you found that there are similarities between county assessor sites, or are you starting from scratch for each instance?
Starting from scratch. Not too many similarities
i tried this but keep getting 403s even with headers any ideas?
Got scraping and good proxy like storm or smart proxy
Have you solved this problem?
How do you adapt this script so it logs the main post's copy + the first-level comments?
I think it does doesn’t it 😅
My account got blocked after I tried to run it. It worked up until it was blocked.
lol