My http module grabs from a website but always ends up with status code 202 and no html, I also receive no error message.
Is there a fix to this?
My http module grabs from a website but always ends up with status code 202 and no html, I also receive no error message.
Is there a fix to this?
Hi Jan, and welcome to the Community.
A 202 error means that the request was accepted by the website, but processing wasnât completed. Which would explain why youâre not seeing any HTML. This is a function of the website itself, not something that Make controls.
Can you post screenshots of your HTTP module?
random commerce website
Welcome to the Make community!
So you basically need to âvisitâ the site yourself to get the content. This is called Web Scraping.
Are you getting NO output from the HTTP âMake a requestâ module? This is because the website has employed anti-scraping measures, and has detected that the visit is not made by a human, and has blocked the request silently by returning no content.
Are you getting NO output from the Text Parser âMatch pattern/elementsâ module? This is because there is NO text content in the HTML! The entire page content you are scraping is hosted in a script tag, which is dynamically generated and placed onto the page using JavaScript when loaded and run on the userâs web browser on the client-side.
Make is a server-side runtime environment, so using the HTTP modules, you get just the script tags, and those script tags are ignored by the Text Parser âHTML to Textâ module because it is NOT a HTML layout element. Using the Make HTTP âMake a requestâ module does NOT run any of those JavaScript, so there is no content on the page other than a default message that tells you to enable JavaScript.
This is NOT a Make platform, apps, or Text Parser, or Regular Expression issue/bug.
You CANNOT use normal scraping integrations like ScrapingBee or HTTP âMake a requestâ module to fetch pages from this website.
You will need to use ScrapeNinjaâs âScrape (Real browser)â module to emulate a real person visiting the site using a web browser, as client-side JavaScript needs to run to parse the JSON data in the script tags, and generate the page structure and content.
For more information and demo using ScrapeNinja, see Scraping Bee Integration Runtime Error 400
For web scraping, a service you can use is ScrapeNinja to get content from the page.
ScrapeNinja allows you to use jQuery-like selectors to extract content from elements by using an extractor function. ScrapeNinja also can run the page in a real web-browser, loading all the content and running the page load scripts so it closely simulates what you see, as opposed to just the raw page HTML fetched from the HTTP module.
If you want an example, take a look at Grab data from page and url - #5 by samliew
You can also use AI-powered web scraping tools like Dumpling AI.
This is probably the easiest and quickest way to set-up, because all you need to do is to describe the content that you want, instead of inspecting the element to create selectors, or having to come up with regular expression patterns.
The plus-side of this is that such services combine BOTH fetching and extracting of the data in a single module (saving operations), and doing away with the lengthy setup from the other methods.
For more information on the different methods of web scraping, see Overview of Different Web Scraping Techniques in Make đ
Hope this helps! If you are still having trouble, please provide more details.
â @samliew
P.S.: investing some effort into the tutorials in the Make Academy will save you lots of time and frustration using Make!
As always, a super comprehensive answer from @samliew !
The Redfin property pages are all built dynamically - thereâs no static HTML. If you right-click on a page and select âView page sourceâ youâll see that itâs all loaded through scripting.
So you need to use a scraper thatâs uses a cloud browser that can mimic a real browser.
Aside from ScrapeNinja you might also want to take a look at Airtop(an innovative LLM-powered scraper that allows you to prompt in plain language for what you want to scrape) or Apify.
Apify even has a specifically built scraper for Redfin.