Issue with Leboncoin.fr Data Extraction - No Output & Missing Input Error

Hi there!

I’m experiencing issues with a flow designed to extract data from leboncoin.fr. Here’s what’s happening:

Problem Description:

  1. No output from the first node: When I click on View Inputs and Outputs for the initial extraction node, no data is displayed

  2. Flow execution fails with the following error:

    'Send Email Notification' node could not be run! Missing input 'None', did not receive a value. Please double check your node connections.
    
    
  3. 1. Google Sheet** remains empty**: The data hasn’t been written to 1. the target spreadsheet

Self-Debug Results: GumLoop’s self-debug feature suggests:

The Send Email Notification node is failing with a missing input error. 
Please check the node's email_body parameter and verify that both 
connections (from Ask AI and Google Sheets Writer) are properly established. 
You may need to manually click on the email_body field and re-add the badge 
references for both the AI response and the sheet link. Also, please verify 
in the run logs that both the Ask AI and Google Sheets Writer nodes completed 
successfully before the email node tried to run.

What I’ve tried:

  • Verified node connections

  • Checked * the Google Sheet (no data present)

  • Reviewed the output from the first node (empty)

Has anyone encountered similar issues with web scraping flows or leboncoin specifically? Any suggestions would be greatly appreciated!

Thanks in advance!

Hello,

The issue you’re running into comes from leboncoin.fr itself. The website uses strong anti-bot protection (DataDome), which blocks most scraping attempts. Even when the Website Scraper node appears to run successfully, the page that’s returned does not contain the actual listings. Because of this, the Extract Data node receives no usable content and returns empty results.

I tested this page with multiple providers and the blocking is consistent. At the moment, full scraping of leboncoin.fr is not reliable with standard scraping methods. Firecrawl is able to retrieve partial results in some cases, but the output can be incomplete and may vary from run to run.

In this situation, Gumloop Agents can be useful – because they let you combine multiple tools. For example, an Agent can try Firecrawl in parallel with other providers, validate or clean partial results, and then directly send the output by email. This makes Agents a good orchestration layer when dealing with unreliable or partially blocked sources.

Here’s an example Agent you can use as a starting point: https://www.gumloop.com/agents/hzPHCArUzhhFEMzdSYqryJ (This agent uses tools like Exa, Firecrawl, Parallel..)

You can read more about how Agents work here: https://docs.gumloop.com/core-concepts/agents

If you have API access to the platform, using a custom node could also be the most stable long-term solution for reliably fetching the information.

This topic was automatically closed after 3 days. New replies are no longer allowed.