Web Scraping behind Login

Can anyone point me to some resources or knowledge regarding web scraping behind a login I am scraping a university alumni directory to build an alumni base for our entrepreneurship center. Using the chrome extension I am able to log in and scrape the first page but the directory only shows 50 alumni per page and cannot figure out how to scrape the whole directory or follow up the login browser replay. Thank you

Hey @George_Sterrett! If you’re reporting an issue with a flow or an error in a run, please include the run link and make sure it’s shareable so we can take a look.

  1. Find your run link on the history page. Format: https://www.gumloop.com/pipeline?run_id={{your_run_id}}&workbook_id={{workbook_id}}

  2. Make it shareable by clicking “Share” → ‘Anyone with the link can view’ in the top-left corner of the flow screen.
    GIF guide

  3. Provide details about the issue—more context helps us troubleshoot faster.

You can find your run history here: https://www.gumloop.com/history

Hey @George_Sterrett – Right now, the best option would be to scrape using the browser extension input. It does mean it’ll be manual and require visiting each page, but it’s the most reliable approach. Browser replay relies on AI to map the steps and can sometimes fall short, especially with pagination and scraping.

Short tutorial for Browser Extension Input Node you can follow.

This topic was automatically closed 5 days after the last reply. New replies are no longer allowed.