Latenode

Automatically scrape and summarize news posts in NocoDB

This automation workflow is designed to help organizations efficiently gather and process news content from websites without RSS feeds.

It uses web scraping techniques to extract the latest news posts, leverages AI-powered summary generation and keyword extraction to distill key information, and saves the results to a NocoDB database for further analysis and reporting. The workflow is triggered on a weekly schedule, ensuring that the most recent news is consistently captured and processed. By automating these repetitive data collection and processing tasks, this solution enables teams to stay informed on industry trends and developments without the overhead of manual monitoring and summarization.

Updated Apr 6, 2026Est. run: 25sEst. cost: $0.0014
How Latenode estimates time and cost

Latenode bills workflow runs in credits: 1 credit = 30 seconds of processing. Minimum charge per run depends on your plan. Plug-and-Play (PnP) AI nodes are billed separately—each PnP token is $1 USD, charged pay-as-you-go at vendor cost plus a small processing fee, with no API keys required.

Full pricing — how credits work →
Scraping & data collection

Workflow preview

What this template does

  • Extracts news content from websites without RSS feeds using web scraping techniques
  • Generates AI-powered summaries and extracts keywords from news posts
  • Stores the processed news data in a NocoDB database for further analysis
  • Schedules the data collection and processing workflow to run weekly
  • Saves time and effort by automating repetitive news monitoring and summarization tasks

How it works

1
Trigger

Weekly Scrape Trigger

This automation workflow is triggered every Wednesday at 4:32 AM to scrape the latest news content from a website without an RSS feed.

2
Action

Fetch News Site

The workflow uses a headless browser integration to load the target news website and retrieve its HTML content.

3
Action

Extract News Links

A JavaScript integration is used to parse the HTML and extract the URLs for all individual news posts on the website.

4
Action

Extract Posting Dates

The JavaScript integration also extracts the publication date for each news post from the website's HTML.

5
Logic

Filter Recent Posts

The workflow then filters the list of news posts to include only those published within the last 7 days, ensuring that the most recent content is processed.

6
Logic

Process Each Recent Post

For each of the recent news posts, the workflow performs the following steps to extract, summarize, and save the content.

7
Action

Fetch Post Content

The headless browser integration is used to load the full content of the current news post.

8
Action

Extract Post Details

A JavaScript integration extracts the title and full content of the current news post from the loaded HTML.

9
AI

Generate Summary

The workflow uses an OpenAI integration to generate a concise summary of the current news post.

10
AI

Extract Keywords

Another OpenAI integration is used to identify the 3 most important keywords in the current news post.

11
Action

Save to NocoDB

Finally, the post details, summary, and keywords are stored in a NocoDB database for further analysis and reporting.

Setup guide

1

Add OpenAI API credential

1. In the Latenode Credentials panel, add a new credential for the OpenAI integration. 2. Enter your OpenAI API key.

2

Configure NocoDB database connection

1. In the Latenode Credentials panel, add a new credential for the NocoDB integration. 2. Enter your NocoDB project ID and API token.

3

Set up web scraping with Headless Browser

1. In the Latenode visual builder, add a Headless Browser node. 2. Configure the node to visit the target news website URL. 3. Set up any necessary login or cookie handling to access the news content.

4

Extract news post links and dates

1. In the Latenode visual builder, add a Code node after the Headless Browser node. 2. Write JavaScript code to extract the links and dates of the news posts from the webpage HTML.

5

Save news post data to NocoDB

1. In the Latenode visual builder, add a NocoDB node after the previous steps. 2. Configure the node to save the news post data (title, date, link, summary, keywords) to the NocoDB database.

Requirements

Set up an OpenAI API key to enable summarization and keyword extraction
Configure a NocoDB database connection to store the processed news content
Provide the target news website URL(s) to be scraped
Grant the Latenode workspace access to the target website through browser automation or HTTP(S) requests, depending on the PRIMARY_PATH_FROM_A1

FAQ

Common questions about this template

Each run uses credits on your Latenode plan. We charge for processing time (1 credit = 30 seconds). Your actual cost depends on your plan and how long the run takes. See pricing plans for plans and how credits work.

More templates

You might also like

Browse all templates →
Scraping & data collection

Automatically Sync Google Maps Business Data to a Spreadsheet

This automation workflow allows users to efficiently scrape business data from Google Maps, including names, contact details, and reviews, and export the structured information into a spreadsheet or database. The workflow uses the SerpAPI service to retrieve Google Maps search results, which are then transformed and appended to a Google Sheet. This enables users to generate leads, conduct market analysis, and gain valuable insights from the collected data in a cost-effective and scalable manner.

26s$0.0703
Scraping & data collection

Scrape Zillow property data and sync to Google Sheets automatically

This Latenode automation extracts real estate listing details from Zillow and automatically populates a Google Sheets spreadsheet with the property data. It leverages the Scrape.do web scraping API to bypass anti-bot protections and fetch the full HTML of Zillow listings, then parses key information like price, address, days on Zillow, and Zestimate, and saves the structured results into a Google Sheet. This solution is designed for real estate professionals, investors, and market analysts who need to collect property data at scale without manual effort, enabling better market research, portfolio tracking, and lead generation.

9s$0.0006
Scraping & data collection

Automatically scrape and store Booking.com hotel data

This workflow automates the search and extraction of hotel data from Booking.com, triggered by a chat message. It uses a combination of web scraping with Bright Data's Web Scraper and AI-powered data processing with OpenRouter to deliver a concise, human-friendly list of hotels, including the title, address, original price, and final price. The final output is a clean and formatted report, making it a valuable tool for travelers, event planners, and business professionals who need to quickly find accommodation.

26s$0.0703