Latenode

Automate news extraction and summarization to NocoDB database

This automation workflow is designed to help businesses or individuals who need to regularly extract and summarize news posts from a website that does not provide an RSS feed. The workflow fetches the latest news posts, generates summaries and technical keywords using AI, and stores the results in a NocoDB database for further analysis and processing.

The key steps involve retrieving the web page, extracting relevant content, merging data, generating summaries and keywords using the ChatGPT API, and finally storing the processed data in the NocoDB database. This solution aims to provide a convenient way to stay informed about the latest news without relying on a dedicated RSS feed.

Updated Apr 2, 2026Est. run: 13sEst. cost: $0.0008
How Latenode estimates time and cost

Latenode bills workflow runs in credits: 1 credit = 30 seconds of processing. Minimum charge per run depends on your plan. Plug-and-Play (PnP) AI nodes are billed separately—each PnP token is $1 USD, charged pay-as-you-go at vendor cost plus a small processing fee, with no API keys required.

Full pricing — how credits work →
Scraping & data collection

Workflow preview

What this template does

  • Fetches the latest news posts from a website without an RSS feed
  • Extracts relevant content from the web pages
  • Generates summaries and technical keywords using the ChatGPT API
  • Stores the processed data in a NocoDB database for further analysis
  • Provides a centralized repository of news post data for the business

How it works

1
Trigger

Schedule the Workflow

The workflow is triggered on a weekly schedule to regularly extract and summarize news posts.

2
Action

Retrieve News Website

The workflow uses a headless browser to fetch the HTML content of the news website that does not provide an RSS feed.

3
Logic

Extract News Links

Custom code is used to extract the individual links to news posts from the retrieved HTML content.

4
Logic

Extract News Dates

Additional custom code is used to extract the publication dates of the news posts.

5
Logic

Merge Links and Dates

The workflow combines the extracted news links and dates into a single data structure for further processing.

6
Logic

Filter Recent Posts

The workflow includes only the most recent news posts from the last 7 days, ensuring the focus is on the latest information.

7
Action

Extract Post Content

The workflow retrieves the HTML content of each individual news post for further processing.

8
AI

Save to NocoDB

The processed news post data, including generated summaries and technical keywords, is stored in a NocoDB database for further analysis and processing.

Setup guide

1

Add ChatGPT API credential in Latenode Credentials panel

1. Go to the Latenode Credentials panel. 2. Click 'Add Credential'. 3. Select the 'OpenAI ChatGPT' credential type. 4. Enter your ChatGPT API key. 5. Save the credential.

2

Configure NocoDB database connection in Latenode Credentials panel

1. Go to the Latenode Credentials panel. 2. Click 'Add Credential'. 3. Select the 'NocoDB' credential type. 4. Enter the connection details for your NocoDB database (host, port, database name, username, password). 5. Save the credential.

3

Add a 'Headless Browser' node and configure website URL and CSS selectors

1. Drag and drop a 'Headless Browser' node onto the canvas. 2. In the node settings, enter the URL of the news website you want to scrape. 3. Use the built-in browser tools to inspect the website and find the CSS selectors for extracting the news post links, dates, titles, and content. 4. Enter the CSS selectors in the node settings. 5. Configure any additional options, such as scrolling behavior or authentication, if needed.

4

Add a 'Code' node to generate summaries and keywords using the ChatGPT API

1. Drag and drop a 'Code' node onto the canvas. 2. In the node settings, select the 'OpenAI ChatGPT' credential you added earlier. 3. Write the JavaScript code to use the ChatGPT API to generate summaries and keywords for each news post. 4. Map the input data (title, content) to the API request and the output data (summary, keywords) to the next nodes in the workflow.

5

Add a 'NocoDB' node to store the processed news post data

1. Drag and drop a 'NocoDB' node onto the canvas. 2. In the node settings, select the NocoDB credential you added earlier. 3. Map the input data (title, date, link, summary, keywords) to the appropriate fields in the NocoDB database. 4. Configure any additional options, such as table name or primary key, if needed.

Requirements

Access to the website without an RSS feed
NocoDB database with write permissions
OpenAI ChatGPT API key with 'text-davinci-003' model access
Latenode workspace with 'headless-browser', 'nocodb', and 'openai-chatgpt' nodes configured

FAQ

Common questions about this template

Each run uses credits on your Latenode plan. We charge for processing time (1 credit = 30 seconds). Your actual cost depends on your plan and how long the run takes. See pricing plans for plans and how credits work.

More templates

You might also like

Browse all templates →
Scraping & data collection

Scrape and export Google Maps business data to Google Sheets

This automation allows users to efficiently scrape business data from Google Maps, including names, contact details, and reviews, and export the structured information into a spreadsheet or database for lead generation and market analysis. The workflow triggers manually or on a scheduled basis, fetching search results from Google Maps using the SerpAPI service, deduplicating the data, and writing it to a Google Sheet. This streamlines the process of gathering valuable business intelligence from Google Maps, empowering users to leverage this data for strategic decision-making and sales prospecting.

11s$0.0007
Scraping & data collection

Scrape Zillow property data and sync to Google Sheets automatically

This Latenode automation extracts real estate listing details from Zillow and automatically populates a Google Sheets spreadsheet with the property data. It leverages the Scrape.do web scraping API to bypass anti-bot protections and fetch the full HTML of Zillow listings, then parses key information like price, address, days on Zillow, and Zestimate, and saves the structured results into a Google Sheet. This solution is designed for real estate professionals, investors, and market analysts who need to collect property data at scale without manual effort, enabling better market research, portfolio tracking, and lead generation.

9s$0.0006
Scraping & data collection

Automate Instagram profile data export to Google Sheets

This automation template allows users to scrape comprehensive Instagram profile data using the Apify platform, and automatically export the results into a Google Sheets spreadsheet for analysis. It is designed to run on a schedule, processing a list of usernames by calling the Apify API, appending the scraped data to a Google Sheet, and marking the usernames as processed. The automation integrates with Google Sheets and the Apify actor, providing a streamlined way to gather and organize Instagram profile information without manual intervention.

15s$0.0009