Description
Overview
This automation workflow streamlines RSS feed processing by filtering and managing article entries related to real estate and restaurants. This orchestration pipeline utilizes batch processing and MongoDB integration to ensure unique article storage and triggers notifications for new content. It is designed for users seeking efficient content ingestion with deduplication, initiated via a manual trigger or an hourly cron schedule.
Key Benefits
- Automates RSS feed reading and filtering with keyword-based conditional branching.
- Implements batch processing to handle articles sequentially, reducing load spikes.
- Performs deduplication by querying MongoDB using case-insensitive link matching.
- Supports both manual and scheduled execution for flexible automation workflows.
- Sends structured notifications via webhooks upon successful article insertion.
Product Overview
This no-code integration workflow begins execution through either a manual trigger node or a cron node configured to run every hour. Upon activation, the workflow initiates an RSS feed read operation from a specified XML source. Articles are extracted and passed through sequential conditional filters based on regular expressions that target keywords in article titles, segregating content into categories related to real estate (“realtors” or “real estate”) and restaurants (“restaurant(s)”). Each category’s articles are processed individually in batches of one to maintain controlled throughput.
For each article, the workflow queries a MongoDB collection named “articles” to detect duplicates by matching the article’s link field using case-insensitive regular expressions. Articles not found in the database are inserted with only their title and link fields. After insertion, a webhook dispatch sends the article link as a JSON payload to a notification endpoint. Batch processing continues until all articles are handled, with conditional nodes monitoring batch completion and terminating the flow appropriately. Error handling relies on n8n’s default retry and failure mechanisms. Authentication to MongoDB is managed via dedicated credentials.
Features and Outcomes
Core Automation
This automation workflow processes RSS feed entries by evaluating article titles against keyword-based conditions, leveraging batch size one for deterministic single-item processing. It branches articles into two categories for specialized handling, ensuring precise content segmentation.
- Single-pass evaluation of RSS feed articles with conditional regex filters.
- Deterministic batch execution with controlled item flow per batch.
- Idempotent MongoDB lookups to prevent duplicate article entries.
Integrations and Intake
The workflow integrates an RSS feed reader node configured with a static XML feed URL, connected to MongoDB for article storage using credential-based access. Conditional branches filter incoming articles by keyword presence in titles.
- RSS Feed Read node fetching articles from a fixed XML feed URL.
- MongoDB nodes querying and inserting documents in the “articles” collection.
- OAuth or API key authentication configured for MongoDB credentials.
Outputs and Consumption
Outputs consist of newly inserted article links sent asynchronously as JSON payloads via HTTP POST requests to configured webhook endpoints. The workflow’s response model is asynchronous, with no direct synchronous client response.
- Article links dispatched as JSON in webhook POST requests.
- MongoDB document insertions restricted to title and link fields.
- Batch processing completion indicated by internal no-items-left flags.
Workflow — End-to-End Execution
Step 1: Trigger
The workflow initiates either manually via the “On clicking ‘execute'” manual trigger node or automatically every hour through the “Cron” node. Both triggers lead to the RSS feed reading step, enabling flexible scheduling or immediate execution.
Step 2: Processing
The “RSS Feed Read” node retrieves articles from a predefined RSS XML URL. Articles are then filtered by two sequential conditional nodes using regex on the article titles to route them into real estate or restaurant branches. Batch nodes split the filtered articles into single-item batches for granular processing. Basic presence checks ensure article data integrity before further handling.
Step 3: Analysis
For each batched article, MongoDB is queried with a case-insensitive regex on the article’s link to identify duplicates. Articles absent from the database proceed to insertion nodes. The workflow uses merge nodes with removeKeyMatches mode to filter out already existing articles, ensuring only new content advances.
Step 4: Delivery
After article insertion, webhook nodes send POST requests containing the article link as JSON to external endpoints for notification or downstream processing. Batch completion is monitored and triggers workflow termination via a no-operation node once all articles are processed.
Use Cases
Scenario 1
Content managers need to aggregate news articles about real estate efficiently. The workflow filters RSS feed items for real estate-related keywords, checks for duplicates in MongoDB, and inserts new articles. This results in a curated, up-to-date real estate article database without redundancy.
Scenario 2
Restaurant marketing teams require timely updates on industry news. The orchestration pipeline filters incoming RSS feed data for restaurant-related articles, inserts unique entries into a database, and triggers notifications via webhooks. This provides a streamlined flow of relevant restaurant content.
Scenario 3
Developers require a no-code integration that automates RSS content ingestion with deduplication. By batching articles and querying MongoDB for existence, the workflow ensures only novel articles are processed and notified, reducing manual checks and improving content freshness.
How to use
To deploy this automation workflow, import it into the n8n environment and configure MongoDB credentials with appropriate permissions for querying and inserting documents. Adjust the RSS Feed Read node URL if a different source is desired. The workflow can be executed manually via the trigger node or scheduled using the existing cron node. Upon execution, monitor logs for inserted articles and webhook dispatches. The expected output is the insertion of new articles into MongoDB and webhook notifications containing article links.
Comparison — Manual Process vs. Automation Workflow
| Attribute | Manual/Alternative | This Workflow |
|---|---|---|
| Steps required | Multiple manual steps: feed reading, filtering, deduplication, insertion, notification | Single automated pipeline combining all steps with batch processing |
| Consistency | Prone to human error and missed duplicates | Deterministic filtering and duplicate checking via MongoDB queries |
| Scalability | Limited by manual processing capacity | Handles articles sequentially in batches with scheduled triggers |
| Maintenance | High effort to maintain keyword filters and database integrity | Low maintenance with configurable regex filters and credential-based DB access |
Technical Specifications
| Environment | n8n automation platform |
|---|---|
| Tools / APIs | RSS Feed Read, MongoDB (query and insert), HTTP Request (webhook) |
| Execution Model | Event-driven via manual trigger or hourly cron schedule |
| Input Formats | RSS XML feed with standard article properties (title, link) |
| Output Formats | MongoDB documents (title, link); JSON payloads via HTTP POST |
| Data Handling | Transient processing with no persistent storage outside MongoDB |
| Known Constraints | Relies on availability of external RSS feed and MongoDB service |
| Credentials | MongoDB credential configured in n8n for DB access |
Implementation Requirements
- Access to n8n platform with ability to import and execute workflows.
- MongoDB instance with read and write permissions on the “articles” collection.
- Valid MongoDB credentials configured within n8n credentials manager.
Configuration & Validation
- Import the workflow JSON into n8n and verify all nodes are connected properly.
- Configure MongoDB credentials with correct access rights for query and insert operations.
- Test execution by triggering manually and confirm articles are inserted and webhook notifications sent.
Data Provenance
- Triggered by “On clicking ‘execute'” manual trigger or “Cron” hourly node.
- Article filtering performed by “IF realtors or real estate” and “IF restaurant(s)” conditional nodes.
- Duplicate detection and insertion handled by “MongoDB: Find Article”, “MongoDB: Find Article1”, “MongoDB: Insert”, and “MongoDB: Insert1” nodes.
FAQ
How is the automation workflow triggered?
The workflow can be triggered manually using the manual trigger node or scheduled automatically every hour via the cron node, enabling flexible execution timing.
Which tools or models does the orchestration pipeline use?
The workflow uses an RSS Feed Read node to ingest articles, MongoDB nodes for querying and inserting data, and HTTP Request nodes to send webhook notifications. Keyword filtering is done through conditional nodes using regex.
What does the response look like for client consumption?
After new articles are inserted into MongoDB, the workflow sends asynchronous webhook POST requests containing the article link as a JSON payload for downstream consumption.
Is any data persisted by the workflow?
Only article titles and links are persisted in the configured MongoDB collection named “articles.” The workflow itself does not store data outside MongoDB.
How are errors handled in this integration flow?
Error handling relies on n8n’s default retry mechanisms. The “MongoDB: Find Article1” node is configured to continue on failure, allowing the workflow to proceed despite individual query errors.
Conclusion
This automation workflow provides a structured method for processing RSS feed articles by filtering, deduplicating, and storing them in MongoDB, followed by webhook notifications. It ensures consistent and scalable handling of real estate and restaurant-related content. While it depends on the availability of external RSS feeds and MongoDB services, it offers deterministic and controlled batch processing. The workflow reduces manual effort in content curation and supports flexible trigger options for continuous or on-demand operation.








Reviews
There are no reviews yet.