Backend Engineer - Data Collection & Processing
We are building a private, licensed “Terminal” for strategists and agencies that surfaces cultural signals, media momentum, and advertising posture in near real time. The frontend is already built in Webflow using CMS collections. We need a backend/data engineer to build the intelligence layer that powers it.
What we need built
We need a lightweight backend service that:
Pulls data from multiple external sources (news, social, advertising)
Normalizes that data into consistent metrics
Computes simple scores (velocity, momentum, posture)
Exposes the results to Webflow (via CMS API or JSON endpoint)
This is not a dashboard build. Webflow handles presentation. This role is focused on data ingestion, transformation, and delivery.
INITIAL DATA SOURCES
- News or narrative data (e.g., GDELT, Event Registry, or similar)
- Social conversation data (e.g., Reddit API or equivalent)
- Advertising posture signals (e.g., Meta Ad Library or public ad transparency sources)
- Exact sources can be discussed, but the system must support multiple APIs and future expansion.
- Core metrics to produce
- Daily and 7-day volume counts per topic
- Velocity (current vs baseline)
- A normalized momentum score (0–100)
- A categorical posture label (e.g., Defensive / Neutral / Offensive)
- Timestamped updates
- Technical requirements
- Node.js or Python backend
- Experience working with REST APIs and rate limits
- Scheduled jobs (hourly or daily refresh)
- Serverless deployment (Vercel, AWS Lambda, or Cloudflare Workers)
- Secure handling of API keys
- Clean JSON schema output
- Optional: Write data directly to Webflow CMS via API
Deliverables
- Deployed backend service
- Working ingestion from at least 3 data sources
- Scoring logic implemented
- JSON endpoint or Webflow CMS integration
- Documentation explaining the data flow and how to extend it
Project constraints
- This is a Phase 1 MVP
- Focus on correctness, stability, and extensibility
- UI and frontend work are out of scope
Ideal experience
- Data engineering or backend engineering background
- Prior experience aggregating third-party APIs
- Comfortable designing simple scoring models
- Familiarity with Webflow CMS API is a plus, not required
Job Qualifications
- News or narrative data (e.g., GDELT, Event Registry, or similar)
- Social conversation data (e.g., Reddit API or equivalent)
- Advertising posture signals (e.g., Meta Ad Library or public ad transparency sources)
- A normalized momentum score (0–100)
- A categorical posture label (e.g., Defensive / Neutral / Offensive)
- Node.js or Python backend
- Experience working with REST APIs and rate limits
- JSON endpoint or Webflow CMS integration
- UI and frontend work are out of scope
- Data engineering or backend engineering background
- Prior experience aggregating third-party APIs
- Comfortable designing simple scoring models
Job Benefits
Job Responsibilities
- Pulls data from multiple external sources (news, social, advertising)
- Normalizes that data into consistent metrics
- Computes simple scores (velocity, momentum, posture)
- Exposes the results to Webflow (via CMS API or JSON endpoint)
- Webflow handles presentation
- This role is focused on data ingestion, transformation, and delivery
- Exact sources can be discussed, but the system must support multiple APIs and future expansion
- Core metrics to produce
- Daily and 7-day volume counts per topic
- Velocity (current vs baseline)
- Scheduled jobs (hourly or daily refresh)
- Serverless deployment (Vercel, AWS Lambda, or Cloudflare Workers)
- Secure handling of API keys
- Clean JSON schema output
- Optional: Write data directly to Webflow CMS via API
- Deployed backend service
- Working ingestion from at least 3 data sources
- Scoring logic implemented
- Documentation explaining the data flow and how to extend it
- This is a Phase 1 MVP
- Focus on correctness, stability, and extensibility