Signal Intelligence · Bespoke Data Engineering

Data built for you

Signal can be contracted to build custom data pipelines, proprietary datasets, and private API endpoints — engineered to your exact specification. If you need data that doesn't exist publicly, or public data assembled in a way nobody has done before, we build it.

Signal Bespoke API · v1.0
# Client: Institutional investor
# Dataset: ESG sentiment + short interest correlation
 
GET api.signal-intel.io/v1/bespoke
  /esg-sentiment-tracker
  ?ticker=MSFT,AMZN,GOOG
  &range=90d
  &x_axis=time
  &y_axis=sentiment_score
  &overlay=short_interest
{
  "status": "200 OK",
  "dataset": "esg_sentiment_v2",
  "records": 2847,
  "cadence": "daily",
  "anomalies": 14
}
The process

From brief to live API in six weeks

Signal's bespoke data service follows a structured four-stage process: scoping the question, identifying and cleaning the data sources, building the pipeline, and delivering a documented, versioned API endpoint with full methodology. Most projects run four to eight weeks from first conversation to live endpoint.

01
🎯
Discovery Brief
We work with you to define the exact question the dataset needs to answer — the variables, the geographic scope, the time horizon, and the end use (API, dashboard, periodic report, or all three). No assumptions, no boilerplate.
02
🔬
Source Audit
We identify every primary source — public, scraped, or purchasable — that bears on your question. We assess data quality, coverage gaps, and latency characteristics. You receive a written source audit before we build anything, with a frank assessment of what the data can and cannot tell you.
03
⚙️
Pipeline Build
We build the ETL pipeline, NLP scoring, geospatial processing, or statistical modelling layer — depending on what the dataset requires. All pipelines are built on Supabase with GitHub Actions for scheduling, and documented inline. You own the infrastructure.
04
🚀
API Delivery
A versioned REST API endpoint with authenticated access, JSON or CSV output, configurable X/Y axis variables, date range filtering, and a full methodology document. Optional: a private Signal dashboard for visual exploration. Ongoing maintenance contracts available.
What we build

Use cases by sector

Signal's methodology — catastrophe modelling, OSINT, NLP, and geospatial analysis — applies across a range of institutional use cases. These are the sectors where we have built or are actively developing bespoke data products.

Financial Services
ESG & Climate Risk Intelligence
Proprietary ESG sentiment scoring, physical climate risk overlays for asset portfolios, insurance exposure modelling using NOAA and WRI Aqueduct data — all delivered as queryable API endpoints.
Physical risk scores for listed equities by watershed
Supply chain disruption early-warning signals
Climate litigation sentiment tracker
Regulatory risk NLP scoring by jurisdiction
Compliance & Legal
Sanctions & Beneficial Ownership
Cross-referenced entity resolution across OFAC SDN, FinCEN BOI, OpenCorporates, and ICIJ Offshore Leaks. Identify sanctioned entities in supply chains and corporate ownership structures automatically.
Automated OFAC screening for counterparty due diligence
Beneficial ownership chain graph API
Sanctions evasion flag scoring for transaction monitoring
Shell company pattern detection (Delaware / Cayman)
Newsrooms & Publishers
Bespoke Editorial Data Feeds
Private versions of Signal's public trackers — tuned to your specific beat. A climate desk gets a daily wildfire risk dashboard. A financial crime correspondent gets a dark money flow feed. Priced for editorial budgets.
Beat-specific sentiment and anomaly alerts
Embargoed data delivery before publication
Custom geospatial layers for interactive maps
Private Story Finder tuned to your coverage area
Energy & Infrastructure
Grid Stress & Resource Risk
Data pipeline combining EIA generation data, Dominion/ERCOT grid filings, WRI Aqueduct watershed stress, and USGS water data — delivering a real-time infrastructure stress index across any specified geography.
Data centre energy constraint mapping by county
Offshore wind project pipeline and stop-work tracker
Water scarcity vs. cooling demand forecast
Grid capacity auction cost attribution model
Public Affairs & Policy
Political Intelligence Feeds
NLP-scored congressional speech, lobbying flow tracking, dark money beneficial ownership mapping, and regulatory sentiment monitoring — built for policy shops, advocacy organisations, and government affairs teams.
Issue salience tracking by congressional chamber
Lobbying ROI model by bill and industry
Dark money flow tracker (FEC + 990 cross-reference)
Regulatory commentary sentiment by agency
Security & Intelligence
Threat Signal & OSINT APIs
Disinformation spread velocity tracking, dark web mention monitoring, geopolitical water/resource conflict early-warning models, and critical infrastructure dependency mapping — built on open-source intelligence frameworks.
Narrative coordination detection (cross-platform NLP)
Critical mineral supply chain disruption signals
State-actor attribution scoring for information operations
Infrastructure dependency graph (cloud outage mapping)
What you receive

Every project delivers the same standard package

Signal treats every bespoke engagement as a published investigation with a named methodology. You are not buying a black box — you are buying a transparent, documented, reproducible data product that you can audit, replicate, and own.

Deliverable 01
Versioned REST API
A private, authenticated REST API endpoint hosted on Signal's infrastructure (or deployable to your own). Supports JSON and CSV output, configurable X/Y axis variables, date range filtering, and pagination. SLA-backed uptime on maintenance contracts.
  • API key authentication with rate limiting
  • JSON + CSV format support
  • Configurable axis and time range parameters
  • Webhook alerts for anomaly detection triggers
  • Interactive Swagger / OpenAPI documentation
Deliverable 02
Methodology Document
A Signal-standard methodology document covering every source, every transformation step, every assumption baked into the model, and every known limitation. Written for a peer review standard. You can publish it, share it with clients, or file it with regulators.
  • Named primary sources with access URLs
  • Data cleaning and normalisation decisions documented
  • Statistical or NLP model parameters disclosed
  • Known coverage gaps and confidence intervals
  • Update cadence and staleness risk assessment
Deliverable 03
Source Code & Infrastructure
You own everything. Full source code in a private GitHub repository, a Supabase schema with documented tables and indexes, GitHub Actions cron files for pipeline scheduling, and a Dockerfile for self-hosting. Signal retains no proprietary claim on bespoke work.
  • Python ETL scripts, fully commented
  • Supabase PostgreSQL schema with migrations
  • GitHub Actions workflow files
  • Docker + deployment documentation
  • Test suite with coverage report
Deliverable 04
Private Dashboard (Optional)
A private instance of Signal's visualizer — pre-loaded with your dataset, your axis variables, your colour palette, and optionally white-labelled to your organisation. Built with the same Canvas and Chart.js stack as Signal's public tools. Exportable to PNG and PDF.
  • Bar chart, pie chart, and map views
  • Configurable X/Y axis variable selectors
  • Anomaly detection with 2σ highlighting
  • Date range slider with preset windows
  • PNG / PDF export and embeddable iframe
Investment

Transparent project pricing

Signal quotes fixed-price projects after the discovery brief, so there are no scope surprises. The ranges below reflect typical project complexity. Maintenance contracts are priced separately from delivery. All prices in GBP; USD invoicing available.

Tier 01
Data Audit
Source identification, quality assessment, and feasibility report. No build — just the honest brief on what's possible.
£2.5K
Fixed fee · 1–2 weeks · No ongoing commitment
  • Source landscape audit (up to 20 datasets reviewed)
  • Data quality and coverage gap assessment
  • Feasibility report with build cost estimate
  • Kickoff call + written brief
  • Option to proceed to Tier 02 at cost
Get a data audit →
Tier 03
Multi-Source Intelligence
Cross-referenced datasets from multiple sources, with entity resolution, NLP scoring, or geospatial analysis. For institutional-grade intelligence products.
£25K+
Fixed price after scoping · 6–12 weeks
  • Full source audit across 10+ datasets
  • Entity resolution or NLP modelling layer
  • Multi-variable API with anomaly detection
  • Peer-reviewed methodology document
  • Private dashboard with white-label option
  • Maintenance contract available (£800/month)
  • On-call support during delivery
Start scoping →
Reference engagements

Anonymised project examples

Client identities are confidential. These are illustrative examples of the type of bespoke data work Signal has been contracted to deliver or has in active development.

Financial Services · Hedge Fund
ESG Regulatory Sentiment Tracker
Weekly NLP scoring of EU and UK regulatory publications (ESMA, FCA, ECB) for ESG-relevant language, cross-referenced with short interest data for FTSE 350 and EuroStoxx 600 companies. Delivered as a daily JSON feed with anomaly alerts via webhook.
NLP · VADER + fine-tuned BERT 8-week build Daily cadence Tier 03
Infrastructure · Energy Developer
US Data Centre Water Risk API
Facility-level water risk scoring for 370+ US hyperscale data centres — combining USGS streamflow, WRI Aqueduct basin stress, and ICPRB Potomac River data. Used for siting decisions and ESG disclosure reporting. Monthly refresh, county-level resolution.
GIS · PostGIS 6-week build Monthly cadence Tier 02
Legal · International Law Firm
Sanctions Evasion Entity Graph
Cross-referencing OFAC SDN with OpenCorporates and ICIJ Offshore Leaks to identify sanctioned-entity exposure within client supply chains. Delivered as a queryable graph database with a Cypher API endpoint and a PDF due diligence report format.
Entity resolution · Neo4j 10-week build Real-time SDN sync Tier 03
Publishing · National Broadsheet
Private Editorial Data Feed
A private version of Signal's sentiment tracker scoped to five beats — climate, cybersecurity, AI policy, housing, and health — delivered weekly to the newsroom's Slack with a visualiser dashboard for interactive exploration. Priced at editorial budget rate.
Reddit NLP + search velocity 3-week build Weekly cadence Tier 02
Start a conversation

Commission a bespoke dataset

Tell us what you need — a rough brief is enough to start. We respond within one business day with an honest assessment of what's possible and an indicative cost range. No sales calls unless you want them.

We respond within one business day · [email protected] for urgent enquiries
Direct contact
[email protected]
For sensitive briefs, use Signal's encrypted
contact form or arrange a secure call.
Typical turnaround
From brief to live API. These are typical ranges — complex entity resolution or satellite data processing adds time.
Audit: 1–2 wks Tier 02: 4–6 wks Tier 03: 6–12 wks
NDA & Confidentiality
Signal signs mutual NDAs before any brief exchange. All bespoke work is confidential by default — we will not reference the engagement publicly without explicit permission. Source code and data belong to the client on delivery.
What we won't build
Signal does not build datasets for disinformation operations, targeted surveillance, or any application that conflicts with our editorial independence or our published methodology standards. We reserve the right to decline any brief without explanation.
Already a Signal Pro subscriber?
20% discount on all bespoke project tiers. Mention your subscription in your brief or email [email protected] for the discount code before enquiring.