Backend Developer – Django / PostgreSQL

Remote Full-time
The system ingests operational data, computes industrial KPIs, generates structured AI insights, and exposes deterministic APIs for a mobile application.

This role is strictly backend-focused. No frontend work is included.

Backend Architecture

The platform is built on:
• Django + Django REST Framework
• PostgreSQL with ELT structure: raw to staging to analytics
• Celery + Redis for task orchestration
• Stripe for billing boundary, already scoped separately
• Docker-based deployment

Core Architectural Principles
• Multi-tenant isolation at organisation and site level
• Deterministic KPI recomputation
• Append-only raw data layer
• Strict schema validation for ingestion
• Versioned KPI logic
• AI outputs must be grounded in stored data
• No autonomous AI actions, advisory only

Backend Responsibilities High-Level

1. Data Ingestion Layer
• Build a robust CSV ingestion pipeline
• Implement header validation and schema enforcement
• Ensure idempotent file handling with no duplicate ingestion
• Transform raw data into the canonical ProductionFact model
• Maintain ingestion logs and validation reports

2. Manufacturing Data Model Refinement

Refactor the ProductionFact schema to support:
• Workcenter context
• SKU and job granularity
• Structured downtime categorisation
• Cost attribution fields

Additionally:
• Implement canonical master data tables
• Enforce referential integrity

3. KPI Engine Industrial-Grade
• Correct OEE computation including availability, performance, and quality
• Implement structured downtime loss logic
• Build reliability metrics foundation using event-based design
• Ensure deterministic recompute capability
• Support time-series aggregation

4. Dashboard APIs
• Expose pre-computed KPI endpoints
• Implement cached read APIs
• Support filtering by site, shift, and workcenter
• Enforce entitlement gating

5. AI Insight Layer Backend Only

Generate and store:
• AI Suggestions
• AI Improvements
• AI Insights

Additionally:
• Ensure traceability to source data
• Cache AI outputs
• No frontend integration required

6. Task Orchestration

Implement Celery task chains:

validate to transform to ingest to compute KPIs to generate AI

Also include:
• Scheduled ingestion support
• Idempotent task handling

Phase 3 – Manufacturing Intelligence Expansion

1. Job-Level Margin Foundation Complete Implementation

Data Model Expansion

Extend the schema with a dedicated JobPerformance model. Do not overload ProductionFact.

The model must include:
• job_id indexed and tenant-scoped
• site_id
• workcenter_id
• sku_id
• quoted_revenue
• quoted_material_cost
• quoted_labour_cost
• quoted_overhead_cost
• actual_material_cost
• actual_labour_cost
• allocated_overhead_cost
• downtime_cost
• scrap_cost
• revenue_recognised
• job_status
• job_start_date
• job_end_date

All monetary fields must use Decimal with currency support.

Margin Calculations Deterministic

Implement:

Actual Margin equals revenue_recognised minus actual_material plus actual_labour plus allocated_overhead plus downtime_cost plus scrap_cost.

Quoted Margin equals quoted_revenue minus quoted_material plus quoted_labour plus quoted_overhead.

Margin Variance percentage equals Actual minus Quoted divided by Quoted.

Margin Erosion Attribution must break down percentage erosion into:
• Scrap contribution
• Downtime contribution
• Labour overrun
• Material price variance

All formulas must be versioned and logged.

---

Margin APIs

Build:
• api margin job job_id
• api margin site site_id
• api margin summary

Responses must include:
• Margin values
• Variance percentage
• Erosion breakdown
• Financial impact
• Data lineage metadata

All results must be cacheable and recomputable.

2. Cost Attribution Logic Production-Grade

Deterministic Cost Model

Implement a cost engine with:

Material per good unit equals actual_material_cost divided by good_units.

Labour per runtime hour equals actual_labour_cost divided by runtime_hours.

Overhead allocation must support configurable methods:
• Per shift
• Per runtime hour
• Per job

A configuration table must define the allocation rule per tenant.

KPI Endpoints

Build:
• api kpi cost-per-unit
• api kpi cost-variance
• api kpi unit-economics

All endpoints must support filtering by:
• site
• workcenter
• sku
• job
• time range

All responses must include formula version and input data range.

3. Cross-Site Normalised Benchmarking Internal

Normalisation Rules

Standardise:
• OEE time-weighted
• Scrap percentage
• Cost per unit

Ensure:
• Comparable time ranges
• Comparable shift hours
• Currency normalisation

Percentile Logic

For each KPI:
• Compute distribution across sites
• Assign percentile rank
• Flag top performer
• Flag bottom performer
• Flag above or below median

Store benchmarking snapshots for reproducibility.

Benchmark APIs

Build:
• api benchmark kpi kpi_name
• api benchmark site site_id

Responses must return:
• Rank
• Percentile
• Group average
• Variance from average
• Financial delta if site matched top quartile

4. Economic Impact Layer Mandatory

Every KPI endpoint must optionally include:
• Economic impact value
• Impact calculation logic
• Time range used

Examples:

Scrap impact equals scrap_units multiplied by material_cost_per_unit.

Downtime impact equals downtime_minutes multiplied by cost_per_minute.

OEE delta impact equals lost throughput multiplied by contribution margin.

Impact values must be stored in the analytics layer for audit.

Add an economic_impact object in API responses.

5. AI Grounding and Traceability Production-Ready

Every AI output must store:
• ai_output_id
• organisation_id
• related_kpi_id
• source_table_names
• source_record_ids
• time_range
• kpi_version
• prompt_snapshot
• structured_input_data_snapshot
• model_name
• generation_timestamp

No AI output may exist without lineage.

Audit Endpoint

Build:
• api ai audit ai_output_id

Return:
• Full citation trail
• KPI inputs used
• Raw data reference
• Formula version
• Economic impact linkage

This ensures defensibility under regulatory scrutiny.

6. Industrial Readiness and Maturity Scoring

Implement a scoring engine with inputs:
• Percentage data completeness
• KPI coverage ratio
• Margin model activation
• Benchmarking availability
• Historical depth of data

Output:
• 0 to 100 maturity score
• Tier classification: Foundational, Structured, Optimised

Expose:
• api readiness organisation

Score must be recomputable and transparent.

Phase 3 Outcome

After completion, Exec App will provide:
• True job-level economic diagnostics
• Deterministic cost engine
• Internal benchmarking
• Financial impact visibility
• Audit-ready AI outputs
• Organisational maturity scoring

Documentation and Validation
• Postman collection
• API documentation
• Proof of idempotency
• Migration discipline with no schema corruption
• Clean README with setup steps

What Is Not Included
• React Native frontend
• Mobile UI
• Website or marketing pages
• App store deployment
• DevOps infrastructure build-out, Docker assumed

Required Experience
• Django + DRF at production level
• PostgreSQL schema design
• Celery + Redis
• Multi-tenant SaaS backend architecture
• Clean migration management
• API design discipline

Timeline and Budget

Timeline: 4 to 6 weeks preferred, milestone-based delivery.

Total Budget: 300 dollars. No negotiation. More work to follow.

Apply tot his job

Apply To this Job
Apply Now

Similar Opportunities

Experienced Registered Behavior Technician for In-Home ABA Therapy - Atlanta, GA

Remote Full-time

Immediate Hiring: Experienced Registered Behavioral Technician (RBT) for Clinic-Based ABA Therapy Services

Remote Full-time

Experienced Registered Behavioral Technician (RBT) - ABA Therapy for Children with Autism Spectrum Disorder

Remote Full-time

Experienced Registered Nurse - Telehealth: Providing Remote Care Coordination and Patient Support

Remote Full-time

Experienced Substitute Teacher for Riverside County Schools - Join Scoot Education's Innovative Team

Remote Full-time

Experienced Substitute Teacher for San Bernardino County - Flexible Schedules & Competitive Pay

Remote Full-time

Experienced School Year Instructional Coach for High-Dosage Tutoring Programs in Edgewater Park, NJ

Remote Full-time

Experienced School Year Tutor for K-8 Students in Math and Literacy - Mickleton, NJ

Remote Full-time

Experienced Secondary Social Studies Teacher for Kansas - Flexible Hybrid Remote Arrangement

Remote Full-time

USPS Office Helper

Remote Full-time

Independent Freight Agents – 1099 - Work from Home - Flexible Schedule

Remote Full-time

Coordinator, Tech Customer Service

Remote Full-time

Business Adjunct Instructor (REMOTE)

Remote Full-time

Contractor Telehealth Nurse Practitioner or Physician Assistant, TX

Remote Full-time

[Remote] Associate Vice President – Sports Public Relations

Remote Full-time

**Experienced Chat Support Agent – Entry-Level, No Degree Required – Flexible Remote Work Opportunity**

Remote Full-time

Experienced Sorting Facility Associate - Mail Processing and Delivery Operations

Remote Full-time

Azure DevOps, GitHub, C# Developer - 100% Remote - Contract (Client in Concord, NH) - B4071B

Remote Full-time

Patient Care Customer Service Representative - Remote Evening and Weekend Opportunities for Delivering Exceptional Patient Experiences

Remote Full-time

VP, Resolution Services

Remote Full-time
← Back to Home