Docupace · 19 hours ago
Palantir Foundry Data Engineer
Docupace is seeking a Senior Data Engineer to design, build, and operate the foundational data infrastructure behind their Palantir Foundry platform. This role is central to enabling analytics, automation, and AI capabilities across the wealth-management ecosystem.
Responsibilities
Design and implement Foundry pipelines from custodian APIs, internal databases, and third-party providers using Pipeline Builder and Code Repositories (Python/PySpark/SQL)
Model and maintain the Foundry Ontology (objects, properties, links) for core domains: Clients, Households, Accounts, Transactions, Advisors, Documents/Forms
Expose curated datasets/objects for Workshop apps, reporting, and programmatic access (e.g., GraphQL/REST)
Define validation rules and monitors for NIGO (Not-In-Good-Order) detection, compliance surveillance, and NAO (New Account Opening) workflows
Implement schema evolution, data lineage, and documentation; ensure traceability from source to Ontology
Enforce data access controls and masking for PII; support audit readiness (FINRA/SEC) and privacy (GDPR/GLBA)
Establish SLAs for freshness and availability; schedule jobs, implement alerting, and manage backfills
Own incident response/runbooks and cost-aware compute usage; use Git-based workflows and code reviews
Partner with Infra/Sec to align IAM, secrets, and environment configuration
Translate product/compliance use cases into data contracts; coordinate with upstream engineering on API quality
Enable downstream builders (Workshop apps, reports, AI features) via semantic layers and clear documentation
Provide office hours and training for analysts and product teams
Qualification
Required
5+ years in data engineering with distributed processing (Spark/PySpark) and advanced SQL
Hands-on Palantir Foundry experience (≥1 year building pipelines and Ontology) or demonstrably fast ramp with strong DE background on comparable stacks (e.g., Databricks/Spark)
Strong proficiency in SQL — Foundry heavily relies on SQL-like operations for data transformations (similar to pipelines in dbt or Snowflake)
Experience with data lineage and versioned transformations
Proven data modeling across relational and graph/relationship domains
Comfort working with Spark, Hadoop, or other distributed compute engines
Knowledge of data partitioning, optimization, and parallelized workloads
API integration (REST/GraphQL), including auth, pagination, retries, and schema change handling
Testing & CI/CD for data (unit tests, data quality checks, code reviews) and strong Git hygiene
Excellent communication with product, compliance, and engineering partners
Preferred
Wealth management/fintech or regulated industry experience; familiarity with FINRA/SEC workflows
Multi-custodian data (e.g., Schwab, Fidelity, Pershing), account opening, NIGO resolution, surveillance
AWS data services and Infra-as-Code (e.g., S3/IAM/Glue/Terraform)
Streaming (Kafka) and/or document/OCR pipelines for forms and unstructured data
Palantir training/certifications
Benefits
Competitive compensation
Growth in a modern data/AI platform environment.
Company
Docupace
Docupace is a solutions provider focused on digitizing and automating operations in the financial advice and investment industry.
Funding
Current Stage
Growth StageTotal Funding
$19MKey Investors
FTV Capital
2024-07-23Acquired
2020-04-28Private Equity
2018-01-09Undisclosed· $2.5M
Recent News
2025-11-04
2025-10-14
Company data provided by crunchbase