- Arganteal accepts applications from direct candidates only.
We do not work with third-party recruiters or staffing agencies - Required Country Location: Costa Rica, Peru, Argentina, Brazil, Columbia, South Africa, Mexico, or Panama
- This is full time work at 40 hours per week
Overview
Our client seeks a motivated
Senior Developer, Data & AI
to join their team in developing a groundbreaking, modular platform built from the ground up.
Our client digitizes and contextualizes
multi-modal sensor data
from both digital and physical environments into specialized
time-series, graph, and vector databases
, powering real-time analytics, compliance, and AI-driven context mapping.
This role is ideal for a creative problem-solver who thrives at the intersection of
data engineering, distributed systems, and applied AI
.
Key Responsibilities
Platform Design & Development
- Architect, develop, and deploy core modules (Data, Access, & Agent's) for end-to-end data ingestion, contextualization, and visualization
- Design and code sensor collection agents across heterogeneous systems (Windows, Linux, macOS, mobile, IoT)
- Implement real-time ingestion pipelines using technologies like Apache Kafka, Apache NiFi, Redis Streams, or AWS Kinesis
- Persist and query multi-modal data across time-series (MongoDB, InfluxDB, TimescaleDB), graph (Neo4j), and vector databases (Qdrant, FAISS, Pinecone, or Weaviate)
API & Data Access Layer
- Build secure, scalable RESTful and GraphQL APIs for exposing platform data models, sensor configuration, and reporting
- Implement a unified Database Access Layer (DBAL) to abstract query logic across multiple databases
- Experiment with or extend Model Context Protocol (MCP) or a similar standardized data interchange for multi-DB, multi-agent interoperability
System Integration & Data Streaming
- Develop low-latency data pipelines for transporting and transforming event streams (syslog, telemetry, keystrokes, IoT feeds, cloud service logs)
- Collaborate with frontend engineers to connect Access (visual mapping UI) with back-end pipelines
Optimization & Scalability
- Optimize database query performance using down-sampling, partitioning, and caching techniques
- Design solutions for horizontal scaling and containerized deployment (Docker, Kubernetes, OpenShift)
- Apply a MacGyver-mindset for rapid prototyping and iterative refinement under real-world constraints
Collaboration & Mentoring
- Work directly with compliance officers, security analysts, and business process owners to refine data models for regulatory and operational needs
- Conduct code reviews, mentor junior developers, and promote best practices across the team
Required Skills & Experience
- Programming: Strong proficiency in and Python (C++ a plus)
- Streaming: Hands-on experience with Kafka, NiFi, Redis Streams, or AWS Kinesis
Databases:
Time-series: MongoDB, InfluxDB, TimescaleDB, or AWS Timestream
- Graph: Neo4j (Cypher, APOC, graph schema design)
Vector: Qdrant, FAISS, Pinecone, or Weaviate
AI/Agents: Experience with—or strong interest in—Agentic AI frameworks, multi-agent orchestration, and context-aware data processing
- Data Interchange: Familiarity with MCP-like protocols or interest in defining standardized APIs for cross-database access
- Cloud/Infra: AWS, Azure, or GCP with containerization (Docker, Kubernetes)
- Software Engineering: Strong grasp of algorithms, distributed systems, microservice design, and API security
- Problem Solving: Strong debugging skills, creative mindset, and ability to balance speed with scalability
Preferred Skills
- Machine Learning/NLP integration into multi-modal pipelines
- CI/CD automation and DevOps practices
- Knowledge of enterprise integration patterns, event-driven systems, and zero-trust security models
- Experience with compliance frameworks (NERC CIP, FedRAMP, GDPR, SOX)
Qualifications
- Bachelor's degree in Computer Science, Engineering, or related field (or equivalent hands-on experience)
- 5+ years professional software development with data-intensive or AI-driven systems
- Proven experience designing, deploying, and scaling modular platforms in production
- Arganteal accepts applications from direct candidates only.
We do not work with third-party recruiters or staffing agencies
Powered by JazzHR
RKYogGysFW