Overview
We are looking for a Software Engineer / Data Engineer to join our team.
What will you do?
Design, build, and operate large-scale Big Data systems, including persistence, partitioning, indexing, and search capabilities. Develop and maintain Java-based applications and APIs. Architect and implement cloud-native solutions using AWS or comparable cloud platforms. Design and optimize high-performance graph databases containing tens of billions of edges. Develop graph traversal capabilities using Apache TinkerPop, Gremlin, JanusGraph, or similar technologies. Build and maintain NoSQL and relational database solutions supporting complex Big Data applications. Design partition and sort key strategies to ensure consistent system performance. Develop aggregation operations to de-duplicate records across continuous data feeds. Build and operate serverless data processing pipelines using AWS Lambda, Step Functions, and PySpark. Design and operate large-scale geospatial indexing solutions using GeoMESA. Develop and maintain Kubernetes-based containerized environments. Implement DevSecOps and agile development practices in production environments. Maintain configuration management using Git-based repositories. Facilitate technical discussions across cross-functional teams to develop mission-aligned implementation strategies. Ensure compliance with federal security, regulatory, and accreditation requirements. Implement data security and governance controls including LDAP integration, encryption, and auditing.
Do you have what it takes?
- Active TS/SCI with Polygraph required.
- Bachelor's degree in Geospatial Intelligence, Geography, Remote Sensing, Intelligence Studies, Engineering, or related field, or equivalent experience
- Demonstrated experience with Java development.
- Experience designing and operating Big Data systems.
- Experience developing and maintaining APIs.
- Experience designing cloud-native architectures using AWS or similar cloud platforms.
- Experience building and optimizing large-scale graph databases using technologies such as Cassandra, DynamoDB, Neo4j, or JanusGraph.
- Experience developing graph traversal capabilities using Apache TinkerPop and Gremlin.
- Experience designing and operating NoSQL solutions.
- Experience in data modeling, partition sharding, stream processing, and metrics gathering.
- Experience developing high-performance data processing pipelines.
- Experience with Kubernetes, Docker, and container orchestration.
- Experience with Apache NiFi.
- Experience implementing DevSecOps and agile methodologies.
- Experience with data security controls including encryption and centralized access management (LDAP).
- Experience working with structured, semi-structured, and unstructured data formats (CSV, JSON, AVRO, Parquet, Protocol Buffers, etc.).
- Experience with relational and NoSQL databases including PostgreSQL, MariaDB, MongoDB, Cassandra, ELK, MinIO, and AWS S3.
- Experience working in Linux environments such as CentOS or Rocky Linux.
- Experience with Python and related libraries.
- Experience supporting large collaboration and development environments
|