Job Requirements
McLean, VA
Intel Agency (NSA, CIA, FBI, etc) Full Scope Polygraph
Mid Level Career (5+ yrs experience)
Salary not specified
Join Premium to unlock estimated salaries
Job Description
Opportunity via QSSHire | Talent-as-a-Service Recruiting
Role: Data & Software Engineer
Location: McLean, VA (Onsite)
Clearance: TS/SCI with Full Scope Polygraph
Rate: Open
________________________________________
Are You a Data & Software Engineer Ready to Build Scalable Data Pipelines and Modern Data Platforms?
QSSHire is seeking a Data & Software Engineer to support a Project, focused on building complex data flows and scalable data solutions for a custom application.
This role requires a highly skilled engineer with strong Python expertise, experience building production-grade ETL pipelines, and deep knowledge of data governance, security, and compliance principles. The ideal candidate will bring experience working with modern data platforms, cloud-native technologies, and large-scale data processing frameworks.
________________________________________
In This Role, You’ll:
• Design, build, and maintain end-to-end data pipelines using Python
• Develop and deploy data workflows using orchestration tools (e.g., Airflow, Spark job orchestration)
• Containerize and deploy applications in AWS cloud environments
• Configure and optimize Spark and PySpark jobs for large-scale data processing
• Work with stakeholders to understand requirements and design scalable data solutions with minimal oversight
• Troubleshoot data quality issues, pipeline failures, and performance bottlenecks
• Support large-scale data migration and platform modernization efforts
• Optimize relational databases (MySQL, PostgreSQL) for analytical workloads, including schema design and query tuning
• Implement and maintain data lineage, cataloging, and governance solutions
• Work with geospatial data formats and tools
• Integrate AI/ML models and services into data pipelines
• Develop automation scripts using Bash for data processing and system tasks
• Contribute to data engineering documentation, standards, and best practices
________________________________________
You’ll Succeed Here If You Have:
• 5+ years of experience in data engineering or software engineering roles
• Strong experience with Apache Spark & PySpark
• Advanced proficiency in Python (Pandas, NumPy)
• Experience building scalable ETL/data pipelines in production environments
• Hands-on experience with AWS services (S3, Lambda, Step Functions)
• Experience with containerization tools (Docker, Podman)
• Strong SQL skills, including experience with Trino
• Experience with NoSQL databases (DynamoDB)
• Familiarity with data lake technologies (Apache Iceberg)
• Experience with data orchestration tools (Airflow or similar)
• Experience using Terraform or CloudFormation for infrastructure as code
• Experience with data lineage and governance tools (OpenLineage, Unity Catalog OSS, Apache Polaris)
• Experience with Apache Superset for data visualization
• Experience with geospatial technologies (H3, PostGIS)
• Strong understanding of version control and DevOps practices (Git, IaC workflows)
• Experience working with data catalogs and diverse data formats
________________________________________
Bonus Points If You Have:
• Experience integrating AI/ML models into data workflows
• Experience supporting data platform modernization initiatives
• Strong background in data governance, privacy, and compliance frameworks
• Experience working in agile, fast-paced environments
________________________________________
Why QSSHire?
As a modern Talent-as-a-Service recruiting partner, QSSHire connects top talent with impactful opportunities. We focus on aligning your expertise with meaningful work, offering transparency, flexibility, and career growth.
Role: Data & Software Engineer
Location: McLean, VA (Onsite)
Clearance: TS/SCI with Full Scope Polygraph
Rate: Open
________________________________________
Are You a Data & Software Engineer Ready to Build Scalable Data Pipelines and Modern Data Platforms?
QSSHire is seeking a Data & Software Engineer to support a Project, focused on building complex data flows and scalable data solutions for a custom application.
This role requires a highly skilled engineer with strong Python expertise, experience building production-grade ETL pipelines, and deep knowledge of data governance, security, and compliance principles. The ideal candidate will bring experience working with modern data platforms, cloud-native technologies, and large-scale data processing frameworks.
________________________________________
In This Role, You’ll:
• Design, build, and maintain end-to-end data pipelines using Python
• Develop and deploy data workflows using orchestration tools (e.g., Airflow, Spark job orchestration)
• Containerize and deploy applications in AWS cloud environments
• Configure and optimize Spark and PySpark jobs for large-scale data processing
• Work with stakeholders to understand requirements and design scalable data solutions with minimal oversight
• Troubleshoot data quality issues, pipeline failures, and performance bottlenecks
• Support large-scale data migration and platform modernization efforts
• Optimize relational databases (MySQL, PostgreSQL) for analytical workloads, including schema design and query tuning
• Implement and maintain data lineage, cataloging, and governance solutions
• Work with geospatial data formats and tools
• Integrate AI/ML models and services into data pipelines
• Develop automation scripts using Bash for data processing and system tasks
• Contribute to data engineering documentation, standards, and best practices
________________________________________
You’ll Succeed Here If You Have:
• 5+ years of experience in data engineering or software engineering roles
• Strong experience with Apache Spark & PySpark
• Advanced proficiency in Python (Pandas, NumPy)
• Experience building scalable ETL/data pipelines in production environments
• Hands-on experience with AWS services (S3, Lambda, Step Functions)
• Experience with containerization tools (Docker, Podman)
• Strong SQL skills, including experience with Trino
• Experience with NoSQL databases (DynamoDB)
• Familiarity with data lake technologies (Apache Iceberg)
• Experience with data orchestration tools (Airflow or similar)
• Experience using Terraform or CloudFormation for infrastructure as code
• Experience with data lineage and governance tools (OpenLineage, Unity Catalog OSS, Apache Polaris)
• Experience with Apache Superset for data visualization
• Experience with geospatial technologies (H3, PostGIS)
• Strong understanding of version control and DevOps practices (Git, IaC workflows)
• Experience working with data catalogs and diverse data formats
________________________________________
Bonus Points If You Have:
• Experience integrating AI/ML models into data workflows
• Experience supporting data platform modernization initiatives
• Strong background in data governance, privacy, and compliance frameworks
• Experience working in agile, fast-paced environments
________________________________________
Why QSSHire?
As a modern Talent-as-a-Service recruiting partner, QSSHire connects top talent with impactful opportunities. We focus on aligning your expertise with meaningful work, offering transparency, flexibility, and career growth.
group id: 91142086