Capstone Integrated Solutions Logo

Capstone Integrated Solutions

Senior Data Engineer (AWS)

Posted 5 Hours Ago
Remote
Hiring Remotely in USA
Senior level
Remote
Hiring Remotely in USA
Senior level
Lead data architecture, develop ETL pipelines using AWS Glue, integrate with CMIC ERP, and ensure data quality for AI-driven projects.
The summary above was generated by AI

Capnexus is a comprehensive services provider. Our team consists of outstanding professionals, highly experienced in designing, building, and supporting retail software. We see ourselves as a build-as-a-service provider who follows a repeatable business pattern that can be applied to a variety of platforms and verticals. Having a culture built on outcomes and delivery at the core of the business, Capnexus is providing its customers with a complete suite of services for software development, system analysis, integration, implementation, and support, as well as the option to engage a single team to perform all the services they require. 

Who You Are and What You'll Do: 

Capnexus is looking for a highly skilled Senior AWS Data Engineer to lead data architecture, pipeline development, and ERP integration for a 12-week AI-powered modernization engagement in the construction industry. This role is focused on designing and implementing the data engineering backbone of an intelligent subcontractor pre-qualification platform, including CMIC ERP API integration, Amazon Textract data extraction pipelines, ETL development using AWS Glue, and data quality validation. This is an exciting opportunity to apply advanced cloud data engineering skills on a platform that leverages generative AI to automate and modernize enterprise workflows. 

Responsibilities: 

  • Lead data architecture design, API assessment, and ETL requirements gathering during the Discovery & Design phase. 
  • Develop and configure CMIC ERP API integration to establish reliable data exchange between the ERP system and the AWS platform. 
  • Design and implement data pipelines using AWS Glue for ETL processing of subcontractor documents and ERP data. 
  • Integrate Amazon Textract to extract structured data from insurance certificates, bonding letters, and financial documents. 
  • Build and maintain data models to support AI-powered validation, risk profiling, and executive reporting. 
  • Configure Amazon S3 data lake architecture to store and manage raw, processed, and curated data assets. 
  • Implement AWS Lambda and AWS Step Functions to orchestrate data workflows and automated processing pipelines. 
  • Develop and expose data through Amazon API Gateway to support application and dashboard consumption. 
  • Ensure data quality, validation, and integrity across all integration points and pipeline outputs. 
  • Conduct data integration testing and support user acceptance testing (UAT) for data-dependent features. 
  • Collaborate with Full Stack, AI/ML, and DevOps team members to ensure seamless end-to-end data flows. 
  • Contribute to knowledge transfer documentation, data pipeline runbooks, and operations guides. 

 

Qualifications: 

  • 5+ years of data engineering experience, with at least 2+ years working in AWS cloud environments. 
  • Strong proficiency in SQL and experience with AWS database services including RDS, DynamoDB, and Aurora. 
  • Hands-on experience with AWS Glue for ETL development and data pipeline orchestration. 
  • Experience integrating with ERP systems or enterprise APIs, with ability to assess and document API capabilities. 
  • Working knowledge of Amazon Textract for automated document data extraction. 
  • Proficiency with Amazon S3 for data lake design, storage management, and lifecycle policies. 
  • Experience with AWS Lambda and AWS Step Functions for serverless data workflow orchestration. 
  • Solid understanding of data modeling, data quality validation, and integration testing practices. 
  • Familiarity with Amazon API Gateway for exposing data services to downstream consumers. 
  • Strong analytical, problem-solving, and communication skills with the ability to work in Agile/Scrum teams. 

 

Nice to Have: 

  • Experience with Amazon Bedrock or AI/ML-integrated data pipelines. 
  • Familiarity with CMIC ERP or similar construction industry enterprise resource planning systems. 
  • Experience with GraphQL APIs or AWS AppSync. 
  • Knowledge of TypeScript or Python for pipeline scripting and Lambda function development. 
  • AWS Certification (Data Analytics Specialty, Database Specialty, or Solutions Architect). 
  • Background in construction, insurance, or financial services industries. 

 

"Our Culture": 

At Capstone, the central principles that we all adhere to, and the glue that holds us together, are our keystones. Our four keystones are: 

"A Customer Obsessed, Delivery Focused, Culture" 

  • We’re driven to exceed our customers’ expectations by listening, leading, solving problems, and delivering what we promise 
  • We aim to be the most dependable and trusted partner serving our customers. TRUST = CONSISTENCY x TIME 

"A Culture of Learning and Sharing" 

  • We value “Lifetime Learners”; those who are hungry, competitive, curious, and self-motivated in their pursuit of knowledge. 
  • Personal and professional growth depends on teamwork and continuous learning. By sharing knowledge, skills, ideas, and effort, we benefit our customers, ourselves, and our communities. 
  • We recognize that the thoughts, feelings, and backgrounds of others are as important as our own. Everyone has something to learn and everyone has something they can teach. 
  • Knowledge and ability are valued. Sharing knowledge and helping others learn new capabilities is valued exponentially. 

"A Culture of Growth and Scalability" 

  • Growth comes from not establishing barriers in your role. “Cross functional skill sets are valued and help us deliver to our customers in a truly agile fashion. It comes with understanding that when asked to do something new, you will need support, have questions, and make some mistakes along the way. 
  • The most elegant solution is a simple solution. Simple doesn’t mean easy. It’s often more difficult to break a complex problem down into simple, scalable terms. We don’t appreciate, or value, over architected solutions or superfluous coding. 
  • Time is one of our most precious commodities. Scalability implies being respectful of this and passionate about making the most efficient use of each and every one of our team members time. 

"All Work is Strategic" 

  • No matter how small a project or assignment appears, every single engagement is an opportunity for us to prove ourselves, build trust, and develop relationships that last and grow 
  • Every task, interaction, and commitment matters 
  • Big or small, we execute our plans and strategies with focus, commitment, and passion 

 

We offer: 

Job Type: Full-time, 1099

Duration: Short-term

Benefits: 

  • Remote work 

 

Capnexus is an equal opportunity employer. We embrace and celebrate diversity and are committed to creating an inclusive and safe environment for all employees. Experience comes in many forms, and we’re dedicated to adding new perspectives to the team. We encourage you to apply even if your experience doesn’t perfectly align with what we have listed. We look forward to hearing from you. 

No Agencies Please! 

Top Skills

Amazon Api Gateway
Amazon S3
Amazon Textract
Aurora
Aws Glue
Aws Lambda
Aws Step Functions
DynamoDB
Rds
SQL

Similar Jobs

10 Days Ago
Remote
USA
Senior level
Senior level
Artificial Intelligence • Information Technology • Software • Database
Build production-grade AI-powered data tooling: extract data from Snowflake, generate and store embeddings, enable semantic search, design enrichment pipelines using LLM APIs, optimize AWS infrastructure, and create reusable services and SDKs for scalable, observable data and AI workflows.
Top Skills: AWSEcsEksEmbeddingsIamLambdaLlm ApisOpenaiPineconePythonS3Semantic SearchSnowflakeVector Databases
7 Days Ago
In-Office or Remote
Senior level
Senior level
Information Technology • Software
Design and operate enterprise-grade data pipelines using Python and AWS services. Responsibilities include ETL development, cloud architecture design, and collaboration with stakeholders.
Top Skills: AirflowAWSDockerEventbridgeFastapiFlaskGlueKinesisKubernetesLambdaNumpyPandasPrefectPysparkPythonS3SqlalchemyStep FunctionsTerraform
15 Days Ago
In-Office or Remote
Mid level
Mid level
Database
As an AWS Data Engineer, you will develop data pipelines, create data models, troubleshoot issues, and collaborate with teams while utilizing tools like PySpark and SQL.
Top Skills: AirflowAmazon AthenaAmazon EmrAmazon GlueAmazon RedshiftPrestoPysparkSnowflakeSQL

What you need to know about the Colorado Tech Scene

With a business-friendly climate and research universities like CU Boulder and Colorado State, Colorado has made a name for itself as a startup ecosystem. The state boasts a skilled workforce and high quality of life thanks to its affordable housing, vibrant cultural scene and unparalleled opportunities for outdoor recreation. Colorado is also home to the National Renewable Energy Laboratory, helping cement its status as a hub for renewable energy innovation.

Key Facts About Colorado Tech

  • Number of Tech Workers: 260,000; 8.5% of overall workforce (2024 CompTIA survey)
  • Major Tech Employers: Lockheed Martin, Century Link, Comcast, BAE Systems, Level 3
  • Key Industries: Software, artificial intelligence, aerospace, e-commerce, fintech, healthtech
  • Funding Landscape: $4.9 billion in VC funding in 2024 (Pitchbook)
  • Notable Investors: Access Venture Partners, Ridgeline Ventures, Techstars, Blackhorn Ventures
  • Research Centers and Universities: Colorado School of Mines, University of Colorado Boulder, University of Denver, Colorado State University, Mesa Laboratory, Space Science Institute, National Center for Atmospheric Research, National Renewable Energy Laboratory, Gottlieb Institute

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account