Capstone Integrated Solutions Logo

Capstone Integrated Solutions

Senior Data Engineer (AWS)

Posted 8 Days Ago
Remote
Hiring Remotely in USA
Senior level
Remote
Hiring Remotely in USA
Senior level
Lead data architecture, develop ETL pipelines using AWS Glue, integrate with CMIC ERP, and ensure data quality for AI-driven projects.
The summary above was generated by AI

Capnexus is a comprehensive services provider. Our team consists of outstanding professionals, highly experienced in designing, building, and supporting retail software. We see ourselves as a build-as-a-service provider who follows a repeatable business pattern that can be applied to a variety of platforms and verticals. Having a culture built on outcomes and delivery at the core of the business, Capnexus is providing its customers with a complete suite of services for software development, system analysis, integration, implementation, and support, as well as the option to engage a single team to perform all the services they require. 

Who You Are and What You'll Do: 

Capnexus is looking for a highly skilled Senior AWS Data Engineer to lead data architecture, pipeline development, and ERP integration for a 12-week AI-powered modernization engagement in the construction industry. This role is focused on designing and implementing the data engineering backbone of an intelligent subcontractor pre-qualification platform, including CMIC ERP API integration, Amazon Textract data extraction pipelines, ETL development using AWS Glue, and data quality validation. This is an exciting opportunity to apply advanced cloud data engineering skills on a platform that leverages generative AI to automate and modernize enterprise workflows. 

Responsibilities: 

  • Lead data architecture design, API assessment, and ETL requirements gathering during the Discovery & Design phase. 
  • Develop and configure CMIC ERP API integration to establish reliable data exchange between the ERP system and the AWS platform. 
  • Design and implement data pipelines using AWS Glue for ETL processing of subcontractor documents and ERP data. 
  • Integrate Amazon Textract to extract structured data from insurance certificates, bonding letters, and financial documents. 
  • Build and maintain data models to support AI-powered validation, risk profiling, and executive reporting. 
  • Configure Amazon S3 data lake architecture to store and manage raw, processed, and curated data assets. 
  • Implement AWS Lambda and AWS Step Functions to orchestrate data workflows and automated processing pipelines. 
  • Develop and expose data through Amazon API Gateway to support application and dashboard consumption. 
  • Ensure data quality, validation, and integrity across all integration points and pipeline outputs. 
  • Conduct data integration testing and support user acceptance testing (UAT) for data-dependent features. 
  • Collaborate with Full Stack, AI/ML, and DevOps team members to ensure seamless end-to-end data flows. 
  • Contribute to knowledge transfer documentation, data pipeline runbooks, and operations guides. 

 

Qualifications: 

  • 5+ years of data engineering experience, with at least 2+ years working in AWS cloud environments. 
  • Strong proficiency in SQL and experience with AWS database services including RDS, DynamoDB, and Aurora. 
  • Hands-on experience with AWS Glue for ETL development and data pipeline orchestration. 
  • Experience integrating with ERP systems or enterprise APIs, with ability to assess and document API capabilities. 
  • Working knowledge of Amazon Textract for automated document data extraction. 
  • Proficiency with Amazon S3 for data lake design, storage management, and lifecycle policies. 
  • Experience with AWS Lambda and AWS Step Functions for serverless data workflow orchestration. 
  • Solid understanding of data modeling, data quality validation, and integration testing practices. 
  • Familiarity with Amazon API Gateway for exposing data services to downstream consumers. 
  • Strong analytical, problem-solving, and communication skills with the ability to work in Agile/Scrum teams. 

 

Nice to Have: 

  • Experience with Amazon Bedrock or AI/ML-integrated data pipelines. 
  • Familiarity with CMIC ERP or similar construction industry enterprise resource planning systems. 
  • Experience with GraphQL APIs or AWS AppSync. 
  • Knowledge of TypeScript or Python for pipeline scripting and Lambda function development. 
  • AWS Certification (Data Analytics Specialty, Database Specialty, or Solutions Architect). 
  • Background in construction, insurance, or financial services industries. 

 

"Our Culture": 

At Capstone, the central principles that we all adhere to, and the glue that holds us together, are our keystones. Our four keystones are: 

"A Customer Obsessed, Delivery Focused, Culture" 

  • We’re driven to exceed our customers’ expectations by listening, leading, solving problems, and delivering what we promise 
  • We aim to be the most dependable and trusted partner serving our customers. TRUST = CONSISTENCY x TIME 

"A Culture of Learning and Sharing" 

  • We value “Lifetime Learners”; those who are hungry, competitive, curious, and self-motivated in their pursuit of knowledge. 
  • Personal and professional growth depends on teamwork and continuous learning. By sharing knowledge, skills, ideas, and effort, we benefit our customers, ourselves, and our communities. 
  • We recognize that the thoughts, feelings, and backgrounds of others are as important as our own. Everyone has something to learn and everyone has something they can teach. 
  • Knowledge and ability are valued. Sharing knowledge and helping others learn new capabilities is valued exponentially. 

"A Culture of Growth and Scalability" 

  • Growth comes from not establishing barriers in your role. “Cross functional skill sets are valued and help us deliver to our customers in a truly agile fashion. It comes with understanding that when asked to do something new, you will need support, have questions, and make some mistakes along the way. 
  • The most elegant solution is a simple solution. Simple doesn’t mean easy. It’s often more difficult to break a complex problem down into simple, scalable terms. We don’t appreciate, or value, over architected solutions or superfluous coding. 
  • Time is one of our most precious commodities. Scalability implies being respectful of this and passionate about making the most efficient use of each and every one of our team members time. 

"All Work is Strategic" 

  • No matter how small a project or assignment appears, every single engagement is an opportunity for us to prove ourselves, build trust, and develop relationships that last and grow 
  • Every task, interaction, and commitment matters 
  • Big or small, we execute our plans and strategies with focus, commitment, and passion 

 

We offer: 

Job Type: Full-time, 1099

Duration: Short-term

Benefits: 

  • Remote work 

 

Capnexus is an equal opportunity employer. We embrace and celebrate diversity and are committed to creating an inclusive and safe environment for all employees. Experience comes in many forms, and we’re dedicated to adding new perspectives to the team. We encourage you to apply even if your experience doesn’t perfectly align with what we have listed. We look forward to hearing from you. 

No Agencies Please! 

Similar Jobs

3 Days Ago
In-Office or Remote
Senior level
Senior level
Information Technology • Software
The Senior Data Platform Engineer leads the design and management of data pipelines using Python and AWS, focusing on data engineering for finance-related analytics and systems.
Top Skills: AirflowAWSCdkCloudFormationDockerEcsEventbridgeFastapiFlaskGlueKinesisKubernetesLambdaNumpyPandasPrefectPysparkPytestPythonS3SqlalchemyStep FunctionsTerraform
18 Days Ago
Remote
USA
Senior level
Senior level
Artificial Intelligence • Information Technology • Software • Database
Build production-grade AI-powered data tooling: extract data from Snowflake, generate and store embeddings, enable semantic search, design enrichment pipelines using LLM APIs, optimize AWS infrastructure, and create reusable services and SDKs for scalable, observable data and AI workflows.
Top Skills: AWSEcsEksEmbeddingsIamLambdaLlm ApisOpenaiPineconePythonS3Semantic SearchSnowflakeVector Databases
23 Days Ago
In-Office or Remote
Mid level
Mid level
Database
As an AWS Data Engineer, you will develop data pipelines, create data models, troubleshoot issues, and collaborate with teams while utilizing tools like PySpark and SQL.
Top Skills: AirflowAmazon AthenaAmazon EmrAmazon GlueAmazon RedshiftPrestoPysparkSnowflakeSQL

What you need to know about the Austin Tech Scene

Austin has a diverse and thriving tech ecosystem thanks to home-grown companies like Dell and major campuses for IBM, AMD and Apple. The state’s flagship university, the University of Texas at Austin, is known for its engineering school, and the city is known for its annual South by Southwest tech and media conference. Austin’s tech scene spans many verticals, but it’s particularly known for hardware, including semiconductors, as well as AI, biotechnology and cloud computing. And its food and music scene, low taxes and favorable climate has made the city a destination for tech workers from across the country.

Key Facts About Austin Tech

  • Number of Tech Workers: 180,500; 13.7% of overall workforce (2024 CompTIA survey)
  • Major Tech Employers: Dell, IBM, AMD, Apple, Alphabet
  • Key Industries: Artificial intelligence, hardware, cloud computing, software, healthtech
  • Funding Landscape: $4.5 billion in VC funding in 2024 (Pitchbook)
  • Notable Investors: Live Oak Ventures, Austin Ventures, Hinge Capital, Gigafund, KdT Ventures, Next Coast Ventures, Silverton Partners
  • Research Centers and Universities: University of Texas, Southwestern University, Texas State University, Center for Complex Quantum Systems, Oden Institute for Computational Engineering and Sciences, Texas Advanced Computing Center

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account