Job Requirements
Merrifield, VA
Secret Polygraph not specified
Senior Level Career (10+ yrs experience)
$175,000 - $190,000
Job Description
Insight Global is seeking a Senior Data Warehouse Modernization Developer to support a long term federal data modernization initiative. These engineers focus on modernizing data warehouse pipeline architecture by gradually reducing reliance on SSIS and SQL-only patterns, introducing Python based ETL, improving modularity, and setting the groundwork for a future move into PySpark, distributed compute, and eventually Databricks or similar cloud platforms. They should have an in depth understanding of the Kimball model in place, and experience incrementally replacing or re platforming components.
Responsibilities include:
- Learn and understand the existing Kimball-modeled SQL Server structure over time.
- Reverse-engineer SSIS packages and convert them into Python-based modular ETL pipelines.
- Build reusable Python frameworks for extraction, transformation, and loading.
- Begin introducing PySpark for scalable transformation patterns.
- Contribute to planning a future transition path (Databricks, Azure Synapse, Snowflake, etc.).
- Improve orchestration, dependency management, and data quality practices.
- Identify technical debt and modernize aging ETL components.
- Collaborate with EDW SMEs to understand business logic and lineage.
- Build documentation and repeatable processes for modernization.
REQUIRED SKILLS AND EXPERIENCE
- Active Secret clearance
- Solid SQL and understanding of Kimball dimensional modeling.
- Strong Python: Pandas, SQLAlchemy, packaging, module structure, unit testing.
- Experience with PySpark or Spark (or willingness to ramp quickly).
- Familiarity with cloud data platforms (Azure, AWS, etc.).
- Experience modernizing legacy ETL systems is a major plus.
- Ability to read, interpret, and redesign SSIS/SQL-based ETL logic.
- Understanding of data ingestion, transformation, and structured warehouse design.
- Experience with Git, CI/CD pipelines, and modern development practices.
NICE TO HAVE SKILLS AND EXPERIENCE
- Exposure to Databricks, Snowflake, or Hadoop ecosystems
- Cloud experience (Azure or AWS)
- Experience working in federal environments (DOJ, DEA, FBI, DHS, etc.)
- Experience supporting both classified and unclassified systems
- Experience with large scale or high volume data environments
- Exposure to AI / data experimentation initiatives
Responsibilities include:
- Learn and understand the existing Kimball-modeled SQL Server structure over time.
- Reverse-engineer SSIS packages and convert them into Python-based modular ETL pipelines.
- Build reusable Python frameworks for extraction, transformation, and loading.
- Begin introducing PySpark for scalable transformation patterns.
- Contribute to planning a future transition path (Databricks, Azure Synapse, Snowflake, etc.).
- Improve orchestration, dependency management, and data quality practices.
- Identify technical debt and modernize aging ETL components.
- Collaborate with EDW SMEs to understand business logic and lineage.
- Build documentation and repeatable processes for modernization.
REQUIRED SKILLS AND EXPERIENCE
- Active Secret clearance
- Solid SQL and understanding of Kimball dimensional modeling.
- Strong Python: Pandas, SQLAlchemy, packaging, module structure, unit testing.
- Experience with PySpark or Spark (or willingness to ramp quickly).
- Familiarity with cloud data platforms (Azure, AWS, etc.).
- Experience modernizing legacy ETL systems is a major plus.
- Ability to read, interpret, and redesign SSIS/SQL-based ETL logic.
- Understanding of data ingestion, transformation, and structured warehouse design.
- Experience with Git, CI/CD pipelines, and modern development practices.
NICE TO HAVE SKILLS AND EXPERIENCE
- Exposure to Databricks, Snowflake, or Hadoop ecosystems
- Cloud experience (Azure or AWS)
- Experience working in federal environments (DOJ, DEA, FBI, DHS, etc.)
- Experience supporting both classified and unclassified systems
- Experience with large scale or high volume data environments
- Exposure to AI / data experimentation initiatives
group id: 10112344
Defining Company Culture