Posted today
Public Trust
$80,001 - $120,000
Unspecified
IT - Software
Remote/Hybrid• (Off-Site/Hybrid)
Description
SAIC is looking for a Java Developer who will be responsible for converting existing PySpark codebases into optimized Java-based Spark applications. This role includes developing, refactoring, and maintaining scalable data processing solutions developed on the Databricks platform (or similar Spark execution environments).
Key Responsibilities:
• Convert existing PySpark applications into equivalent, efficient Java Spark implementations
• Design, develop, and maintain scalable Spark-based data pipelines
• Implement data processing logic using Java 8+ with best practices in OOP and functional programming
• Integrate solutions with IRS datasets including IRMF, BMF, and IMF
• Optimize Spark jobs for performance, maintainability, and cost-efficiency
• Collaborate across development, data engineering, and architecture teams
• Troubleshoot and debug Spark workloads in distributed environments
• Ensure compliance with IRS data handling, security, and governance policies
Qualifications
Required QualificationsRequired:
• IRS datasets (IRMF, BMF, IMF) and tax system data structures.
• Java 8+ (experience with functional programming, Streams API, Lambdas).
• Apache Spark (Spark Core, Spark SQL, DataFrame APIs, performance tuning).
• Big data ecosystems (HDFS, Hive, Kafka, S3).
• Working with batch and streaming ETL pipelines for data processing.
Preferred Qualifications:
Target salary range: $80,001 - $120,000. The estimate displayed represents the typical salary range for this position based on experience and other factors.
SAIC is looking for a Java Developer who will be responsible for converting existing PySpark codebases into optimized Java-based Spark applications. This role includes developing, refactoring, and maintaining scalable data processing solutions developed on the Databricks platform (or similar Spark execution environments).
Key Responsibilities:
• Convert existing PySpark applications into equivalent, efficient Java Spark implementations
• Design, develop, and maintain scalable Spark-based data pipelines
• Implement data processing logic using Java 8+ with best practices in OOP and functional programming
• Integrate solutions with IRS datasets including IRMF, BMF, and IMF
• Optimize Spark jobs for performance, maintainability, and cost-efficiency
• Collaborate across development, data engineering, and architecture teams
• Troubleshoot and debug Spark workloads in distributed environments
• Ensure compliance with IRS data handling, security, and governance policies
Qualifications
Required QualificationsRequired:
- Bachelor's degree in Computer Science, Information Systems, or a related field.
- Active MBI Clearance
- 5+ years of professional experience in a data engineering or software development role.
- Advanced expertise in:
• IRS datasets (IRMF, BMF, IMF) and tax system data structures.
• Java 8+ (experience with functional programming, Streams API, Lambdas).
• Apache Spark (Spark Core, Spark SQL, DataFrame APIs, performance tuning).
• Big data ecosystems (HDFS, Hive, Kafka, S3).
• Working with batch and streaming ETL pipelines for data processing.
- Proficient with Git, Maven/Gradle, and DevOps tools.
- Expertise in debugging Spark transformations and ensuring performance.
Preferred Qualifications:
- Hands-on experience converting PySpark workloads into Java Spark.
- Familiarity with ecosystems such as Databricks, Google Dataproc, or similar.
- Knowledge of Delta Lake or Apache Iceberg.
- Proven experience in big data performance modeling and tuning.
Target salary range: $80,001 - $120,000. The estimate displayed represents the typical salary range for this position based on experience and other factors.
group id: 10111346