Job Details

Data Engineer (ADF, Orchestration, Data Flows, Snowflake) | REMOTE

  2025-10-01     Skill Mine     all cities,AK  
Description:

[Minneapolis, MN, 55423] | 2025-05-12 15:08:31

  • Background candidates should ideally have come from SSIS /SQL transitioning to Azure Data Factory (Data Integration) Snowflake (Source and Target End Point)
  • Azure has a lot of components to it such as Fabric, DataBricks, ADF, Pyspark
  • They need candidates that can use ADF as both orchestration and integration
  • many candidates claim to have ADF but for orchestration really just use Pyspark to trigger the jobs through ADF
  • They need to be advanced at Building the Integration through data flows logic as well

100% Telecommute

Work Hours:9am-5pm CST

Project:

  • As a member of the Data Management team, the Data Engineer supports the Alabama EDS by developing and maintaining workflows, identifying, and resolving data quality issues, and optimizing processes to improve performance.
  • The Data Engineer will also support intrastate agencies by monitoring automated data extracts and working directly with state partners to create new extracts based on business specifications.

Responsibilities:

  • Develop and manage effective working relationships with other departments, groups, and personnel with whom work must be coordinated or interfaced
  • Efficiently communicate with ETL architect while understanding the requirements and business process knowledge in order to transform the data in a way thats geared towards the needs of end users
  • Assist in the overall architecture of the ETL Design, and proactively provide inputs in designing, implementing, and automating the ETL flows
  • Investigate and mine data to identify potential issues within ETL pipelines, notify end-users and propose adequate solutions
  • Developing ETL pipelines and data flows in and out of the data warehouse using a combination of Azure Data Factory and Snowflake toolsets
  • Developing idempotent ETL process design so that interrupted, incomplete, or failed processes can be rerun without errors using ADF dataflows and Pipelines
  • Ability to work in Snowflake Virtual Warehouses as needed in Snowflake and automate data pipelines using Snowpipe for tedious ETL problems
  • Capturing changes in data dimensions and maintaining versions of them using Stream sets in snowflake and scheduling them using Tasks
  • Optimize every step of the data movement not only limited to source and during travel but also when it's at rest in the database for accelerated responses
  • Must have the ability to build a highly efficient orchestrator that can schedule jobs, execute workflows, perform Data quality checks, and coordinate dependencies among tasks
  • Responsible for testing of ETL system code, data design, and pipelines and data flows. Root cause analysis on all processes and resolving production issues are also a part of the process and routine tests on databases and data flow and pipeline testing
  • Responsible for documenting the implementations, and test cases as well as responsible for building deployment documents needed for CI/CD

Ideal Background: Data Engineer with Healthcare (Medicaid) and Microsoft Azure based experience with Snowflake and Azure Data Factory

TOP REQUIREMENTS:

  • 5+ years of Data engineering experience with a focus on Data Warehousing
  • 2+ years of experience creating pipelines in Azure Data Factory (ADF)
  • 3+ years of experience creating stored procedures with Oracle PL/SQL, SQL Server T-SQL, or Snowflake SQL

Required:

  • 5+ years of Data engineering experience with a focus on Data Warehousing
  • 2+ years of experience creating pipelines in Azure Data Factory (ADF)
  • 5+ years developing ETL using Informatica PowerCenter, SSIS, Azure Data Factory, or similar tools.
  • 5+ years of experience with Relational Databases, such as Oracle, Snowflake, SQL Server, etc.
  • 3+ years of experience creating stored procedures with Oracle PL/SQL, SQL Server T-SQL, or Snowflake SQL
  • 2+ years of experience with GitHub, SVN, or similar source control systems
  • 2+ years of experience processing structured and un-structured data.
  • Experience with HL7 and FHIR standards, and processing files in these formats.
  • 3+ years analyzing project requirements and developing detailed specifications for ETL requirements.
  • Excellent problem-solving and analytical skills, with the ability to troubleshoot and optimize data pipelines.
  • Ability to adapt to evolving technologies and changing business requirements.
  • Bachelors or Advanced Degree in a related field such as Information Technology/Computer Science, Mathematics/Statistics, Analytics, Business

Preferred:

  • 2+ years of batch or PowerShell scripting
  • 2+ years of experience with Python scripting.
  • 3+ years of data modeling experience in a data warehouse environment
  • Experience or familiarity with Informatica Intelligent Cloud Services (specifically Data Integration)
  • Experience designing and building APIs in Snowflake and ADF (e.g. REST, RPC)
  • Experience with State Medicaid / Medicare / Healthcare applications
  • Azure certifications related to data engineering or data analytics.

Required Skills : Data Analysis

Basic Qualification :

Additional Skills :

This is a high PRIORITY requisition. This is a PROACTIVE requisition

Drug Screen : No

PSS Technology Inc, established in 2012, is a U.S.-based company specializing in Information Technology and Services. The company is dedicated to delivering value to clients by providing professional treatment and high-quality services. Their commitment to client satisfaction is evident in their efficient, effective, and cost-effective solutions.

The company offers a range of services, including:

  • Digital marketing
  • Web design and development
  • Business Process Outsourcing (BPO)
  • Staffing and recruitment
  • Talent acquisition

PSS Technology Inc operates from its headquarters in Cupertino, California.

The company's management team boasts over 15 years of combined experience, ensuring a deep understanding of client needs and a focus on delivering tailored solutions.

For more information about PSS Technology Inc and their services, you can visit their official website at www.psstechnologyinc.com .

Job Code

JPC - 2313

Job Start Date

2025-06-02

City

Minneapolis

Primary Skills

Data Analysis

Posted Date

2025-05-12 15:08:31

Job End Date

2026-12-31

Number of Positions

1

#J-18808-Ljbffr


Apply for this Job

Please use the APPLY HERE link below to view additional details and application instructions.

Apply Here

Back to Search