Cloud ETL Developer

job
  • Rose International
Job Summary
Location
Richmond ,VA 23214
Job Type
Contract
Visa
Any Valid Visa
Salary
PayRate
Qualification
BCA
Experience
2Years - 10Years
Posted
25 Jan 2025
Share
Job Description

Date Posted: 01/22/2025

Hiring Organization: Rose International

Position Number: 476952

Job Title: Cloud ETL Developer

Job Location: Richmond, VA, USA, 23219

Work Model: Hybrid

Shift: This position requires onsite 3 days a week with 2 remote

Employment Type: Temporary

Estimated Duration (In months): 5

Min Hourly Rate($): 75.00

Max Hourly Rate($): 85.00

Must Have Skills/Attributes: Azure, Business Intelligence, Cloud, ESRI, ETL, MS Project, MS Visio, Warehouse

Job Description

***Only qualified Cloud ETL Developer candidates located in the Richmond, VA area to be considered due to the position requiring an onsite presence***


Required Skills:

• Designs and develops systems for the maintenance of the Data Asset Program, ETL processes, and business intelligence (10 Years)

• Design and supports the DW database and table schemas for new and existent data sources for the data hub and warehouse. Design and development of Data (10 Years)

• Work closely with data analysts, data scientists, and other data consumers within the business in an attempt to gather and populate data hub and data (10 Years)

• Advanced understanding of data integrations. Strong knowledge of database architectures, strong understanding of ingesting spatial data (10 Years)

• Ability to negotiate and resolve conflicts, Ability to effectively prioritize and handle multiple tasks and projects (10 Years)

• Experience with key data warehousing architectures including Kimball and Inmon, and has a broad experience designing solutions using a broad set of da (10 Years)

• Expertise in Data Factory v2, Data Lake Store, Data Lake Analytics, Azure Analysis Services, Azure Synapse (10 Years)

• IBM DataStage, Erwin, SQL Server (SSIS, SSRS, SSAS), ORACLE, T-SQL, Azure SQL Database, Azure SQL Datawarehouse (10 Years)

• Operating System Environments (Windows, Unix, etc.) Scripting experience with Windows and/or Python, Linux Shell scripting (10 Years)

• Experience in AZURE Cloud engineering (10 Years)

• The candidate must have a minimum of 10 years of experience delivering business data analysis artifacts

• 5+ years of experience as an Agile Business Analyst; strong understanding of Scrum concepts and methodology

• Experience with ESRI ArcGIS

• Experience with Azure Databricks, Azure data factory, snowflake

• Experience with enterprise data management

• Experience organizing and maintaining Product and Sprint backlogs

• Experience translating client and product strategy requirements into dataset requirements and user stories

• Proficient with defining acceptance criteria and managing acceptance process

• Experience with reporting systems – operational data stores, data warehouses, data lakes, data marts


Preferred Skills:

• Advanced understanding of data integrations

• Strong knowledge of database architectures

• Strong analytical and problem-solving skills

• Ability to build strong relationships both internally and externally

• Ability to negotiate and resolve conflicts

• Ability to effectively prioritize and handle multiple tasks and projects

• Strong written and verbal communication skills

• Desire to learn, innovate and evolve technology


Technologies Required:

• Data Factory v2, Data Lake Store, Data Lake Analytics, Azure Analysis Services, AZURE Synapse

• IBM DataStage, Erwin, SQL Server (SSIS, SSRS, SSAS), ORACLE, T-SQL, Azure SQL Database, Azure SQL Datawarehouse

• Operating System Environments (Windows, Unix, etc.)

• Scripting experience with Windows and/or Python, Linux Shell scripting


Computer Skills/MS Office/Software:

• Excellent computer skills and be highly proficient in the use of MS Word, PowerPoint, MS Excel, MS Project, MS Visio, and MS Team Foundation Server, which will all be necessary in the creation of visually and verbally engaging ETL, data designs and tables as well as the communication of documentation and reporting

• Deep passion for data analytics technologies as well as analytical and dimensional modeling. The candidate must be extensively familiar with ETL(Extraction, Transformation & Load), data warehousing, and business intelligence tools such as business objects, Power BI and Tableau

• The candidate must also have vast knowledge of database design and modeling in the context of data warehousing

• Experience with key data warehousing architectures including Kimball and Inmon, and has a broad experience designing solutions using a broad set of data stores (e.g., HDFS, Azure Data Lake Store, Azure Blob Storage, Azure SQL Data Warehouse, Azure Cosmos DB


Client is seeking a Master Data Analyst with demonstrated experience in data analytics to work as a key member of Enterprise Data Asset team. This analyst will support teams working in Agile (Sprint) to analyze datasets to be made available in a cloud-based data management platform that will support the agency to produce master data with data governance Responsibilities include analyzing source systems which contain a spatial component for candidate datasets; documenting business processes and data lifecycle; developing data requirements, user stories and acceptance criteria; and testing strategies. Develop ETL to extract business data and spatial data and load it into a data warehousing environment. Design and test the performance of the system. Consult with various teams to understand the company’s data storage needs and develop data warehousing options. Deep knowledge of coding languages, such as python, Java, XML, and SQL. Well-versed in warehousing architecture techniques such as MOLAP, ROLAP, ODS, DM, and EDW


Client is a fast-paced organization with very high standards for work quality and efficiency. This position is expected to handle multiple projects, and remain flexible and productive, despite changing priorities and processes. Ongoing improvement and efficiency are a part of our culture, and each team member is expected to proactively contribute to process improvements


Job Responsibilities:

• Work with the Project team members and business stakeholders to understand business processes and pain points

• Develop expertise in source system datasets and data lifecycle Profile source data which may contain a spatial component; review source data and compare content and structure to dataset requirements; identify conflicts and determine recommendations for resolution Conduct entity resolution to identify matching and merging and semantic conflicts Elicit, record, and manage metadata

• Diagram current processes and proposed modifications using process flows, context diagrams and data flow diagrams

• Decompose requirements into Epics and Features and create clear and concise user stories that are easy to understand and implement by technical staff Utilize progressive elaboration; map stories to data models and architectures to be used by internal staff to facilitate master data management Identify and group related user stories into themes, document dependencies and associated business processes

• Discover and document requirements and user stories with a focus on improving both business and technical processing

• Assist Product Owner in maintaining the product backlog

• Create conceptual prototypes and mock-ups

• Collaborate with staff, vendors, consultants, and contractors as they are engaged on tasks to formulate, detail and test potential and implemented solutions

• Perform Quality Analyst functions such as defining test objectives, test plans and test cases, and executing test cases

• Coordinate and Facilitate User Acceptance Testing with Business and ensure Project Managers/Scrum Masters are informed of the progress

• Designs and develops systems for the maintenance of the Data Asset Program (Data Hub), ETL processes, ETL processes for spatial data, and business intelligence

• Develop a new data engineering process that leverage a new cloud architecture and will extend or migrate our existing data pipelines to this architecture as needed

• Design and supports the DW database and table schemas for new and existent data sources for the data hub and warehouse. Design and development of Data Marts

• Work closely with data analysts, data scientists, and other data consumers within the business in an attempt together and populate data hub and data warehouse table structure, which is optimized for reporting

• The Data developers' partners with Data modeler and Data architect in an attempt to refine the business’s data requirements, which must be met for building and maintaining Data Assets


Estimated Start Date: 02/03/2025

Interview Process: Both Phone and In Person



Benefits:

For information and details on employment benefits offered with this position, please visit here . Should you have any questions/concerns, please contact our HR Department via our secure website .

California Pay Equity:

For information and details on pay equity laws in California, please visit the State of California Department of Industrial Relations' website here .

Rose International is an Equal Opportunity Employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, age, sex, sexual orientation, gender (expression or identity), national origin, arrest and conviction records, disability, veteran status or any other characteristic protected by law. Positions located in San Francisco and Los Angeles, California will be administered in accordance with their respective Fair Chance Ordinances.

If you need assistance in completing this application, or during any phase of the application, interview, hiring, or employment process, whether due to a disability or otherwise, please contact our HR Department .

Rose International has an official agreement (ID #132522), effective June 30, 2008, with the U.S. Department of Homeland Security, U.S. Citizenship and Immigration Services, Employment Verification Program (E-Verify). (Posting required by OCGA 13/10-91.).

Other Smiliar Jobs
 
  • Oakland, CA
  • 3 Days ago
  • Fairfield, CA
  • 3 Days ago
  • Minneapolis, MN
  • 3 Days ago
  • New York, NY
  • 3 Days ago
  • Trenton, NJ
  • 3 Days ago
  • Baton Rouge, LA
  • 3 Days ago
  • Baton Rouge, LA
  • 3 Days ago
  • Staunton, VA
  • 2 Days ago
  • Carson City, NV
  • 2 Days ago
  • O'Fallon, MO
  • 2 Days ago
  • El Segundo, CA
  • 1 Days ago
  • Muscatine, IA
  • 1 Days ago
  • Jeffersonville, OH
  • 1 Days ago