General Responsibilities
- Review business requirements, familiarize with and understand business rules and transactional data model
- Define conceptual, logical model and physical model mapping from data source to curated model and data mart.
a) Analyze requirements and recommend changes to the physical model.
b) Develop scripts for the physical model, create database and / or delta lake file structure.
c) Access Oracle DB environments, set necessary tools for developing solution.
Implement data design methodologies, historical and dimensional models
a) Develop curated model to store historical data captured incrementally from source
b) Design dimensional data mart models, create source-to-target-mapping documentation, design and document data transformation from curated model to data mart
c) Perform data profiling, assess data accuracy, design and document data quality and master data management rules
Functionality Review, Data Load review, Performance Review, Data Consistency checks.
a) Help troubleshooting data mart design issues
b) Review performance of ETL with developers and suggest improvements
- Participate in end-to-end integrated testing for Full Load and Incremental Load and advise on issues
- Plan for Go Live, Production Deployment.
a) Work with system administrator, ETL developers and ministry team to define production deployment steps.
b) Configure parameters, scripts for go live. Test and review the instructions.
c) Review release documentation
Go Live Support and Review after Go Live.
a) Review data models, ETL process, tools and provide recommendation on improving performance and reduce ETL timelines.
b) Review Infrastructure and any performance issues for overall process improvement
- Proactively communicate with stakeholders on any changes required to conceptual, logical and physical models, communicate and review dependencies and risks.
- Knowledge Transfer to Ministry staff, development of documentation on the work completed.
a) Document share and work on the architecture end-to end-working knowledge, Troubleshooting steps, configuration and scripts review.
b) Transfer documents, scripts and review of documents.
Data Modeler Requirements :
- 7+ years BI Data Architect experience in enterprise applications and solutions design / development and related with data warehousing, data lake implementations and dimensional modelling.
- Collect business-level questions and propose approaches to address business needs and provide data insights.
- Expand documentation and knowledge of business processes relative to available data to provide contextual guidance for operation / project, reporting and insights generation.
- Ability to design and articulate complex technical concepts into executable development work packages.
- Knowledge of BI tools for metadata modeling and report design (e.g. Power BI)
- MS SQL Server Technology, Azure Data Lake, Azure Databricks
- Expert knowledge developing data warehouse solutions on MS Stack (Azure Data Lake, SQL, ADF, Databrciks, PowerBI) to store and retrieve centralized information.
Experience designing the data warehouse using dimensional and delta lake concepts.
- Create / maintain enterprise data model and data dictionary. Help development te am to optimize database performance. Coordinate with the Integration department to identify future needs and requirements.
- Extensive knowledge of data modelling tools (e.g. SAP PowerDesigner, Visio)
- Review, install and configure information systems to ensure functionality and security. Analyze structural requirements for new data warehouse and applications
- Experience using Oracle database server and tools (12c, 19c), PL / SQL for development of Business Intelligence applications.
- Demonstrated skills in writing SQL stored procedures and packages for datamarts and reporting.
- Demonstrated experience in Azure DevOps
- Demonstrated experience in performance tuning of Business Intelligence applications, including data model and schema optimization
Skills :
- 7+ years in data modelling and data warehouse design (Must Have)
- 2+ years Azure Data Lake and Azure Databricks SQL Warehouse (Must Have)
- 5+ years SQL (Must Have)
Assets :
- Knowledge of Curam IBM COTS solutions (Social Assistance Management System)
- ETL design concepts
- Knowledge of Enterprise Architecture tools and frameworks (ArchiMate, TOGAF, Zachmann)
Evaluation Criteria :
Design Documentation and Analysis Skills (30 points)
- Demonstrated experience in creating both Functional Design Documents (FDD) and amp; Detailed Design Documents (DDD).
- Experience in Fit-Gap analysis, system use case reviews, requirements reviews, coding exercises and reviews.
- Experience in the development and maintaining a plan to address contract deliverables, through the identification of significant milestones and expected results with weekly status reporting.
- Work with the Client and amp; Developer(s) assigned to refine / confirm Business Requirements
- Participate in defect fixing, testing support and development activities for Informatica / ETL tool. Assist with defect fixing and testing support for Cognos reports.
- Analyze and document solution complexity and interdependencies
BI Data Modelling and Technical Skills (40 points)
Understanding of Data Modelling for Business Intelligence including :
a. Expert Knowledge of data warehouse design methodologies, delta lake and dimensional modeling in particular
b. Understanding of Extract / Transform / Load processes to transform data for reporting / BI purposes
c. Ability to define schema for reporting databases
d. Experience with advanced modeling tools,
- Knowledge of BI tools for metadata modeling and report design (e.g. PowerBI, Cognos 10 / 11)
- Extensive knowledge and experience in MS SQL Server Technology, Azure Databricks SQL Warehouse, Azure Data Lake
- Experience using Oracle database server and tools (12c, 19c), PL / SQL for development of Business Intelligence applications.
Demonstrated skills in writing and reverse engineering SQL stored procedures and packages for datamarts and reporting.
Demonstrated experience in performance tuning of Business Intelligence applications, including data model and schema optimization
Quality Assurance (20 points)
Demonstrated experience in defining and executing tests across the development lifecycle (unit testing, system testing, user acceptance testing) and using results to refine database design
Knowledge Transfer (10 points)
The Architect / Modeler must have previous work experience in conducting Knowledge Transfer and training sessions, ensuring the resources will receive the required knowledge to support the system.
The resource must develop learning activities using review-watch-do methodology and amp; demonstrate the ability to prepare and present.
- Development of documentation and materials as part of a review and knowledge transfer to other members
- Development of specific activities as part of a review (hand over to ministry staff)and building block approach which, build on knowledge transfer and skills development from the previous stage to the next
- Development and facilitation of classroom based or virtual instructor demo led sessions for Developers
- Monitor identified milestones and submission of status reports to ensure Knowledge Transfer is fully completed