Proficient with data warehouse/ data lake development methodologies and the ability to apply best practices in building scalable and maintainable data warehouses/data lake
Strong SQL skills including Data Definition Language (DDL), Data Manipulation Language (DML), views, functions and stored procedures
Proficiency in ETL tools like Azure Data Factory, Azure Synapse, or Microsoft SSIS (SQL Server Integration Services) to extract data from various sources and APIs, transform it to fit the target schema, and load it into the data warehouse/data lake
Ability to design and implement data warehouse data models, including star schema, snowflake schema, and dimension hierarchies for optimized data retrieval and analysis
Expert in data integration techniques and data quality processes to ensure data accuracy, consistency, and reliability in the data warehouse
Expert on data warehouse architecture principles, such as data staging areas, data marts, data lakes, and the overall data flow
Proficient in Python or Scala scripting for automating ETL processes and data manipulation
Proficient with Git and DevOps deployment technologies
Understanding of data security principles and compliance regulations to protect sensitive information in the data warehouses/data lake
Skills in optimizing data warehouse/data lake performance, including query optimization, index creation, and partitioning