Hello, Hope you are doing well.
Please have a look at the below job description and share some relevant profiles for this opportunity. *Role : Snowflake Developer* *Location : NYC, NY (Onsite from Day 1 with Hybrid)* *Experience : 11+ Years* *Duration : Contract* *Note: No H1b and Looking for Locals candidates of NY.* *Primary Responsibilities:* · Architecture, design, implementation and operationalization of large-scale data and analytics solutions on Snowflake Cloud Data Warehouse. · Hands-on experience with Snowflake utilities such as SnowSQL, SnowPipe, Python, Tasks, Streams, Time travel, Optimizer, Metadata Manager, data sharing, and stored procedures. · Experience in Data warehousing - OLTP, OLAP, Dimensions, Facts, and Data modeling. · Need to have working knowledge of MS Azure configuration items with respect to Snowflake. · Developing EL pipelines in and out of data warehouse using combination of Databricks, Python and SnowSQL. · Developing scripts UNIX, Python etc. to do Extract, Load and Transform data. · Provide production support for Data Warehouse issues such data load problems, transformation translation problems · Translate mapping specifications to data transformation design and development strategies and code, incorporating standards and best practices for optimal execution. · Understanding data pipelines and modern ways of automating data pipeline using cloud based testing and clearly document implementations, so others can easily understand the requirements, implementation, and test conditions. · Perform code reviews to ensure fit to requirements, optimal execution patterns and adherence to established standards. · Establishing and monitoring Operational Level Agreements for the health and performance/cost of the warehouse environment (Loads, queries, data quality) *Preferred Skills:* · Bachelor’s degree in computer science, Engineering, Technical Science · 3 years of technical architecture and build *experience with large-scale data warehouse solutions.* · *Code optimization expertise* · *Experience building data ingestion pipelines using Python and Databricks in working with MS Azure.* · *3 years’ experience in Finance / Banking industry – some understanding of Securities and Banking products and their data footprints.* *Basic Qualifications:* · Minimum 3 years of designing and implementing an operational production grade large-scale data solution on *MA Azure Snowflake Data Warehouse.* · Including hands on experience with productionized data ingestion and processing pipelines using *Python, Databricks, SnowSql* · Excellent understanding of Snowflake Internals and integration of Snowflake with other data processing and reporting technologies · Excellent presentation and communication skills, both written and verbal ability to problem solve and design in an environment with unclear requirements. · Ability to lead and drive performance of a team of developers – local and offshore. -- *Warm Regards!* *Sandeep Aggarwal* *Direct: +1 (848-668-9626)* *Email: **[email protected]* <[email protected]> *Connect with me: www.linkedin.com/in/sandeep-aggarwal-434413245 <http://www.linkedin.com/in/sandeep-aggarwal-434413245>* -- You received this message because you are subscribed to the Google Groups "project managment" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/project-managment/CAJ-2PJW_qT9fRTbsqMUmdipOcp8%3DxnycmowUYrUn31zpXd9B-Q%40mail.gmail.com.
