Location: New York,NY, USA
Design and implement standardized data management procedures around data staging, data ingestion, data preparation, data provisioning, and data destruction. Ensure quality of technical solutions as data moves across multiple zones and environments. Provide insight into the changing data environment, data processing, data storage and utilization requirements for the company, and offer suggestions for solutions. Ensure managed analytic assets to support the company's strategic goals by creating and verifying data acquisition requirements and strategy. Develop, construct, test, and maintain architectures. Align architecture with business requirements and use programming language and tools. Identify ways to improve data reliability, efficiency, and quality. Conduct research for industry and business questions. Deploy sophisticated analytics programs, machine learning, and statistical methods to efficiently implement solutions. Prepare data for predictive and prescriptive modeling and find hidden patterns using data in support of the Data Science team. Use data to discover tasks to automate. Create data monitoring capabilities for each business process and works with data consumers on updates. Align data architecture to the solution architecture; contributes to overall solution architecture. Develop patterns for standardizing the environment technology stack. Help maintain the integrity and security of company data.Requirements: Bachelor's degree or foreign education equivalent in Computer Science, Computer Engineering or related field plus five (5) years of experience performing data engineering. Alternatively, Master's degree or foreign education equivalent in Computer Science, Computer Engineering or related field plus three (3) years of experience performing data engineering. Specific skills/other requirements - Must also possess the following (quantitative experience not applicable to this section): (1) Demonstrated expertise (DE) working in data warehousing ecosystem leveraging python and Pyspark to extract, transform and load heterogeneous data ranging from flat files to databases and handle versioning, deployment and maintenance of artifacts; (2) DE working in a ‘Big Data' ecosystem process data to file systems, data structures/data bases, automation, security, messaging, and movement; and (3) DE working in a production cloud infrastructure.Remote work permitted anywhere in U.S. except in DC, Alaska, Arkansas, Hawaii, Idaho, Kentucky, Mississippi, Missouri, Nebraska, North Dakota, Oklahoma, Oregon, Rhode Island, South Dakota, Vermont, and Wyoming.Minimum Salary: 163238 Maximum Salary: 180000 Salary Unit: Yearly