Data Engineer (Human Capital) | TELECOMMUTE
: Job Details :


Data Engineer (Human Capital) | TELECOMMUTE

Efficus

Location: all cities,AK, USA

Date: 2024-12-02T19:13:27Z

Job Description:
100% telecommute Hours: Regular day working hours with some flexibility, but must be available for most of Central Time working hours with a heavier focus on availability in the mornings. Project:
  • Working with customers to design data extracts based on their reporting needs
  • Collaborating with IT Dev Ops partners to ingest extracts into a data warehouse
  • Creating scalable data pipelines for reporting teams from the data warehouse to our BI tool
Team:
  • We are a team of 9 Data Engineers and Analysts + Engineering Managers/Leaders
  • We have 1 product owner on the team and work with several others directly
  • We have a Scrum Master aligned as well
Responsibilities:
  • Architect all phases of software engineering including requirements analysis, application design, code development and testing with a focus on business intelligence dataset development
  • Design reusable components, frameworks, and libraries
  • Contribute and make recommendations to the design and architecture to enable secure, scalable, and maintainable solutions
  • Clearly articulate the implications of design/architectural decisions, issues and plans to technology leadership
  • Work collaboratively with People Analytics on the development and production of standard datasets to drive actionable decision making and reporting stability
  • Conduct design and code reviews to ensure code developed meets business needs, coding best practices guidelines, unit testing, security, and scalability and maintainability guidelines
  • Work very closely with architecture groups and drive solutions
  • Assist with updating infrastructure components
  • Use engineering best practices following an Agile methodology to deliver high-quality emerging tech solutions
  • Communicate with impact - influence and negotiate effectively with all internal and external stakeholders to achieve win-win solutions that advance organizational goals
  • Grow and maintain knowledge of and leverage emerging technologies
  • Develop and analyze highly complex system standards, thresholds, and recommendations to maximize system performance
  • Analyze project requirements and develop detailed specifications for new data warehouse reporting requirements
  • Research API calls and make necessary changes to meet the business, contractual, security, performance needs
  • Assesses and interprets customer requests for feasibility, priority, and complexity
  • Create and maintain internal process documentation
  • Support projects and change initiatives aligned to key priorities of People Analytics and People Analytics customers
  • Understands priorities and organizes prescribed and non-prescribed work to meet or exceed deadlines and expectations
  • Proactively keeps data secure and decommissions legacy content in our environment
  • Serve as a resource to others within the People Analytics community; mentors other data engineers; provides explanations and information to others on difficult issues, problems, and solutions
  • Works with minimal guidance; seeks guidance on only the most complex tasks
  • Coaches, provides feedback, and guides others within the People Analytics community
Ideal Background: Ideal candidate would have experience in the Human Capital space, working with the backend data warehousing of human capital systems Required:
  • 3+ years of data engineering experience
  • 3+ years of full lifecycle application, software development experience
  • 3+ years of modern programming language such as Python, Java, and Scala
  • 2+ years of SDLC experience in an Agile environment
  • Working knowledge of the following business and technology concepts: APIs, CI/CD, Big Data, data architecture and governance
  • Experience with Cloud technologies and platforms such as Docker, OSE, Kubernetes, AWS, Snowflake, and Azure
  • Experience with Jenkins, GitHub, Big Data technologies like Spark
  • Experience using IDEs such as Eclipse, JBoss, IntelliJ
  • Relational database experience
  • Experience ingesting and working with large and complex datasets
  • Experience gathering requirements from end users
Preferred:
  • Experience in data extraction, manipulation, and management
  • Experience researching issues, investigating data accuracy, and performing quality checks
  • Project Management skills
  • Ability to work independently and collaborate effectively with others
Apply Now!

Similar Jobs (0)