Unlock Your Career Potential: Technology at ADP. Do you enjoy exploring, identifying and inspiring the future of the workplace and the lives of millions of people? At ADP, the world’s largest B2B cloud company, our Technology team is comprised of brilliant engineers, architects, data scientists, infrastructure experts, and more. We were first in our industry to offer a SaaS solution and continue to push the envelope utilizing the latest operating platforms to deliver the highly automated, intelligent and predictive solutions that are redefining what is possible. Named one of Forbes’ “Most Innovative Companies” and one of Computerworld’s “100 Best Places to Work in IT”, we are committed to leading the way in product development and research, empowering you to bring to life the latest innovations that will forever change the way businesses manage their most vital asset, employees.
The Lead Application Developer will lead the development of building Data Integration solutions using Informatica Powercenter and Big Data technologies. The candidate will also provide design expertise and consultation to other developers in the team. The candidate will work with clients, analysts, architects and engineers to plan, design, develop, test and implement Data integration solutions consistent with business objective. The candidate will help team to move into big data technologies.
The Lead Application Developer primary responsibility will be to design and develop the data integration processes used to extract, transform and load using ETL tools. The candidate is expected to be results-oriented, have knowledge of new and legacy technologies and relevant platforms.
Design, develop, implement and maintain programs to load, aggregate and extract the data to meet a wide range of business requirements using Informatica Powercenter
Provide leadership and guidance to the other ETL(Informatica) developers as needed
Develop process and code to offload some of the ETL work to Big Data platform using Python/Scala
Provide data engineering solutions to help Data Science community
Responsible for integrating different ETL code modules
Review requirements, perform data analysis and design ETL flow
Create data models using star and snowflake schemas, de-normalized data structures, slowly changing dimensions
Understanding of enterprise data warehouse, data federation, data virtualization, hub & spoke architectures
ETL and database performance tuning and optimization
Develop LINUX scripts to automate the load process
Create and execute Unit test plans
Participate effectively in relevant aspects of SDLC including Planning, development, testing, reviews and implementation
Translate client requirements into Informatica Powercenter mappings and workflows
Investigate and resolve data issues as needed
Prepare Technical design documentation
Participate in project planning and release management. Higher level jobs would be expected to manage broader or multiple projects at a time.
Maintain and support weekly/monthly load process
10+ years of experience in ETL using Informatica Powercenter
7-10 years of experience working with Oracle databases
5 years of experience writing UNIX shell scripts
3 years of experience writing Python/Scala/Pig scripts
3 years of experience working with Big Data platforms like Cloudera, Horton
Experience in working with AWS Redshift is a plus
Working knowledge of RStudio
Expert in optimizing Informatica Powercenter mappings
3-5 years of experience in building Data Models (Star schema and Snowflake schema)
Expert in tuning Oracle databases by adding Indexes and creating Materialized views
Demonstrate in-depth understanding of Data Warehousing and ETL Concepts, ETL Loading strategy, Data Archiving, ETL Error Handling
A self-starting team player able to work independently
Knowledge of Master Data Management, Tableau, OBIEE, Salesforce, Informatica IDQ, Address Doctor, Oracle ERP