Work with the brightest minds at one of the largest financial institutions in the world. This is long-term contract opportunity that includes a competitive benefit package!
Our client has been around for over 150 years and is continuously innovating in today's digital age. If you want to work for a company that is not only a household name, but also truly cares about satisfying customers' financial needs and helping people succeed financially, apply today.
Position: Big Data Developer
Location: CHARLOTTE, North Carolina
Term: 12 months
Design high performing data models on big-data architecture as data services.
Design and build high performing and scalable data pipeline platform using Hadoop, Apache Spark and Amazon S3 based object storage architecture.
Work with business analysts, development teams and project managers for requirements and business rules.
Collaborate with source system and approved provisioning point (APP) teams, Architects, Data Analysts and Modelers to build scalable and performant data solutions.
Effectively work in a hybrid environment where legacy ETL and Data Warehouse applications and new big-data applications co-exist.
Work with Infrastructure Engineers and System Administrators as appropriate in designing the big-data infrastructure.
Support ongoing data management efforts for Development, QA and Production environments.
Leverage knowledge of industry trends to build best in class technology to provide competitive advantage.
Lead moderately complex initiatives and deliverables within technical domain environments.
Contribute to large scale planning of strategies.
Design, code, test, debug, and document for projects and programs associated with technology domain, including upgrades and deployments.
Review moderately complex technical challenges that require an in-depth evaluation of technologies and procedures.
Resolve moderately complex issues and lead a team to meet existing client needs or potential new clients needs while leveraging solid understanding of the function, policies, procedures, or compliance requirements.
Collaborate and consult with peers, colleagues, and mid-level managers to resolve technical challenges and achieve goals.
Lead projects and act as an escalation point, provide guidance and direction to less experienced staff.
Is this a good fit? (Requirements):
7+ years of application development and implementation experience.
7+ years of experience delivering complex enterprise wide information technology solutions.
Experience in Hadoop ecosystem tools for real-time batch data ingestion, processing and provisioning such as Apache Flume, Apache Kafka, Apache Sqoop, Apache Flink, Apache Spark or Apache Storm.
5+ years of experience delivering ETL, data warehouse and data analytics capabilities on big-data architecture such as Hadoop 5+ years of Java or Python experience.
7+ years of ETL (Extract, Transform, Load) Programming experience.
7+ years of reporting experience, analytics experience or a combination of both 5+ years of Hadoop experience 5+ years of operational risk, conduct risk or compliance domain experience.
Excellent verbal, written, and interpersonal communication skills.
Ability to work effectively in virtual environment where key team members and partners are in various time zones and locations.
Knowledge and understanding of project management methodologies: used in waterfall or Agile development projects.
Knowledge and understanding of DevOps principles Ability to interact effectively and confidently with senior management.
4+ years of Software Engineering experience, or equivalent demonstrated through one or a combination of the following: work experience, training, military experience, or education.