Data Engineering - AWS Lake House at JPMorgan Chase & Co.

Posted on: 11/10/2021

Location: Plano (ON-SITE)

full time

Original Source

Tags: spark sql nosql etl oracle hadoop unix kafka aws python

As a member of our Software Engineering Group we look first and foremost for people who are passionate around solving business problems through innovation & engineering practices. You will be required to apply your depth of knowledge and expertise to all aspects of the software development lifecycle, as well as partner continuously with your many stakeholders on a daily basis to stay focused on common goals. We embrace a culture of experimentation and constantly strive for improvement and learning. You'll work in a collaborative, trusting, thought-provoking environment-one that encourages diversity of thought and creative solutions that are in the best interests of our customers globally. Commercial Banking IT is looking for a Big Data Software Engineer with skills and experience with large-scale Hadoop based data platforms who will be responsible for design, development and testing of a next generation enterprise data hub and reporting and analytic applications. This individual will work with an existing development team to create the new Hadoop-based platform and migrate the existing data platforms and provide production support. The current platform uses many tools including Oracle SQL, SQL Server, SSIS, and SSRS/SSAS. The candidate will be accountable for design, development, implementation and post-implementation maintenance and support. The candidate will develop and test new interfaces, enhancements/changes to existing interfaces, new data structures, and new reporting capabilities. Responsibilities: - Acquire data from primary or secondary data sources - Identify, analyze, and interpret trends or patterns in complex data sets - Transforming existing ETL logic into AWS Platform - Innovate new ways of managing, transforming and validating data - Establish and enforce guidelines to ensure consistency, quality and completeness of data assets - Apply quality assurance best practices to all work products - Analyze, design and code business-related solutions, as well as core architectural changes, using an Agile programming approach resulting in software delivered on time and in budget; - Experience of working in a development teams, using agile techniques and Object Oriented development and scripting languages, is preferred - Comfortable learning cutting edge technologies and applications to greenfield projectQualifications - Internal This role requires a wide variety of strengths and capabilities, including: - BS/BA degree or equivalent experience - Advanced knowledge of application, data and infrastructure architecture disciplines - 3+ Experience in a Big Data technologies on the AWS platform specifically on at least 2-3 of these - Lake Formation, EMR, Scoop, Spark, Hive, Athena, Redshift, Airflow, Kafka - 5+ years of Experience in Java/Python Development on Big Data Platforms - Experience developing/designing on AWS Lake formation based platforms is preferred - Strong Experience in writing SQL's - Experience in implementing complex ETL transformations on AWS platforms(EMR/Spark) - Strong Experience with UNIX shell scripting to automate file preparation and database loads - Experience with ETL tools is a plus - Experience with AWS Redshift is a plus - Experience in data quality testing; adept at writing test cases and scripts, presenting and resolving data issues - Strong analytical skills with the ability to collect, organize, analyze, and disseminate significant amounts of information with attention to detail and accuracy - Strong development discipline and adherence to best practices and standards. - Ability to manage multiple priorities and projects coupled with the flexibility to quickly adapt to ever-evolving business needs - Demonstrated independent problem solving skills and ability to develop solutions to complex analytical/data-driven problems - Must be able to communicate complex issues in a crisp and concise fashion to multiple levels of management - Excellent interpersonal skills necessary to work effectively with colleagues at various levels of the organization and across multiple locations - Familiarity with NoSQL database platforms is a plus - Proficiency across the full range of database and business intelligence tools; publishing and presenting information in an engaging way is a plus - Experience with multiple reporting tools (QlikView/QlikSense, Tableau, Cognos) is a plus - Financial Services and Commercial banking experience is a plusJPMorgan Chase & Co., one of the oldest financial institutions, offers innovative financial solutions to millions of consumers, small businesses and many of the world's most prominent corporate, institutional and government clients under the J.P. Morgan and Chase brands. Our history spans over 200 years and today we are a leader in investment banking, consumer and small business banking, commercial banking, financial transaction processing and asset management.We recognize that our people are our strength and the diverse talents they bring to our global workforce are directly linked to our success. We are an equal opportunity employer and place a high value on diversity and inclusion at our company. We do not discriminate on the basis of any protected attribute, including race, religion, color, national origin, gender, sexual orientation, gender identity, gender expression, age, marital or veteran status, pregnancy or disability, or any other basis protected under applicable law. In accordance with applicable law, we make reasonable accommodations for applicants' and employees' religious practices and beliefs, as well as any mental health or physical disability needs.Equal Opportunity Employer/Disability/Veterans