Do you want your voice heard and your actions to count? Discover your opportunity with Mitsubishi UFJ Financial Group (MUFG), one of the world’s leading financial groups. Across the globe, we’re 150,000 colleagues, striving to make a difference for every client, organization, and community we serve. We stand for our values, building long-term relationships, serving society, and fostering shared and sustainable growth for a better world. With a vision to be the world’s most trusted financial group, it’s part of our culture to put people first, listen to new and diverse ideas and collaborate toward greater innovation, speed and agility. This means investing in talent, technologies, and tools that empower you to own your career. Join MUFG, where being inspired is expected and making a meaningful impact is rewarded. The selected colleague will work at an MUFG office or client sites four days per week and work remotely one day. A member of our recruitment team will provide more details. Architecting and executing new cloud-based solutions aligned with enterprise standards; overseeing decommissioning of legacy platforms and seamless migration to modern infrastructure; and collaborating with infrastructure, security, and application teams to ensure scalable and secure deployments. Supporting and implementing Tier 1 and Tier 2 reports with high availability and accuracy; leading BCBS 239 compliance efforts, ensuring data aggregation and reporting meet regulatory expectations; and partnering with business stakeholders to align reporting outputs with decision-making needs. Ensuring Data Management and Control Implementation Framework (DMCIF) controls are properly implemented, documented, and tested; driving periodic control assessments and remediation plans in coordination with audit and risk teams; and maintaining traceability and transparency across control execution and reporting pipelines. Driving end-to-end cloud implementations, managing legacy platform migrations to a modern tech stack, and ensuring robust reporting and regulatory compliance across Tier 1 and Tier 2 domains. Designing, developing and implementing data warehouse solutions for the Bank utilizing expertise in enterprise data architecture, designs, solutions and technologies. Functioning as a Subject Matter Expert (SME) for technical applications relating to banking, including loans, deposits, treasury. Performing API based data integration between data consumers and Enterprise Data Platform in Cloud. This Enterprise Data Platform build out involves bringing in the Bank’s commercial banking products to support regulatory reporting, and BCBS 239 Risk Aggregation Framework. Leading, planning, designing, documenting, developing, testing, implementing, monitoring, maintaining and supporting enterprise data warehouse and data mart solutions, including subject area marts and interfaces to downstream applications. Liaising with internal stakeholders to assess business, technical, quality and security requirements to achieve intended outcomes in accordance with Bank processes, standards and procedures. Developing and maintaining complex ETL mappings, workflows and Unix shell scripts in a normalized/denormalized data warehouse/data mart environment based on technical specifications and other supporting documentation, utilizing Informatica PowerCenter, Unix Shell Scripts, advanced SQL and Tidal Enterprise Scheduler. Supporting Business initiatives and day to day activities across the entire spectrum of our Bank, including all Front Office, Middle Office and Back Office users as well as Credit Risk, Finance, Compliance, Comptroller and Tokyo Head Office personnel, where it is used for reporting and investigation purposes. Performing Device Under Test, System Integration Testing, User Acceptance Testing, Disaster Recovery and regression testing activities. Ensuring best practices, standards, processes and procedures are followed. Supporting Information Management and other Information Technology Department initiatives. Directing, monitoring and assigning work to those on the team. Leading the team in the preparation of functional requirements, design documents and conducting feasibility and cost benefit studies and a comprehensive knowledge of the Bank's business and system flow within the organization. Working on large, complex projects that have enterprise wide impact and require subject matter expertise of multiple process improvement areas and mastery of process improvement tools. Leading the API based data integration between data consumers and the Enterprise Data Platform in Cloud using AWS API Gateway, Python, PostgresSQL, Kafka, AWS Active MQ, Informatica, WhereScape. Coordinating with the development groups to ensure data accuracy to business analysts, leadership groups, and other end users to aid in ongoing operational insights. Conducting training/knowledge sharing sessions, Ensure best practices and standards. Leading and designing Data Migration process and activities from current warehouses to Data Lake. Supervising 8-10 Data Developers and Analysts and ensuring timely delivery of migration milestones, cloud onboarding and reporting enhancements.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level