This site uses cookies. To find out more, see our Cookies Policy

VP, Big Data Engineer, GTO in Singapore at UOB Group

Date Posted: 4/19/2018

Job Snapshot

Job Description

Functional area: Business Technology Services
Employment type: Full-time
Job Type: Permanent

  • Relevant work experience of 3-4 years of experience (overall 8-12 years of experience)
  • Experience in building production live applications involving high-performance algorithms, predictive models machine learning
  • Proficient with Big Data programming languages such as Map Reduce, Scala, Java, Pyspark, Python, Microsoft R, Pig, etc.
  • Experience in developing dimensional data model in Hadoop cluster with Facts and Dimension using Hbase, Accumulo, Hive, etc..
  • Ability to create new data pipeline consuming or publishing API's , processing Real time feeds, e.g. Twitter, Facebook
  • Experience in creating customer software components (eg: specialized UDF’s) and analytics applications
  • Working closely with business stakeholders to prepare large amounts of data in structured, semi structured and unstructured database, SQL & No SQL databases.
  • Experience in memory/NoSQL databases like mongo dB, Cassandra, redshift etc.
  • Experience with Big Data visualization tools like Qlik, Microsoft Power BI
  • Technology – Cloudera Distribution Hub (CDH 5.8 and above) set of tools (eg: Hive, impala, HDFS, Oozie, Pig, Hbase)
  • Hands on experience on Real time Data transfer using Sqoop, flume, Kafka.
  • Hands on experience on Hadoop command line interface and data transfer / disk copy utilities
  • Expertise of scheduling resources using YARN
  • Good knowledge of using Cloudera Manager to start up and bring down services
  • Prior experience of working closely with stakeholders to design and define security rules and their permissions
  • Good understanding and working experience with Kerberos enabled cluster and sentry based authorization for end user data access.
  • Experience of working with ETL, data preparation and data governance tools – Informatica Power Center, BDM, metadata manager, Informatica Data Quality, Intelligent Data Lake
  • Rich Experience in developing Teradata BTECH scripts, data loading utilities with fast load and multi load scripts and GCFR framework.
  • Understanding of Master Data Management/Reference data management principles and process
  • Design and Develop data integration service and delivering source-to-target mapping
  • Experience with data extract-transform-load (ETL) design specification