Highest Paid Jobs In The GTA

Big Data Delivery Architect

Amyantek Inc.

This is a Full-time position in Toronto, ON posted April 13, 2017.

Accountabilities

•Leverage Lambda Architecture principles for mixed workload data processing in Hadoop to establish the application solution architecture;

•Partner with the HDFS administrators to optimize the Hadoop cluster; proactively monitor and measure cluster utilization and make recommendations for expansions to support SLAs;

•Explore and recommend alternative fit-for-purpose data integration interfaces between the Enterprise Hadoop platform and external systems (API-based, REST and other interface architecture) based on the use-case;

•Design, build, test and optimize semantic objects in Hadoop for OLAP consumption via COTS BI & visualization tools;

•Work collaboratively with all levels of business and IT stakeholders to understand complex data relations and business requirements and implement and test Big Data solutions;

•Apply DEVOPS & Agile concepts to the development of Big Data solutions;

•Support PoC initiatives to explore the viability of open-source and COTS products relevant to the Hadoop ecosystem;

•Occasional off-hours work required to troubleshoot production;

•Support the development of data mining and machine learning applications

Required Qualifications, Knowledge, & Skills

Only candidates with experience working on a live production-grade Hadoop cluster (on-premise or on VPC) will be considered. Experience working with a Hadoop sandbox on a single-node VM is not ideal.

Key technical proficiencies:

•2+ years Big Data implementation experience leveraging a myriad of high-compression data formats – parquet, avro, json, orc, etc.

•Bachelor’s degree in Computer Science, Engineering, Mathematics, or related mathematical and/or technical fields with Master’s degree preferred

•3+ years hands-on work experience on a production-grade on-premise or VPC-hosted Hadoop cluster;

•2+ years architecting Hadoop data ingestion leveraging Lambda Architecture principles for mix workload processing;

•1+ year implementing real-time data replication from relational databases (RDBMS) to Apache HBase leveraging COTS data replication middleware products;

•3+ years developing data streaming algorithm leveraging kafka queues and storm, flume, spark or flink streaming engine; batch processing using sqoop and oozie or other means;

•5+ years hands-on experience working with UNIX, RHEL Linux, or similar filesystem OS;

•5+ years working with RDBMS and DW products (Oracle & Teradata preferred);

•Experience working with Informatica ETL and IDR for data replication; knowledge of DR architecture leveraging falcon is an asset;

•Building visuals and BI dashboards using Qlik, SAS, Tableau and/or Excel)

Skills Required

  • Education level: None
  • Work experience (years): None

Package

Salary: N/D

Montreal Part Time Jobs | Jobs Sherbrooke | Jobs Laval | Jobs Gatineau | Job Search Montreal | Mississauga Jobs | Laval Emploi | Gatineau Emplois | Emploi Saint-Jérôme | Emplois Sherbrooke | Mississauga Jobs | Emploi Directeur | Emploi de Ventes | Emploi d’étudiant | Emploi Trois-Rivières | Emploi Chauffeur | Emploi Restauration | Healthcare Jobs | Hospitality Job