Highest Paid Jobs In The GTA

Big Data Delivery Architect

Amyantek Inc.

This is a Full-time position in Toronto, ON posted April 13, 2017.

Accountabilities

•Leverage Lambda Architecture principles for mixed workload data processing in Hadoop to establish the application solution architecture;

•Partner with the HDFS administrators to optimize the Hadoop cluster; proactively monitor and measure cluster utilization and make recommendations for expansions to support SLAs;

•Explore and recommend alternative fit-for-purpose data integration interfaces between the Enterprise Hadoop platform and external systems (API-based, REST and other interface architecture) based on the use-case;

•Design, build, test and optimize semantic objects in Hadoop for OLAP consumption via COTS BI & visualization tools;

•Work collaboratively with all levels of business and IT stakeholders to understand complex data relations and business requirements and implement and test Big Data solutions;

•Apply DEVOPS & Agile concepts to the development of Big Data solutions;

•Support PoC initiatives to explore the viability of open-source and COTS products relevant to the Hadoop ecosystem;

•Occasional off-hours work required to troubleshoot production;

•Support the development of data mining and machine learning applications

Required Qualifications, Knowledge, & Skills

Only candidates with experience working on a live production-grade Hadoop cluster (on-premise or on VPC) will be considered. Experience working with a Hadoop sandbox on a single-node VM is not ideal.

Key technical proficiencies:

•2+ years Big Data implementation experience leveraging a myriad of high-compression data formats – parquet, avro, json, orc, etc.

•Bachelor’s degree in Computer Science, Engineering, Mathematics, or related mathematical and/or technical fields with Master’s degree preferred

•3+ years hands-on work experience on a production-grade on-premise or VPC-hosted Hadoop cluster;

•2+ years architecting Hadoop data ingestion leveraging Lambda Architecture principles for mix workload processing;

•1+ year implementing real-time data replication from relational databases (RDBMS) to Apache HBase leveraging COTS data replication middleware products;

•3+ years developing data streaming algorithm leveraging kafka queues and storm, flume, spark or flink streaming engine; batch processing using sqoop and oozie or other means;

•5+ years hands-on experience working with UNIX, RHEL Linux, or similar filesystem OS;

•5+ years working with RDBMS and DW products (Oracle & Teradata preferred);

•Experience working with Informatica ETL and IDR for data replication; knowledge of DR architecture leveraging falcon is an asset;

•Building visuals and BI dashboards using Qlik, SAS, Tableau and/or Excel)

Skills Required

  • Education level: None
  • Work experience (years): None

Package

Salary: N/D

All the network: Emploi Adjoint Administratif | Emploi Commis de Bureau | Emploi Commis d’Entrepôt | Emploi Électromécanicien | Emploi Chauffeur | Emploi Construction | Emploi Construction | Emploi en Ventes | Emploi Directeur | Emploi Étudiant | Emploi Restauration | Emploi Réceptionniste | Emploi Sherbrooke | Emploi St-Jérôme | Emploi Trois-Rivières | Emploi Gatineau | Jobs Health Care | Jobs Hospitality | Jobs Ottawa | Job Search Montreal | Jobs Gatineau | Jobs in Sales | Jobs Laval | Jobs Sherbrooke | Jobs Toronto | Emploi Journalier de Production | Emploi Laval | Jobs Mississauga | Jobs Mississauga | Part-Time Jobs Montreal | Part-Time Jobs Retail | Emploi Préposé à l’Entretien Ménager | Jobs Receptionist | Jobs Scarborough | Trabaja en Canada | Emploi Camionneur | Jobs Hamilton | Emploi Montréal | Jobs Markham | Emploi Longueuil | Jobs Vaughan | Jobs Brampton | Emploi Saint-Jean-sur-Richelieu | Emploi Drummondville | Emploi Administration | Emploi Assurances | Emploi Journaliste | Emploi Informatique