Inside this Business Group
Responsibilities Duties -Navigate the Hadoop ecosystem and know how to leverage or optimize the use of what Hadoop has to offer. (Hive, Pig, Sqoop, Flume, Oozie, Mahout, etc.) -Write MapReduce processing and HDFS storage code -Hadoop development, debugging, and implementation of workflows and common algorithms -Writing and executing joins to link data sets in MapReduce -Integrating Hadoop into an Existing Enterprise. Loading Data from an RDBMS into HDFS by Using Sqoop. Managing Real-Time Data Using Flume. -Use the HDFS architecture, including how HDFS implements file sizes, block sizes, and block abstraction. Understand default replication values and storage requirements for replication. Determine how HDFS stores, reads, and writes files. -Construct proper job configuration parameters, including using JobConf and appropriate properties. Identify the correct procedures for MapReduce job submission. -Analyze the order of operations in a MapReduce job, how data moves from place to place, how partitioners and combiners function, and the sort and shuffle process -Analyze and determine which of Hadoop's data types for keys and values are appropriate for a job. Understand common key and value types in the MapReduce framework and the interfaces they implement -The use cases and usage occasions for HBase, Hadoop, and RDBMS -Designing optimal HBase schemas for efficient data storage and recovery -How to connect to HBase using the Java API, configure the HBase cluster, and administer an HBase cluster -Apache Hadoop and data ETL (extract, transform, load), ingestion, and processing with Hadoop tools -Joining multiple data sets and analyzing disparate data with Pig -Organizing data into tables, performing transformations, and simplifying complex queries with Hive -Performing real-time interactive analyses on massive data sets stored in HDFS or HBase using SQL with Impala -How to pick the best tool for a given task in Hadoop, achieve interoperability, and manage recurring workflows Qualifications Experience -Bachelor's degree in Engineering/Science in Computer Science or Information Technology/Systems. - 10+ years of professional IT experience -Strong knowledge of software development life cycle, software testing techniques, test methodologies and framework, strong coding/scripting instincts, as test automation is more similar to developing code -Strong software logic design, analytical/analysis and problem solving skills -Domain knowledge in Data Analysis, BI and Analytics methodologies using SQL, Microsoft BI, SAP BOBJ and Tableau would be an added advantage. -Strong communication skills -Problem solving capabilities -Team work and maintain good attitude -Excellent team player and ability to work in a highly-dynamic environment.
Intel's Information Technology Group (IT) designs, deploys and supports the information technology architecture and hardware/software applications for Intel. This includes the LAN, WAN, telephony, data centers, client PCs, backup and restore, and enterprise applications. IT is also responsible for e-Commerce development, data hosting and delivery of Web content and services.