IBM Open Platform (IOP) with Apache Hadoop is the first premiere collaborative platform to enable Big Data solutions to be developed on the common set of Apache Hadoop technologies. The Open Data Platform initiative (ODP) is a shared industry effort focused on promoting and advancing the state of Apache Hadoop and Big Data technologies for the enterprise. The current ecosystem is challenged and slowed by fragmented and duplicated efforts between different groups. The ODP Core will take the guesswork out of the process and accelerate many use cases by running on a common platform. It allows enterprises to focus on building business driven applications.This module provides an in-depth introduction to the main components of the ODP core --namely Apache Hadoop (inclusive of HDFS, YARN, and MapReduce) and Apache Ambari -- as well as providing a treatment of the main open-source components that are generally made available with the ODP core in a production Hadoop cluster.
Standorte und Zeitplan
Lage
Beginn
Düsseldorf
(Nordrhein-Westfalen, NRW)
Beginn
auf AnfrageAnmeldung möglich
Hinweise zu diesem Kurs
This intermediate training course is for those who want a foundation of IBM BigInsights. This includes: Big data engineers, data scientist, developers or programmers, administrators who are interested in learning about IBM’s Open Platform with Apache Hadoop.
None, however, knowledge of Linux would be beneficial.
Fragen & Antworten
Ihre Frage hinzufügen
Unsere Berater und andere Nutzer werden Ihnen antworten können
Wir überprüfen Ihre Frage, um sicherzustellen, dass sie an die Veröffentlichungsstandards anpasst. Nach Ihren Antworten haben wir auch entdeckt, dass Sie für diesen Kurs möglicherweise nicht anmelden können. Entweder das wegen Ihrer Ausbildung sein können oder Ihrer Lage und so weiter. Auf jedem Fall wird es besser wenn Sie es mit Ihrer Ausbildungsstätte erkären.
Vielen Dank!
Wir überprüfen Ihre Frage und werden diese in Kürze veröffentlichen.
Oder bevorzugen Sie, dass das Zentrum Sie kontaktiert?
Meinungen
Haben Sie diesen Kurs belegt? Teilen Sie Ihre Meinung
Erfolge dieses Bildungszentrums
2020
Sämtlich Kurse sind auf dem neuesten Stand
Die Durchschnittsbewertung liegt über 3,7
Mehr als 50 Meinungen in den letzten 12 Monaten
Dieses Bildungszentrum ist seit 15 Mitglied auf Emagister
Themen
Access
Apache
IBM
Inhalte
Key Topics:- Unit 1: IBM Open Platform with Apache HadoopExercise 1: Exploring the HDFS- Unit 2: Apache AmbariExercise 2: Managing Hadoop clusters with Apache Ambari- Unit 3: Hadoop Distributed File SystemExercise 3: File access and basic commands with HDFS- Unit 4: MapReduce and YarnTopic 1: Introduction to MapReduce based on MR1Topic 2: Limitations of MR1Topic 3: YARN and MR2Exercise 4: Creating and coding a simple MapReduce jobPossibly a more complex second Exercise- Unit 5: Apache SparkExercise 5: Working with Spark’s RDD to a Spark job- Unit 6: Coordination, management, and governanceExercise 6: Apache ZooKeeper, Apache Slider, Apache Knox- Unit 7: Data MovementExercise 7: Moving data into Hadoop with Flume and Sqoop- Unit 8: Storing and Accessing DataTopic 1: Representing Data: CSV, XML, JSON, and YAMLTopic 2: Open Source Programming Languages: Pig, Hive, and Other [R, Python, etc]Topic 3: NoSQL ConceptsTopic 4: Accessing Hadoop data using HiveExercise 8: Performing CRUD operations using the HBase shellTopic 5: Querying Hadoop data using HiveExercise 9: Using Hive to Access Hadoop / HBase Data- Unit 9: Advanced TopicsTopic 1: Controlling job workflows with OozieTopic 2: Search using Apache SolrNo lab exercisesObjectives:- List and describe the major components of the open-source Apache Hadoop stack and the approach taken by the Open Data Foundation.- Manage and monitor Hadoop clusters with Apache Ambari and related components- Explore the Hadoop Distributed File System (HDFS) by running Hadoop commands.- Understand the differences between Hadoop 1 (with MapReduce 1) and Hadoop 2 (with YARN and MapReduce 2).- Create and run basic MapReduce jobs using command line.- Explain how Spark integrates into the Hadoop ecosystem.- Execute iterative algorithms using Spark’s RDD.- Explain the role of coordination, management, and governance in the Hadoop ecosystem using Apache Zookeeper, Apache Slider, and Apache Knox.- Explore common methods for performing data movement (Configure Flume for data loading of log files, Move data into the HDFS from relational databases using Sqoop)- Understand when to use various data storage formats (flat files, CSV/delimited, Avro/Sequence files, Parquet, etc.).- Review the differences between the available open-source programming languages typically used with Hadoop (Pig, Hive) and for Data Science (Python, R)- Query data from Hive.- Perform random access on data stored in HBase.- Explore advanced concepts, including Oozie and Solr
Zusätzliche Informationen
Förderung durch z.B. Bildungsprämie und NRW-Bildungsscheck möglichUnterrichtsmethodepresentation, discussion, hands-on exercises, demonstrations on the system.Beginn am 1.Tag:Beginn: 09:00Dieses Training bieten wir in Kooperation mit der Integrata AG an.