Data engineering foundations Part 2, Building data pipelines with Kafka and Nifi Part 2, Building data pipelines with Kafka and Nifi.

4+ Hours of Video Instruction The Perfect Way to Get Started with Data Pipelines, Kafka, and NiFi Data Engineering Foundations Part 2: Building Data Pipelines with Kafka and NiFi provides over four hours of video introducing you to creating data pipelines at scale with Kafka and NiFi. You learn to w...

Descripción completa

Detalles Bibliográficos
Autor Corporativo: Addison-Wesley Professional (Firm), publisher (publisher)
Otros Autores: Eadline, Doug, 1956- presenter (presenter)
Formato: Video
Idioma:Inglés
Publicado: [Place of publication not identified] : Addison-Wesley Professional [2022]
Edición:[First edition]
Colección:LiveLessons (Indianapolis, Ind.)
Materias:
Ver en Biblioteca Universitat Ramon Llull:https://discovery.url.edu/permalink/34CSUC_URL/1im36ta/alma991009707501606719
Descripción
Sumario:4+ Hours of Video Instruction The Perfect Way to Get Started with Data Pipelines, Kafka, and NiFi Data Engineering Foundations Part 2: Building Data Pipelines with Kafka and NiFi provides over four hours of video introducing you to creating data pipelines at scale with Kafka and NiFi. You learn to work with the Kafka message broker and discover how to establish NiFi dataflow. You also learn about data movement and storage. All software used in videos is open source and freely available for your use and experimentation on the included virtual machine. About the Instructor Doug Eadline, PhD, began his career as a practitioner and a chronicler of the Linux Cluster HPC revolution and now documents big data analytics. Starting with the first Beowulf How To document, Dr. Eadline has written hundreds of articles, white papers, and instructional documents covering virtually all aspects of HPC computing. Prior to starting and editing the popular ClusterMonkey.net website in 2005, he served as editor-in-chief for ClusterWorld Magazine and was Senior HPC Editor for Linux Magazine. Currently, he is a consultant to the HPC industry and writes a monthly column in HPC Admin Magazine. He has practical hands-on experience in many aspects of HPC, including hardware and software design, benchmarking, storage, GPU, cloud, and parallel computing. He is the co-author of the Apache Hadoop YARN book and author of Hadoop Fundamentals LiveLessons and Apache Hadoop YARN LiveLessons. Skill Level: Beginner Intermediate Learn How To: Understand Kafka topics, brokers, and partitions Implement basic Kafka usage modes Use Kafka producers and consumers with Python Utilize the KafkaEsque graphical user interface Understand the core concepts of NiFi Understand NiFi flow and web UI components Understand direct data movement with HDFS Use HBase with Python Happybase Use Sqoop for database movement Who Should Take This Course: Users, developers, and administrators interested in learning the fundamental aspects and operations of date engineering and scalable systems Course Requirements: Basic understanding of programming and development A working knowledge of Linux systems and tools Familiarity with Python Lesson Descriptions: Lesson 7: Working with the Kafka Message Broker In Lesson 7, Doug introduces introduce the Kafka message broker concept and describes the producer-consumer model that enables input data to be reliably decoupled from output requests. Kafka producers and consumers are developed using Python, and internal broker operations are displayed using the Kafkaesque graphical user interface. Lesson 8: Working with NiFi Dataflow Lesson 8 begins with a description of NiFi flow-based programming and then provides several examples that include writing pipeline data to the local file system, then to the Hadoop Distributed File System, and finally to Hadoop Hive tables. The entire flow process is constructed using the NiFi web Graphical User Interface. The creation of portable flow templates for all examples is also presented. Lesson 9: Big Data Movement and Storage Lesson 9 provides you with several methods for moving data to and from the Hadoop Distributed File System. Hands-on examples include direct web downloads and using Python Pydoop to move data. Basic data movement between Apache HBase, Hive, and Spark using Python Happybase and Hive-SQL is also presented. Finally, movement of relational data to and from the Hadoop Distributed File System is demonstrated using Apache Sqoop. About Pearson Video Training: Pearson publishes expert-led video tutorials covering a wide selection of technology topics designed to teach you the skills you need to succeed. These professional and personal technology videos feature world-leading author instructors published by your trusted technology brands: Addison-Wesley, Cisco Press, Pearson IT Certification, Sams, and Que Topics include: IT Certification, Network Security, Cisco Technology, Programming, Web Development, Mobile Development, and more. Learn more about Pearson Video training at http://www.informit.com/video.
Descripción Física:1 online resource (1 video file (4 hr., 29 min.)) : sound, color
ISBN:9780138087029