This repository is going to update based on my challenges in installing and using the Hadoop's tools Spark
-
Updated
Mar 29, 2020
This repository is going to update based on my challenges in installing and using the Hadoop's tools Spark
This project focuses on analyzing movie data using Pyspark tailored for efficient data processing on Hadoop Distributed File System (HDFS)
[Work in progress] Client library for simplified access to Apache Accumulo
Processing and transforming data via Hadoop Ecosystem
HDFS、MapReduce、Hive、Zookeeper原理以及实践操作
資料平行批次與串流處理以及搭建機器學習環境會用到的container
Ambiente com o objetivo de praticar o uso das ferramentas Ansible e Hadoop usando uma única instância
Learn and implement the Hadoop Ecosystem to drive Big Data Analytics.
Apache Hadoop Components Installation Guide on Windows
Some basic procedures for parallel computing in the Hadoop environment
Learning Spark 2 on Cloudera, programming with scala 2.10.
Big Data is Stored and analyzed of various Customer using Hadoop and other tools like Hive, Zookeeper, Hbase and sqoop and all details of the customer is analyzed then result are given.This result is very useful for companies.
Getting tweets using Flume service and analyzing tweets
Add a description, image, and links to the hadoop-ecosystem topic page so that developers can more easily learn about it.
To associate your repository with the hadoop-ecosystem topic, visit your repo's landing page and select "manage topics."