Standard Hadoop MapReduce Tasks using Java
-
Updated
Nov 16, 2018 - Java
Standard Hadoop MapReduce Tasks using Java
Collection of homework (mostly Spark-based) from the course "Big Data Computing" - University of Padua.
SUTD 2021 50.043 Database and Big Data Systems Code Dump
Flink SQL 实战 -中文博客专栏
Collection of homework (mostly Spark-based) from the course "Big Data Computing" - University of Padua.
Fetch data from Twitter and push it through Kafka to Spark then HDFS
Large-scale data computing word count project
MapReduce Job Development, RDDs Programming, Medical Data Management, Sales Analysis, And Efficient Data Integration For Big Data Analysis. Spark: Big Data Processing, SQOOP Integration, And Spark Structured Streaming For Real-Time Data.
Reservoir Sampling for Group-By Queries in Flink Platform. Answering effectively Single Aggregate.
Eskimo is a state of the art Big Data Infrastructure and Management Web Console to build, manage and operate Big Data 2.0 Analytics clusters on Kubernetes. This is the git repository of Eskimo Community Edition.
This repository contains an Apache Flink application for real-time sales analytics built using Docker Compose to orchestrate the necessary infrastructure components, including Apache Flink, Elasticsearch, and Postgres
Add a description, image, and links to the big-data-processing topic page so that developers can more easily learn about it.
To associate your repository with the big-data-processing topic, visit your repo's landing page and select "manage topics."