PolarDB-X is a cloud native distributed SQL Database designed for high concurrency, massive storage, complex querying scenarios.
-
Updated
Jun 6, 2025 - Java
PolarDB-X is a cloud native distributed SQL Database designed for high concurrency, massive storage, complex querying scenarios.
Distributed tensors and Machine Learning framework with GPU and MPI acceleration in Python
PolarDB-X is a cloud native distributed SQL Database designed for high concurrency, massive storage, complex querying scenarios.
Command line tool to quickly generate a lot of files in a lot of directories
Building a Bloom Filter on English dictionary words
The project is based on the analysis of the "IBM Transactions for Anti Money Laundering" dataset published on Kaggle. The task is to implement a model which predicts whether or not a transaction is illicit, using the attribute "Is Laundering" as a label to be predicted.
gipa -- compression/decompression tool to package compress and encode massive archive files with floating-point data
Building PageRank algorithm on Web Graph around Stanford.edu using NetworkX python library
This repository contains a LaTeX file that generates a PDF document comprising comprehensive notes for the course "Algorithms for Massive Datasets"
Building node2vec algorithm
Permite abrir e manipular arquivos massivos de texto/dados cujo seria impossivel abrir em um computador, por exemplo um arquivo de texto de +20gb, permite manipular o arquivo pegando apenas as linhas necessárias sem travar o computador por falta de memória.
Scalable, chunk-wise K-anonymization tool based on the Optimal Lattice Anonymization (OLA) algorithm. It is designed to handle large datasets by processing them in manageable chunks, ensuring data privacy while maintaining utility.
Calculate statistical measures of one column in big data Datasets with these simply Hadoop Application
📺 Content Recommendation System for the Netflix Prize Challenge with Collaborative Filtering.
Automated massive geolocator of addresses with parallel processing.
TF-Package: Multiple-Input Multiple-Output Keras Data-Generator for massive and complex datasets
Training the MASSIVE dataset by Amazon(english-US, German-DE and Swahili-KE)
word count in Spark
Stream, parse, manipulate and transform extremly large data ( can be 1 GB or 1TB ) in NodeJS without any process block, memory overflow or bottle neck with peak performance. And also show it in UI with the help of webStreams
Add a description, image, and links to the massive-datasets topic page so that developers can more easily learn about it.
To associate your repository with the massive-datasets topic, visit your repo's landing page and select "manage topics."