Apache Spark is a unified analytics engine for large-scale data processing. It provides high-level APIs in Java, Scala, Python and R, and an optimized engine that supports general execution graphs.
URL: Visit APIs.json URL
- big data, analytics, distributed computing, machine learning, streaming, apache
- Created: 2024-01-01
- Modified: 2024-01-01
REST API for monitoring Spark applications, accessing cluster information, and managing Spark jobs through the Spark UI backend.
Human URL: https://spark.apache.org/docs/latest/monitoring.html
- monitoring, metrics, jobs, stages
Spark module for structured data processing with DataFrame and Dataset APIs. Provides SQL interface and supports various data sources.
Human URL: https://spark.apache.org/sql/
- sql, dataframes, structured data
Scalable, high-throughput, fault-tolerant stream processing of live data streams.
Human URL: https://spark.apache.org/streaming/
- streaming, real-time, data processing
Spark's scalable machine learning library consisting of common learning algorithms and utilities, including classification, regression, clustering, and collaborative filtering.
Human URL: https://spark.apache.org/mllib/
- machine learning, algorithms, ml, data science
Spark API for graphs and graph-parallel computation with a collection of graph algorithms and builders.
Human URL: https://spark.apache.org/graphx/
- graphs, graph processing, analytics
- GitHub Repository
- Downloads
- Getting Started
- Community
- Blog
- Issue Tracker
- License
- Stack Overflow
FN: Apache Software Foundation
Email: dev@spark.apache.org