
apache/sparkApache Spark™ is a multi-language engine for executing data engineering, data science, and machine learning on single-node machines or clusters. It provides high-level APIs in Scala, Java, Python, and R, and an optimized engine that supports general computation graphs for data analysis. It also supports a rich set of higher-level tools including Spark SQL for SQL and DataFrames, pandas API on Spark for pandas workloads, MLlib for machine learning, GraphX for graph processing, and Structured Streaming for stream processing.
<[***]>
You can find the latest Spark documentation, including a programming guide, on the project web page. This README file only contains basic setup instructions.
The easiest way to start using Spark is through the Scala shell:
bashdocker run -it apache/spark /opt/spark/bin/spark-shell
Try the following command, which should return 1,000,000,000:
scalascala> spark.range(1000 * 1000 * 1000).count()
<[***]>
Use the images on <[***]>
Use the images on <[***]>






manifest unknown 错误
TLS 证书验证失败
DNS 解析超时
410 错误:版本过低
402 错误:流量耗尽
身份认证失败错误
429 限流错误
凭证保存错误
来自真实用户的反馈,见证轩辕镜像的优质服务