<?xml version="1.0" encoding="utf-8" ?><rss version="2.0"><channel><title>Bing: Spark SQL Book</title><link>http://www.bing.com:80/search?q=Spark+SQL+Book</link><description>Search results</description><image><url>http://www.bing.com:80/s/a/rsslogo.gif</url><title>Spark SQL Book</title><link>http://www.bing.com:80/search?q=Spark+SQL+Book</link></image><copyright>Copyright © 2026 Microsoft. All rights reserved. These XML results may not be used, reproduced or transmitted in any manner or for any purpose other than rendering Bing results within an RSS aggregator for your personal, non-commercial use. Any other use of these results requires express written permission from Microsoft Corporation. By accessing this web page or using these results in any manner whatsoever, you agree to be bound by the foregoing restrictions.</copyright><item><title>Apache Spark™ - Unified Engine for large-scale data analytics</title><link>https://spark.apache.org/</link><description>Apache Spark is a multi-language engine for executing data engineering, data science, and machine learning on single-node machines or clusters.</description><pubDate>Thu, 26 Mar 2026 02:19:00 GMT</pubDate></item><item><title>Overview - Spark 4.1.1 Documentation</title><link>https://spark.apache.org/docs/latest/</link><description>Spark Connect is a new client-server architecture introduced in Spark 3.4 that decouples Spark client applications and allows remote connectivity to Spark clusters.</description><pubDate>Wed, 08 Apr 2026 01:42:00 GMT</pubDate></item><item><title>Documentation | Apache Spark</title><link>https://spark.apache.org/documentation.html</link><description>Hands-On Exercises Hands-on exercises from Spark Summit 2014. These let you install Spark on your laptop and learn basic concepts, Spark SQL, Spark Streaming, GraphX and MLlib. Hands-on exercises from Spark Summit 2013. These exercises let you launch a small EC2 cluster, load a dataset, and query it with Spark, Shark, Spark Streaming, and MLlib.</description><pubDate>Tue, 07 Apr 2026 04:57:00 GMT</pubDate></item><item><title>Quick Start - Spark 4.1.1 Documentation</title><link>https://spark.apache.org/docs/latest/quick-start.html</link><description>Quick Start Interactive Analysis with the Spark Shell Basics More on Dataset Operations Caching Self-Contained Applications Where to Go from Here This tutorial provides a quick introduction to using Spark. We will first introduce the API through Spark’s interactive shell (in Python or Scala), then show how to write applications in Java, Scala, and Python. To follow along with this guide ...</description><pubDate>Wed, 08 Apr 2026 13:45:00 GMT</pubDate></item><item><title>PySpark Overview — PySpark 4.1.1 documentation - Apache Spark</title><link>https://spark.apache.org/docs/latest/api/python/index.html</link><description>PySpark Overview # Date: Jan 02, 2026 Version: 4.1.1 Useful links: Live Notebook | GitHub | Issues | Examples | Community | Stack Overflow | Dev Mailing List | User Mailing List PySpark is the Python API for Apache Spark. It enables you to perform real-time, large-scale data processing in a distributed environment using Python. It also provides a PySpark shell for interactively analyzing your ...</description><pubDate>Wed, 08 Apr 2026 21:09:00 GMT</pubDate></item><item><title>Examples - Apache Spark</title><link>https://spark.apache.org/examples.html</link><description>Apache Spark ™ examples This page shows you how to use different Apache Spark APIs with simple examples. Spark is a great engine for small and large datasets. It can be used with single-node/localhost environments, or distributed clusters. Spark’s expansive API, excellent performance, and flexibility make it a good option for many analyses.</description><pubDate>Wed, 08 Apr 2026 12:19:00 GMT</pubDate></item><item><title>Spark SQL &amp; DataFrames | Apache Spark</title><link>https://spark.apache.org/sql/</link><description>Spark SQL is Spark's module for working with structured data, either within Spark programs or through standard JDBC and ODBC connectors.</description><pubDate>Tue, 07 Apr 2026 20:20:00 GMT</pubDate></item><item><title>Spark SQL and DataFrames - Spark 4.1.1 Documentation</title><link>https://spark.apache.org/docs/latest/sql-programming-guide.html</link><description>Spark SQL, DataFrames and Datasets Guide Spark SQL is a Spark module for structured data processing. Unlike the basic Spark RDD API, the interfaces provided by Spark SQL provide Spark with more information about the structure of both the data and the computation being performed.</description><pubDate>Wed, 08 Apr 2026 06:07:00 GMT</pubDate></item><item><title>Spark Streaming - Spark 4.1.1 Documentation - Apache Spark</title><link>https://spark.apache.org/docs/latest/streaming-programming-guide.html</link><description>Spark Streaming is an extension of the core Spark API that enables scalable, high-throughput, fault-tolerant stream processing of live data streams. Data can be ingested from many sources like Kafka, Kinesis, or TCP sockets, and can be processed using complex algorithms expressed with high-level functions like map, reduce, join and window.</description><pubDate>Thu, 26 Mar 2026 19:44:00 GMT</pubDate></item><item><title>Quickstart: DataFrame — PySpark 4.1.1 documentation - Apache Spark</title><link>https://spark.apache.org/docs/latest/api/python/getting_started/quickstart_df.html</link><description>Quickstart: DataFrame # This is a short introduction and quickstart for the PySpark DataFrame API. PySpark DataFrames are lazily evaluated. They are implemented on top of RDD s. When Spark transforms data, it does not immediately compute the transformation but plans how to compute later. When actions such as collect() are explicitly called, the computation starts. This notebook shows the basic ...</description><pubDate>Tue, 07 Apr 2026 11:38:00 GMT</pubDate></item></channel></rss>