Spark It is a necessary skill for big data development . The question often asked in an interview is Spark What is it? , Or please introduce
Spark, Today's article mainly explains this problem . Many people's answers are not accurate enough , The most accurate description of this problem can be found on the official website .
<>1. Overall introduction
Open the official website to see a line of eye-catching tables and :
Unified engine for large-scale data analytics
Translate it : Unified engine for large-scale data analysis . Keep looking down :
What is Apache Spark™?
Apache Spark™ is a multi-language engine for executing data engineering, data
science, and machine learning on single-node machines or clusters.
Here is the answer to our question :Apache Spark™ Is a multilingual engine , Used to perform data engineering on single node machines or clusters , Data science and machine learning .
Summarize the main points :Spark Is a computing engine , For the calculation of large-scale data , Support multiple programming languages .
<>2. features
The above is a general description , Introduction to some more specific features , The official website also made an answer :
Key features
Simple. Fast. Scalable. Unified.
Spark The characteristics of are summarized in four words : simple , Fast , Scalable , Unity . A more specific description is also given on the official website :
Batch/streaming data
Unify the processing of your data in batches and real-time streaming, using
your preferred language: Python, SQL, Scala, Java or R.
Batch processing / Stream processing : have access to Python,SQL,Scala,Java or R, Unified data processing through batch processing and real-time streaming processing .
SQL analytics
Execute fast, distributed ANSI SQL queries for dashboarding and ad-hoc
reporting. Runs faster than most data warehouses.
SQL analysis : Fast execution for dashboards and interim reports , Distributed ANSI SQL query . Faster than most data warehouses .
Data science at scale
Perform Exploratory Data Analysis (EDA) on petabyte-scale data without having
to resort to downsampling
Large scale data science : yes PB Series for exploratory data analysis (EDA), Without down sampling
Machine learning
Train machine learning algorithms on a laptop and use the same code to scale
to fault-tolerant clusters of thousands of machines.
machine learning : Training machine learning algorithm on notebook computer , And use the same code to expand to a fault-tolerant cluster of thousands of machines .
<>3. ecology
Apache Spark™ integrates with your favorite frameworks, helping to scale them
to thousands of machines.
Data science and Machine learning
SQL analytics and BI
Storage and Infrastructure
Spark Integrates multiple frameworks , Ability to extend these frameworks to thousands of machines . These frameworks include :
* Data science and machine learning :scikit-learn,pandas,TensorFlow,PyTorch,mlflow,R
* SQL Analysis and BI:Superset,Power BI,Looker,redash,tableau,dbt
* Storage and infrastructure :Elasticsearch,MongoDB,Kafka,delta
lake,kubernetes,Airflow,Parquet,SQL Server,cassandra,orc
<>4. Core module
Spark Core: Provided Spark The most basic and core functions ,Spark Other functions such as :Spark SQL,Spark
Streaming,GraphX,MLlib All in Spark Core Based on .
Spark SQL:Spark Components used to manipulate structured data . adopt Spark SQL, Users can use SQL perhaps Apache Hive Version SQL
dialect (HQL) To query data .
Spark Streaming:Spark Components of streaming computing for real-time data on the platform , Provides a rich way to process data streams API.
Spark MLlib:MLlib yes Spark A machine learning algorithm library provided by .MLlib
Not only provides model evaluation , Additional functions such as data import , Some lower level machine learning primitives are also provided .
Spark GraphX:GraphX yes Spark Framework and algorithm library for graph computing .
<>5. summary
At the end of the article Spark What is this problem to make a summary :
* Spark Is a memory based fast , currency , Scalable big data analysis and calculation engine .
* Spark Core Provided in Spark The most basic and core functions .
* Spark SQL yes Spark Components used to manipulate structured data . adopt Spark SQL, Users can use SQL perhaps Apache Hive Version
SQL dialect (HQL) To query data .
* Spark Streaming yes Spark Components of streaming computing for real-time data on the platform , Provides a rich way to process data streams API.
Technology
Daily Recommendation