Wei Zhang is a renowned figure in the CAD (Computer-Aided Design) industry in Canada, with over 30 years of experience spanning his native China and Canada. As the founder of a CAD training center, Wei has been instrumental in shaping the skills of hundreds of technicians and engineers in technical drawing and CAD software applications. He is a certified developer with Autodesk, demonstrating his deep expertise and commitment to staying at the forefront of CAD technology. Wei’s passion for education and technology has not only made him a respected educator but also a key player in advancing CAD methodologies in various engineering sectors. His contributions have significantly impacted the way CAD is taught and applied in the professional world, bridging the gap between traditional drafting techniques and modern digital solutions.
Beginning Apache Spark 3 Pdf Link
from pyspark.sql import SparkSession spark = SparkSession.builder .appName("MyApp") .config("spark.sql.adaptive.enabled", "true") .getOrCreate() 3.1 RDD – The Original Foundation RDDs (Resilient Distributed Datasets) are low‑level, immutable, partitioned collections. They provide fault tolerance via lineage. However, they are not recommended for new projects because they lack optimization.
df = spark.read.parquet("sales.parquet") df.filter("amount > 1000").groupBy("region").count().show() You can register DataFrames as temporary views and run SQL: beginning apache spark 3 pdf
spark.stop()
squared_udf = udf(squared, IntegerType()) df.withColumn("squared_val", squared_udf(df.value)) from pyspark
Example:
spark-submit first_spark_app.py spark-submit \ --master yarn \ --deploy-mode cluster \ --num-executors 10 \ --executor-memory 8G \ --executor-cores 4 \ my_etl_job.py Chapter 10: Common Pitfalls and Best Practices | Pitfall | Solution | |----------------------------------|----------------------------------------------| | Using RDDs unnecessarily | Prefer DataFrames + Catalyst optimizer | | Too many shuffles | Use repartition sparingly; leverage bucketing | | Ignoring AQE | Enable it; let Spark 3 optimize dynamically | | Collecting large DataFrames | Use take() or sample() instead of collect() | | Not handling skew | Enable AQE skewJoin or salt the join key | | Long‑running streaming without watermark | Always set watermarks for event‑time processing | Conclusion Apache Spark 3 represents a mature, powerful, and developer‑friendly engine for all data processing needs. Its unified approach – from batch to streaming, from SQL to machine learning – reduces complexity while delivering industry‑leading performance. df = spark
# Read df = spark.read.option("header", "true").csv("path/to/file.csv") df.write.parquet("output.parquet") 4.2 Common Transformations | Operation | Example | |------------------|-------------------------------------------| | Select columns | df.select("name", "age") | | Filter rows | df.filter(df.age > 21) | | Add column | df.withColumn("new", df.value * 2) | | Group and aggregate | df.groupBy("dept").avg("salary") | | Join | df1.join(df2, "id", "inner") | 4.3 Handling Missing Data df.dropna(how="any", subset=["important_col"]) df.fillna("age": 0, "name": "unknown") 4.4 User‑Defined Functions (UDFs) When built‑in functions are insufficient:
