Spark SQL for Data Engineering 1 : I am going to start spark sql sessions as series. #sparksql

Spark SQL Part 1 : I am going to start spark sql sessions as series. #sparksql #deltalake #pyspark ' Databricks Notebooks code for ... source

https://quadexcel.com/wp/spark-sql-for-data-engineering-1-i-am-going-to-start-spark-sql-sessions-as-series-sparksql/

๐ŸŒŸ Just wrapped up the homework for Batch 5 of the Zoomcamp!
I processed and analyzed the yellow_tripdata_2024-10.parquet and taxi_zone_lookup.csv datasets using PySpark and Spark SQL. Feels great to finish a hands-on project! ๐Ÿ†
#DataEngineering #Zoomcamp #DataTalks #ETL #PySpark #SparkSQL
๐Ÿ“ˆ Spark SQL is amazing!
Today I worked on SQL queries within PySpark to analyze and transform large datasets. This is such a powerful tool for data engineering! ๐Ÿš€
#DataEngineering #Zoomcamp #PySpark #DataTalks #SparkSQL

program to be used, for example, in a #streaming environment.

Other MEOS bindings include #Java with #JMEOS, for C# with #MEOS.NET and for #SparkSQL.

I feel like a #sparksql #databricks for the SQL Server professional talk is going to come out of this client engagement. Which will be my first talk in 5? years

Basic premise is much of what you're writing today for SQL Server translates just fine to Databricks.

There's little differences like limit vs top, bigger differences (truncate doesn't reset the identity seed) and Toto, we're not in Kansas any more with stuff like GROUP BY ALL (which is hot for lazy typists like me)

my #spark cluster
is bigger than
your spark cluster

#nerdboast

(hopefully no correlation to my #pyspark #sparksql inefficiency)