Training Requirement: Freelance Trainer โ€“ Big Data & Spark

Location: Pune, Mumbai | Duration: Project-Based / Part-Time
Experience: 10+ years

๐Ÿ“ฉ Email: amritk1@overturerede.com

๐Ÿ“ž Call/WhatsApp: +91 9289118667

You can also explore and apply to current openings here:
๐Ÿ”— https://zurl.co/3fAbr

#BigData #ApacheSpark #FreelanceTrainer #DataEngineering #HiringNow #RemoteJobs #SparkSQL #DataFrames #KafkaStreaming #TechTraining #HadoopEcosystem #MLlib #RealTimeAnalytics

๐ŸŒŸ Just wrapped up the homework for Batch 5 of the Zoomcamp!
I processed and analyzed the yellow_tripdata_2024-10.parquet and taxi_zone_lookup.csv datasets using PySpark and Spark SQL. Feels great to finish a hands-on project! ๐Ÿ†
#DataEngineering #Zoomcamp #DataTalks #ETL #PySpark #SparkSQL
๐Ÿ“ˆ Spark SQL is amazing!
Today I worked on SQL queries within PySpark to analyze and transform large datasets. This is such a powerful tool for data engineering! ๐Ÿš€
#DataEngineering #Zoomcamp #PySpark #DataTalks #SparkSQL

program to be used, for example, in a #streaming environment.

Other MEOS bindings include #Java with #JMEOS, for C# with #MEOS.NET and for #SparkSQL.

I feel like a #sparksql #databricks for the SQL Server professional talk is going to come out of this client engagement. Which will be my first talk in 5? years

Basic premise is much of what you're writing today for SQL Server translates just fine to Databricks.

There's little differences like limit vs top, bigger differences (truncate doesn't reset the identity seed) and Toto, we're not in Kansas any more with stuff like GROUP BY ALL (which is hot for lazy typists like me)

my #spark cluster
is bigger than
your spark cluster

#nerdboast

(hopefully no correlation to my #pyspark #sparksql inefficiency)