Now that I got a #datalakehouse in a #Europeancloud I want more: some actual data to query, #pyspark and notebooks and #datavisualisation. In this blogpost I get Project Jupyter Hub running, I get open data from the #EuropeanParliament and try to visualise that data in #PowerBI. Getting the connection working turned out to be an expensive affair.
https://marcel-jan.eu/datablog/2026/01/16/data-engineering-in-the-european-cloud-part-3-adding-data-and-visualisations/
Data engineering in the European cloud – Part 3: Adding data and visualisations – Expedition Data

ADF Interview Questions | Cloud Data Engineer #databricks #pyspark #adf #datafactory #microsoft

Q2. What are the main components of Azure Data Factory? Learn about the main components of Azure Data Factory! This video ... source

https://quadexcel.com/wp/adf-interview-questions-cloud-data-engineer-databricks-pyspark-adf-datafactory-microsoft-5/

ADF Interview Questions | Cloud Data Engineer #databricks #pyspark #adf #datafactory #microsoft - QuadExcel.com

Q2. What are the main components of Azure Data Factory? Learn about the main components of Azure Data Factory! This video ... source

QuadExcel.com

Spark SQL for Data Engineering 1 : I am going to start spark sql sessions as series. #sparksql

Spark SQL Part 1 : I am going to start spark sql sessions as series. #sparksql #deltalake #pyspark ' Databricks Notebooks code for ... source

https://quadexcel.com/wp/spark-sql-for-data-engineering-1-i-am-going-to-start-spark-sql-sessions-as-series-sparksql/

ADF Interview Questions | Cloud Data Engineer #databricks #pyspark #adf #datafactory #microsoft

Q13. How can you optimize the performance of an Azure Data Factory pipeline? Boost the performance of your Azure Data ... source

https://quadexcel.com/wp/adf-interview-questions-cloud-data-engineer-databricks-pyspark-adf-datafactory-microsoft-4/

ADF Interview Questions | Cloud Data Engineer #databricks #pyspark #adf #datafactory #microsoft - QuadExcel.com

Q13. How can you optimize the performance of an Azure Data Factory pipeline? Boost the performance of your Azure Data ... source

QuadExcel.com

ADF Interview Questions | Cloud Data Engineer #databricks #pyspark #adf #datafactory #microsoft

Q23. How did you handle a situation where a pipeline needed to run based on specific business events? Running pipelines ... source

https://quadexcel.com/wp/adf-interview-questions-cloud-data-engineer-databricks-pyspark-adf-datafactory-microsoft-3/

ADF Interview Questions | Cloud Data Engineer #databricks #pyspark #adf #datafactory #microsoft - QuadExcel.com

Q23. How did you handle a situation where a pipeline needed to run based on specific business events? Running pipelines ... source

QuadExcel.com

Ya esta disponible La Experimental #14

🌐 Tendencias #Internet
💻 Gestión de #Git hooks
🧑🏻‍💻 Diseño #TUI con #GoLang
🐍 #Python sin GIL
💾 Guía de #PySpark SQL
🤖 Agente #IA local
🐧 Guía de seguridad #Linux
🌩️ Servicio de monitorización SelfHosted
💼 Informe laboral Tech en España de #manfred

Link: https://open.substack.com/pub/laexperimental/p/le-14?r=54y3on&utm_campaign=post&utm_medium=web&showWelcomeOnShare=true

#tech #technology #selfhosting #gnulinux #cli #trabajo #opensource #development

La Experimental #14

Analisis web, gestor de Git hooks, diseño TUI en Go, Python sin GIL, guía PySpark SQL, agente de IA local, seguridad en Linux, monitorización de servidores self-hosted, informe de empleo tech España.

La Experimental

GCP Interview Questions | Cloud Data Engineer #gcpdataengineer #databricks #pyspark #gcp

Q12. How to handle schema evolution and versioning in a data lake architecture on GCP? Learn how to manage schema ... source

https://quadexcel.com/wp/gcp-interview-questions-cloud-data-engineer-gcpdataengineer-databricks-pyspark-gcp/

GCP Interview Questions | Cloud Data Engineer #gcpdataengineer #databricks #pyspark #gcp - QuadExcel.com

Q12. How to handle schema evolution and versioning in a data lake architecture on GCP? Learn how to manage schema ... source

QuadExcel.com

ADF Interview Questions | Cloud Data Engineer #databricks #pyspark #adf #datafactory #microsoft

Q10. How do you implement parameterization in an Azure Data Factory pipeline? Learn how to implement parameterization in ... source

https://quadexcel.com/wp/adf-interview-questions-cloud-data-engineer-databricks-pyspark-adf-datafactory-microsoft-2/

ADF Interview Questions | Cloud Data Engineer #databricks #pyspark #adf #datafactory #microsoft - QuadExcel.com

Q10. How do you implement parameterization in an Azure Data Factory pipeline? Learn how to implement parameterization in ... source

QuadExcel.com

Продвинутый анализ на PySpark: учимся работать с рекуррентными соотношениями

Обработка и анализ временных последовательностей (временных рядов) достаточно часто встречающаяся задача. Обычно она решается с помощью идентичных подходов и методов. Однако когда анализ временного ряда предполагает выражение каждого последующего элемента через предыдущие, возникают проблемы с эффективностью реализации такого анализа. Это особенно актуально в контексте больших данных. В данной статье я продемонстрирую подход к анализу и вычислению рекуррентных соотношений. В качестве примера будет представлена реализация на базе Apache Spark и Python метода экспоненциальной скользящей средней с использованием DataFrame API. Мы рассмотрим метод агрегации данных, совместимый со Spark Connect, который был добавлен в версию 3.1 (для Scala - начиная с версии фреймворка 3.0), а именно – функцию aggregate.

https://habr.com/ru/companies/axenix/articles/952278/

#apache_spark #pyspark #python #рекуррентные_соотношения #временные_ряды #анализ_данных #spark_connect

Продвинутый анализ на PySpark: учимся работать с рекуррентными соотношениями

Всем привет! Обработка и анализ временных последовательностей (временных рядов) достаточно часто встречающаяся задача. Обычно она решается с помощью идентичных подходов и методов. Однако когда анализ...

Хабр

Part 1 : #PySpark Data Pre-processing Essentials #filtering || #Deduplication || Data Cleansing.

Learn PySpark data pre-processing with our tutorial! Learn the art of filtering and deduplication, essential techniques for cleaning ... source

https://quadexcel.com/wp/part-1-pyspark-data-pre-processing-essentials-filtering-deduplication-data-cleansing/

Part 1 : #PySpark Data Pre-processing Essentials #filtering || #Deduplication || Data Cleansing. - QuadExcel.com

Learn PySpark data pre-processing with our tutorial! Learn the art of filtering and deduplication, essential techniques for cleaning ... source

QuadExcel.com