Gallery of Apache Zeppelin notebooks using Enth-Spark-AI.
-
Updated
Jun 15, 2017
Apache Spark is an open source distributed general-purpose cluster-computing framework. It provides an interface for programming entire clusters with implicit data parallelism and fault tolerance.
Gallery of Apache Zeppelin notebooks using Enth-Spark-AI.
An image for running Scala Jupyter notebooks and Apache Spark in the cloud on OpenShift
Big Data Management related Zeppelin notebooks
A repository for ipython notebook backup
python rdd notebook in apache spark
pyspark notebook with movie lens dataset
text analysis ipython notebooks for text analysis
Sample notebooks on Azure Databricks for ETL
Heart disease classification with data mining(Zeppelin Notebook)
Notebooks for Python and Spark for Big Data
Apache Spark cluster connected to a Jupyter Notebook instance
Juypiter Notebooks to demonstrate the Lambda Architecture with Kafka Streams and Apache Spark
2019 Canadian Federal Election: Calculating the results using Apache Spark (Databricks notebook in Scala)
Deploy apache spark in client mode on Kubernetes cluster, integrate with Jupyter notebook through Jupyterhub server.
Jupyter notebook portant portant sur l'analyse des bases de données annuelles des accidents corporels de la circulation routière
Pyspark and Spark [ My Notes and all practise Notebook ]
Simulating a consultancy project for Repsol, the repository contains both the code notebook and the analysis.
Performed Big Data Analysis on Bundesliga Football League Dataset using tools PySpark, spark-SQL, and numpy and done in Jupyter Notebook.
Performing tasks on 5 million records CSV data using PySpark and Elastisearch (No Logstash!). NOTE: The tasks have been performed in both Jupyter Notebook as well as in Python script and may differ slightly.
Created by Matei Zaharia
Released May 26, 2014