- Spark Design Patterns
Spark Design Patterns is intended to provide various use case scenarios and architectural concepts on prototyping adaptive applications.
mkdocs.yml # Configuration file.
docs/
readme.md # General information
summary.md # Table of contents
units/
design-patterns.md # Design patterns and architectural concepts
scenarios.md # Scenarios in practice
use-cases.md # Typical use cases
user-guide/
changelog.md # Changelog page
code_of_conduct.md # Code of conduct page
code_owners.md # Code owners page
dependency.md # Dependency page
description.md # Description page
faq.md # FAQ page
getting-started.md # Getting started page
troubleshooting.md # Troubleshooting page
upgrade.md # Upgrade page
reporting/
info.md # Informational page
issue_template.md # Issue template page
bug_report.md # Bug report template
custom_report.md # Custom report template
feature_request_template.md # Feature request template
pull_request_template.md # Pull request template
contributing/
info.md # Contributing page
links/
info.md # Links page
The Website documentation is provided by gitbook and can be discovered at url:
To enable documentation serving at localhost you can run the following script
./scripts/build-docs.sh
otherwise, the next steps should be considered:
1) Install/upgrade pip command-line utility:
pip install --upgrade pip
or if you need to install pip for the first time:
wget --no-check-certificate https://bootstrap.pypa.io/get-pip.py
python get-pip.py
2) Install the mkdocs package using pip:
pip install mkdocs
or more conveniently for Windows subsystem:
python -m pip install mkdocs
python -m mkdocs
3) Install the material theme for the website:
python -m pip install mkdocs-material
python -m pip install markdown-include
...
(see requirements.txt)
...
4) Start the server at localhost by running the following command:
python -m mkdocs build --clean --config-file mkdocs.yml
python -m mkdocs serve --verbose --dirtyreload
5) Deploy to GitHub pages:
python -m mkdocs --verbose gh-deploy --force --remote-branch gh-pages
5) Open up localhost in browser to get the default documentation homepage.
Running k8s cluster with tilt command by acquiring k8s deployment configuration:
tilt up
Shutting down k8s cluster with provisioned resources removal:
tilt down --delete-namespaces
The project uses SemVer for versioning. For the versions available, see the tags on this repository.
Spark Patterns is maintained by the following GitHub team-members:
with community support please contact with us if you have some question or proposition.
Please read CONTRIBUTING.md for details on our code of conduct, and the process for submitting pull requests to us (emoji key).
This project follows the all-contributors specification. Contributions of any kind are welcome!
See also the list of contributors who participated in this project.
Spark Patterns is distributed under LGPL version 3 or later, [License]. LGPLv3 is additional permissions on top of GPLv3.
Like
and Spark Patterns ? Consider buying me a coffee :)
- Hey there I really need help with spark well m new to this so it would be nice if someone was down to help
- GCC 14 release brings Ada/GNAT/SPARK improvements
- How to run Ada and SPARK code on NVIDIA GPUs and CUDA
- Co-Developing Programs and Their Proof of Correctness (AdaCore blog)
- CACM article about SPARK...
- [FTSCS23] Does Rust SPARK Joy? Safe Bindings from Rust to SPARK, Applied to the BBQueue Li...
- Memory Safety with Formal Proof Webinar
- SPARK Pro for Proven Memory Safety Webinar - Jan 31st
- Rust and SPARK: Software Reliability for Everyone (2020)
- [VIDEO] SPARK Pro For Embedded System Programming
- What is Apache Spark? The big data platform that crushed Hadoop
- BI meets data science in Microsoft Fabric
- AWS Glue upgrades Spark engines, backs Ray framework
- A Beginnerβs Guide to Learning Hadoop
- Decoding HDFS: Essential Terms Every Data Engineer Should Know
- Apache Hadoop Yarn: Scheduler and ApplicationsManager ηθ¨
- Apache Hive : A primer
- Data Engineering Series 6: Batch Processing with Apache Spark
- Mastering Ninja Resource Management
- HADOOP INSTALLATION ON UBUNTU COMPLETE GUIDE
- Cross Account S3 Reads Using AWS Glue Spark
- Classic Papers in the Field of Big Data
- Introduction To Big Data
- Sunset: Hadoop Developer In Real World cluster
- How to kill a running Spark application?
- How does a consumer know the offset to read after restart in Kafka?
- What is the default number of executors in Spark?
- What is the default number of cores and amount of memory allocated to an application in Spark?
- How to find the number of objects in an S3 bucket?
- Improving Performance with Adaptive Query Execution in Apache Spark 3.0
- Exploring the Power of Apache Spark 3.0: Adaptive Query Execution and More
- What is the difference between spark.sql.shuffle.partitions and spark.default.parallelism in Spark?
- What is the difference between client and cluster deploy modes in Spark?
- Rewriting filters for partition columns
- Cloudera and Hortonworks β why the merger?
- Query Hive tables directly from Kognitio
- Whatβs New in Kognitio 8.2.2
- Using Kognitio 8.2.1 and later versions on HDP Hortonworks
- Reading AVRO data into Kognitio
- Installing Kognitio on Azure HDInsight
- Iβve put data in Hadoop so analytics will be quick, right?
- How will Kognitio on Hadoop work for you?
- Need faster analytics on data held in MapR? Try Kognitio.
- Fixing Hadoop Namenodes in Safe Mode
- Change Data Capture with Kafka Connect and Cassandra
- Moving Data from Confluent Cloud to Cosmos DB
- Generating Artificial Data with Databricks Generator
- Kafka and SIEM/SOAR Tools
- Exporting a Hive Table to CSV
- How Spark Determines Task Numbers and Parallelism
- Helpful Tools for Apache Kafka Developers
- A Warning: VPCs and Distributed Database Platforms
- Tips for Decreasing the Impact of Rebalancing in Kafka Streams
- Common Hadoop questions, errors and there answers
- How to Convert Python script to shell script online
- how to read HDFS file in spark or Scala
- Python script to check disk space Windows
- What are real-time industry applications of Hadoop?
- List of Top Development Courses from Udemy
- Useful Command for cluster management
- How to Run Word Count program in Eclipse with screenshot
- MapReduce | Hadoop Developer Self Learning Outline
- HDFS2 | Hadoop Developer Self Learning Outline
- CRUNCH YOUR WAY IN HADOOP
- HIVE INTERVIEW RELATED PREPARATION
- A USECASE ON TRAVEL APP
- HIVE ON RESCUE- A HEALTHCARE USE_CASE ON STRUCTURED DATA
- WAYS TO BULK LOAD DATA IN HBASE
- MULTIPLE OUTPUT WITH MULTIPLE INPUT FILE NAME
- XML FILE PROCESSING IN HADOOP
- HADOOP POC ON EXCEL DATA WEATHER REPORT ANALYSIS
- HADOOP (PROOF OF CONCEPTS) WEATHER REPORT ANALYSIS
- HIVE 2.1.1 INSTALLATION IN HADOOP 2.7.3 IN UBUNTU 16