Architecture
What is a good Big Data Architecture?
A good data lake enables fast querying of data, in a positive ROI and low maintenance.
Basically, Faster Cheaper and simpler should be the motto for a good Architecture.
Our Architecture Blogs
- JutoLake – Analytics with no boundaries
- Big Data Architecture Best Practices
- 16 Tips to reduce costs on AWS SQL Athena
- DFP Data Transfer Files Use Case | Airflow and BigQuery 93% Cost Reduction
- Big Data Jargon | FAQ’s and everything you wanted to know and didn’t ask about Big Data …
- Recommendations for a new AWS account
- Access management in AWS SageMaker
- Questions and answers on AWS EMR Jupiter
- How to work with maximize resource allocation and Spark dynamic allocation [ AWS EMR Spark ]
- When should we use EMR and When should we use Redshift? EMR VS Redshift
- Flume VS Kafka , basic comparison
- AWS EMR and Hadoop Demystified – Comprehensive training program suggestion for Data Engineers in 200KM/h
Meetups which includes Architecture
English
- Rise, Fall and re-Rise of the Semantic Layer
- My First Petabyte Scale Architecture- Part1
- Architectural Evolution of Amazon internal Data Platform
- Data integration, ETL, ELT, … challenges, and complexities
- Data Mesh: Experimentation to Industrialisation
- Airflow Distributed work loads vs AWS Lambda vs Multi Threaded Python Script
- Fast Analytics with Presto and Alluxio
- A gentle Introduction to Alluxio
- Advanced ETL Demystified
- An Intro to Rivery Through Kits – Plug and Play Data Models
- 80% Cost Reduction in Google Cloud BigQuery
- AWS Big Data Demystified – Part 1
- AWS Big Data Demystified – Part 4
- GCP Big Data Demystified | Investing.com
- Serverless Data Pipelines
- Unify Data Analytics: Any Stack Any Cloud
- Data Pipeline Observability
- Kafka + MQTT : make IoT data available for processing
Hebrew
- The lack of communication between data consumers
- Alluxio – The good, the bad and the ugly
- My First Architecture Parts 1-3
- How Skai leverages Snowflake to move faster
- When was the last time you analyzed 10 trillion records?
- Data Engineering Challenges Demystified
- Version Control for Your Object Storage
- From Block storage to Object storage
- How to create your own SaaS
- Data Engineering Demystified
- Data Engineering Use Cases Demystified
- 200KM/h overview on Big Data in AWS | Part 1
- 200KM/h overview on Big Data in AWS | Part 2
- AWS Big Data in 200KM/h
- AWS Big Data Demystified – Part 2
- AWS Big Data Demystified – Part 3
- Walla’s Migration to AWS Cloud
- Automate all your EMR related activities