Building Scalable AI with Big Data and Distributed Intelligence

Building Scalable AI with Big Data Algorithms Updated

A code-first course on building scalable data systems with distributed algorithms. Part of the Hands-On AI Science series, designed around Innovation-First Learning principles.

Scalable AI Tasks

Modern AI has outgrown a single machine. Training frontier models, processing web-scale data, and coordinating many agents that act and learn in parallel all demand distributed infrastructure. This course fuses theoretical concepts with engineering ideas, so students can design systems that stay correct, fast, and cost-efficient as problems grow.

Big Data & Distributed Intelligence

Core concepts, models, and ideas behind scaling intelligence: the MapReduce computation model, data partitioning and shuffling, distributed SGD and parameter servers, model and data parallelism, and multi-agent coordination, communication, and emergent behavior.

Tools & Platforms

PySpark, Databricks, Hadoop MapReduce, Horovod, DeepSpeed, Ray, PettingZoo, Gymnasium, Stable Baselines3, and cloud compute (AWS/Azure).

Modular Syllabus

A specific course syllabus is built for each audience: graduate or undergraduate, across engineering, digital health, or computer science.

Innovation Through Tools Mastery

As AI and mature libraries handle standard tasks, professional developers must focus on innovation. Student projects tackle new use cases by generating unique data and fine-tuning task-specific scalable AI systems.

Guided Student Projects

Students begin their projects while learning the material and enrich them as new concepts arrive. Each team gives several in-class presentations for discussion and feedback.

Typical Weekly Schedule

Week 1

MapReduce Fundamentals

Hadoop, word count, mapper/reducer

Week 2

Spark & Distributed DataFrames

PySpark, Databricks, transformations

Week 3

Advanced MapReduce Algorithms

Joins, sorting, graph algorithms at scale

Week 4

Distributed ML Pipelines

Spark MLlib, feature engineering

Week 5

Project Proposal Presentations

Student proposals, peer feedback

Week 6

Distributed Deep Learning Training

Horovod, DeepSpeed, data parallelism

Week 7

Parameter Servers & Model Parallelism

Ray, gradient synchronization

Week 8

Interim Project Presentations

Progress demos, instructor feedback

Week 9

Multi-Agent RL Foundations

Gymnasium, Stable Baselines3

Week 10

Multi-Agent Environments

PettingZoo, cooperative/competitive

Week 11

Communication & Emergent Behavior

Agent coordination, shared rewards

Week 12

Scalable Inference & Optimization

Model distillation, quantization, ONNX

Week 13

Final Project Presentations

Live demos, peer evaluation

Building Language AI

Language AI

LLMs and Agents

Building Vision AI

Vision AI

Foundation and Generative Models

Building Temporal AI

Temporal AI

Sequential Intelligence and RL