Perform multihost inference using Pathways

Multihost inference is a method of running model inference that distributes the model across multiple accelerators hosts. This enables the inference of large models that don't fit on a single host. Pathways can be deployed for both batch and real time multihost inference use cases.

Before you begin

Make sure you have:

Run Batch inference using JetStream

JetStream is a throughput and memory-optimized engine for large language model (LLM) inference on XLA devices, primarily Tensor Processing Units (TPUs) written in JAX.

You can use a prebuilt JetStream Docker image to run a batch inference workload, as shown in the following YAML. This container is built from the OSS JetStream project. For more information about MaxText-JetStream flags, see JetStream MaxText server flags. The following example uses Trillium chips (v6e-16) to load the Llama3.1-405b int8 checkpoint and perform inference over it. This example assumes you already have a GKE cluster with at least one v6e-16 nodepool inside it.

Start model server and Pathways

  1. Get credentials to the cluster and add them to your local kubectl context.
          gcloud container clusters get-credentials $CLUSTER \       --zone=$ZONE \       --project=$PROJECT \       && kubectl config set-context --current --namespace=default     
  2. Deploy LeaderWorkerSet (LWS) API.
          VERSION=v0.4.0       kubectl apply --server-side -f "https://github.com/kubernetes-sigs/lws/releases/download/${VERSION}/manifests.yaml"     
  3. Copy and paste the following YAML into a file named pathways-job.yaml: This YAML has been optimized for the v6e-16 slice shape. For more information on how to convert a Meta checkpoint into a JAX compatible checkpoint, follow the checkpoint creation guide in Creating inference checkpoints. As an example, instructions for Llama3.1-405B are provided here Checkpoint conversion for Llama3.1-405B.
        apiVersion: leaderworkerset.x-k8s.io/v1     kind: LeaderWorkerSet     metadata:       name: jetstream-pathways       annotations:         leaderworkerset.sigs.k8s.io/exclusive-topology: cloud.google.com/gke-nodepool     spec:       replicas: 1       leaderWorkerTemplate:         leaderTemplate:           metadata:             labels:                app: jetstream-pathways           spec:             nodeSelector:               cloud.google.com/gke-tpu-accelerator: TPU_ACCELERATOR_TYPE # Example: tpu-v6e-slice               cloud.google.com/gke-tpu-topology: TPU_TOPOLOGY # Example: 4x4             tolerations:             - key: "google.com/tpu"               operator: "Exists"               effect: "NoSchedule"             containers:             - name: pathways-proxy               image: us-docker.pkg.dev/cloud-tpu-v2-images/pathways/proxy_server:latest               args:               - --resource_manager_address=$(LWS_LEADER_ADDRESS):38677               - --server_port=38681               - --gcs_scratch_location=gs://cloud-pathways-staging/tmp               imagePullPolicy: Always               ports:               - containerPort: 38681             - name: pathways-rm               env:                      - name: HOST_ADDRESS                 value: "$(LWS_LEADER_ADDRESS)"               - name: TPU_SKIP_MDS_QUERY                 value: "true"               image: us-docker.pkg.dev/cloud-tpu-v2-images/pathways/server:latest               args:               - --server_port=38677               - --gcs_scratch_location=gs://cloud-pathways-staging/tmp               - --node_type=resource_manager               - --instance_count=1               - --instance_type=tpuv6e:TPU_TOPOLOGY # Example: 4x4               - --temporary_flags_for_debugging=temporary_flag_for_debugging_worker_expected_tpu_chip_config=megachip_tccontrol               imagePullPolicy: Always               ports:               - containerPort: 38677             - name: jax-tpu               image: us-docker.pkg.dev/cloud-tpu-images/inference/jetstream-pathways:v0.2.0               # Optimized settings used to serve Llama3.1-405b.               args:               - MaxText/configs/v5e/inference/llama3_405b_v5e-64.yml               - model_name=llama3.1-405b               - load_parameters_path=GCS_CHECKPOINT_PATH               - max_prefill_predict_length=1024               - max_target_length=2048               - async_checkpointing=false               - steps=1               - ici_fsdp_parallelism=1               - ici_autoregressive_parallelism=2               - ici_tensor_parallelism=8               - scan_layers=false               - weight_dtype=bfloat16               - per_device_batch_size=10               - enable_single_controller=true               - quantization=int8               - quantize_kvcache=true               - checkpoint_is_quantized=true               - enable_model_warmup=true               imagePullPolicy: Always               ports:                - containerPort: 9000               startupProbe:                 httpGet:                   path: /healthcheck                   port: 8000                   scheme: HTTP                 periodSeconds: 1                 initialDelaySeconds: 900                 failureThreshold: 10000               livenessProbe:                 httpGet:                   path: /healthcheck                   port: 8000                   scheme: HTTP                 periodSeconds: 60                 failureThreshold: 10               readinessProbe:                 httpGet:                   path: /healthcheck                   port: 8000                   scheme: HTTP                 periodSeconds: 60                 failureThreshold: 10             - name: jetstream-http               image: us-docker.pkg.dev/cloud-tpu-images/inference/jetstream-http:v0.2.3               imagePullPolicy: Always               ports:               - containerPort: 8000         # The size variable defines the number of worker nodes to be created.         # It must be equal to the number of hosts + 1 (for the leader node).         size: 5         workerTemplate:           spec:             nodeSelector:               cloud.google.com/gke-tpu-accelerator: TPU_ACCELERATOR_TYPE # Example: tpu-v6e-slice               cloud.google.com/gke-tpu-topology: TPU_TOPOLOGY # Example: 4x4             tolerations:             - key: "google.com/tpu"               operator: "Exists"               effect: "NoSchedule"             containers:             - name: worker               args:               - --server_port=38679               - --resource_manager_address=$(LWS_LEADER_ADDRESS):38677               - --gcs_scratch_location=gs://cloud-pathways-staging/tmp               image: us-docker.pkg.dev/cloud-tpu-v2-images/pathways/server:latest               imagePullPolicy: Always               ports:               - containerPort: 38679               resources:                 limits:                   google.com/tpu: "4"     ---      apiVersion: v1     kind: Service     metadata:       name: jetstream-svc     spec:       selector:         app: jetstream-pathways       ports:       - protocol: TCP         name: jetstream-http         port: 8000         targetPort: 8000     
    Replace the following:
    • TPU_ACCELERATOR_TYPE: The TPU accelerator type. For example, tpu-v6e-slice.
    • TPU_TOPOLOGY: The TPU topology. For example, 2x4.
    • GCS_CHECKPOINT_PATH: The GCS path to the checkpoint.
    Apply this YAML. Wait for the PathwaysJob to be scheduled. Once scheduled, the model server may take some time to restore the checkpoint. For the 405B model, this takes ~7 minutes.
  4. Look at the Kubernetes logs to see if the JetStream model server is ready: The workload was named `jetstream-pathways` in the previous YAML, and `0` is the head node.
          kubectl logs -f jetstream-pathways-0 -c jax-tpu       
    The output is similar to the following which indicates the JetStream model server is ready to serve requests:
        2025-03-02 02:15:07,682 - JetstreamLogger - INFO - Initializing the driver with 1 prefill engines and 1 generate engines in interleaved mode     2025-03-02 02:15:07,683 - JetstreamLogger - INFO - Spinning up prefill thread 0.     2025-03-02 02:15:07,683 - JetstreamLogger - INFO - Spinning up transfer thread 0.     2025-03-02 02:15:07,684 - JetstreamLogger - INFO - Spinning up generate thread 0.     2025-03-02 02:15:07,684 - JetstreamLogger - INFO - Spinning up detokenize thread 0.     2025-03-02 02:15:07,685 - JetstreamLogger - INFO - Driver initialized.     ...     ...     ...     INFO:     Started server process [7]     INFO:     Waiting for application startup.     INFO:     Application startup complete.     INFO:     Uvicorn running on http://0.0.0.0:9999 (Press CTRL+C to quit)     

Connect to the model server

You can access the JetStream Pathways deployment using GKE's ClusterIP service. The ClusterIP service is only reachable from within the cluster. Therefore, to access the service from outside the cluster, you must first establish a port-forwarding session by running the following command:

kubectl port-forward pod/${HEAD_POD} 8000:8000

Verify that you can access the JetStream HTTP server by opening a new terminal and running the following command:

curl --request POST \ --header "Content-type: application/json" \ -s \ localhost:8000/generate \ --data \ '{     "prompt": "What are the top 5 programming languages",     "max_tokens": 200 }' 

The initial request can take several seconds to complete due to model warmup. The output should be similar to the following:

{     "response": " for web development?\nThe top 5 programming languages for web development are:\n1. **JavaScript**: JavaScript is the most popular language for web development, used by over 90% of websites for client-side scripting. It's also popular for server-side programming with technologies like Node.js.\n2. **HTML/CSS**: HTML (Hypertext Markup Language) and CSS (Cascading Style Sheets) are not programming languages, but are essential for building websites. HTML is used for structuring content, while CSS is used for styling and layout.\n3. **Python**: Python is a popular language for web development, especially with frameworks like Django and Flask. It's known for its simplicity, flexibility, and large community of developers.\n4. **Java**: Java is a popular language for building enterprise-level web applications, especially with frameworks like Spring and Hibernate. It's known for its platform independence, strong security features, and large community of developers.\n5. **PHP**: PHP is a mature language for web" } 

Disaggregated inference

Disaggregated serving is a technique for running large language models (LLMs) that separates the prefill and decode stages into different processes, potentially on different machines. This allows for better utilization of resources and can lead to improvements in performance and efficiency, especially for large models.

  • Prefill: this stage processes the input prompt and generates an intermediate representation (like a key-value cache). It's often compute intensive.
  • Decode: this stage generates the output tokens, one by one, using the prefill representation. It is typically memory-bandwidth bound.

By separating these stages, disaggregated serving allows for prefill and decode to run in parallel, improving throughput and latency.

To enable disaggregated serving, modify the following YAML to utilize two v6e-8 slices: one for prefill and the other for generate. Before proceeding, ensure your GKE cluster has at least two node pools configured with this v6e-8 topology. For optimal performance, specific XLA flags have been configured.

Create a llama2-70b checkpoint following the same process as llama3.1-405b checkpoint creation, detailed in the previous section.

  1. To launch the JetStream server in disaggregated mode using Pathways, copy and paste the following YAML into a file named pathways-job.yaml:
    apiVersion: leaderworkerset.x-k8s.io/v1 kind: LeaderWorkerSet metadata:   name: jetstream-pathways   annotations:     leaderworkerset.sigs.k8s.io/subgroup-exclusive-topology: cloud.google.com/gke-nodepool spec:   replicas: 1   leaderWorkerTemplate:     subGroupPolicy:       subGroupSize: 2     leaderTemplate:       metadata:         labels:           app: jetstream-pathways       spec:         nodeSelector:           cloud.google.com/gke-tpu-accelerator: TPU_ACCELERATOR_TYPE # Example: tpu-v6e-slice           cloud.google.com/gke-tpu-topology: TPU_TOPOLOGY # Example: 2x4         tolerations:         - key: "google.com/tpu"           operator: "Exists"           effect: "NoSchedule"         containers:         - name: pathways-proxy           image: us-docker.pkg.dev/cloud-tpu-v2-images/pathways/proxy_server:latest           args:           - --resource_manager_address=$(LWS_LEADER_ADDRESS):38677           - --server_port=38681           - --gcs_scratch_location=gs://cloud-pathways-staging/tmp           imagePullPolicy: Always           ports:           - containerPort: 38681         - name: pathways-rm           env:                  - name: HOST_ADDRESS             value: "$(LWS_LEADER_ADDRESS)"           - name: TPU_SKIP_MDS_QUERY             value: "true"           image: us-docker.pkg.dev/cloud-tpu-v2-images/pathways/server:latest           args:           - --server_port=38677           - --gcs_scratch_location=gs://cloud-pathways-staging/tmp           - --node_type=resource_manager           - --instance_count=2           - --instance_type=tpuv6e:TPU_TOPOLOGY # Example: 2x4           imagePullPolicy: Always           ports:           - containerPort: 38677         - name: jax-tpu           image: us-docker.pkg.dev/cloud-tpu-images/inference/jetstream-pathways:v0.2.0           # Optimized settings used to serve Llama2-70b.           args:           - MaxText/configs/base.yml           - tokenizer_path=assets/tokenizer.llama2           - load_parameters_path=GCS_CHECKPOINT_PATH           - max_prefill_predict_length=1024           - max_target_length=2048           - model_name=llama2-70b           - ici_fsdp_parallelism=1           - ici_autoregressive_parallelism=1           - ici_tensor_parallelism=-1           - scan_layers=false           - weight_dtype=bfloat16           - per_device_batch_size=1           - checkpoint_is_quantized=true            - quantization=int8           - quantize_kvcache=true           - compute_axis_order=0,2,1,3           - ar_cache_axis_order=0,2,1,3           - stack_prefill_result_cache=True           # Specify disaggregated mode to run Jetstream           - inference_server=ExperimentalMaxtextDisaggregatedServer_8           - inference_benchmark_test=True           - enable_model_warmup=True           env:           - name: LOG_LEVEL             value: "INFO"           imagePullPolicy: Always           securityContext:             capabilities:               add: ["SYS_PTRACE", "NET_ADMIN", "SYS_TIME"]           ports:            - containerPort: 9000           startupProbe:             httpGet:               path: /healthcheck               port: 8000               scheme: HTTP             periodSeconds: 1             initialDelaySeconds: 240             failureThreshold: 10000           livenessProbe:             httpGet:               path: /healthcheck               port: 8000               scheme: HTTP             periodSeconds: 60             failureThreshold: 100           readinessProbe:             httpGet:               path: /healthcheck               port: 8000               scheme: HTTP             periodSeconds: 60             failureThreshold: 100         - name: jetstream-http           image: us-docker.pkg.dev/cloud-tpu-images/inference/jetstream-http:v0.2.3           imagePullPolicy: Always           ports:           - containerPort: 8000     # The size variable defines the number of worker nodes to be created.     # It must be equal to the number of hosts + 1 (for the leader node).     size: 5     workerTemplate:       spec:         nodeSelector:           cloud.google.com/gke-tpu-accelerator: TPU_ACCELERATOR_TYPE # Example: tpu-v6e-slice           cloud.google.com/gke-tpu-topology: TPU_TOPOLOGY # Example: 2x4         containers:         - name: worker           args:           - --server_port=38679           - --resource_manager_address=$(LWS_LEADER_ADDRESS):38677           - --gcs_scratch_location=gs://cloud-pathways-staging/tmp           image: us-docker.pkg.dev/cloud-tpu-v2-images/pathways/server:latest           imagePullPolicy: Always           ports:           - containerPort: 38679           resources:             limits:               google.com/tpu: "4" ---  apiVersion: v1 kind: Service metadata:   name: jetstream-svc spec:   selector:     app: jetstream-pathways   ports:   - protocol: TCP     name: jetstream-http     port: 8000     targetPort: 8000   
    Replace the following:
    • TPU_ACCELERATOR_TYPE: The TPU accelerator type. For example, tpu-v6e-slice.
    • TPU_TOPOLOGY: The TPU topology. For example, 2x4.
    • GCS_CHECKPOINT_PATH: The GCS path to the checkpoint.
  2. Apply this YAML, the model server will take some time to restore the checkpoint. For the 70B model, this may take ~2 minutes.
      kubectl apply -f pathways-job.yaml       
  3. Look at the Kubernetes logs to see if the JetStream model server is ready:
        kubectl logs -f jetstream-pathways-0 -c jax-tpu     
    You will see output similar to the following which indicates the JetStream model server is ready to serve requests:
        2025-03-02 02:15:07,682 - JetstreamLogger - INFO - Initializing the driver with 1 prefill engines and 1 generate engines in interleaved mode     2025-03-02 02:15:07,683 - JetstreamLogger - INFO - Spinning up prefill thread 0.     2025-03-02 02:15:07,683 - JetstreamLogger - INFO - Spinning up transfer thread 0.     2025-03-02 02:15:07,684 - JetstreamLogger - INFO - Spinning up generate thread 0.     2025-03-02 02:15:07,684 - JetstreamLogger - INFO - Spinning up detokenize thread 0.     2025-03-02 02:15:07,685 - JetstreamLogger - INFO - Driver initialized.     ...     ...     ...     INFO:     Started server process [7]     INFO:     Waiting for application startup.     INFO:     Application startup complete.     INFO:     Uvicorn running on http://0.0.0.0:9999 (Press CTRL+C to quit)   

Connect to the model server

You can access the JetStream Pathways deployment through the GKE's ClusterIP service. The ClusterIP service is only reachable from within the cluster. Therefore, to access the service from outside the cluster, establish a port-forwarding session by running the following command:

kubectl port-forward pod/${HEAD_POD} 8000:8000

Verify that you can access the JetStream HTTP server by opening a new terminal and running the following command:

curl --request POST \ --header "Content-type: application/json" \ -s \ localhost:8000/generate \ --data \ '{     "prompt": "What are the top 5 programming languages",     "max_tokens": 200 }' 

The initial request can take several seconds to complete due to model warmup. The output should be similar to the following:

{     "response": " used in software development?\nThe top 5 programming languages used in software development are:\n\n1. Java: Java is a popular programming language used for developing enterprise-level applications, Android apps, and web applications. Its platform independence and ability to run on any device that has a Java Virtual Machine (JVM) installed make it a favorite among developers.\n2. Python: Python is a versatile language that is widely used in software development, data analysis, artificial intelligence, and machine learning. Its simplicity, readability, and ease of use make it a popular choice among developers.\n3. JavaScript: JavaScript is a widely used programming language for web development, allowing developers to create interactive client-side functionality for web applications. It is also used for server-side programming, desktop and mobile application development, and game development.\n4. C++: C++ is a high-performance programming language used for developing operating systems, games, and other high-performance applications." } 

What's next