Skip to content

TPU-related docs are somewhat confusing #14330

@dinhanhx

Description

@dinhanhx

📚 Documentation

In Trainer class api,
image
it said not to use accelerator parameter and will be removed in 1.7.0 (current stable version is 1.7.1), BUT tpu_cores said to use accelator.
image
In Accelerator: TPU Training (also stable version), all the examples use accelerator='tpu'.
image

I don't really know which one to follow. And when I head to, What is a Strategy?
image
I don't really understand why we need strategy="ddp_spawn" here when without it, the model is still trained on 8 cores?
Moreover, Learn more of TPU links to examples that don't use any TPU strategy?
image

This example should use precision='bf16'
image


If you enjoy Lightning, check out our other projects! ⚡

  • Metrics: Machine learning metrics for distributed, scalable PyTorch applications.

  • Lite: enables pure PyTorch users to scale their existing code on any kind of device while retaining full control over their own loops and optimization logic.

  • Flash: The fastest way to get a Lightning baseline! A collection of tasks for fast prototyping, baselining, fine-tuning, and solving problems with deep learning.

  • Bolts: Pretrained SOTA Deep Learning models, callbacks, and more for research and production with PyTorch Lightning and PyTorch.

  • Lightning Transformers: Flexible interface for high-performance research using SOTA Transformers leveraging PyTorch Lightning, Transformers, and Hydra.

cc @Borda @carmocca @JackCaoG @steventk-g @Liyang90 @rohitgr7 @kaushikb11

Metadata

Metadata

Assignees

No one assigned

    Labels

    accelerator: tpuTensor Processing UnitdocsDocumentation relatedquestionFurther information is requestedwon't fixThis will not be worked on

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions