From d22104cd0fabc2f3ff5f76fc6f7cfe56dc165ba5 Mon Sep 17 00:00:00 2001 From: Ricardo-shuo-liu <13838152117@139.com> Date: Mon, 3 Nov 2025 23:11:53 +0800 Subject: [PATCH 1/4] fix-c19-c23 --- _typos.toml | 7 ------- docs/design/data_type/float16.md | 4 ++-- docs/design/dynamic_rnn/rnn_design.md | 2 +- docs/design/dynamic_rnn/rnn_design_en.md | 2 +- docs/design/mkldnn/int8/QAT/C++.md | 2 +- docs/design/motivation/api.md | 2 +- docs/dev_guides/sugon/paddle_c86_cn.md | 2 +- docs/guides/model_convert/update_en.md | 2 +- 8 files changed, 8 insertions(+), 15 deletions(-) diff --git a/_typos.toml b/_typos.toml index 92ea00cff15..64560a9bec4 100644 --- a/_typos.toml +++ b/_typos.toml @@ -40,12 +40,6 @@ Simle = "Simle" Sovler = "Sovler" Successed = "Successed" classfy = "classfy" -contxt = "contxt" -convertion = "convertion" -convinience = "convinience" -correponding = "correponding" -corresonding = "corresonding" -correspoinding = "correspoinding" corss = "corss" creatation = "creatation" creats = "creats" @@ -135,7 +129,6 @@ similary = "similary" simplier = "simplier" skiped = "skiped" softwares = "softwares" -sould = "sould" specail = "specail" sperated = "sperated" splited = "splited" diff --git a/docs/design/data_type/float16.md b/docs/design/data_type/float16.md index 4081fd6903b..af9d2e1d888 100644 --- a/docs/design/data_type/float16.md +++ b/docs/design/data_type/float16.md @@ -93,7 +93,7 @@ To support the above features, two fundamental conversion functions are provided float16 float_to_half_rn(float f); // convert to half precision in round-to-nearest-even mode float half_to_float(float16 h); ``` -which provides one-to-one conversion between float32 and float16. These twos functions will do different conversion routines based on the current hardware. CUDA/ARM instrinsics will be used when the corresonding hardware is available. If the hardware or compiler level does not support float32 to float16 conversion, software emulation will be performed to do the conversion. +which provides one-to-one conversion between float32 and float16. These twos functions will do different conversion routines based on the current hardware. CUDA/ARM instrinsics will be used when the corresponding hardware is available. If the hardware or compiler level does not support float32 to float16 conversion, software emulation will be performed to do the conversion. ## float16 inference In Fluid, a neural network is represented as a protobuf message called [ProgramDesc](https://github.com/PaddlePaddle/docs/blob/develop/docs/design/concepts/program.md), whose Python wrapper is a [Program](https://github.com/PaddlePaddle/docs/blob/develop/docs/design/modules/python_api.md#program). The basic structure of a program is some nested [blocks](https://github.com/PaddlePaddle/docs/blob/develop/docs/design/modules/python_api.md#block), where each block consists of some [variable](https://github.com/PaddlePaddle/docs/blob/develop/docs/design/modules/python_api.md#variable) definitions and a sequence of [operators](https://github.com/PaddlePaddle/docs/blob/develop/docs/design/modules/python_api.md#operator). An [executor](https://github.com/PaddlePaddle/docs/blob/develop/docs/design/concepts/executor.md) will run a given program desc by executing the sequence of operators in the entrance block of the program one by one. @@ -112,7 +112,7 @@ Operators including convolution and multiplication (used in fully-connected laye When these operators are running in float16 mode, the float16 kernel requires those parameter variables to contain weights of Fluid float16 data type. Thus, we need a convenient way to convert the original float weights to float16 weights. -In Fluid, we use tensor to hold actual data for a variable on the c++ end. [Pybind](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/fluid/pybind/tensor_py.h) is used to bind c++ tensors of certain data type with numpy array of the correponding numpy data type on the Python end. Each common c++ built-in data type has a corresponding numpy data type of the same name. However, since there is no built-in float16 type in c++, we cannot directly bind numpy float16 data type with the Fluid float16 class. Since both Fluid float16 and numpy float16 use uint16 as the internal data storage type, we use c++ built-in type `uint16_t` and the corresponding numpy uint16 data type to bridge the gap via [Pybind](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/fluid/pybind/tensor_py.h). +In Fluid, we use tensor to hold actual data for a variable on the c++ end. [Pybind](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/fluid/pybind/tensor_py.h) is used to bind c++ tensors of certain data type with numpy array of the corresponding numpy data type on the Python end. Each common c++ built-in data type has a corresponding numpy data type of the same name. However, since there is no built-in float16 type in c++, we cannot directly bind numpy float16 data type with the Fluid float16 class. Since both Fluid float16 and numpy float16 use uint16 as the internal data storage type, we use c++ built-in type `uint16_t` and the corresponding numpy uint16 data type to bridge the gap via [Pybind](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/fluid/pybind/tensor_py.h). The following code demonstrates how to do the tensor conversion. ```Python diff --git a/docs/design/dynamic_rnn/rnn_design.md b/docs/design/dynamic_rnn/rnn_design.md index 1c5fde8f403..b04b6f624d9 100644 --- a/docs/design/dynamic_rnn/rnn_design.md +++ b/docs/design/dynamic_rnn/rnn_design.md @@ -62,7 +62,7 @@ public: LODTensor LODSliceShared(int level, int elem_begin, int elem_end) const; // copy other's lod_start_pos_, to share LOD info. - // NOTE the LOD info sould not be changed. + // NOTE the LOD info could not be changed. void ShareConstLODFrom(const LODTensor &other) { lod_start_pos_ = other.lod_start_pos_; } diff --git a/docs/design/dynamic_rnn/rnn_design_en.md b/docs/design/dynamic_rnn/rnn_design_en.md index 31153595f0b..4a011cd7ace 100644 --- a/docs/design/dynamic_rnn/rnn_design_en.md +++ b/docs/design/dynamic_rnn/rnn_design_en.md @@ -53,7 +53,7 @@ public: LODTensor LODSliceShared(int level, int elem_begin, int elem_end) const; // copy other's lod_start_pos_, to share LOD info. - // NOTE the LOD info sould not be changed. + // NOTE the LOD info could not be changed. void ShareConstLODFrom(const LODTensor &other) { lod_start_pos_ = other.lod_start_pos_; } diff --git a/docs/design/mkldnn/int8/QAT/C++.md b/docs/design/mkldnn/int8/QAT/C++.md index 3203e3c5bdd..d47ab61802a 100644 --- a/docs/design/mkldnn/int8/QAT/C++.md +++ b/docs/design/mkldnn/int8/QAT/C++.md @@ -51,7 +51,7 @@ To download other Quant models, set the `QUANT_MODEL_NAME` variable to on of the - `ResNet50_qat_channelwise`, with input/output scales in `fake_quantize_range_abs_max` operators and the `out_threshold` attributes, with weight scales in `fake_channel_wise_dequantize_max_abs` operators -### Model convertion +### Model conversion To run this quantiozation approach, first you need to set `AnalysisConfig` first and use `EnableMkldnnInt8` function that converts fake-quant model to INT8 OneDNN one. Examples: diff --git a/docs/design/motivation/api.md b/docs/design/motivation/api.md index bc222564e3e..87eca5bd72a 100644 --- a/docs/design/motivation/api.md +++ b/docs/design/motivation/api.md @@ -54,7 +54,7 @@ def f(in): return o # Create 3 topologies (subnets), they share parameters because all -# correspoinding layers have the same parameter names. +# corresponding layers have the same parameter names. fA = f(paddle.layer.data(input_name="A")) fB = f(paddle.layer.data(input_name="B")) fQ = f(paddle.layer.data(input_name="Q")) diff --git a/docs/dev_guides/sugon/paddle_c86_cn.md b/docs/dev_guides/sugon/paddle_c86_cn.md index 80d11e1ff72..fa35c4b36d1 100644 --- a/docs/dev_guides/sugon/paddle_c86_cn.md +++ b/docs/dev_guides/sugon/paddle_c86_cn.md @@ -33,7 +33,7 @@ ROCm 软件栈本身具备较高的成熟度与完备性,用户根据 ROCm 提 - 动态库加载: 在 [paddle/phi/backends/dynload](https://github.com/PaddlePaddle/Paddle/tree/develop/paddle/phi/backends/dynload) 目录下动态加载 ROCm 加速库及所需 API,如 [hiprand.h](https://github.com/PaddlePaddle/Paddle/tree/develop/paddle/phi/backends/dynload/hiprand.h) [miopen.h](https://github.com/PaddlePaddle/Paddle/tree/develop/paddle/phi/backends/dynload/miopen.h) [rocblas.h](https://github.com/PaddlePaddle/Paddle/tree/develop/paddle/phi/backends/dynload/rocblas.h)等 - Driver/Runtime 适配:主要在 [paddle/fluid/platform/device/gpu](https://github.com/PaddlePaddle/Paddle/tree/develop/paddle/fluid/platform/device/gpu) 目录下对 HIP 和 CUDA 进行了相关 API 的封装,其中在 [gpu_types.h](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/phi/core/platform/device/gpu/gpu_types.h) 少量封装了部分与 CUDA 差异较小的数据类型定义,部分 ROCm 独有代码位于[paddle/phi/core/platform/device/gpu/rocm](https://github.com/PaddlePaddle/Paddle/tree/develop/paddle/phi/core/platform/device/gpu/rocm)目录 - Memory 管理:利用上一步封装好的 Driver/Runtime API 对 [memcpy.cc](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/phi/core/memory/memcpy.cc#L574) 与 [paddle/phi/core/memory/allocation](https://github.com/PaddlePaddle/Paddle/tree/develop/paddle/phi/core/memory/allocation) 目录下的多种 Memory Allocator 进行实现 - - Device Context 管理:利用封装好的 API 实现对设备上下文的管理及设备池的初始化,位于 [device_contxt.h](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/phi/core/platform/device_context.h) + - Device Context 管理:利用封装好的 API 实现对设备上下文的管理及设备池的初始化,位于 [device_context.h](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/phi/core/platform/device_context.h) - 其他设备管理相关的适配接入,如 Profiler, Tracer, Error Message, NCCL 等,代码主要位于 [Paddle/platform](https://github.com/PaddlePaddle/Paddle/tree/develop/paddle/fluid/platform) 目录下 3. 算子注册:主要包括 HIP Kernel 的算子注册,以及 MIOpen 的算子在 ROCm 平台上的注册 - 数据类型支持:除通用数据类型外,还需适配 Paddle 支持的特殊数据类型包括 [float16.h](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/phi/common/float16.h#L144) [complex.h](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/phi/common/complex.h#L88) [bfloat16.h](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/phi/common/bfloat16.h#L65) 等 diff --git a/docs/guides/model_convert/update_en.md b/docs/guides/model_convert/update_en.md index 9526d05b7a7..8ddb743634c 100644 --- a/docs/guides/model_convert/update_en.md +++ b/docs/guides/model_convert/update_en.md @@ -67,7 +67,7 @@ In order to make the API organization more concise and clear, the original direc ### API alias rule -- APIs are created with aliases in different paths for better convinience: +- APIs are created with aliases in different paths for better convenience: - All APIs under device, framework, and tensor directories are aliased in the paddle root directory; all APIs are not aliased in the paddle root directory except a few special APIs. - All APIs in the paddle.nn directory except for the functional directory have aliases in the paddle.nn directory; all APIs in the functional directory have no aliases in the paddle.nn directory. - ** **It is recommended to give preference to aliases with shorter paths**, for example `paddle.add -> paddle.tensor.add`; `paddle.add` is recommended. From 98a9c13b365bf58b5474816567fa302ff4c187ba Mon Sep 17 00:00:00 2001 From: Ricardo-shuo-liu <13838152117@139.com> Date: Tue, 4 Nov 2025 21:39:43 +0800 Subject: [PATCH 2/4] fix-c6-c7-c24-c26 --- _typos.toml | 4 ---- docs/design/motivation/api.md | 2 +- docs/design/others/simple_op_design.md | 4 ++-- docs/dev_guides/custom_device_docs/stream_api_en.md | 2 +- .../convert_from_pytorch/nlp_migration_experiences_cn.md | 4 ++-- 5 files changed, 6 insertions(+), 10 deletions(-) diff --git a/_typos.toml b/_typos.toml index 64560a9bec4..b8a765bd74f 100644 --- a/_typos.toml +++ b/_typos.toml @@ -39,10 +39,6 @@ Similarily = "Similarily" Simle = "Simle" Sovler = "Sovler" Successed = "Successed" -classfy = "classfy" -corss = "corss" -creatation = "creatation" -creats = "creats" dafault = "dafault" datas = "datas" decribe = "decribe" diff --git a/docs/design/motivation/api.md b/docs/design/motivation/api.md index 87eca5bd72a..b2f407f78a4 100644 --- a/docs/design/motivation/api.md +++ b/docs/design/motivation/api.md @@ -61,7 +61,7 @@ fQ = f(paddle.layer.data(input_name="Q")) topology = paddle.layer.less_than( paddle.layer.cross_entropy(fA, fQ), - paddle.layer.corss_entropy(fB, fQ)) + paddle.layer.cross_entropy(fB, fQ)) # Derive parameters required in topology and create them in model. parameters = paddle.parameters.create(topology) diff --git a/docs/design/others/simple_op_design.md b/docs/design/others/simple_op_design.md index c93c21affe9..13c1ec35b0b 100644 --- a/docs/design/others/simple_op_design.md +++ b/docs/design/others/simple_op_design.md @@ -64,7 +64,7 @@ message OpProto { To generate Python code automatically: ```python -def create_python_ops_creatation_functions(): +def create_python_ops_creation_functions(): op_protos = paddle.framework.OpRegistry.get_all_op_proto() for type_name in op_protos: op_proto = op_protos[type_name] @@ -77,7 +77,7 @@ def create_python_ops_creatation_functions(): __impl__.__doc__ = create_doc_string(op_proto) globals()[type_name] = __impl__ -create_python_ops_creatation_functions() +create_python_ops_creation_functions() ``` ### Message from Python to C++ diff --git a/docs/dev_guides/custom_device_docs/stream_api_en.md b/docs/dev_guides/custom_device_docs/stream_api_en.md index 9c854052773..eea401432fa 100644 --- a/docs/dev_guides/custom_device_docs/stream_api_en.md +++ b/docs/dev_guides/custom_device_docs/stream_api_en.md @@ -10,7 +10,7 @@ C_Status (*create_stream)(const C_Device device, C_Stream* stream) ### Description -It creats a stream, which is used to execute asynchronous tasks within the framework. In the stream, tasks are done in order. When the device does not support asynchronous execution, the API is required to be implemented with an empty method. +It creates a stream, which is used to execute asynchronous tasks within the framework. In the stream, tasks are done in order. When the device does not support asynchronous execution, the API is required to be implemented with an empty method. ### Parameter diff --git a/docs/guides/model_convert/convert_from_pytorch/nlp_migration_experiences_cn.md b/docs/guides/model_convert/convert_from_pytorch/nlp_migration_experiences_cn.md index a9de43df157..1642f47c404 100644 --- a/docs/guides/model_convert/convert_from_pytorch/nlp_migration_experiences_cn.md +++ b/docs/guides/model_convert/convert_from_pytorch/nlp_migration_experiences_cn.md @@ -354,7 +354,7 @@ A:有的,例如: 下面以 BERT 模型为例介绍转换的过程: - 如果 PyTorch repo 中已经提供权重,那么可以直接下载并进行后续的转换。huggingface 的 transformers 中提供了大部分模型参数,使用模型权重名称`model_name_or_path`即可加载(如`bert-base-uncased`)。或者从 huggingface 官网直接下载:https://huggingface.co/bert-base-uncased/tree/main -- 如果没有提供,则可以基于 PyTorch 代码,随机生成一个初始化权重(定义完 model 以后,使用`torch.save()` API 保存模型权重),然后将生成的 PyTorch 权重 (`bert_sequence_classfy.pth`) 转换为飞桨模型权重。 +- 如果没有提供,则可以基于 PyTorch 代码,随机生成一个初始化权重(定义完 model 以后,使用`torch.save()` API 保存模型权重),然后将生成的 PyTorch 权重 (`bert_sequence_classify.pth`) 转换为飞桨模型权重。 ```python from transformers import BertModel @@ -366,7 +366,7 @@ PATH = './torch_weight.bin' torch.save(hf_model.state_dict(), PATH) ``` -然后将生成的 PyTorch 权重 (`bert_sequence_classfy.pth`) 转换为飞桨模型权重,转换代码如下(代码解释详见代码后的 FAQ) +然后将生成的 PyTorch 权重 (`bert_sequence_classify.pth`) 转换为飞桨模型权重,转换代码如下(代码解释详见代码后的 FAQ) ```python def convert_pytorch_checkpoint_to_paddle( From d92d74ca6da4d64e730bdd4f6b7bf8df2f8d798b Mon Sep 17 00:00:00 2001 From: Ricardo-shuo-liu <13838152117@139.com> Date: Wed, 5 Nov 2025 22:46:29 +0800 Subject: [PATCH 3/4] fix-d6-d10 --- _typos.toml | 12 +++++++----- docs/design/concepts/tensor.md | 4 ++-- .../error_message_writing_specification_cn.md | 2 +- docs/guides/flags/memory_en.rst | 4 ++-- .../nlp/transformer_in_English-to-Spanish.ipynb | 6 +++--- 5 files changed, 15 insertions(+), 13 deletions(-) diff --git a/_typos.toml b/_typos.toml index 1faa0f64f36..b7317bbfa7b 100644 --- a/_typos.toml +++ b/_typos.toml @@ -3,6 +3,13 @@ extend-exclude = [ # Skip `Accuray ` check in these files "docs/practices/cv/3D_image_classification_from_CT_scans.ipynb", + # Skip `datas ` check in these files + "docs/practices/time_series/autoencoder.ipynb", + "docs/practices/nlp/transformer_in_English-to-Spanish.ipynb", + "docs/guides/06_distributed_training/data_parallel/amp_cn.rst", + "docs/practices/nlp/addition_rnn.ipynb", + "docs/practices/cv/pointnet.ipynb" + ] [default] @@ -36,11 +43,6 @@ Similarily = "Similarily" Simle = "Simle" Sovler = "Sovler" Successed = "Successed" -dafault = "dafault" -datas = "datas" -decribe = "decribe" -decribes = "decribes" -deocder = "deocder" desgin = "desgin" desginated = "desginated" desigin = "desigin" diff --git a/docs/design/concepts/tensor.md b/docs/design/concepts/tensor.md index 17c6628a035..990e4bfa6dd 100644 --- a/docs/design/concepts/tensor.md +++ b/docs/design/concepts/tensor.md @@ -116,12 +116,12 @@ Before writing code, please make sure you already look through Majel Source Code ### Memory Management -`Allocation` manages a block of memory in device(CPU/GPU). We use `Place` to decribe memory location. The details of memory allocation and deallocation are implemented in `Allocator` and `DeAllocator`. Related low-level API such as `hl_malloc_device()` and `hl_malloc_host()` are provided by Paddle. +`Allocation` manages a block of memory in device(CPU/GPU). We use `Place` to describe memory location. The details of memory allocation and deallocation are implemented in `Allocator` and `DeAllocator`. Related low-level API such as `hl_malloc_device()` and `hl_malloc_host()` are provided by Paddle. ### Dim and Array #### Dim -`Dim` decribes the dimension information of an array. +`Dim` describes the dimension information of an array. `DDimVar` is an alias of a specializd class of boost.variant class template. diff --git a/docs/dev_guides/style_guide_and_references/error_message_writing_specification_cn.md b/docs/dev_guides/style_guide_and_references/error_message_writing_specification_cn.md index ba0ca92ce2a..3f4935b2dd1 100644 --- a/docs/dev_guides/style_guide_and_references/error_message_writing_specification_cn.md +++ b/docs/dev_guides/style_guide_and_references/error_message_writing_specification_cn.md @@ -266,7 +266,7 @@ PADDLE_ENFORCE_EQ( ```c++ PADDLE_ENFORCE( - tmp == *data_type || *data_type == dafault_data_type, + tmp == *data_type || *data_type == default_data_type, phi::errors::InvalidArgument( "The DataType of %s Op's duplicable Variable %s must be " "consistent. The current variable type is (%s), but the " diff --git a/docs/guides/flags/memory_en.rst b/docs/guides/flags/memory_en.rst index 74ad5a3910c..8c348eec57d 100644 --- a/docs/guides/flags/memory_en.rst +++ b/docs/guides/flags/memory_en.rst @@ -205,7 +205,7 @@ FLAGS_initial_gpu_memory_in_mb=4096 will allocate 4 GB as initial GPU chunk. Note ------- If you set this flag, the memory size set by FLAGS_fraction_of_gpu_memory_to_use will be overrided by this flag, PaddlePaddle will allocate the initial gpu memory with size specified by this flag. -If you don't set this flag, the dafault value 0 will disable this GPU memory strategy. PaddlePaddle will use FLAGS_fraction_of_gpu_memory_to_use to allocate the initial GPU chunk. +If you don't set this flag, the default value 0 will disable this GPU memory strategy. PaddlePaddle will use FLAGS_fraction_of_gpu_memory_to_use to allocate the initial GPU chunk. @@ -246,7 +246,7 @@ FLAGS_reallocate_gpu_memory_in_mb=1024 will re-allocate 1 GB if run out of GPU m Note ------- If this flag is set, the memory size set by FLAGS_fraction_of_gpu_memory_to_use will be overrided by this flag, PaddlePaddle will re-allocate the gpu memory with size specified by this flag. -If you don't set this flag, the dafault value 0 will disable this GPU memory strategy. PaddlePaddle will use FLAGS_fraction_of_gpu_memory_to_use to re-allocate GPU memory. +If you don't set this flag, the default value 0 will disable this GPU memory strategy. PaddlePaddle will use FLAGS_fraction_of_gpu_memory_to_use to re-allocate GPU memory. FLAGS_use_pinned_memory diff --git a/docs/practices/nlp/transformer_in_English-to-Spanish.ipynb b/docs/practices/nlp/transformer_in_English-to-Spanish.ipynb index 6a44992e642..ea845f3789c 100644 --- a/docs/practices/nlp/transformer_in_English-to-Spanish.ipynb +++ b/docs/practices/nlp/transformer_in_English-to-Spanish.ipynb @@ -1325,7 +1325,7 @@ }, { "cell_type": "code", - "execution_count": 33, + "execution_count": null, "metadata": { "execution": { "iopub.execute_input": "2023-02-03T07:29:14.998922Z", @@ -1424,8 +1424,8 @@ " encoder_outputs = self.encoder(encoder_emb)\n", "\n", " # 解码器\n", - " deocder_emb = self.ps2(decoder_inputs)\n", - " decoder_outputs = self.decoder(deocder_emb, encoder_outputs)\n", + " decoder_emb = self.ps2(decoder_inputs)\n", + " decoder_outputs = self.decoder(decoder_emb, encoder_outputs)\n", "\n", " # dropout\n", " out = self.drop(decoder_outputs)\n", From bea67d8974edc0c6cc954a1a18dbf56c9cefd8bf Mon Sep 17 00:00:00 2001 From: Ricardo-shuo-liu <13838152117@139.com> Date: Thu, 6 Nov 2025 13:04:51 +0800 Subject: [PATCH 4/4] fix-some-qe --- _typos.toml | 7 +------ docs/practices/nlp/transformer_in_English-to-Spanish.ipynb | 2 +- 2 files changed, 2 insertions(+), 7 deletions(-) diff --git a/_typos.toml b/_typos.toml index b7317bbfa7b..4a6fb17b022 100644 --- a/_typos.toml +++ b/_typos.toml @@ -3,12 +3,6 @@ extend-exclude = [ # Skip `Accuray ` check in these files "docs/practices/cv/3D_image_classification_from_CT_scans.ipynb", - # Skip `datas ` check in these files - "docs/practices/time_series/autoencoder.ipynb", - "docs/practices/nlp/transformer_in_English-to-Spanish.ipynb", - "docs/guides/06_distributed_training/data_parallel/amp_cn.rst", - "docs/practices/nlp/addition_rnn.ipynb", - "docs/practices/cv/pointnet.ipynb" ] @@ -30,6 +24,7 @@ Clas = "Clas" arange = "arange" unsupport = "unsupport" Nervana = "Nervana" +datas = "datas" # These words need to be fixed Creenshot = "Creenshot" diff --git a/docs/practices/nlp/transformer_in_English-to-Spanish.ipynb b/docs/practices/nlp/transformer_in_English-to-Spanish.ipynb index ea845f3789c..2a8718a8c4e 100644 --- a/docs/practices/nlp/transformer_in_English-to-Spanish.ipynb +++ b/docs/practices/nlp/transformer_in_English-to-Spanish.ipynb @@ -1325,7 +1325,7 @@ }, { "cell_type": "code", - "execution_count": null, + "execution_count": 33, "metadata": { "execution": { "iopub.execute_input": "2023-02-03T07:29:14.998922Z",