From dfc60f6b7d9a2bd94ffa95ae3f0f7d3c4fb4efcb Mon Sep 17 00:00:00 2001 From: Ikko Eltociear Ashimine Date: Thu, 20 Jul 2023 01:16:16 +0900 Subject: [PATCH] [LayerNorm] Fix typo in ln_api.cpp unintialized -> uninitialized --- csrc/layer_norm/ln_api.cpp | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/csrc/layer_norm/ln_api.cpp b/csrc/layer_norm/ln_api.cpp index cadf078..abb2e07 100644 --- a/csrc/layer_norm/ln_api.cpp +++ b/csrc/layer_norm/ln_api.cpp @@ -132,7 +132,7 @@ std::vector dropout_add_ln_fwd(const at::Tensor &x0, // Input: TORCH_CHECK(x0.is_contiguous()); // c10::IntArrayRef does not own the storage, so we need to construct a vector. - // Otherwise just constructing IntArrayRef({blah}) will cause unintialized memory because + // Otherwise just constructing IntArrayRef({blah}) will cause uninitialized memory because // blah is then deallocated. std::vector sizes_vec {!x0_subset_.has_value() ? x0.size(0) : x0_subset_.value().size(0), x0.size(1)}; auto sizes = c10::IntArrayRef(sizes_vec); @@ -331,7 +331,7 @@ std::vector dropout_add_ln_bwd(const at::Tensor &dz, // BxSxhidd TORCH_CHECK(hidden_size == cols); // c10::IntArrayRef does not own the storage, so we need to construct a vector. - // Otherwise just constructing IntArrayRef({blah}) will cause unintialized memory because + // Otherwise just constructing IntArrayRef({blah}) will cause uninitialized memory because // blah is then deallocated. std::vector x0_sizes_vec {!x0_subset_.has_value() ? rows : x0_numrows, cols}; auto x0_sizes = c10::IntArrayRef(x0_sizes_vec);