Skip to content

Commit

Permalink
Fix merge conflicts
Browse files Browse the repository at this point in the history
Signed-off-by: Vasudevan Rengasamy <[email protected]>
  • Loading branch information
vasunvidia committed Apr 15, 2024
1 parent 5cf6413 commit 360bd3e
Show file tree
Hide file tree
Showing 2 changed files with 4 additions and 4 deletions.
2 changes: 1 addition & 1 deletion nemo/core/optim/distributed_adam.py
Original file line number Diff line number Diff line change
Expand Up @@ -69,7 +69,7 @@ def __init__(
):

# Initialize process groups
if 'process_group' not in kwargs and parallel_state.is_initialized():
if 'process_group' not in kwargs and not parallel_state.is_unitialized():
kwargs['process_group'] = parallel_state.get_data_parallel_group(with_context_parallel=True)
if disable_distributed_parameters:
world_size = torch.distributed.get_world_size()
Expand Down
6 changes: 3 additions & 3 deletions nemo/package_info.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,10 +13,10 @@
# limitations under the License.


MAJOR = 2
MINOR = 0
MAJOR = 1
MINOR = 23
PATCH = 0
PRE_RELEASE = 'rc0.beta'
PRE_RELEASE = 'rc0'

# Use the following formatting: (major, minor, patch, pre-release)
VERSION = (MAJOR, MINOR, PATCH, PRE_RELEASE)
Expand Down

0 comments on commit 360bd3e

Please sign in to comment.