Skip to content
This repository was archived by the owner on Nov 19, 2025. It is now read-only.

Commit 34cab21

Browse files
committed
fix typos
Signed-off-by: omahs <73983677+omahs@users.noreply.github.com>
1 parent 35e5bd7 commit 34cab21

5 files changed

Lines changed: 5 additions & 5 deletions

File tree

CONTRIBUTING.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@ We use ``black`` as our style guide. To fix your format run `pip install pre-com
5050
1. Methods should be atomic. A method shouldn't be longer than 75 lines, e.g. can be fit into the computer screen without scrolling.
5151
1. If a method has arguments that don't fit into one line, each argument should be in its own line for readability.
5252
1. Add ``__init__.py`` for every folder.
53-
1. F-strings are prefered to formatted strings.
53+
1. F-strings are preferred to formatted strings.
5454
1. Loggers are preferred to print. Use the logger from NeMo via ``from nemo.utils import logging``
5555
1. Private functions (functions start with ``_``) shouldn't be called outside its host file.
5656
1. If a comment lasts multiple lines, use ``'''`` instead of ``#``.

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ NeMo-Aligner is a scalable toolkit for efficient model alignment. The toolkit ha
1111

1212
The NeMo-Aligner toolkit is built using the NeMo Framework, which enables scalable training across thousands of GPUs using tensor, data, and pipeline parallelism for all alignment components. Additionally, our checkpoints are cross-compatible with the NeMo ecosystem, facilitating inference deployment and further customization (https://github.com/NVIDIA/NeMo-Aligner).
1313

14-
The toolkit is currently in it's early stages. We are committed to improving the toolkit to make it easier for developers to pick and choose different alignment algorithms to build safe, helpful, and reliable models.
14+
The toolkit is currently in its early stages. We are committed to improving the toolkit to make it easier for developers to pick and choose different alignment algorithms to build safe, helpful, and reliable models.
1515

1616
## Key Features
1717

docs/RLHFTraining.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,5 +49,5 @@ We have many optimizations available for performance that you can enable.
4949
* `model.ppo.length_params.max_length`: This sets the max tokens to generate in rollouts, it defaults to half the sequence length of the model. But if you know your model is not too verbose (e.g. after supervised fine tuning) then you can set it to a lower number.
5050

5151
#### Critic performance optimization hyperparameters
52-
* `trainer.ppo.combine_rm_and_critic_server`: When enabled, inference requests to the critic server will also return the rewards. This saves the need of having to run a seperate reward model server..
52+
* `trainer.ppo.combine_rm_and_critic_server`: When enabled, inference requests to the critic server will also return the rewards. This saves the need of having to run a separate reward model server..
5353
* `model.offload_adam_states`: When enabled, offload the distributed adam optimizer states onto CPU during inference. This allows us to save memory during inference for a bigger `trainer.ppo.inference_micro_batch_size`. No effect if the optimizer is not distributed adam.

docs/user-guide/rlhf.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -407,7 +407,7 @@ We test the scaling of our TRT-LLM integration by running Llama3 70B Actor and L
407407
+------------------+-------------------+-----------------------------+----------------------+--------------------+
408408
409409
.. note::
410-
for 64x32 config we used a ``rollout_micro_batch_size`` of 16 instead of 8 due to the additional memory from the the distributed optimizer.
410+
for 64x32 config we used a ``rollout_micro_batch_size`` of 16 instead of 8 due to the additional memory from the distributed optimizer.
411411
412412
413413
We also support running RLHF on Llama3.1 405B Actor and Reward Model. The following numbers are generated with ``num_rollout_samples=128``, ``global_batch_size=128``, reshard turned off, engine offloading set to False.

docs/user-guide/steerlm.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -395,7 +395,7 @@ Run Inference
395395
web_server=False \
396396
port=1427
397397
398-
Please wait for the server to be ready before proceeeding.
398+
Please wait for the server to be ready before proceeding.
399399

400400
#. Create Python helper functions:
401401

0 commit comments

Comments
 (0)