Batch Size

Batch Size refers to the number of data samples processed by the model in one iteration before updating its internal parameters (weights).

Implementation

In LLM training, data is processed in batches (e.g., 4, 8, 32 sequences at a time) rather than one by one or all at once.

Trade-offs

    Mike 3.0

    Send a message to start the chat!

    You can ask the bot anything about me and it will help to find the relevant information!

    Try asking: