Across
- 6. Dense vector representation of text
- 8. Algorithm for computing gradients in neural nets
- 11. Neural architecture that made modern LLMs possible
- 13. Method of generating text from probabilities
- 14. Sampling method that keeps cumulative probability mass p
- 17. Approach where no examples are provided
- 18. Group of training examples processed together
- 20. Random selection method used during generation
Down
- 1. Weights inside a neural network
- 2. Conversational application of LLMs
- 3. Basic unit of text input to a model
- 4. Adapting a pretrained model to a specific task
- 5. Process that splits text into tokens
- 7. To run a model to get predictions (verb)
- 9. Set of token strings a model recognizes
- 10. Learning approach using a small number of examples
- 12. Mechanism that lets models focus on important tokens
- 15. User input that guides the model's response
- 16. Objective function minimized during training
- 19. GPU platform commonly used to accelerate training
