RLHF
Reinforcement Learning from Human Feedback (RLHF) is a machine learning technique that uses human input to guide the training of AI models.
Reinforcement Learning from Human Feedback (RLHF) is a machine learning technique that uses human input to guide the training of AI models.
An iterative design process that focuses on the users and their needs at every phase of the design process.
A test proposed by Alan Turing to determine if a machine's behavior is indistinguishable from that of a human.
Qualitative data that provides insights into the context and human aspects behind quantitative data.
The idea that self-control or willpower draws upon a limited pool of mental resources that can be used up.
A cognitive bias where the perception of one positive trait influences the perception of other unrelated traits.
The phenomenon where taking a test on material improves long-term retention of that material more than additional study sessions.
A statistical theory that states that the distribution of sample means approximates a normal distribution as the sample size becomes larger, regardless of the population's distribution.
A psychological phenomenon where people develop a preference for things simply because they are familiar with them.