Reinforcement Learning from Human Feedback (RLHF)
Mina Parham
AI Engineer
prompt_data = load_dataset("center-for-humans-and-machines/rlhf-hackathon-prompts",
split="train")
prompt_data['prompt'][0]
'How important is climate change?'
Input=
, {{Text}}:
, ###Human:
from datasets import load_dataset
preference_data = load_dataset("trl-internal-testing/hh-rlhf-helpful-base-trl-style",
split="train")
def extract_prompt(text):
# Extract the prompt as the first element in the list
prompt = text[0]["content"]
return prompt
# Apply the extraction function to the dataset
preference_data_with_prompt = preference_data.map(
lambda example: {**example, 'prompt': extract_prompt(example['chosen'])}
)
sample = preference_data_with_prompt.select(range(1))
sample['prompt']
'What vitamins are essential for the body to function?'
sample['chosen']
[ { "content": "What vitamins are essential for the body to function?", "role":
"user" }, { "content": "There are some very important vitamins that ensure the
proper functioning of the body, including Vitamins A, C, D, E, and K along ...}]
Reinforcement Learning from Human Feedback (RLHF)