Update README.md
Browse files
README.md
CHANGED
|
@@ -76,24 +76,6 @@ outputs = model.generate(
|
|
| 76 |
# Decode and print the response
|
| 77 |
response = tokenizer.decode(outputs[0][inputs.input_ids.shape[1]:], skip_special_tokens=True)
|
| 78 |
print(response)
|
| 79 |
-
|
| 80 |
-
# Another example: design assistance
|
| 81 |
-
messages_design = [
|
| 82 |
-
{"role": "user", "content": "Explain the key considerations for designing a stable feedback amplifier."}
|
| 83 |
-
]
|
| 84 |
-
text_design = tokenizer.apply_chat_template(
|
| 85 |
-
messages_design,
|
| 86 |
-
tokenize=False,
|
| 87 |
-
add_generation_prompt=True
|
| 88 |
-
)
|
| 89 |
-
inputs_design = tokenizer(text_design, return_tensors='pt').to(model.device)
|
| 90 |
-
outputs_design = model.generate(
|
| 91 |
-
inputs=inputs_design.input_ids,
|
| 92 |
-
attention_mask=inputs_design.attention_mask,
|
| 93 |
-
generation_config=generation_config
|
| 94 |
-
)
|
| 95 |
-
response_design = tokenizer.decode(outputs_design[0][inputs_design.input_ids.shape[1]:], skip_special_tokens=True)
|
| 96 |
-
print(response_design)
|
| 97 |
```
|
| 98 |
|
| 99 |
## Training procedure
|
|
|
|
| 76 |
# Decode and print the response
|
| 77 |
response = tokenizer.decode(outputs[0][inputs.input_ids.shape[1]:], skip_special_tokens=True)
|
| 78 |
print(response)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 79 |
```
|
| 80 |
|
| 81 |
## Training procedure
|