analogllm commited on
Commit
1809c8e
·
verified ·
1 Parent(s): 9cd551b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -18
README.md CHANGED
@@ -76,24 +76,6 @@ outputs = model.generate(
76
  # Decode and print the response
77
  response = tokenizer.decode(outputs[0][inputs.input_ids.shape[1]:], skip_special_tokens=True)
78
  print(response)
79
-
80
- # Another example: design assistance
81
- messages_design = [
82
- {"role": "user", "content": "Explain the key considerations for designing a stable feedback amplifier."}
83
- ]
84
- text_design = tokenizer.apply_chat_template(
85
- messages_design,
86
- tokenize=False,
87
- add_generation_prompt=True
88
- )
89
- inputs_design = tokenizer(text_design, return_tensors='pt').to(model.device)
90
- outputs_design = model.generate(
91
- inputs=inputs_design.input_ids,
92
- attention_mask=inputs_design.attention_mask,
93
- generation_config=generation_config
94
- )
95
- response_design = tokenizer.decode(outputs_design[0][inputs_design.input_ids.shape[1]:], skip_special_tokens=True)
96
- print(response_design)
97
  ```
98
 
99
  ## Training procedure
 
76
  # Decode and print the response
77
  response = tokenizer.decode(outputs[0][inputs.input_ids.shape[1]:], skip_special_tokens=True)
78
  print(response)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
79
  ```
80
 
81
  ## Training procedure