Text Generation
Transformers
PyTorch
olmo
Inference Endpoints
upiter commited on
Commit
228b5e4
·
verified ·
1 Parent(s): b9de43f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -3,6 +3,7 @@ license: apache-2.0
3
  datasets:
4
  - bigcode/the-stack
5
  - HuggingFaceFW/fineweb
 
6
  ---
7
 
8
 
@@ -59,4 +60,4 @@ TinyCodeLM models were pretrained from scratch on a single H100 node (four GPUs)
59
  ```
60
 
61
  # Safety
62
- This work explores data-driven mechanisms for improving the quality of language model-generated code. Our synthetic data generation method relies on open-source data and our experiments leverage open-source software and resources. It is important to acknowledge that all language models for code synthesis have the potential to be misused – whether intentionally or unintentionally – for generation of code with vulnerabilities and/or malicious behaviors. Any and all model generated code has the potential to be harmful and must not be executed without precautions.
 
3
  datasets:
4
  - bigcode/the-stack
5
  - HuggingFaceFW/fineweb
6
+ library_name: transformers
7
  ---
8
 
9
 
 
60
  ```
61
 
62
  # Safety
63
+ This work explores data-driven mechanisms for improving the quality of language model-generated code. Our synthetic data generation method relies on open-source data and our experiments leverage open-source software and resources. It is important to acknowledge that all language models for code synthesis have the potential to be misused – whether intentionally or unintentionally – for generation of code with vulnerabilities and/or malicious behaviors. Any and all model generated code has the potential to be harmful and must not be executed without precautions.