File size: 1,088 Bytes
49cf832
 
c02139b
49cf832
 
 
 
 
f3e12fe
 
 
49cf832
0e38adb
244fc1f
3ab57b2
 
aa053f8
244fc1f
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
---
title: README
emoji: 
colorFrom: gray
colorTo: red
sdk: static
pinned: false
---
<p>
    <img src="https://huggingface.co/datasets/loubnabnl/repo-images/resolve/main/bigcode_light.png" alt="drawing" width="440"/>
</p>

<p>
BigCode is an open scientific collaboration working on responsible training of large language models for coding applications.

You can find more information on the main website at <a href="https://www.bigcode-project.org/" class="underline">https://www.bigcode-project.org</a>. You can also follow Big Code on Twitter at <a href="https://twitter.com/BigCodeProject" class="underline">https://twitter.com/BigCodeProject</a>.

In this organization, you can find <a href="https://huggingface.co/datasets/bigcode/the-stack" class="underline">The Stack</a>, a 6.4TB of source code in 358 programming languages from permissive licenses.
You can also find <a href="https://huggingface.co/bigcode/santacoder" class="underline">SantaCoder</a>, a strong 1.1B code generation model trained on Java, JavaScript & Python. In addition to some data governance tools.
</p>