Datasets:

Modalities:
Text
Formats:
webdataset
Size:
< 1K
ArXiv:
Libraries:
Datasets
WebDataset
huyang0905 commited on
Commit
c4f10a3
·
1 Parent(s): ceaa1e2

upload readme

Browse files
Files changed (1) hide show
  1. README.md +47 -0
README.md ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # DataBack: Dataset of CNF Formulas and Backbone Variable Phases
2
+
3
+ ## What is DataBack
4
+ `DataBack` is a dataset that consists of SAT CNF formulas, each labeled with the phases of its backbone variables. Within `DataBack`, there are two distinct subsets: the pre-training set, named `DataBack-PT`, and the fine-tuning set, named `DataBack-FT`. The state-of-the-art backbone extractor, `CadiBack`, has been employed to obtain the backbone labels. Due to the increased complexity of the fine-tuning formulas, we have allocated a timeout of 1,000 seconds for the pre-training formulas and 5,000 seconds for the fine-tuning ones.
5
+
6
+ The `DataBack` dataset has been employed to both pre-train and fine-tune our `NeuroBack` model, which has demonstrated significant improvements in SAT solving efficiency. For an in-depth exploration of `DataBack`, please refer to [our `NeuroBack` paper](https://arxiv.org/pdf/2110.14053.pdf).
7
+
8
+ ## Authors
9
+ Wenxi Wang, Yang Hu, Mohit Tiwari, Sarfraz Khurshid, Ken McMillan, Risto Miikkulainen
10
+
11
+ ## Publication
12
+ If you use `DataBack` in your research, please kindly cite [our paper](https://arxiv.org/pdf/2110.14053.pdf):
13
+ ```bib
14
+ @article{wang2023neuroback,
15
+ author = {Wang, Wenxi and
16
+ Hu, Yang and
17
+ Tiwari, Mohit and
18
+ Khurshid, Sarfraz and
19
+ McMillan, Kenneth L. and
20
+ Miikkulainen, Risto},
21
+ title = {NeuroBack: Improving CDCL SAT Solving using Graph Neural Networks},
22
+ journal={arXiv preprint arXiv:2110.14053},
23
+ year={2021}
24
+ }
25
+ ```
26
+ ## Directory Structure
27
+ ```
28
+ |- original # Original CNFs and their backbone variable phases
29
+ | |- cnf_pt.tar.gz # CNFs for model pre-training
30
+ | |- bb_pt.tar.gz # Backbone phases for pre-training CNFs
31
+ | |- cnf_ft.tar.gz # CNFs for model fine-tuning
32
+ | |- bb_ft.tar.gz # Backbone phases for fine-tuning CNFs
33
+ |
34
+ |- dual # Dual CNFs and their backbone variable phases
35
+ | |- dual_cnf_pt.tar.gz # Dual CNFs for model pre-training
36
+ | |- dual_bb_pt.tar.gz # Backbone phases for dual pre-training CNFs
37
+ | |- dual_cnf_ft.tar.gz # Dual CNFs for model fine-tuning
38
+ | |- dual_bb_ft.tar.gz # Backbone phases for dual fine-tuning CNFs
39
+ ```
40
+
41
+ ## File Naming Convention
42
+ In the original directory, each CNF tar file contains compressed CNF files named: `[cnf_name].cnf.[compression_format]`, where `[compression_format]` could be bz2, lzma, xz, gz, etc. Correspondingly, each backbone tar file comprises compressed backbone files named: `[cnf_name].cnf.backbone.xz`. It is important to note that a compressed CNF file will always share its `[cnf_name]` with its associated compressed backbone file.
43
+
44
+ In the dual directory, the naming convention remains consistent, but with an added `d_` prefix for each compressed CNF or backbone file to indicate it pertains to a dual CNF formula.
45
+
46
+ ## Contact
47
+ Wenxi Wang ([email protected]), Yang Hu ([email protected])