enas commited on
Update README.md
Browse files
README.md
CHANGED
|
@@ -162,7 +162,7 @@ This is the **default model** used in the paper.
|
|
| 162 |
|
| 163 |
## Dataset corpus
|
| 164 |
### **Training TABERTA Corpus**
|
| 165 |
-
TABERTA is fine-tuned on WikiDBs, a large-scale corpus of relational databases automatically extracted from Wikidata, containing over 100K databases and 1.6M tables spanning diverse domains and realistic schema designs. WikiDBs is used only for representation learning
|
| 166 |
|
| 167 |
|
| 168 |
### **Experiencing and evaluating Corpus**
|
|
|
|
| 162 |
|
| 163 |
## Dataset corpus
|
| 164 |
### **Training TABERTA Corpus**
|
| 165 |
+
TABERTA is fine-tuned on WikiDBs, a large-scale corpus of relational databases automatically extracted from Wikidata, containing over 100K databases and 1.6M tables spanning diverse domains and realistic schema designs. WikiDBs is used only for representation learning no downstream benchmark queries or relevance labels are observed during training by making the learned embeddings reusable across tasks.
|
| 166 |
|
| 167 |
|
| 168 |
### **Experiencing and evaluating Corpus**
|