ksennr commited on
Commit
3adbc5d
·
verified ·
1 Parent(s): fd82d42

updated appendix

Browse files
Files changed (1) hide show
  1. README.md +1 -3
README.md CHANGED
@@ -108,7 +108,7 @@ For any questions or issues with the dataset, please contact the author at kilia
108
 
109
  ## Appendix
110
 
111
- ##### Code to Generate the Train-Test Split on full_data in Python
112
  ```
113
  from sklearn.model_selection import train_test_split
114
 
@@ -119,7 +119,6 @@ template_names = lexicographicDataWikidataSPARQL['template_name'].unique()
119
  test_set = pd.DataFrame()
120
  train_set = pd.DataFrame()
121
 
122
-
123
  for template_name in template_names:
124
  # get the samples for the template_name
125
  samples = lexicographicDataWikidataSPARQL[lexicographicDataWikidataSPARQL['template_name'] == template_name]
@@ -135,7 +134,6 @@ for template_name in template_names:
135
  print(f"{template_name} has more than 20 samples")
136
  train, test = train_test_split(samples, test_size=20)
137
 
138
-
139
  test_set = pd.concat([test_set, test])
140
  train_set = pd.concat([train_set, train])
141
  ```
 
108
 
109
  ## Appendix
110
 
111
+ #### 1. Code for Generating the Train-Test Split on full_data in Python
112
  ```
113
  from sklearn.model_selection import train_test_split
114
 
 
119
  test_set = pd.DataFrame()
120
  train_set = pd.DataFrame()
121
 
 
122
  for template_name in template_names:
123
  # get the samples for the template_name
124
  samples = lexicographicDataWikidataSPARQL[lexicographicDataWikidataSPARQL['template_name'] == template_name]
 
134
  print(f"{template_name} has more than 20 samples")
135
  train, test = train_test_split(samples, test_size=20)
136
 
 
137
  test_set = pd.concat([test_set, test])
138
  train_set = pd.concat([train_set, train])
139
  ```