damerajee commited on
Commit
f9197bc
·
verified ·
1 Parent(s): 3144f8d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -24,7 +24,7 @@ task_categories:
24
  language:
25
  - hi
26
  - en
27
- pretty_name: 'long-context '
28
  size_categories:
29
  - 100K<n<1M
30
  ---
@@ -33,11 +33,11 @@ size_categories:
33
 
34
  This dataset was filtered from AI4BHarat dataset [sangraha](https://huggingface.co/datasets/ai4bharat/sangraha),which is the largest high-quality, cleaned Indic language pretraining data containing 251B tokens summed up over 22 languages, extracted from curated sources, existing multilingual corpora and large scale translations.
35
 
36
- This dataset only Hindi as of now
37
 
38
  # Information
39
  * First this dataset is mainly for long context training
40
- * The minimum len is and maximum len is
41
 
42
  # Getting started
43
 
@@ -46,6 +46,7 @@ For downloading the entire dataset:
46
  from datasets import load_dataset
47
  dataset = load_dataset("damerajee/long_context_hindi")
48
  ```
 
49
  If dataset is too big you can simply stream:
50
  ```python
51
  from datasets import load_dataset
 
24
  language:
25
  - hi
26
  - en
27
+ pretty_name: long_context
28
  size_categories:
29
  - 100K<n<1M
30
  ---
 
33
 
34
  This dataset was filtered from AI4BHarat dataset [sangraha](https://huggingface.co/datasets/ai4bharat/sangraha),which is the largest high-quality, cleaned Indic language pretraining data containing 251B tokens summed up over 22 languages, extracted from curated sources, existing multilingual corpora and large scale translations.
35
 
36
+ This dataset contains only Hindi as of now
37
 
38
  # Information
39
  * First this dataset is mainly for long context training
40
+ * The minimum len is `6000` and maximum len is `3754718`
41
 
42
  # Getting started
43
 
 
46
  from datasets import load_dataset
47
  dataset = load_dataset("damerajee/long_context_hindi")
48
  ```
49
+
50
  If dataset is too big you can simply stream:
51
  ```python
52
  from datasets import load_dataset