सन्दर्भ:
psgs_w100.nq.exact
इस डेटासेट को TFDS में लोड करने के लिए निम्नलिखित कमांड का उपयोग करें:
ds = tfds.load('huggingface:wiki_dpr/psgs_w100.nq.exact')
- विवरण :
This is the wikipedia split used to evaluate the Dense Passage Retrieval (DPR) model.
It contains 21M passages from wikipedia along with their DPR embeddings.
The wikipedia articles were split into multiple, disjoint text blocks of 100 words as passages.
- लाइसेंस : कोई ज्ञात लाइसेंस नहीं
- संस्करण : 0.0.0
- विभाजन :
विभाजित करना | उदाहरण |
---|---|
'train' | 21015300 |
- विशेषताएँ :
{
"id": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"text": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"title": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"embeddings": {
"feature": {
"dtype": "float32",
"id": null,
"_type": "Value"
},
"length": -1,
"id": null,
"_type": "Sequence"
}
}
psgs_w100.nq.संपीड़ित
इस डेटासेट को TFDS में लोड करने के लिए निम्नलिखित कमांड का उपयोग करें:
ds = tfds.load('huggingface:wiki_dpr/psgs_w100.nq.compressed')
- विवरण :
This is the wikipedia split used to evaluate the Dense Passage Retrieval (DPR) model.
It contains 21M passages from wikipedia along with their DPR embeddings.
The wikipedia articles were split into multiple, disjoint text blocks of 100 words as passages.
- लाइसेंस : कोई ज्ञात लाइसेंस नहीं
- संस्करण : 0.0.0
- विभाजन :
विभाजित करना | उदाहरण |
---|---|
'train' | 21015300 |
- विशेषताएँ :
{
"id": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"text": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"title": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"embeddings": {
"feature": {
"dtype": "float32",
"id": null,
"_type": "Value"
},
"length": -1,
"id": null,
"_type": "Sequence"
}
}
psgs_w100.nq.no_index
इस डेटासेट को TFDS में लोड करने के लिए निम्नलिखित कमांड का उपयोग करें:
ds = tfds.load('huggingface:wiki_dpr/psgs_w100.nq.no_index')
- विवरण :
This is the wikipedia split used to evaluate the Dense Passage Retrieval (DPR) model.
It contains 21M passages from wikipedia along with their DPR embeddings.
The wikipedia articles were split into multiple, disjoint text blocks of 100 words as passages.
- लाइसेंस : कोई ज्ञात लाइसेंस नहीं
- संस्करण : 0.0.0
- विभाजन :
विभाजित करना | उदाहरण |
---|---|
'train' | 21015300 |
- विशेषताएँ :
{
"id": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"text": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"title": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"embeddings": {
"feature": {
"dtype": "float32",
"id": null,
"_type": "Value"
},
"length": -1,
"id": null,
"_type": "Sequence"
}
}
psgs_w100.multiset.exact
इस डेटासेट को TFDS में लोड करने के लिए निम्नलिखित कमांड का उपयोग करें:
ds = tfds.load('huggingface:wiki_dpr/psgs_w100.multiset.exact')
- विवरण :
This is the wikipedia split used to evaluate the Dense Passage Retrieval (DPR) model.
It contains 21M passages from wikipedia along with their DPR embeddings.
The wikipedia articles were split into multiple, disjoint text blocks of 100 words as passages.
- लाइसेंस : कोई ज्ञात लाइसेंस नहीं
- संस्करण : 0.0.0
- विभाजन :
विभाजित करना | उदाहरण |
---|---|
'train' | 21015300 |
- विशेषताएँ :
{
"id": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"text": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"title": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"embeddings": {
"feature": {
"dtype": "float32",
"id": null,
"_type": "Value"
},
"length": -1,
"id": null,
"_type": "Sequence"
}
}
psgs_w100.multiset.compressed
इस डेटासेट को TFDS में लोड करने के लिए निम्नलिखित कमांड का उपयोग करें:
ds = tfds.load('huggingface:wiki_dpr/psgs_w100.multiset.compressed')
- विवरण :
This is the wikipedia split used to evaluate the Dense Passage Retrieval (DPR) model.
It contains 21M passages from wikipedia along with their DPR embeddings.
The wikipedia articles were split into multiple, disjoint text blocks of 100 words as passages.
- लाइसेंस : कोई ज्ञात लाइसेंस नहीं
- संस्करण : 0.0.0
- विभाजन :
विभाजित करना | उदाहरण |
---|---|
'train' | 21015300 |
- विशेषताएँ :
{
"id": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"text": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"title": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"embeddings": {
"feature": {
"dtype": "float32",
"id": null,
"_type": "Value"
},
"length": -1,
"id": null,
"_type": "Sequence"
}
}
psgs_w100.multiset.no_index
इस डेटासेट को TFDS में लोड करने के लिए निम्नलिखित कमांड का उपयोग करें:
ds = tfds.load('huggingface:wiki_dpr/psgs_w100.multiset.no_index')
- विवरण :
This is the wikipedia split used to evaluate the Dense Passage Retrieval (DPR) model.
It contains 21M passages from wikipedia along with their DPR embeddings.
The wikipedia articles were split into multiple, disjoint text blocks of 100 words as passages.
- लाइसेंस : कोई ज्ञात लाइसेंस नहीं
- संस्करण : 0.0.0
- विभाजन :
विभाजित करना | उदाहरण |
---|---|
'train' | 21015300 |
- विशेषताएँ :
{
"id": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"text": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"title": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"embeddings": {
"feature": {
"dtype": "float32",
"id": null,
"_type": "Value"
},
"length": -1,
"id": null,
"_type": "Sequence"
}
}