Hi!
I'm trying to have iterations to a parallel bulk operations, basically each operation is ~100K docs, and it's finishing too soon, I want to load it for a longer period of time
So I tried iterations and also target-throughput and also time-period but no success
It just finished 1 iteration and that's it, doesn't continue again, over and over.
how can I achieve that ?
{
"version": 2,
"title": "a",
"description": "a",
"indices": [
{
"name": "a1",
"body": "mapping.json"
},
{
"name": "a2",
"body": "mapping.json"
},
{
"name": "a3 ",
"body": "mapping.json"
},
{
"name": "a4",
"body": "mapping.json"
},
{
"name": "a5",
"body": "mapping.json"
},
{
"name": "a6",
"body": "mapping.json"
},
{
"name": "a7",
"body": "mapping.json"
},
{
"name": "a8",
"body": "mapping.json"
},
{
"name": "a9",
"body": "mapping.json"
}
],
"corpora": [
{
"name": "data1",
"documents": [
{
"source-file": "data.json",
"target-index": "a1",
"document-count": 100117
}
]
},
{
"name": "data2",
"documents": [
{
"source-file": "data.json",
"target-index": "a2",
"document-count": 100117
}
]
},{
"name": "data3",
"documents": [
{
"source-file": "data.json",
"target-index": "a3",
"document-count": 100117
}
]
},{
"name": "data4",
"documents": [
{
"source-file": "data.json",
"target-index": "a4",
"document-count": 100117
}
]
},{
"name": "data5",
"documents": [
{
"source-file": "data.json",
"target-index": "a5",
"document-count": 100117
}
]
},{
"name": "data6",
"documents": [
{
"source-file": "data.json",
"target-index": "a6",
"document-count": 100117
}
]
},{
"name": "data7",
"documents": [
{
"source-file": "data.json",
"target-index": "a7",
"document-count": 100117
}
]
},{
"name": "data8",
"documents": [
{
"source-file": "data.json",
"target-index": "a8",
"document-count": 100117
}
]
},{
"name": "data9",
"documents": [
{
"source-file": "data.json",
"target-index": "a9",
"document-count": 100117
}
]
}
],
"schedule": [
{
"parallel": {
"iterations": 10000000,
"tasks": [
{
"name": "bulk1",
"clients": 3,
"target-throughput": 50,
"operation": {
"operation-type": "bulk",
"corpora": "data1",
"bulk-size": 100
}
},
{
"name": "bulk",
"clients": 3,
"operation": {
"operation-type": "bulk",
"corpora": "data2",
"bulk-size": 100
}
},
{
"name": "bulk3",
"clients": 3,
"operation": {
"operation-type": "bulk",
"corpora": "data3",
"bulk-size": 100
}
},
{
"name": "bulk4",
"clients": 3,
"operation": {
"operation-type": "bulk",
"corpora": "data4",
"bulk-size": 100
}
},
{
"name": "bulk5",
"clients": 3,
"operation": {
"operation-type": "bulk",
"corpora": "data5",
"bulk-size": 100
}
},
{
"name": "bulk6",
"clients": 3,
"operation": {
"operation-type": "bulk",
"corpora": "data6",
"bulk-size": 100
}
},
{
"name": "bulk7",
"clients": 3,
"operation": {
"operation-type": "bulk",
"corpora": "data7",
"bulk-size": 100
}
},
{
"name": "bulk8",
"clients": 3,
"operation": {
"operation-type": "bulk",
"corpora": "data8",
"bulk-size": 100
}
},
{
"name": "bulk9",
"clients": 3,
"operation": {
"operation-type": "bulk",
"corpora": "data9",
"bulk-size": 100
}
}
]
}
}
]
}
Tried upgrading to 1.3.0 same deal, it basically does only 1 iteration, even though the data file include only raw data, without metadata id's (from my understanding esrally parses the data and add metadata header to each request with id)
did some internet search, and found similar cases, and you told that esrally doesn't support more than 1 iteration on bulk operations, and suggested people duplicate the corpora or make bigger files, hope there is a new way to handle this more elegant.
would love to get some help !
thank you all very much !