[MINOR] Fixing spark long running yaml for non-partitioned (#5607)
This commit is contained in:
committed by
GitHub
parent
99555c897a
commit
f8b9399615
@@ -14,24 +14,24 @@
|
|||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
dag_name: cow-spark-deltastreamer-long-running-multi-partitions.yaml
|
dag_name: cow-spark-deltastreamer-long-running-multi-partitions.yaml
|
||||||
dag_rounds: 6
|
dag_rounds: 10
|
||||||
dag_intermittent_delay_mins: 0
|
dag_intermittent_delay_mins: 1
|
||||||
dag_content:
|
dag_content:
|
||||||
first_insert:
|
first_insert:
|
||||||
config:
|
config:
|
||||||
record_size: 200
|
record_size: 200
|
||||||
num_partitions_insert: 1
|
num_partitions_insert: 1
|
||||||
repeat_count: 1
|
repeat_count: 1
|
||||||
num_records_insert: 10000
|
num_records_insert: 1000
|
||||||
type: SparkInsertNode
|
type: SparkInsertNode
|
||||||
deps: none
|
deps: none
|
||||||
first_upsert:
|
first_upsert:
|
||||||
config:
|
config:
|
||||||
record_size: 200
|
record_size: 200
|
||||||
num_partitions_insert: 1
|
num_partitions_insert: 1
|
||||||
num_records_insert: 300
|
num_records_insert: 1000
|
||||||
repeat_count: 1
|
repeat_count: 1
|
||||||
num_records_upsert: 3000
|
num_records_upsert: 1000
|
||||||
num_partitions_upsert: 1
|
num_partitions_upsert: 1
|
||||||
type: SparkUpsertNode
|
type: SparkUpsertNode
|
||||||
deps: first_insert
|
deps: first_insert
|
||||||
@@ -43,7 +43,6 @@ dag_content:
|
|||||||
deps: first_upsert
|
deps: first_upsert
|
||||||
second_validate:
|
second_validate:
|
||||||
config:
|
config:
|
||||||
validate_once_every_itr : 3
|
|
||||||
validate_hive: false
|
validate_hive: false
|
||||||
delete_input_data: true
|
delete_input_data: true
|
||||||
type: ValidateDatasetNode
|
type: ValidateDatasetNode
|
||||||
|
|||||||
Reference in New Issue
Block a user