-
Notifications
You must be signed in to change notification settings - Fork 679
TFX Evaluator takes too long in "TFX_Pipeline_for_Bert_Preprocessing" #71
Comments
Hi @deep-diver,
Regarding GCP AI Platform: Can you share a bit more info around your configuration? I hope this helps. |
I am not sure how to set smaller step number? eval_config = tfma.EvalConfig(
model_specs=[tfma.ModelSpec(label_key='label')],
slicing_specs=[tfma.SlicingSpec()],
metrics_specs=[
tfma.MetricsSpec(metrics=[
tfma.MetricConfig(
class_name='CategoricalAccuracy',
threshold=tfma.MetricThreshold(
value_threshold=tfma.GenericValueThreshold(
lower_bound={'value': 0.5}),
change_threshold=tfma.GenericChangeThreshold(
direction=tfma.MetricDirection.HIGHER_IS_BETTER,
absolute={'value': -1e-2})))
])
]
)
evaluator = Evaluator(
examples=example_gen.outputs['examples'],
model=trainer.outputs['model'],
baseline_model=model_resolver.outputs['model'],
eval_config=eval_config
)
components.append(evaluator) I will try the ALBERT version! Thanks! 👍 👍 But I want to know if there is a way to allocate GPU machine only for TFX Evaluator component. Since this is a simple personal project, I don't want have multiple k8s GPU nodes. (can I set one node with GPU, and two nodes without GPU?) |
Configuring nodes for specific components is in development now, and should be available soon. |
Or is there a way to wrap TFX component in |
@hanneshapke
It seems like 'Evaluator' component takes too long time (more than 2 hours, and it hadn't done yet) in Kubeflow environment on GCP AI Platform Pipeline. It is very unexpected behaviour when comparing the notebook version which took about less than 5 minutes with GPU.
I am assuming that environments with and without GPU behaves differently (since Evaluator tries to evaluate two models[blessing, current] by inferencing inputs). If that is the case, the problem is that I want to allocate one GPU k8s node for one specific TFX component. Otherwise I have to equip every single nodes with GPU which is not desirable.
Any possible thoughts?
The text was updated successfully, but these errors were encountered: