pdfcoffee
Chapter 12In a TFX pipeline, a unit of data, called an artifact, is passed between components.Normally a component has one input artifact and one output artifact. Every artifacthas an associated metadata that defines its type and properties. The artifact typedefines the ontology of artifacts in the entire TFX system, while the artifact propertyspecifies the ontology specific to an artifact type. Users have the option to extend theontology globally or locally.TFX pipeline componentsThe following diagram shows the flow of data between different TFX components:Flow of data between TFX componentsAll the images in the TFX section have been adapted from theTensorFlow Extended official guide: https://www.tensorflow.org/tfx/guide.To begin with we have ExampleGen, which ingests the input data, and can also splitthe input dataset. The data then flows to StatisticsGen, which calculates the statisticsof the dataset. Then comes SchemaGen, which examines the statistics and createsa data schema; then an ExampleValidator, which looks for anomalies and missingvalues in the data; and Transform, which performs feature engineering in thedataset. The transformed dataset is then fed to the Trainer, which trains the model.The performance of the model is evaluated using Evaluator and ModelValidator.Finally, if all is well, the Pusher deploys the model on the serving infrastructure.[ 457 ]
TensorFlow and CloudTFX librariesTFX provides several Python packages that are used to create pipeline components.Quoting from the TensorFlow Extended User Guide (https://www.tensorflow.org/tfx/guide).These packages are the libraries which you will use to create the components ofyour pipelines so that your code can focus on the unique aspects of your pipeline.Different libraries included in TFX are:• TensorFlow Data Validation (TFDV) is a library for analyzing andvalidating machine learning data• TensorFlow Transform (TFT) is a library for preprocessing data withTensorFlow• TensorFlow is used for training models with TFX• TensorFlow Model Analysis (TFMA) is a library for evaluating TensorFlowmodels• TensorFlow Metadata (TFMD) provides standard representations formetadata that are useful when training machine learning models withTensorFlow• ML Metadata (MLMD) is a library for recording and retrieving metadataassociated with ML developers and data scientists' workflowsThe following diagram demonstrates the relationship between TFX libraries andpipeline components:Figure 7: Relationships between TFX libraries and pipeline components, visualized[ 458 ]
- Page 442 and 443: Reinforcement LearningThis chapter
- Page 444 and 445: Chapter 11And unlike unsupervised l
- Page 446 and 447: Chapter 11Normally, the value is de
- Page 448 and 449: Chapter 11• The next question tha
- Page 450 and 451: Chapter 11This neural network takes
- Page 452 and 453: Chapter 11The MuJoCo environment re
- Page 454 and 455: Chapter 11We will first import the
- Page 456 and 457: Chapter 11The αα is the learning
- Page 458 and 459: Chapter 11We set up the global valu
- Page 460 and 461: Chapter 11else:return np.argmax(sel
- Page 462 and 463: Chapter 11DQN to play a game of Ata
- Page 464 and 465: Chapter 11self.model.add( Conv2D(64
- Page 466 and 467: Chapter 11Here the action A was sel
- Page 468 and 469: Chapter 11Image source: https://arx
- Page 470 and 471: Chapter 11A neural network is used
- Page 472: Chapter 1111. Details regarding ins
- Page 475 and 476: TensorFlow and Cloud• Scalability
- Page 477 and 478: TensorFlow and Cloud• Azure DevOp
- Page 479 and 480: TensorFlow and Cloud• Lambda: The
- Page 481 and 482: TensorFlow and Cloud• Deep Learni
- Page 483 and 484: TensorFlow and CloudEC2 on AmazonTo
- Page 485 and 486: TensorFlow and CloudCompute Instanc
- Page 487 and 488: TensorFlow and CloudYou just share
- Page 489 and 490: TensorFlow and CloudIn case you req
- Page 491: TensorFlow and CloudIt starts with
- Page 495 and 496: TensorFlow and CloudReferences1. To
- Page 497 and 498: TensorFlow for Mobile and IoT and T
- Page 499 and 500: TensorFlow for Mobile and IoT and T
- Page 501 and 502: TensorFlow for Mobile and IoT and T
- Page 503 and 504: TensorFlow for Mobile and IoT and T
- Page 505 and 506: TensorFlow for Mobile and IoT and T
- Page 507 and 508: TensorFlow for Mobile and IoT and T
- Page 509 and 510: TensorFlow for Mobile and IoT and T
- Page 511 and 512: TensorFlow for Mobile and IoT and T
- Page 513 and 514: TensorFlow for Mobile and IoT and T
- Page 515 and 516: TensorFlow for Mobile and IoT and T
- Page 517 and 518: TensorFlow for Mobile and IoT and T
- Page 519 and 520: TensorFlow for Mobile and IoT and T
- Page 521 and 522: TensorFlow for Mobile and IoT and T
- Page 523 and 524: TensorFlow for Mobile and IoT and T
- Page 525 and 526: TensorFlow for Mobile and IoT and T
- Page 527 and 528: An introduction to AutoMLThat is pr
- Page 529 and 530: An introduction to AutoMLFeature co
- Page 531 and 532: An introduction to AutoMLThis Effic
- Page 533 and 534: An introduction to AutoMLGoogle Clo
- Page 535 and 536: An introduction to AutoMLThen, we c
- Page 537 and 538: An introduction to AutoMLOnce the d
- Page 539 and 540: An introduction to AutoMLIf your mo
- Page 541 and 542: An introduction to AutoMLClicking o
Chapter 12
In a TFX pipeline, a unit of data, called an artifact, is passed between components.
Normally a component has one input artifact and one output artifact. Every artifact
has an associated metadata that defines its type and properties. The artifact type
defines the ontology of artifacts in the entire TFX system, while the artifact property
specifies the ontology specific to an artifact type. Users have the option to extend the
ontology globally or locally.
TFX pipeline components
The following diagram shows the flow of data between different TFX components:
Flow of data between TFX components
All the images in the TFX section have been adapted from the
TensorFlow Extended official guide: https://www.tensorflow.
org/tfx/guide.
To begin with we have ExampleGen, which ingests the input data, and can also split
the input dataset. The data then flows to StatisticsGen, which calculates the statistics
of the dataset. Then comes SchemaGen, which examines the statistics and creates
a data schema; then an ExampleValidator, which looks for anomalies and missing
values in the data; and Transform, which performs feature engineering in the
dataset. The transformed dataset is then fed to the Trainer, which trains the model.
The performance of the model is evaluated using Evaluator and ModelValidator.
Finally, if all is well, the Pusher deploys the model on the serving infrastructure.
[ 457 ]