Module: tfm.nlp.models
Stay organized with collections
Save and categorize content based on your preferences.
Models are combinations of tf.keras
layers and models that can be trained.
Several pre-built canned models are provided to train encoder networks.
These models are intended as both convenience functions and canonical examples.
Classes
class BertClassifier
: Classifier model based on a BERT-style transformer-based encoder.
class BertPretrainer
: BERT pretraining model.
class BertPretrainerV2
: BERT pretraining model V2.
class BertSpanLabeler
: Span labeler model based on a BERT-style transformer-based encoder.
class BertTokenClassifier
: Token classifier model based on a BERT-style transformer-based encoder.
class DualEncoder
: A dual encoder model based on a transformer-based encoder.
class ElectraPretrainer
: ELECTRA network training model.
class Seq2SeqTransformer
: Transformer model with Keras.
class T5Transformer
: Transformer Encoder+Decoder for sequence to sequence.
class T5TransformerParams
: Transformer parameters.
class TransformerDecoder
: Transformer decoder.
class TransformerEncoder
: Transformer encoder.
class XLNetClassifier
: Classifier model based on XLNet.
class XLNetPretrainer
: XLNet-based pretrainer.
class XLNetSpanLabeler
: Span labeler model based on XLNet.
Functions
attention_initializer(...)
: Initializer for attention layers in Seq2SeqTransformer.
Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License, and code samples are licensed under the Apache 2.0 License. For details, see the Google Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates. Some content is licensed under the numpy license.
Last updated 2024-02-02 UTC.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Missing the information I need","missingTheInformationINeed","thumb-down"],["Too complicated / too many steps","tooComplicatedTooManySteps","thumb-down"],["Out of date","outOfDate","thumb-down"],["Samples / code issue","samplesCodeIssue","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2024-02-02 UTC."],[],[],null,["# Module: tfm.nlp.models\n\n\u003cbr /\u003e\n\n|---------------------------------------------------------------------------------------------------------------------|\n| [View source on GitHub](https://github.com/tensorflow/models/blob/v2.15.0/official/nlp/modeling/models/__init__.py) |\n\nModels are combinations of [`tf.keras`](https://www.tensorflow.org/api_docs/python/tf/keras) layers and models that can be trained.\n\nSeveral pre-built canned models are provided to train encoder networks.\nThese models are intended as both convenience functions and canonical examples.\n\nClasses\n-------\n\n[`class BertClassifier`](../../tfm/nlp/models/BertClassifier): Classifier model based on a BERT-style transformer-based encoder.\n\n[`class BertPretrainer`](../../tfm/nlp/models/BertPretrainer): BERT pretraining model.\n\n[`class BertPretrainerV2`](../../tfm/nlp/models/BertPretrainerV2): BERT pretraining model V2.\n\n[`class BertSpanLabeler`](../../tfm/nlp/models/BertSpanLabeler): Span labeler model based on a BERT-style transformer-based encoder.\n\n[`class BertTokenClassifier`](../../tfm/nlp/models/BertTokenClassifier): Token classifier model based on a BERT-style transformer-based encoder.\n\n[`class DualEncoder`](../../tfm/nlp/models/DualEncoder): A dual encoder model based on a transformer-based encoder.\n\n[`class ElectraPretrainer`](../../tfm/nlp/models/ElectraPretrainer): ELECTRA network training model.\n\n[`class Seq2SeqTransformer`](../../tfm/nlp/models/Seq2SeqTransformer): Transformer model with Keras.\n\n[`class T5Transformer`](../../tfm/nlp/models/T5Transformer): Transformer Encoder+Decoder for sequence to sequence.\n\n[`class T5TransformerParams`](../../tfm/nlp/models/T5TransformerParams): Transformer parameters.\n\n[`class TransformerDecoder`](../../tfm/nlp/models/TransformerDecoder): Transformer decoder.\n\n[`class TransformerEncoder`](../../tfm/nlp/models/TransformerEncoder): Transformer encoder.\n\n[`class XLNetClassifier`](../../tfm/nlp/models/XLNetClassifier): Classifier model based on XLNet.\n\n[`class XLNetPretrainer`](../../tfm/nlp/models/XLNetPretrainer): XLNet-based pretrainer.\n\n[`class XLNetSpanLabeler`](../../tfm/nlp/models/XLNetSpanLabeler): Span labeler model based on XLNet.\n\nFunctions\n---------\n\n[`attention_initializer(...)`](../../tfm/nlp/models/attention_initializer): Initializer for attention layers in Seq2SeqTransformer.\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Other Members ------------- ||\n|--------|-----|\n| EOS_ID | `1` |\n\n\u003cbr /\u003e"]]