public static final class InputDataConfig.Builder extends com.google.protobuf.GeneratedMessageV3.Builder<InputDataConfig.Builder> implements InputDataConfigOrBuilder
Specifies Vertex AI owned input data to be used for training, and possibly evaluating, the Model.Protobuf type
google.cloud.aiplatform.v1beta1.InputDataConfig| Modifier and Type | Method and Description |
|---|---|
InputDataConfig.Builder |
addRepeatedField(com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) |
InputDataConfig |
build() |
InputDataConfig |
buildPartial() |
InputDataConfig.Builder |
clear() |
InputDataConfig.Builder |
clearAnnotationSchemaUri()
Applicable only to custom training with Datasets that have DataItems and
Annotations.
|
InputDataConfig.Builder |
clearAnnotationsFilter()
Applicable only to Datasets that have DataItems and Annotations.
|
InputDataConfig.Builder |
clearBigqueryDestination()
Only applicable to custom training with tabular Dataset with BigQuery
source.
|
InputDataConfig.Builder |
clearDatasetId()
Required.
|
InputDataConfig.Builder |
clearDestination() |
InputDataConfig.Builder |
clearField(com.google.protobuf.Descriptors.FieldDescriptor field) |
InputDataConfig.Builder |
clearFilterSplit()
Split based on the provided filters for each set.
|
InputDataConfig.Builder |
clearFractionSplit()
Split based on fractions defining the size of each set.
|
InputDataConfig.Builder |
clearGcsDestination()
The Cloud Storage location where the training data is to be
written to.
|
InputDataConfig.Builder |
clearOneof(com.google.protobuf.Descriptors.OneofDescriptor oneof) |
InputDataConfig.Builder |
clearPredefinedSplit()
Supported only for tabular Datasets.
|
InputDataConfig.Builder |
clearSavedQueryId()
Only applicable to Datasets that have SavedQueries.
|
InputDataConfig.Builder |
clearSplit() |
InputDataConfig.Builder |
clearStratifiedSplit()
Supported only for tabular Datasets.
|
InputDataConfig.Builder |
clearTimestampSplit()
Supported only for tabular Datasets.
|
InputDataConfig.Builder |
clone() |
String |
getAnnotationSchemaUri()
Applicable only to custom training with Datasets that have DataItems and
Annotations.
|
com.google.protobuf.ByteString |
getAnnotationSchemaUriBytes()
Applicable only to custom training with Datasets that have DataItems and
Annotations.
|
String |
getAnnotationsFilter()
Applicable only to Datasets that have DataItems and Annotations.
|
com.google.protobuf.ByteString |
getAnnotationsFilterBytes()
Applicable only to Datasets that have DataItems and Annotations.
|
BigQueryDestination |
getBigqueryDestination()
Only applicable to custom training with tabular Dataset with BigQuery
source.
|
BigQueryDestination.Builder |
getBigqueryDestinationBuilder()
Only applicable to custom training with tabular Dataset with BigQuery
source.
|
BigQueryDestinationOrBuilder |
getBigqueryDestinationOrBuilder()
Only applicable to custom training with tabular Dataset with BigQuery
source.
|
String |
getDatasetId()
Required.
|
com.google.protobuf.ByteString |
getDatasetIdBytes()
Required.
|
InputDataConfig |
getDefaultInstanceForType() |
static com.google.protobuf.Descriptors.Descriptor |
getDescriptor() |
com.google.protobuf.Descriptors.Descriptor |
getDescriptorForType() |
InputDataConfig.DestinationCase |
getDestinationCase() |
FilterSplit |
getFilterSplit()
Split based on the provided filters for each set.
|
FilterSplit.Builder |
getFilterSplitBuilder()
Split based on the provided filters for each set.
|
FilterSplitOrBuilder |
getFilterSplitOrBuilder()
Split based on the provided filters for each set.
|
FractionSplit |
getFractionSplit()
Split based on fractions defining the size of each set.
|
FractionSplit.Builder |
getFractionSplitBuilder()
Split based on fractions defining the size of each set.
|
FractionSplitOrBuilder |
getFractionSplitOrBuilder()
Split based on fractions defining the size of each set.
|
GcsDestination |
getGcsDestination()
The Cloud Storage location where the training data is to be
written to.
|
GcsDestination.Builder |
getGcsDestinationBuilder()
The Cloud Storage location where the training data is to be
written to.
|
GcsDestinationOrBuilder |
getGcsDestinationOrBuilder()
The Cloud Storage location where the training data is to be
written to.
|
PredefinedSplit |
getPredefinedSplit()
Supported only for tabular Datasets.
|
PredefinedSplit.Builder |
getPredefinedSplitBuilder()
Supported only for tabular Datasets.
|
PredefinedSplitOrBuilder |
getPredefinedSplitOrBuilder()
Supported only for tabular Datasets.
|
String |
getSavedQueryId()
Only applicable to Datasets that have SavedQueries.
|
com.google.protobuf.ByteString |
getSavedQueryIdBytes()
Only applicable to Datasets that have SavedQueries.
|
InputDataConfig.SplitCase |
getSplitCase() |
StratifiedSplit |
getStratifiedSplit()
Supported only for tabular Datasets.
|
StratifiedSplit.Builder |
getStratifiedSplitBuilder()
Supported only for tabular Datasets.
|
StratifiedSplitOrBuilder |
getStratifiedSplitOrBuilder()
Supported only for tabular Datasets.
|
TimestampSplit |
getTimestampSplit()
Supported only for tabular Datasets.
|
TimestampSplit.Builder |
getTimestampSplitBuilder()
Supported only for tabular Datasets.
|
TimestampSplitOrBuilder |
getTimestampSplitOrBuilder()
Supported only for tabular Datasets.
|
boolean |
hasBigqueryDestination()
Only applicable to custom training with tabular Dataset with BigQuery
source.
|
boolean |
hasFilterSplit()
Split based on the provided filters for each set.
|
boolean |
hasFractionSplit()
Split based on fractions defining the size of each set.
|
boolean |
hasGcsDestination()
The Cloud Storage location where the training data is to be
written to.
|
boolean |
hasPredefinedSplit()
Supported only for tabular Datasets.
|
boolean |
hasStratifiedSplit()
Supported only for tabular Datasets.
|
boolean |
hasTimestampSplit()
Supported only for tabular Datasets.
|
protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable |
internalGetFieldAccessorTable() |
boolean |
isInitialized() |
InputDataConfig.Builder |
mergeBigqueryDestination(BigQueryDestination value)
Only applicable to custom training with tabular Dataset with BigQuery
source.
|
InputDataConfig.Builder |
mergeFilterSplit(FilterSplit value)
Split based on the provided filters for each set.
|
InputDataConfig.Builder |
mergeFractionSplit(FractionSplit value)
Split based on fractions defining the size of each set.
|
InputDataConfig.Builder |
mergeFrom(com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry) |
InputDataConfig.Builder |
mergeFrom(InputDataConfig other) |
InputDataConfig.Builder |
mergeFrom(com.google.protobuf.Message other) |
InputDataConfig.Builder |
mergeGcsDestination(GcsDestination value)
The Cloud Storage location where the training data is to be
written to.
|
InputDataConfig.Builder |
mergePredefinedSplit(PredefinedSplit value)
Supported only for tabular Datasets.
|
InputDataConfig.Builder |
mergeStratifiedSplit(StratifiedSplit value)
Supported only for tabular Datasets.
|
InputDataConfig.Builder |
mergeTimestampSplit(TimestampSplit value)
Supported only for tabular Datasets.
|
InputDataConfig.Builder |
mergeUnknownFields(com.google.protobuf.UnknownFieldSet unknownFields) |
InputDataConfig.Builder |
setAnnotationSchemaUri(String value)
Applicable only to custom training with Datasets that have DataItems and
Annotations.
|
InputDataConfig.Builder |
setAnnotationSchemaUriBytes(com.google.protobuf.ByteString value)
Applicable only to custom training with Datasets that have DataItems and
Annotations.
|
InputDataConfig.Builder |
setAnnotationsFilter(String value)
Applicable only to Datasets that have DataItems and Annotations.
|
InputDataConfig.Builder |
setAnnotationsFilterBytes(com.google.protobuf.ByteString value)
Applicable only to Datasets that have DataItems and Annotations.
|
InputDataConfig.Builder |
setBigqueryDestination(BigQueryDestination.Builder builderForValue)
Only applicable to custom training with tabular Dataset with BigQuery
source.
|
InputDataConfig.Builder |
setBigqueryDestination(BigQueryDestination value)
Only applicable to custom training with tabular Dataset with BigQuery
source.
|
InputDataConfig.Builder |
setDatasetId(String value)
Required.
|
InputDataConfig.Builder |
setDatasetIdBytes(com.google.protobuf.ByteString value)
Required.
|
InputDataConfig.Builder |
setField(com.google.protobuf.Descriptors.FieldDescriptor field,
Object value) |
InputDataConfig.Builder |
setFilterSplit(FilterSplit.Builder builderForValue)
Split based on the provided filters for each set.
|
InputDataConfig.Builder |
setFilterSplit(FilterSplit value)
Split based on the provided filters for each set.
|
InputDataConfig.Builder |
setFractionSplit(FractionSplit.Builder builderForValue)
Split based on fractions defining the size of each set.
|
InputDataConfig.Builder |
setFractionSplit(FractionSplit value)
Split based on fractions defining the size of each set.
|
InputDataConfig.Builder |
setGcsDestination(GcsDestination.Builder builderForValue)
The Cloud Storage location where the training data is to be
written to.
|
InputDataConfig.Builder |
setGcsDestination(GcsDestination value)
The Cloud Storage location where the training data is to be
written to.
|
InputDataConfig.Builder |
setPredefinedSplit(PredefinedSplit.Builder builderForValue)
Supported only for tabular Datasets.
|
InputDataConfig.Builder |
setPredefinedSplit(PredefinedSplit value)
Supported only for tabular Datasets.
|
InputDataConfig.Builder |
setRepeatedField(com.google.protobuf.Descriptors.FieldDescriptor field,
int index,
Object value) |
InputDataConfig.Builder |
setSavedQueryId(String value)
Only applicable to Datasets that have SavedQueries.
|
InputDataConfig.Builder |
setSavedQueryIdBytes(com.google.protobuf.ByteString value)
Only applicable to Datasets that have SavedQueries.
|
InputDataConfig.Builder |
setStratifiedSplit(StratifiedSplit.Builder builderForValue)
Supported only for tabular Datasets.
|
InputDataConfig.Builder |
setStratifiedSplit(StratifiedSplit value)
Supported only for tabular Datasets.
|
InputDataConfig.Builder |
setTimestampSplit(TimestampSplit.Builder builderForValue)
Supported only for tabular Datasets.
|
InputDataConfig.Builder |
setTimestampSplit(TimestampSplit value)
Supported only for tabular Datasets.
|
InputDataConfig.Builder |
setUnknownFields(com.google.protobuf.UnknownFieldSet unknownFields) |
getAllFields, getField, getFieldBuilder, getOneofFieldDescriptor, getParentForChildren, getRepeatedField, getRepeatedFieldBuilder, getRepeatedFieldCount, getUnknownFields, hasField, hasOneof, internalGetMapField, internalGetMutableMapField, isClean, markClean, newBuilderForField, onBuilt, onChanged, setUnknownFieldsProto3findInitializationErrors, getInitializationErrorString, internalMergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, newUninitializedMessageException, toStringaddAll, addAll, mergeDelimitedFrom, mergeDelimitedFrom, mergeFrom, newUninitializedMessageExceptionequals, finalize, getClass, hashCode, notify, notifyAll, wait, wait, waitfindInitializationErrors, getAllFields, getField, getInitializationErrorString, getOneofFieldDescriptor, getRepeatedField, getRepeatedFieldCount, getUnknownFields, hasField, hasOneofpublic static final com.google.protobuf.Descriptors.Descriptor getDescriptor()
protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()
internalGetFieldAccessorTable in class com.google.protobuf.GeneratedMessageV3.Builder<InputDataConfig.Builder>public InputDataConfig.Builder clear()
clear in interface com.google.protobuf.Message.Builderclear in interface com.google.protobuf.MessageLite.Builderclear in class com.google.protobuf.GeneratedMessageV3.Builder<InputDataConfig.Builder>public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()
getDescriptorForType in interface com.google.protobuf.Message.BuildergetDescriptorForType in interface com.google.protobuf.MessageOrBuildergetDescriptorForType in class com.google.protobuf.GeneratedMessageV3.Builder<InputDataConfig.Builder>public InputDataConfig getDefaultInstanceForType()
getDefaultInstanceForType in interface com.google.protobuf.MessageLiteOrBuildergetDefaultInstanceForType in interface com.google.protobuf.MessageOrBuilderpublic InputDataConfig build()
build in interface com.google.protobuf.Message.Builderbuild in interface com.google.protobuf.MessageLite.Builderpublic InputDataConfig buildPartial()
buildPartial in interface com.google.protobuf.Message.BuilderbuildPartial in interface com.google.protobuf.MessageLite.Builderpublic InputDataConfig.Builder clone()
clone in interface com.google.protobuf.Message.Builderclone in interface com.google.protobuf.MessageLite.Builderclone in class com.google.protobuf.GeneratedMessageV3.Builder<InputDataConfig.Builder>public InputDataConfig.Builder setField(com.google.protobuf.Descriptors.FieldDescriptor field, Object value)
setField in interface com.google.protobuf.Message.BuildersetField in class com.google.protobuf.GeneratedMessageV3.Builder<InputDataConfig.Builder>public InputDataConfig.Builder clearField(com.google.protobuf.Descriptors.FieldDescriptor field)
clearField in interface com.google.protobuf.Message.BuilderclearField in class com.google.protobuf.GeneratedMessageV3.Builder<InputDataConfig.Builder>public InputDataConfig.Builder clearOneof(com.google.protobuf.Descriptors.OneofDescriptor oneof)
clearOneof in interface com.google.protobuf.Message.BuilderclearOneof in class com.google.protobuf.GeneratedMessageV3.Builder<InputDataConfig.Builder>public InputDataConfig.Builder setRepeatedField(com.google.protobuf.Descriptors.FieldDescriptor field, int index, Object value)
setRepeatedField in interface com.google.protobuf.Message.BuildersetRepeatedField in class com.google.protobuf.GeneratedMessageV3.Builder<InputDataConfig.Builder>public InputDataConfig.Builder addRepeatedField(com.google.protobuf.Descriptors.FieldDescriptor field, Object value)
addRepeatedField in interface com.google.protobuf.Message.BuilderaddRepeatedField in class com.google.protobuf.GeneratedMessageV3.Builder<InputDataConfig.Builder>public InputDataConfig.Builder mergeFrom(com.google.protobuf.Message other)
mergeFrom in interface com.google.protobuf.Message.BuildermergeFrom in class com.google.protobuf.AbstractMessage.Builder<InputDataConfig.Builder>public InputDataConfig.Builder mergeFrom(InputDataConfig other)
public final boolean isInitialized()
isInitialized in interface com.google.protobuf.MessageLiteOrBuilderisInitialized in class com.google.protobuf.GeneratedMessageV3.Builder<InputDataConfig.Builder>public InputDataConfig.Builder mergeFrom(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws IOException
mergeFrom in interface com.google.protobuf.Message.BuildermergeFrom in interface com.google.protobuf.MessageLite.BuildermergeFrom in class com.google.protobuf.AbstractMessage.Builder<InputDataConfig.Builder>IOExceptionpublic InputDataConfig.SplitCase getSplitCase()
getSplitCase in interface InputDataConfigOrBuilderpublic InputDataConfig.Builder clearSplit()
public InputDataConfig.DestinationCase getDestinationCase()
getDestinationCase in interface InputDataConfigOrBuilderpublic InputDataConfig.Builder clearDestination()
public boolean hasFractionSplit()
Split based on fractions defining the size of each set.
.google.cloud.aiplatform.v1beta1.FractionSplit fraction_split = 2;hasFractionSplit in interface InputDataConfigOrBuilderpublic FractionSplit getFractionSplit()
Split based on fractions defining the size of each set.
.google.cloud.aiplatform.v1beta1.FractionSplit fraction_split = 2;getFractionSplit in interface InputDataConfigOrBuilderpublic InputDataConfig.Builder setFractionSplit(FractionSplit value)
Split based on fractions defining the size of each set.
.google.cloud.aiplatform.v1beta1.FractionSplit fraction_split = 2;public InputDataConfig.Builder setFractionSplit(FractionSplit.Builder builderForValue)
Split based on fractions defining the size of each set.
.google.cloud.aiplatform.v1beta1.FractionSplit fraction_split = 2;public InputDataConfig.Builder mergeFractionSplit(FractionSplit value)
Split based on fractions defining the size of each set.
.google.cloud.aiplatform.v1beta1.FractionSplit fraction_split = 2;public InputDataConfig.Builder clearFractionSplit()
Split based on fractions defining the size of each set.
.google.cloud.aiplatform.v1beta1.FractionSplit fraction_split = 2;public FractionSplit.Builder getFractionSplitBuilder()
Split based on fractions defining the size of each set.
.google.cloud.aiplatform.v1beta1.FractionSplit fraction_split = 2;public FractionSplitOrBuilder getFractionSplitOrBuilder()
Split based on fractions defining the size of each set.
.google.cloud.aiplatform.v1beta1.FractionSplit fraction_split = 2;getFractionSplitOrBuilder in interface InputDataConfigOrBuilderpublic boolean hasFilterSplit()
Split based on the provided filters for each set.
.google.cloud.aiplatform.v1beta1.FilterSplit filter_split = 3;hasFilterSplit in interface InputDataConfigOrBuilderpublic FilterSplit getFilterSplit()
Split based on the provided filters for each set.
.google.cloud.aiplatform.v1beta1.FilterSplit filter_split = 3;getFilterSplit in interface InputDataConfigOrBuilderpublic InputDataConfig.Builder setFilterSplit(FilterSplit value)
Split based on the provided filters for each set.
.google.cloud.aiplatform.v1beta1.FilterSplit filter_split = 3;public InputDataConfig.Builder setFilterSplit(FilterSplit.Builder builderForValue)
Split based on the provided filters for each set.
.google.cloud.aiplatform.v1beta1.FilterSplit filter_split = 3;public InputDataConfig.Builder mergeFilterSplit(FilterSplit value)
Split based on the provided filters for each set.
.google.cloud.aiplatform.v1beta1.FilterSplit filter_split = 3;public InputDataConfig.Builder clearFilterSplit()
Split based on the provided filters for each set.
.google.cloud.aiplatform.v1beta1.FilterSplit filter_split = 3;public FilterSplit.Builder getFilterSplitBuilder()
Split based on the provided filters for each set.
.google.cloud.aiplatform.v1beta1.FilterSplit filter_split = 3;public FilterSplitOrBuilder getFilterSplitOrBuilder()
Split based on the provided filters for each set.
.google.cloud.aiplatform.v1beta1.FilterSplit filter_split = 3;getFilterSplitOrBuilder in interface InputDataConfigOrBuilderpublic boolean hasPredefinedSplit()
Supported only for tabular Datasets. Split based on a predefined key.
.google.cloud.aiplatform.v1beta1.PredefinedSplit predefined_split = 4;hasPredefinedSplit in interface InputDataConfigOrBuilderpublic PredefinedSplit getPredefinedSplit()
Supported only for tabular Datasets. Split based on a predefined key.
.google.cloud.aiplatform.v1beta1.PredefinedSplit predefined_split = 4;getPredefinedSplit in interface InputDataConfigOrBuilderpublic InputDataConfig.Builder setPredefinedSplit(PredefinedSplit value)
Supported only for tabular Datasets. Split based on a predefined key.
.google.cloud.aiplatform.v1beta1.PredefinedSplit predefined_split = 4;public InputDataConfig.Builder setPredefinedSplit(PredefinedSplit.Builder builderForValue)
Supported only for tabular Datasets. Split based on a predefined key.
.google.cloud.aiplatform.v1beta1.PredefinedSplit predefined_split = 4;public InputDataConfig.Builder mergePredefinedSplit(PredefinedSplit value)
Supported only for tabular Datasets. Split based on a predefined key.
.google.cloud.aiplatform.v1beta1.PredefinedSplit predefined_split = 4;public InputDataConfig.Builder clearPredefinedSplit()
Supported only for tabular Datasets. Split based on a predefined key.
.google.cloud.aiplatform.v1beta1.PredefinedSplit predefined_split = 4;public PredefinedSplit.Builder getPredefinedSplitBuilder()
Supported only for tabular Datasets. Split based on a predefined key.
.google.cloud.aiplatform.v1beta1.PredefinedSplit predefined_split = 4;public PredefinedSplitOrBuilder getPredefinedSplitOrBuilder()
Supported only for tabular Datasets. Split based on a predefined key.
.google.cloud.aiplatform.v1beta1.PredefinedSplit predefined_split = 4;getPredefinedSplitOrBuilder in interface InputDataConfigOrBuilderpublic boolean hasTimestampSplit()
Supported only for tabular Datasets. Split based on the timestamp of the input data pieces.
.google.cloud.aiplatform.v1beta1.TimestampSplit timestamp_split = 5;hasTimestampSplit in interface InputDataConfigOrBuilderpublic TimestampSplit getTimestampSplit()
Supported only for tabular Datasets. Split based on the timestamp of the input data pieces.
.google.cloud.aiplatform.v1beta1.TimestampSplit timestamp_split = 5;getTimestampSplit in interface InputDataConfigOrBuilderpublic InputDataConfig.Builder setTimestampSplit(TimestampSplit value)
Supported only for tabular Datasets. Split based on the timestamp of the input data pieces.
.google.cloud.aiplatform.v1beta1.TimestampSplit timestamp_split = 5;public InputDataConfig.Builder setTimestampSplit(TimestampSplit.Builder builderForValue)
Supported only for tabular Datasets. Split based on the timestamp of the input data pieces.
.google.cloud.aiplatform.v1beta1.TimestampSplit timestamp_split = 5;public InputDataConfig.Builder mergeTimestampSplit(TimestampSplit value)
Supported only for tabular Datasets. Split based on the timestamp of the input data pieces.
.google.cloud.aiplatform.v1beta1.TimestampSplit timestamp_split = 5;public InputDataConfig.Builder clearTimestampSplit()
Supported only for tabular Datasets. Split based on the timestamp of the input data pieces.
.google.cloud.aiplatform.v1beta1.TimestampSplit timestamp_split = 5;public TimestampSplit.Builder getTimestampSplitBuilder()
Supported only for tabular Datasets. Split based on the timestamp of the input data pieces.
.google.cloud.aiplatform.v1beta1.TimestampSplit timestamp_split = 5;public TimestampSplitOrBuilder getTimestampSplitOrBuilder()
Supported only for tabular Datasets. Split based on the timestamp of the input data pieces.
.google.cloud.aiplatform.v1beta1.TimestampSplit timestamp_split = 5;getTimestampSplitOrBuilder in interface InputDataConfigOrBuilderpublic boolean hasStratifiedSplit()
Supported only for tabular Datasets. Split based on the distribution of the specified column.
.google.cloud.aiplatform.v1beta1.StratifiedSplit stratified_split = 12;hasStratifiedSplit in interface InputDataConfigOrBuilderpublic StratifiedSplit getStratifiedSplit()
Supported only for tabular Datasets. Split based on the distribution of the specified column.
.google.cloud.aiplatform.v1beta1.StratifiedSplit stratified_split = 12;getStratifiedSplit in interface InputDataConfigOrBuilderpublic InputDataConfig.Builder setStratifiedSplit(StratifiedSplit value)
Supported only for tabular Datasets. Split based on the distribution of the specified column.
.google.cloud.aiplatform.v1beta1.StratifiedSplit stratified_split = 12;public InputDataConfig.Builder setStratifiedSplit(StratifiedSplit.Builder builderForValue)
Supported only for tabular Datasets. Split based on the distribution of the specified column.
.google.cloud.aiplatform.v1beta1.StratifiedSplit stratified_split = 12;public InputDataConfig.Builder mergeStratifiedSplit(StratifiedSplit value)
Supported only for tabular Datasets. Split based on the distribution of the specified column.
.google.cloud.aiplatform.v1beta1.StratifiedSplit stratified_split = 12;public InputDataConfig.Builder clearStratifiedSplit()
Supported only for tabular Datasets. Split based on the distribution of the specified column.
.google.cloud.aiplatform.v1beta1.StratifiedSplit stratified_split = 12;public StratifiedSplit.Builder getStratifiedSplitBuilder()
Supported only for tabular Datasets. Split based on the distribution of the specified column.
.google.cloud.aiplatform.v1beta1.StratifiedSplit stratified_split = 12;public StratifiedSplitOrBuilder getStratifiedSplitOrBuilder()
Supported only for tabular Datasets. Split based on the distribution of the specified column.
.google.cloud.aiplatform.v1beta1.StratifiedSplit stratified_split = 12;getStratifiedSplitOrBuilder in interface InputDataConfigOrBuilderpublic boolean hasGcsDestination()
The Cloud Storage location where the training data is to be
written to. In the given directory a new directory is created with
name:
`dataset-<dataset-id>-<annotation-type>-<timestamp-of-training-call>`
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
All training input data is written into that directory.
The Vertex AI environment variables representing Cloud Storage
data URIs are represented in the Cloud Storage wildcard
format to support sharded data. e.g.: "gs://.../training-*.jsonl"
* AIP_DATA_FORMAT = "jsonl" for non-tabular data, "csv" for tabular data
* AIP_TRAINING_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/training-*.${AIP_DATA_FORMAT}"
* AIP_VALIDATION_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/validation-*.${AIP_DATA_FORMAT}"
* AIP_TEST_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/test-*.${AIP_DATA_FORMAT}"
.google.cloud.aiplatform.v1beta1.GcsDestination gcs_destination = 8;hasGcsDestination in interface InputDataConfigOrBuilderpublic GcsDestination getGcsDestination()
The Cloud Storage location where the training data is to be
written to. In the given directory a new directory is created with
name:
`dataset-<dataset-id>-<annotation-type>-<timestamp-of-training-call>`
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
All training input data is written into that directory.
The Vertex AI environment variables representing Cloud Storage
data URIs are represented in the Cloud Storage wildcard
format to support sharded data. e.g.: "gs://.../training-*.jsonl"
* AIP_DATA_FORMAT = "jsonl" for non-tabular data, "csv" for tabular data
* AIP_TRAINING_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/training-*.${AIP_DATA_FORMAT}"
* AIP_VALIDATION_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/validation-*.${AIP_DATA_FORMAT}"
* AIP_TEST_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/test-*.${AIP_DATA_FORMAT}"
.google.cloud.aiplatform.v1beta1.GcsDestination gcs_destination = 8;getGcsDestination in interface InputDataConfigOrBuilderpublic InputDataConfig.Builder setGcsDestination(GcsDestination value)
The Cloud Storage location where the training data is to be
written to. In the given directory a new directory is created with
name:
`dataset-<dataset-id>-<annotation-type>-<timestamp-of-training-call>`
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
All training input data is written into that directory.
The Vertex AI environment variables representing Cloud Storage
data URIs are represented in the Cloud Storage wildcard
format to support sharded data. e.g.: "gs://.../training-*.jsonl"
* AIP_DATA_FORMAT = "jsonl" for non-tabular data, "csv" for tabular data
* AIP_TRAINING_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/training-*.${AIP_DATA_FORMAT}"
* AIP_VALIDATION_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/validation-*.${AIP_DATA_FORMAT}"
* AIP_TEST_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/test-*.${AIP_DATA_FORMAT}"
.google.cloud.aiplatform.v1beta1.GcsDestination gcs_destination = 8;public InputDataConfig.Builder setGcsDestination(GcsDestination.Builder builderForValue)
The Cloud Storage location where the training data is to be
written to. In the given directory a new directory is created with
name:
`dataset-<dataset-id>-<annotation-type>-<timestamp-of-training-call>`
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
All training input data is written into that directory.
The Vertex AI environment variables representing Cloud Storage
data URIs are represented in the Cloud Storage wildcard
format to support sharded data. e.g.: "gs://.../training-*.jsonl"
* AIP_DATA_FORMAT = "jsonl" for non-tabular data, "csv" for tabular data
* AIP_TRAINING_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/training-*.${AIP_DATA_FORMAT}"
* AIP_VALIDATION_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/validation-*.${AIP_DATA_FORMAT}"
* AIP_TEST_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/test-*.${AIP_DATA_FORMAT}"
.google.cloud.aiplatform.v1beta1.GcsDestination gcs_destination = 8;public InputDataConfig.Builder mergeGcsDestination(GcsDestination value)
The Cloud Storage location where the training data is to be
written to. In the given directory a new directory is created with
name:
`dataset-<dataset-id>-<annotation-type>-<timestamp-of-training-call>`
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
All training input data is written into that directory.
The Vertex AI environment variables representing Cloud Storage
data URIs are represented in the Cloud Storage wildcard
format to support sharded data. e.g.: "gs://.../training-*.jsonl"
* AIP_DATA_FORMAT = "jsonl" for non-tabular data, "csv" for tabular data
* AIP_TRAINING_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/training-*.${AIP_DATA_FORMAT}"
* AIP_VALIDATION_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/validation-*.${AIP_DATA_FORMAT}"
* AIP_TEST_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/test-*.${AIP_DATA_FORMAT}"
.google.cloud.aiplatform.v1beta1.GcsDestination gcs_destination = 8;public InputDataConfig.Builder clearGcsDestination()
The Cloud Storage location where the training data is to be
written to. In the given directory a new directory is created with
name:
`dataset-<dataset-id>-<annotation-type>-<timestamp-of-training-call>`
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
All training input data is written into that directory.
The Vertex AI environment variables representing Cloud Storage
data URIs are represented in the Cloud Storage wildcard
format to support sharded data. e.g.: "gs://.../training-*.jsonl"
* AIP_DATA_FORMAT = "jsonl" for non-tabular data, "csv" for tabular data
* AIP_TRAINING_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/training-*.${AIP_DATA_FORMAT}"
* AIP_VALIDATION_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/validation-*.${AIP_DATA_FORMAT}"
* AIP_TEST_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/test-*.${AIP_DATA_FORMAT}"
.google.cloud.aiplatform.v1beta1.GcsDestination gcs_destination = 8;public GcsDestination.Builder getGcsDestinationBuilder()
The Cloud Storage location where the training data is to be
written to. In the given directory a new directory is created with
name:
`dataset-<dataset-id>-<annotation-type>-<timestamp-of-training-call>`
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
All training input data is written into that directory.
The Vertex AI environment variables representing Cloud Storage
data URIs are represented in the Cloud Storage wildcard
format to support sharded data. e.g.: "gs://.../training-*.jsonl"
* AIP_DATA_FORMAT = "jsonl" for non-tabular data, "csv" for tabular data
* AIP_TRAINING_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/training-*.${AIP_DATA_FORMAT}"
* AIP_VALIDATION_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/validation-*.${AIP_DATA_FORMAT}"
* AIP_TEST_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/test-*.${AIP_DATA_FORMAT}"
.google.cloud.aiplatform.v1beta1.GcsDestination gcs_destination = 8;public GcsDestinationOrBuilder getGcsDestinationOrBuilder()
The Cloud Storage location where the training data is to be
written to. In the given directory a new directory is created with
name:
`dataset-<dataset-id>-<annotation-type>-<timestamp-of-training-call>`
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
All training input data is written into that directory.
The Vertex AI environment variables representing Cloud Storage
data URIs are represented in the Cloud Storage wildcard
format to support sharded data. e.g.: "gs://.../training-*.jsonl"
* AIP_DATA_FORMAT = "jsonl" for non-tabular data, "csv" for tabular data
* AIP_TRAINING_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/training-*.${AIP_DATA_FORMAT}"
* AIP_VALIDATION_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/validation-*.${AIP_DATA_FORMAT}"
* AIP_TEST_DATA_URI =
"gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/test-*.${AIP_DATA_FORMAT}"
.google.cloud.aiplatform.v1beta1.GcsDestination gcs_destination = 8;getGcsDestinationOrBuilder in interface InputDataConfigOrBuilderpublic boolean hasBigqueryDestination()
Only applicable to custom training with tabular Dataset with BigQuery source. The BigQuery project location where the training data is to be written to. In the given project a new dataset is created with name `dataset_<dataset-id>_<annotation-type>_<timestamp-of-training-call>` where timestamp is in YYYY_MM_DDThh_mm_ss_sssZ format. All training input data is written into that dataset. In the dataset three tables are created, `training`, `validation` and `test`. * AIP_DATA_FORMAT = "bigquery". * AIP_TRAINING_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.training" * AIP_VALIDATION_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.validation" * AIP_TEST_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.test"
.google.cloud.aiplatform.v1beta1.BigQueryDestination bigquery_destination = 10;hasBigqueryDestination in interface InputDataConfigOrBuilderpublic BigQueryDestination getBigqueryDestination()
Only applicable to custom training with tabular Dataset with BigQuery source. The BigQuery project location where the training data is to be written to. In the given project a new dataset is created with name `dataset_<dataset-id>_<annotation-type>_<timestamp-of-training-call>` where timestamp is in YYYY_MM_DDThh_mm_ss_sssZ format. All training input data is written into that dataset. In the dataset three tables are created, `training`, `validation` and `test`. * AIP_DATA_FORMAT = "bigquery". * AIP_TRAINING_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.training" * AIP_VALIDATION_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.validation" * AIP_TEST_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.test"
.google.cloud.aiplatform.v1beta1.BigQueryDestination bigquery_destination = 10;getBigqueryDestination in interface InputDataConfigOrBuilderpublic InputDataConfig.Builder setBigqueryDestination(BigQueryDestination value)
Only applicable to custom training with tabular Dataset with BigQuery source. The BigQuery project location where the training data is to be written to. In the given project a new dataset is created with name `dataset_<dataset-id>_<annotation-type>_<timestamp-of-training-call>` where timestamp is in YYYY_MM_DDThh_mm_ss_sssZ format. All training input data is written into that dataset. In the dataset three tables are created, `training`, `validation` and `test`. * AIP_DATA_FORMAT = "bigquery". * AIP_TRAINING_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.training" * AIP_VALIDATION_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.validation" * AIP_TEST_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.test"
.google.cloud.aiplatform.v1beta1.BigQueryDestination bigquery_destination = 10;public InputDataConfig.Builder setBigqueryDestination(BigQueryDestination.Builder builderForValue)
Only applicable to custom training with tabular Dataset with BigQuery source. The BigQuery project location where the training data is to be written to. In the given project a new dataset is created with name `dataset_<dataset-id>_<annotation-type>_<timestamp-of-training-call>` where timestamp is in YYYY_MM_DDThh_mm_ss_sssZ format. All training input data is written into that dataset. In the dataset three tables are created, `training`, `validation` and `test`. * AIP_DATA_FORMAT = "bigquery". * AIP_TRAINING_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.training" * AIP_VALIDATION_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.validation" * AIP_TEST_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.test"
.google.cloud.aiplatform.v1beta1.BigQueryDestination bigquery_destination = 10;public InputDataConfig.Builder mergeBigqueryDestination(BigQueryDestination value)
Only applicable to custom training with tabular Dataset with BigQuery source. The BigQuery project location where the training data is to be written to. In the given project a new dataset is created with name `dataset_<dataset-id>_<annotation-type>_<timestamp-of-training-call>` where timestamp is in YYYY_MM_DDThh_mm_ss_sssZ format. All training input data is written into that dataset. In the dataset three tables are created, `training`, `validation` and `test`. * AIP_DATA_FORMAT = "bigquery". * AIP_TRAINING_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.training" * AIP_VALIDATION_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.validation" * AIP_TEST_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.test"
.google.cloud.aiplatform.v1beta1.BigQueryDestination bigquery_destination = 10;public InputDataConfig.Builder clearBigqueryDestination()
Only applicable to custom training with tabular Dataset with BigQuery source. The BigQuery project location where the training data is to be written to. In the given project a new dataset is created with name `dataset_<dataset-id>_<annotation-type>_<timestamp-of-training-call>` where timestamp is in YYYY_MM_DDThh_mm_ss_sssZ format. All training input data is written into that dataset. In the dataset three tables are created, `training`, `validation` and `test`. * AIP_DATA_FORMAT = "bigquery". * AIP_TRAINING_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.training" * AIP_VALIDATION_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.validation" * AIP_TEST_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.test"
.google.cloud.aiplatform.v1beta1.BigQueryDestination bigquery_destination = 10;public BigQueryDestination.Builder getBigqueryDestinationBuilder()
Only applicable to custom training with tabular Dataset with BigQuery source. The BigQuery project location where the training data is to be written to. In the given project a new dataset is created with name `dataset_<dataset-id>_<annotation-type>_<timestamp-of-training-call>` where timestamp is in YYYY_MM_DDThh_mm_ss_sssZ format. All training input data is written into that dataset. In the dataset three tables are created, `training`, `validation` and `test`. * AIP_DATA_FORMAT = "bigquery". * AIP_TRAINING_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.training" * AIP_VALIDATION_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.validation" * AIP_TEST_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.test"
.google.cloud.aiplatform.v1beta1.BigQueryDestination bigquery_destination = 10;public BigQueryDestinationOrBuilder getBigqueryDestinationOrBuilder()
Only applicable to custom training with tabular Dataset with BigQuery source. The BigQuery project location where the training data is to be written to. In the given project a new dataset is created with name `dataset_<dataset-id>_<annotation-type>_<timestamp-of-training-call>` where timestamp is in YYYY_MM_DDThh_mm_ss_sssZ format. All training input data is written into that dataset. In the dataset three tables are created, `training`, `validation` and `test`. * AIP_DATA_FORMAT = "bigquery". * AIP_TRAINING_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.training" * AIP_VALIDATION_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.validation" * AIP_TEST_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.test"
.google.cloud.aiplatform.v1beta1.BigQueryDestination bigquery_destination = 10;getBigqueryDestinationOrBuilder in interface InputDataConfigOrBuilderpublic String getDatasetId()
Required. The ID of the Dataset in the same Project and Location which data will be used to train the Model. The Dataset must use schema compatible with Model being trained, and what is compatible should be described in the used TrainingPipeline's [training_task_definition] [google.cloud.aiplatform.v1beta1.TrainingPipeline.training_task_definition]. For tabular Datasets, all their data is exported to training, to pick and choose from.
string dataset_id = 1 [(.google.api.field_behavior) = REQUIRED];getDatasetId in interface InputDataConfigOrBuilderpublic com.google.protobuf.ByteString getDatasetIdBytes()
Required. The ID of the Dataset in the same Project and Location which data will be used to train the Model. The Dataset must use schema compatible with Model being trained, and what is compatible should be described in the used TrainingPipeline's [training_task_definition] [google.cloud.aiplatform.v1beta1.TrainingPipeline.training_task_definition]. For tabular Datasets, all their data is exported to training, to pick and choose from.
string dataset_id = 1 [(.google.api.field_behavior) = REQUIRED];getDatasetIdBytes in interface InputDataConfigOrBuilderpublic InputDataConfig.Builder setDatasetId(String value)
Required. The ID of the Dataset in the same Project and Location which data will be used to train the Model. The Dataset must use schema compatible with Model being trained, and what is compatible should be described in the used TrainingPipeline's [training_task_definition] [google.cloud.aiplatform.v1beta1.TrainingPipeline.training_task_definition]. For tabular Datasets, all their data is exported to training, to pick and choose from.
string dataset_id = 1 [(.google.api.field_behavior) = REQUIRED];value - The datasetId to set.public InputDataConfig.Builder clearDatasetId()
Required. The ID of the Dataset in the same Project and Location which data will be used to train the Model. The Dataset must use schema compatible with Model being trained, and what is compatible should be described in the used TrainingPipeline's [training_task_definition] [google.cloud.aiplatform.v1beta1.TrainingPipeline.training_task_definition]. For tabular Datasets, all their data is exported to training, to pick and choose from.
string dataset_id = 1 [(.google.api.field_behavior) = REQUIRED];public InputDataConfig.Builder setDatasetIdBytes(com.google.protobuf.ByteString value)
Required. The ID of the Dataset in the same Project and Location which data will be used to train the Model. The Dataset must use schema compatible with Model being trained, and what is compatible should be described in the used TrainingPipeline's [training_task_definition] [google.cloud.aiplatform.v1beta1.TrainingPipeline.training_task_definition]. For tabular Datasets, all their data is exported to training, to pick and choose from.
string dataset_id = 1 [(.google.api.field_behavior) = REQUIRED];value - The bytes for datasetId to set.public String getAnnotationsFilter()
Applicable only to Datasets that have DataItems and Annotations. A filter on Annotations of the Dataset. Only Annotations that both match this filter and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on (for the auto-assigned that role is decided by Vertex AI). A filter with same syntax as the one used in [ListAnnotations][google.cloud.aiplatform.v1beta1.DatasetService.ListAnnotations] may be used, but note here it filters across all Annotations of the Dataset, and not just within a single DataItem.
string annotations_filter = 6;getAnnotationsFilter in interface InputDataConfigOrBuilderpublic com.google.protobuf.ByteString getAnnotationsFilterBytes()
Applicable only to Datasets that have DataItems and Annotations. A filter on Annotations of the Dataset. Only Annotations that both match this filter and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on (for the auto-assigned that role is decided by Vertex AI). A filter with same syntax as the one used in [ListAnnotations][google.cloud.aiplatform.v1beta1.DatasetService.ListAnnotations] may be used, but note here it filters across all Annotations of the Dataset, and not just within a single DataItem.
string annotations_filter = 6;getAnnotationsFilterBytes in interface InputDataConfigOrBuilderpublic InputDataConfig.Builder setAnnotationsFilter(String value)
Applicable only to Datasets that have DataItems and Annotations. A filter on Annotations of the Dataset. Only Annotations that both match this filter and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on (for the auto-assigned that role is decided by Vertex AI). A filter with same syntax as the one used in [ListAnnotations][google.cloud.aiplatform.v1beta1.DatasetService.ListAnnotations] may be used, but note here it filters across all Annotations of the Dataset, and not just within a single DataItem.
string annotations_filter = 6;value - The annotationsFilter to set.public InputDataConfig.Builder clearAnnotationsFilter()
Applicable only to Datasets that have DataItems and Annotations. A filter on Annotations of the Dataset. Only Annotations that both match this filter and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on (for the auto-assigned that role is decided by Vertex AI). A filter with same syntax as the one used in [ListAnnotations][google.cloud.aiplatform.v1beta1.DatasetService.ListAnnotations] may be used, but note here it filters across all Annotations of the Dataset, and not just within a single DataItem.
string annotations_filter = 6;public InputDataConfig.Builder setAnnotationsFilterBytes(com.google.protobuf.ByteString value)
Applicable only to Datasets that have DataItems and Annotations. A filter on Annotations of the Dataset. Only Annotations that both match this filter and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on (for the auto-assigned that role is decided by Vertex AI). A filter with same syntax as the one used in [ListAnnotations][google.cloud.aiplatform.v1beta1.DatasetService.ListAnnotations] may be used, but note here it filters across all Annotations of the Dataset, and not just within a single DataItem.
string annotations_filter = 6;value - The bytes for annotationsFilter to set.public String getAnnotationSchemaUri()
Applicable only to custom training with Datasets that have DataItems and Annotations. Cloud Storage URI that points to a YAML file describing the annotation schema. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). The schema files that can be used here are found in gs://google-cloud-aiplatform/schema/dataset/annotation/ , note that the chosen schema must be consistent with [metadata][google.cloud.aiplatform.v1beta1.Dataset.metadata_schema_uri] of the Dataset specified by [dataset_id][google.cloud.aiplatform.v1beta1.InputDataConfig.dataset_id]. Only Annotations that both match this schema and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on. When used in conjunction with [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter], the Annotations used for training are filtered by both [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter] and [annotation_schema_uri][google.cloud.aiplatform.v1beta1.InputDataConfig.annotation_schema_uri].
string annotation_schema_uri = 9;getAnnotationSchemaUri in interface InputDataConfigOrBuilderpublic com.google.protobuf.ByteString getAnnotationSchemaUriBytes()
Applicable only to custom training with Datasets that have DataItems and Annotations. Cloud Storage URI that points to a YAML file describing the annotation schema. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). The schema files that can be used here are found in gs://google-cloud-aiplatform/schema/dataset/annotation/ , note that the chosen schema must be consistent with [metadata][google.cloud.aiplatform.v1beta1.Dataset.metadata_schema_uri] of the Dataset specified by [dataset_id][google.cloud.aiplatform.v1beta1.InputDataConfig.dataset_id]. Only Annotations that both match this schema and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on. When used in conjunction with [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter], the Annotations used for training are filtered by both [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter] and [annotation_schema_uri][google.cloud.aiplatform.v1beta1.InputDataConfig.annotation_schema_uri].
string annotation_schema_uri = 9;getAnnotationSchemaUriBytes in interface InputDataConfigOrBuilderpublic InputDataConfig.Builder setAnnotationSchemaUri(String value)
Applicable only to custom training with Datasets that have DataItems and Annotations. Cloud Storage URI that points to a YAML file describing the annotation schema. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). The schema files that can be used here are found in gs://google-cloud-aiplatform/schema/dataset/annotation/ , note that the chosen schema must be consistent with [metadata][google.cloud.aiplatform.v1beta1.Dataset.metadata_schema_uri] of the Dataset specified by [dataset_id][google.cloud.aiplatform.v1beta1.InputDataConfig.dataset_id]. Only Annotations that both match this schema and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on. When used in conjunction with [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter], the Annotations used for training are filtered by both [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter] and [annotation_schema_uri][google.cloud.aiplatform.v1beta1.InputDataConfig.annotation_schema_uri].
string annotation_schema_uri = 9;value - The annotationSchemaUri to set.public InputDataConfig.Builder clearAnnotationSchemaUri()
Applicable only to custom training with Datasets that have DataItems and Annotations. Cloud Storage URI that points to a YAML file describing the annotation schema. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). The schema files that can be used here are found in gs://google-cloud-aiplatform/schema/dataset/annotation/ , note that the chosen schema must be consistent with [metadata][google.cloud.aiplatform.v1beta1.Dataset.metadata_schema_uri] of the Dataset specified by [dataset_id][google.cloud.aiplatform.v1beta1.InputDataConfig.dataset_id]. Only Annotations that both match this schema and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on. When used in conjunction with [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter], the Annotations used for training are filtered by both [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter] and [annotation_schema_uri][google.cloud.aiplatform.v1beta1.InputDataConfig.annotation_schema_uri].
string annotation_schema_uri = 9;public InputDataConfig.Builder setAnnotationSchemaUriBytes(com.google.protobuf.ByteString value)
Applicable only to custom training with Datasets that have DataItems and Annotations. Cloud Storage URI that points to a YAML file describing the annotation schema. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). The schema files that can be used here are found in gs://google-cloud-aiplatform/schema/dataset/annotation/ , note that the chosen schema must be consistent with [metadata][google.cloud.aiplatform.v1beta1.Dataset.metadata_schema_uri] of the Dataset specified by [dataset_id][google.cloud.aiplatform.v1beta1.InputDataConfig.dataset_id]. Only Annotations that both match this schema and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on. When used in conjunction with [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter], the Annotations used for training are filtered by both [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter] and [annotation_schema_uri][google.cloud.aiplatform.v1beta1.InputDataConfig.annotation_schema_uri].
string annotation_schema_uri = 9;value - The bytes for annotationSchemaUri to set.public String getSavedQueryId()
Only applicable to Datasets that have SavedQueries. The ID of a SavedQuery (annotation set) under the Dataset specified by [dataset_id][google.cloud.aiplatform.v1beta1.InputDataConfig.dataset_id] used for filtering Annotations for training. Only Annotations that are associated with this SavedQuery are used in respectively training. When used in conjunction with [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter], the Annotations used for training are filtered by both [saved_query_id][google.cloud.aiplatform.v1beta1.InputDataConfig.saved_query_id] and [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter]. Only one of [saved_query_id][google.cloud.aiplatform.v1beta1.InputDataConfig.saved_query_id] and [annotation_schema_uri][google.cloud.aiplatform.v1beta1.InputDataConfig.annotation_schema_uri] should be specified as both of them represent the same thing: problem type.
string saved_query_id = 7;getSavedQueryId in interface InputDataConfigOrBuilderpublic com.google.protobuf.ByteString getSavedQueryIdBytes()
Only applicable to Datasets that have SavedQueries. The ID of a SavedQuery (annotation set) under the Dataset specified by [dataset_id][google.cloud.aiplatform.v1beta1.InputDataConfig.dataset_id] used for filtering Annotations for training. Only Annotations that are associated with this SavedQuery are used in respectively training. When used in conjunction with [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter], the Annotations used for training are filtered by both [saved_query_id][google.cloud.aiplatform.v1beta1.InputDataConfig.saved_query_id] and [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter]. Only one of [saved_query_id][google.cloud.aiplatform.v1beta1.InputDataConfig.saved_query_id] and [annotation_schema_uri][google.cloud.aiplatform.v1beta1.InputDataConfig.annotation_schema_uri] should be specified as both of them represent the same thing: problem type.
string saved_query_id = 7;getSavedQueryIdBytes in interface InputDataConfigOrBuilderpublic InputDataConfig.Builder setSavedQueryId(String value)
Only applicable to Datasets that have SavedQueries. The ID of a SavedQuery (annotation set) under the Dataset specified by [dataset_id][google.cloud.aiplatform.v1beta1.InputDataConfig.dataset_id] used for filtering Annotations for training. Only Annotations that are associated with this SavedQuery are used in respectively training. When used in conjunction with [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter], the Annotations used for training are filtered by both [saved_query_id][google.cloud.aiplatform.v1beta1.InputDataConfig.saved_query_id] and [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter]. Only one of [saved_query_id][google.cloud.aiplatform.v1beta1.InputDataConfig.saved_query_id] and [annotation_schema_uri][google.cloud.aiplatform.v1beta1.InputDataConfig.annotation_schema_uri] should be specified as both of them represent the same thing: problem type.
string saved_query_id = 7;value - The savedQueryId to set.public InputDataConfig.Builder clearSavedQueryId()
Only applicable to Datasets that have SavedQueries. The ID of a SavedQuery (annotation set) under the Dataset specified by [dataset_id][google.cloud.aiplatform.v1beta1.InputDataConfig.dataset_id] used for filtering Annotations for training. Only Annotations that are associated with this SavedQuery are used in respectively training. When used in conjunction with [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter], the Annotations used for training are filtered by both [saved_query_id][google.cloud.aiplatform.v1beta1.InputDataConfig.saved_query_id] and [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter]. Only one of [saved_query_id][google.cloud.aiplatform.v1beta1.InputDataConfig.saved_query_id] and [annotation_schema_uri][google.cloud.aiplatform.v1beta1.InputDataConfig.annotation_schema_uri] should be specified as both of them represent the same thing: problem type.
string saved_query_id = 7;public InputDataConfig.Builder setSavedQueryIdBytes(com.google.protobuf.ByteString value)
Only applicable to Datasets that have SavedQueries. The ID of a SavedQuery (annotation set) under the Dataset specified by [dataset_id][google.cloud.aiplatform.v1beta1.InputDataConfig.dataset_id] used for filtering Annotations for training. Only Annotations that are associated with this SavedQuery are used in respectively training. When used in conjunction with [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter], the Annotations used for training are filtered by both [saved_query_id][google.cloud.aiplatform.v1beta1.InputDataConfig.saved_query_id] and [annotations_filter][google.cloud.aiplatform.v1beta1.InputDataConfig.annotations_filter]. Only one of [saved_query_id][google.cloud.aiplatform.v1beta1.InputDataConfig.saved_query_id] and [annotation_schema_uri][google.cloud.aiplatform.v1beta1.InputDataConfig.annotation_schema_uri] should be specified as both of them represent the same thing: problem type.
string saved_query_id = 7;value - The bytes for savedQueryId to set.public final InputDataConfig.Builder setUnknownFields(com.google.protobuf.UnknownFieldSet unknownFields)
setUnknownFields in interface com.google.protobuf.Message.BuildersetUnknownFields in class com.google.protobuf.GeneratedMessageV3.Builder<InputDataConfig.Builder>public final InputDataConfig.Builder mergeUnknownFields(com.google.protobuf.UnknownFieldSet unknownFields)
mergeUnknownFields in interface com.google.protobuf.Message.BuildermergeUnknownFields in class com.google.protobuf.GeneratedMessageV3.Builder<InputDataConfig.Builder>Copyright © 2022 Google LLC. All rights reserved.