Train Vowpal Wabbit Version 7-10 Model
Important
Support for Machine Learning Studio (classic) will end on 31 August 2024. We recommend you transition to Azure Machine Learning by that date.
Beginning 1 December 2021, you will not be able to create new Machine Learning Studio (classic) resources. Through 31 August 2024, you can continue to use the existing Machine Learning Studio (classic) resources.
- See information on moving machine learning projects from ML Studio (classic) to Azure Machine Learning.
- Learn more about Azure Machine Learning.
ML Studio (classic) documentation is being retired and may not be updated in the future.
Trains a model using version 7-10 of the Vowpal Wabbit machine learning system
Category: Text Analytics
Note
Applies to: Machine Learning Studio (classic) only
Similar drag-and-drop modules are available in Azure Machine Learning designer.
Module overview
This article describes how to use the Train Vowpal Wabbit Version 7-10 module in Machine Learning Studio (classic), to create a machine learning model using an instance of Vowpal Wabbit (version 7-10).
To use Vowpal Wabbit for machine learning, format your input according to Vowpal Wabbit requirements, and save the data in an Azure blob. Use this module to specify Vowpal Wabbit command-line arguments.
When the experiment is run, an instance of Vowpal Wabbit is loaded into the experiment run-time, together with the specified data. When training is complete, the model is serialized back to the workspace. You can use the model immediately to score data. The trained model is also persisted in Azure storage so that you can use it later without having to reprocess the training data.
To incrementally train an existing model on new data, connect a saved model to the Pre-trained model input, and add the new data to the other input.
Note
Machine Learning Studio (classic) hosts multiple versions of the Vowpal Wabbit framework. This module uses the version 7-10 of the Vowpal Wabbit framework.
If you need to build or score a model based on a previous version (7-4 or 7-6), use these modules: Train Vowpal Wabbit 7-4 Model and Score Vowpal Wabbit 7-4 Model.
For the latest version, use Train Vowpal Wabbit Version 8 Model, together with its scoring module, Score Vowpal Wabbit 8 Model.
What is Vowpal Wabbit?
Vowpal Wabbit (VW) is a fast, parallel machine learning framework that was developed for distributed computing by Yahoo! Research. Later it was ported to Windows and adapted by John Langford (Microsoft Research) for scientific computing in parallel architectures.
Features of Vowpal Wabbit that are important for machine learning include continuous learning (online learning), dimensionality reduction, and interactive learning. Vowpal Wabbit is also a solution for problems when you cannot fit the model data into memory.
The primary users of Vowpal Wabbit in Machine Learning are data scientists who have previously used the framework for machine learning tasks such as classification, regression, topic modeling or matrix factorization. The Azure wrapper for Vowpal Wabbit has very similar performance characteristics to the on-premise version, which means that users can continue to build models, retrain, and score using the powerful features and native performance of Vowpal Wabbit, while gaining the ability to easily publish the trained model as an operationalized service.
The Feature Hashing module also includes functionality provided by Vowpal Wabbit, that lets you transform text datasets into binary features using a hashing algorithm.
How to configure Vowpal Wabbit Version 7-10 Model
This section describes how to train a new model, and how to add new data to an existing model.
Unlike other modules in Studio (classic), this module both specifies the module parameters, and trains the model. If you have an existing model, you can add it as an optional input, to incrementally train the model.
- Prepare input data in one of the required formats
- Train a new model
- Incrementally train an existing model
Use of this module requires authentication to an Azure storage account.
Prepare the input data
To train a model using this module, the input dataset must consist of a single text column in one of the two supported formats: LibSVM or VW.
This doesn't mean that Vowpal Wabbit analyzes only text data, just that the features and values must be prepared in the required text file format.
The data must be read from Azure storage. It is not possible to use Export Data to directly save the input file to Azure for use with Vowpal Wabbit, because the format requires some additional modification. You must ensure the data is in the correct format and then upload the data to Azure blob storage.
However, as a shortcut, you can use the Convert to SVMLight module to generate an SVMLight format file. Then, you can either upload the SVMLight format file to Azure blob storage and use it as the input, or you can modify the file slightly to conform to the Vowpal Wabbit input file requirements.
The Vowpal Wabbit data format has the advantage that it does not require a columnar format, which saves space when dealing with sparse data. For more information about this format, see the Vowpal Wabbit wiki page.
Create and train a Vowpal Wabbit model
Add the Train Vowpal Wabbit Version 7-10 module to your experiment.
Specify the account where the training data is stored. The trained model and hashing file are stored in the same location.
For Azure storage account name, type the name of the Azure storage account.
For Azure storage key, copy and paste the key that is provided for accessing the storage account,
If you don’t have a key, see How to regenerate storage access keys
For Azure container name, type the name of a single container in the specified Azure storage account where the model training data is stored. Do not type the account name or any protocol prefix.
For example, if the full container path and name is
https://myaccount.blob.core.windows.net/vwmodels
, you should type justvwmodels
. For more information about container names, see Naming and Referencing Containers, Blobs, and Metadata.In the VW arguments text box, type the command-line arguments for the Vowpal Wabbit executable.
For example, you might add
–l
to specify the learning rate, or-b
to indicate the number of hashing bits.For more information, see the parameters section.
Name of the input VW file: Type the name of the file that contains the input data. The file must be an existing file in Azure blob storage, located in the previously specified storage account and container. The file must have been prepared using one of the supported formats.
Name of the output readable model (--readable_model) file: Type the name of a file where the trained model should be saved. The file must be saved within the same storage account and container as the input file.
This argument corresponds to the
--readable_model
parameter in the VW command line.Name of the output inverted hash (--invert_hash) file: Type the name of the file in which the inverted hashing function should be saved. The file must be saved within the same storage account and container as the input file.
This argument corresponds to the
--invert_hash
parameter in the VW command line.Please specify file type: Indicate which format your training data uses. Vowpal Wabbit supports these two input file formats:
VW represents the internal format used by Vowpal Wabbit . See the Vowpal Wabbit wiki page for details.
SVMLight is a format used by some other machine learning tools.
Select the option, Use cached results, if you don't want to load the data from storage each time the experiment is re-reun. Assuming no other parameters have changed and a valid cache can be found, Studio (classic) uses a cached version of the data.
If this option is deselected, the module always reads the data from storage.
Run the experiment.
After the model has been generated, right-click the output and select Save as trained model, so that you can re-use and re-train the model later.
Retrain an existing Vowpal Wabbit model
Vowpal Wabbit supports incremental training by adding new data to an existing model. There are two ways to get an existing model for retraining:
Use the output of another Train Vowpal Wabbit Version 8 module in the same experiment.
Locate a saved model in the Trained Models group in Studio (classic), and drag it in to your experiment.
Add the Train Vowpal Wabbit Version 8 module to your experiment.
Connect the previously trained model to the input port of Train Vowpal Wabbit Version 8:
In the Properties pane of Train Vowpal Wabbit Version 8, specify the location and format of the new training data.
Specify a name for the human-readable model output file, and another name for the hash file associated with the updated model.
Note
If there is an existing Vowpal Wabbit model or hash file in the specified location, the files are silently overwritten by the new trained model. To preserve intermediate models when retraining, you must change the storage location or make a local copy of the model files.
Run the experiment.
Right-click the module and select Save as Trained Model to preserve the updated model in your Machine Learning workspace. If you don't specify a new name, the updated model overwrites the existing saved model.
Examples
For examples of how Vowpal Wabbit can be used in machine learning, see the Azure AI Gallery:
-
This experiment demonstrates data preparation, training, and operationalization of a VW model.
Also, see these resources:
Blog describing Vowpal Wabbit implementation and roadmap
Technical notes
This section contains implementation details, tips, and answers to frequently asked questions.
Advantages of Vowpal Wabbit
Vowpal Wabbit provides extremely fast learning over non-linear features like n-grams.
Vowpal Wabbit uses online learning techniques such as stochastic gradient descent (SGD) to fit a model one record at a time. Thus it iterates very quickly over raw data and can develop a good predictor faster than most other models. This approach also avoids having to read all training data into memory.
Vowpal Wabbit converts all data to hashes, not just text data but other categorical variables. Using hashes makes lookup of regression weights more efficient, which is critical for effective stochastic gradient descent.
During training, the module makes calls into a Vowpal Wabbit wrapper developed for Azure. The training data is downloaded in blocks from Azure, utilizing the high bandwidth between the store and the worker roles executing the computations, and is streamed to the VW learners. The resulting model is generally very compact due to the internal compression done by VW. The model is copied back to the experiment workspace where it can be utilized like other models in Machine Learning.
Supported and unsupported parameters
This section describes support for Vowpal Wabbit command line parameters in Machine Learning Studio (classic).
You cannot use the following command-line arguments in Machine Learning Studio (classic).
The input/output options specified in Vowpal Wabbit Wiki - Command-line-arguments
These properties are configured automatically by the module.
Any option that generates multiple outputs or takes multiple inputs is disallowed. These include:
--cbt
,--lda
,--wap
Only supervised learning algorithms are supported. Therefore, options such as these are not supported:
–active
,--rank
,--search
All arguments other than those described above are allowed.
For a complete list of arguments, use the Vowpal Wabbit wiki page.
Restrictions
Because the goal of the service is to support experienced users of Vowpal Wabbit, input data must be prepared ahead of time using the Vowpal Wabbit native text format, rather than the dataset format used by other modules.
Rather than using data in the Azure ML workspace, the training data is directly streamed from Azure, for maximal performance and minimal parsing overhead. For this reason, there is only limited interoperability between the VW modules and other modules in Azure ML.
Module parameters
Name | Range | Type | Default | Description |
---|---|---|---|---|
Azure storage account name | any | String | Type the Azure storage account name | |
Azure storage key | any | SecureString | Provide the Azure storage key | |
Azure container name | any | String | Type the Azure container name | |
VW arguments | any | String | Specify any Vowpal Wabbit arguments. The argument –f is not supported. |
|
Name of the input VW file | any | String | Specify the name of an input file in the Vowpal Wabbit format | |
Name of the output readable model (--readable_model) file | any | String | If specified, outputs a readable model back to the Azure container. This argument is optional. |
|
Name of the output inverted hash (--invert_hash) file | any | String | If specified, outputs a file containing the inverted hash function back to the Azure container. This argument is optional. |
|
Please specify file type | VW SVMLight |
DataType | VW | Indicate whether the file type uses the SVMLight format or the Vowpal Wabbit format. |
Outputs
Name | Type | Description |
---|---|---|
Trained model | ILearner interface | Trained learner |
Exceptions
Exception | Description |
---|---|
Error 0001 | Exception occurs if one or more specified columns of data set couldn't be found. |
Error 0003 | Exception occurs if one or more of inputs are null or empty. |
Error 0004 | Exception occurs if parameter is less than or equal to specific value. |
Error 0017 | Exception occurs if one or more specified columns have type unsupported by current module. |
For a list of errors specific to Studio (classic) modules, see Machine Learning Error codes.
For a list of API exceptions, see Machine Learning REST API Error Codes.
See also
Text Analytics
Feature Hashing
Named Entity Recognition
Score Vowpal Wabbit 7-4 Model
Score Vowpal Wabbit 7-10 Model
Train Vowpal Wabbit 7-4 Model
A-Z Module List