ImageAnalysisResult Class
Represents the outcome of an Image Analysis operation.
- Inheritance
-
azure.ai.vision.imageanalysis._model_base.ModelImageAnalysisResult
Constructor
ImageAnalysisResult(*args: Any, **kwargs: Any)
Variables
Name | Description |
---|---|
caption
|
The generated phrase that describes the content of the analyzed image. |
dense_captions
|
The up to 10 generated phrases, the first describing the content of the whole image, and the others describing the content of different regions of the image. |
metadata
|
Metadata associated with the analyzed image. Required. |
model_version
|
The cloud AI model used for the analysis. Required. |
objects
|
A list of detected physical objects in the analyzed image, and their location. |
people
|
A list of detected people in the analyzed image, and their location. |
read
|
The extracted printed and hand-written text in the analyze image. Also knows as OCR. |
smart_crops
|
A list of crop regions at the desired as aspect ratios (if provided) that can be used as image thumbnails. These regions preserve as much content as possible from the analyzed image, with priority given to detected faces. |
tags
|
A list of content tags in the analyzed image. |
Methods
as_dict |
Return a dict that can be JSONify using json.dump. |
clear | |
copy | |
get | |
items | |
keys | |
pop | |
popitem | |
setdefault | |
update | |
values |
as_dict
Return a dict that can be JSONify using json.dump.
as_dict(*, exclude_readonly: bool = False) -> Dict[str, Any]
Keyword-Only Parameters
Name | Description |
---|---|
exclude_readonly
|
Whether to remove the readonly properties. |
Returns
Type | Description |
---|---|
A dict JSON compatible object |
clear
clear() -> None
copy
copy() -> Model
get
get(key: str, default: Any = None) -> Any
Parameters
Name | Description |
---|---|
key
Required
|
|
default
Required
|
Default value: None
|
items
items() -> ItemsView[str, Any]
keys
keys() -> KeysView[str]
pop
pop(key: str, default: ~typing.Any = <object object>) -> Any
Parameters
Name | Description |
---|---|
key
Required
|
|
default
Required
|
|
popitem
popitem() -> Tuple[str, Any]
setdefault
setdefault(key: str, default: ~typing.Any = <object object>) -> Any
Parameters
Name | Description |
---|---|
key
Required
|
|
default
Required
|
|
update
update(*args: Any, **kwargs: Any) -> None
values
values() -> ValuesView[Any]
Attributes
caption
The generated phrase that describes the content of the analyzed image.
caption: _models.CaptionResult | None
dense_captions
The up to 10 generated phrases, the first describing the content of the whole image, and the others describing the content of different regions of the image.
dense_captions: _models.DenseCaptionsResult | None
metadata
Metadata associated with the analyzed image. Required.
metadata: _models.ImageMetadata
model_version
The cloud AI model used for the analysis. Required.
model_version: str
objects
A list of detected physical objects in the analyzed image, and their location.
objects: _models.ObjectsResult | None
people
A list of detected people in the analyzed image, and their location.
people: _models.PeopleResult | None
read
The extracted printed and hand-written text in the analyze image. Also knows as OCR.
read: _models.ReadResult | None
smart_crops
A list of crop regions at the desired as aspect ratios (if provided) that can be used as image thumbnails. These regions preserve as much content as possible from the analyzed image, with priority given to detected faces.
smart_crops: _models.SmartCropsResult | None
tags
A list of content tags in the analyzed image.
tags: _models.TagsResult | None
Azure SDK for Python