dlpy.image_captioning.ImageCaptioning

dlpy.image_captioning.ImageCaptioning(conn, model_name='image_captioning', num_blocks=3, neurons=50, rnn_type='LSTM', max_output_len=15)

Builds an RNN to be used for image captioning

Parameters:
conn : CAS

Specifies the CAS connection object.

model_name : string, optional

Specifies output name of the model
Default: ‘image_captioning’

num_blocks : int, optional

Specifies number of samelength recurrent layers Default : 3

neurons : int, optional

Specifies number of neurons in each layer Default : 50

rnn_type : string, optional

Specifies the type of the rnn layer. Possible Values: RNN, LSTM, GRU
Default: LSTM

max_output_len : int, optional

Specifies max number of tokens to generate in the final layer (i.e. max caption length) Default : 15

Returns
——-
CASTable