Professional Documents
Culture Documents
Open World Plant Image Identification Based On Convolution Neural Networks
Open World Plant Image Identification Based On Convolution Neural Networks
OR
(cropped) (warped)
convolution layers with ReLU activations and pooling layers convolution layers with PReLU activations and pooling layers
hidden layers with ReLU activations hidden layers with PReLU activations
class
probability Omitted? OR
A B C A B C
Fig. 1. The VGG 16-layers model Fig. 2. Modified model with Unseen Category Query Identification
Thus, any image of unseen category associated with high score REFERENCES
is likely to degrade the MAP. In addition to this, MAP without
images of unseen category is also evaluated.
[1] A. Krizhevsky, I. Sutskever and G. E. Hinton, "ImageNet
B. Evaluation Results Classification with Deep Convolutional," in Advances in Neural
Our first method uses threshold vector 𝒕 to identify and omit Information Processing Systems 25, 2012.
images of unseen category while our second method uses [2] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov,
threshold vector 𝒕′ . We compare our results with the top D. Erhan, V. Vanhoucke and A. Rabinovich, "Going deeper with
performing teams (runs) of PlantCLEF 2016. Evaluation results convolutions," CoRR, vol. 1409.4842, 2014.
are summarized in Table I. [3] K. Simonyan and A. Zisserman, "Very Deep Convolutional
Networks for Large-Scale Image Recognition," CoRR, vol.
TABLE I 1409.1556, 2014.
EVALUATION RESULTS
[4] K. He, X. Zhang, S. Ren and J. Sun, "Deep Residual Learning for
MAP with images of MAP without images of Image Recognition," CoRR, vol. 1512.03385, 2015.
Method
unseen category unseen category
Our method (with 𝒕) 0.736 0.820 [5] H. Goeau, P. Bonnet and A. Joly, "LifeCLEF Plant Identification
Our method (with 𝒕') 0.742 0.827 Task," in CLEF working notes 2015, 2015.
SabanciUGebzeTU Run 1 0.738 0.806 [6] H. Goëau, P. Bonnet and A. Joly, "Plant Identification In An
CMP Run 1 0.710 0.790 Open-World (LifeCLEF 2016)," in CLEF working notes 2016,
LIIR KUL Run 3 0.703 0.761 2016.
UM Run 4 0.669 0.598
[7] K. He, X. Zhang, S. Ren and J. Sun, "Spatial Pyramid Pooling in
QUT Run 3 0.629 0.610
Deep Convolutional Networks for Visual Recognition," CoRR,
vol. 1406.4729, 2014.
Our method with threshold vector 𝒕′ shows the highest MAP.
[8] K. He, X. Zhang, S. Ren and J. Sun, "Delving Deep into
Usage of threshold vector 𝒕′ considering both training and
Rectifiers: Surpassing Human-Level Performance on ImageNet
validation images yields a slight improvement.
Classification," CoRR, vol. 1502.01852, 2015.
V. CONCLUSION [9] B. Xu, N. Wang, T. Chen and M. Li, "Empirical Evaluation of
Rectified Activations in Convolution," CoRR, vol. 1505.00853,
In this paper, we have presented our approach towards open 2015.
world plant image identification, focusing on model [10] V. N. Vapnik, Statistical Learning Theory, Wiley, 1998.
enhancements and an unseen category image identification
[11] X. Glorot and Y. Bengio, "Understanding the difficulty of
algorithm. We still leave some rooms for future improvements,
training deep feedforward neural networks," in Proceedings of
which are itemized as follows:
AISTATS 2010, 2010.