SlideShare a Scribd company logo
1 of 5
Download to read offline
IceBreaker: Solving Cold Start Problem for Video
Recommendation Engines
Agniv Sharma∗
DTU
New Delhi
agnivsharma96@gmail.com
Yaman Kumar∗
Adobe Systems
Noida
ykumar@adobe.com
Abhigyan Khaund∗
IIT Mandi
Mandi
khaundabhigyan@gmail.com
Akash Kumar∗
DTU
New Delhi
akash_bt2k15@dtu.ac.in
Ponnurangam Kumaraguru
IIIT-Delhi
New Delhi
pk@iitd.ac.in
Rajiv Ratn Shah
IIIT-Delhi
New Delhi
rajivratn@iiitd.ac.in
ABSTRACT
Internet has brought about a tremendous increase in content of all
forms and, in that, video content constitutes the major backbone
of the total content being published as well as watched. Thus it
becomes imperative for video recommendation engines such as
Hulu to look for novel and innovative ways to recommend the
newly added videos to their users. However, the problem with new
videos is that they lack any sort of metadata and user interaction so
as to be able to rate the videos for the consumers. To this effect, this
paper introduces the several techniques we develop for the Content
Based Video Relevance Prediction (CBVRP) Challenge being hosted
by Hulu for the ACM Multimedia Conference 2018. We employ
different architectures on the CBVRP dataset to make use of the
provided frame and video level features and generate predictions
of videos that are similar to the other videos. We also implement
several ensemble strategies to explore complementarity between
both the types of provided features. The obtained results are en-
couraging and will impel the boundaries of research for multimedia
based video recommendation systems.
KEYWORDS
Video Recommendation System, Implicit Features, Video Relevance
Prediction, Content based Recommendation, LDA
ACM Reference Format:
Agniv Sharma, Yaman Kumar, Abhigyan Khaund, Akash Kumar, Ponnu-
rangam Kumaraguru, and Rajiv Ratn Shah. 2018. IceBreaker: Solving Cold
Start Problem for Video Recommendation Engines . In Proceedings of ACM
Conference (Conference’17). ACM, New York, NY, USA, 5 pages. https://doi.
org/10.1145/nnnnnnn.nnnnnnn
1 INTRODUCTION
The recent years have seen a huge surge in online content, and, con-
tent in the form of videos has led the way. This stupendous wave
has been steered by online video recommendation systems like
Hulu, Youtube, Netflix, etc. At the same time, the content supplied
by traditional media houses and other licensed broadcasters has
seen a plunge [7] with respect to amateurs and non-professionals
∗Equal Contribution
Conference’17, July 2017, Washington, DC, USA
2018. ACM ISBN 978-x-xxxx-xxxx-x/YY/MM...$15.00
https://doi.org/10.1145/nnnnnnn.nnnnnnn
uploading home-made content on various platforms. The popu-
larity of video streaming can be judged from the fact that sixty
hours of video content is generated every minute on Youtube [5].
A tremendous amount of new content is being generated at a very
fast pace and, on top of that, the content metadata itself ranges
from incorrect to either none or highly limited information [10, 11].
Additionally, the content thus generated is short (lesser than 10
minutes) and un-reviewed. The user interaction with the horde of
these videos is minimal at the very best [10]. Due to these reason,
video recommendation systems suffer from a multi-dimensional
challenge where they have to deal with un-audited, un-rated and
completely new content of which they know nothing about. This
leads to the infamous cold start problem where recommendation en-
gines have to tackle new content and then process and recommend
it to the users in such a manner so as to keep grabbing the eyeballs
and dissuade them from shifting to the competitor’s platform.
Video recommendation has traditionally been inspired from two
types of data: implicit and explicit. While on one hand, video rec-
ommendation on the basis of implicit content deals with features
like light, mood, color, shape, motion, plot and other aesthetics;
on the other hand, explicit content incorporates movie genre, di-
rector, actors, description, previous views and likes. Researchers
have spent considerable time and effort building recommendation
engines based on both the types of data [2, 3, 11, 23, 38]. But as
noted in the previous section, for today’s content explosion, the
recommendation on the basis of explicit content fails utterly. This
leaves one with implicit features. According to Applied Media Aes-
thetic [36], media features such as light, color, camera motion etc.
serve as crucial elements which render emotional, aesthetic and
informative effects. Thus a new artistic work, can be appropriately
evaluated and recommended on the basis of intrinsic features.
In this paper, we present our extensive experiments in building
a video recommendation system which deals with the cold-start
problem and uses intrinsic features to recommend videos. Here, we
portray our experiments and their results over multiple architec-
tures for Content Based Video Relevance Prediction Challenge of
ACM Multimedia 2018, sponsored by Hulu.
2 RELATED WORK
Building recommendation engines in general, and video recom-
mendation engines in particular has always been a hot topic of
arXiv:1808.05636v1
[cs.IR]
16
Aug
2018
Conference’17, July 2017, Washington, DC, USA Yaman Kumar et al.
research. Right from the development of the first recommender sys-
tem, Tapestry[17], recommender systems have undergone drastic
developmental changes. There have been several works which have
employed the low level features, also known as stylistic features
for the task of video recommendation[6, 20, 34, 37]. But none of
these approaches use only low-level features, thus they cannot solve
the cold-start problem in its full glory[11, 26, 28]. However, a few
authors have realized this problem and have tried to tackle this
problem head-on using various methods [11, 29, 32]. Thus one of
the novelties introduced in this work is exploring the recommenda-
tion problem using only stylistic and artistic features for solving
the daunting cold-start dilemma.
3 METHODOLOGY
In this section, we present the dataset and different architectures
i.e., random forest regression on video pairs, deep learning based
regression and neural network with DeepLDA for the CBCRP chal-
lenge.
3.1 Dataset
In this paper, we have used the dataset from the CBVRP challenge[23].
It consists of two tracks of data: (1) TV shows and (2) Movies. Track
one consists of pre-extracted features from 7536 TV shows trailers
and has been divided into three sets - training (3000 shows), vali-
dation (864 shows), and testing (3000 shows). Track two consists
of pre-extracted features from 10826 movie trailers and has been
divided into the following sets - training (4500 movies), validation
(1188 movies), and testing (4500 movies). The final results were
evaluated on the test set by the organizers.
In both the given tracks, the pre-extracted feature vectors were
composed of frame-level and video-level features. For frame-level
feature vectors, frames of the video at 1 fps were passed into the In-
ceptionV3 network[1] trained on ImageNet dataset[13], and output
of the ReLU[25] activation of the last hidden layer of dimension
2048 was taken as the frame level feature. For video level features,
each video at 8 fps was passed into the C3D network[31] and the
activation of the pool5 layer of dimension 512 was taken as the
feature vector. Here we use these pre-extracted features from the
training dataset as the input to our models.
3.2 Random Forest Regression on Video Pairs
To determine the closeness of two given C3D feature vectors, we
consider several simplistic distance metrics - Euclidean[14], Cityblock[19],
Chebyshev[14], Correlation[8, 14], Square Euclidean, Braycurtis[4,
14] and Cosine Similarity[14, 30] .
For a particular video, its positive pairs are formed with the
videos that are in its ground truth. Negative pairs are formed by
choosing videos randomly from the videos not present in the list
of similar videos. To make the model not biased towards any type
of pair[14], we keep the number of positive pairs equal to the
number of negative pairs. For each such pair, we make a vector
X = [x1,x2,x3, ...x7] where each xi is one of the distance types
mentioned above calculated between the C3D vectors of each pair.
The ground truth value corresponding to each pair is taken as 0
for a positive pair and 1 for a negative pair. We then use Random
Forest [22] to regress and choose the best features.
3.3 Deep Learning based Regression
3.3.1 Architecture Overview: This model consists of two parallel
networks as shown in Figure 1. The first network consists of time
distributed dense layers followed by LSTM layers and the second
one contains only dense layers. The final layers concatenate the
outputs of the two models and then output the target values for all
the TV shows and Movies. The optimal number of layers in each
network was found out through experimentation. We used the first
network to process the frame-level features from the dataset. In the
second network, we used video level features from the dataset. As
the experiments panned out, creating a wider network rather than
a deeper one proved to be more beneficial [24, 33, 35].
The size of the last layer is equal to the number of videos to
be considered for that model. For a given video, it outputs the
probability of the similarity of the other video to the given video.
Subsequently, the top K predictions are taken for the evaluation of
results. In our model, we used the efficient ADAM[18] (lr=0.001)
optimization algorithm and ReLU[25] activation function[9]. The
same model was trained with losses as cosine proximity loss and
poisson loss. The results for both the experiments are presented in
Table 1.
3.4 Neural Network with DeepLDA
This architecture is composed of a deep neural network which
uses the output of a C3D model i.e. the video-level features, as
inputs. These are passed through a neural network containing two
hidden fully connected layers, followed by a final output layer. A
modified version of Linear Discriminant Analysis(LDA), is used
as an objective function. Figure 3 shows a high level view of this
model.
3.4.1 Model Architecture: In this model, the video level feature
set is used as the input to a set of three fully connected layers. The
number of fully connected layers was decided empirically. Every
fully connected layer involves, linear transformation of the input
features, followed by a non-linear activation function. The first
two layers, use the ReLU activation. The final layer serves as the
output layer. As the C3D embeddings are limited to 512 dimensional
feature vector, a smaller model with fewer parameters seem to give
better results than a deep model with large number of hidden layers.
3.4.2 Linear Discriminant Analysis Objective Function:
Classic Linear Discriminant Analysis: Classic LDA, is a dimen-
sionality reduction algorithm, which takes the higher dimensional
data, and projects it onto a lower dimensional space such that the
separation within data points of the same class is minimized and
separation between data points of different classes is maximized[16].
Let x1, ..., xn = X ∈ RN xb be a set of N data points which belong
to C different classes. LDA tries to find a transformation matrix
A ∈ Rlxb which projects the input to a lower dimensional space
L where, l = C-1. The linear combination of features, obtained by
multiplying the X with AT has maximum separation in this lower
dimensional subspace. The LDA formulation to find A is given as:
argmax
A
|ASb AT |
|ASw AT |
(1)
IceBreaker: Solving Cold Start Problem for Video Recommendation Engines Conference’17, July 2017, Washington, DC, USA
Figure 1: Deep Learning based Regression
Figure 2: Random Forest based regression on video pairs
where, Sb is the between class scatter matrix and Sw is the within
class scatter matrix. St denotes the overall scatter matrix. Xc =
Xc − mc are the mean-centered observations for each class, the
vector mc represents the mean of each class similarly X represents
mean centered observation of the complete population. The formula
for St , Sw and Sb is given as:
Sc =
1
Nc − 1
X
T
c Xc (2)
Sw =
1
C
Õ
c
Sc (3)
St =
1
N − 1
X
T
X (4)
The value of transformation matrix A that maximizes Eq. 1, max-
imizes the ratio of the ‘between class scatter’ and ‘within class
scatter’. Thereby, giving projections with low intraclass and high
inter-class variance. To find optimum solution for Eq. 1, we need
to solve the eigenvalue problem of Sb e = vSw e, where e and v
represent the eigenvector and their corresponding eigenvalues. The
matrix A is the set of eigenvectors e. In the paper [15], authors
have extrapolated this concept to introduce the DeepLDA objective
function.
In [15], the LDA function has been modified to be used as an
objective function for deep neural networks by maximizing the
individual eigenvalues. Each value gives the degree of separation
along a particular eigenvector i.e. higher eigenvalues correspond to
a greater amount of separation. Maximizing these eigenvalues will
help in obtaining an embedding in which classes are well separated.
However, if we consider all the eigenvalues in the loss function, it
can lead to a trivial solution in which only largest of the eigenvalue
is maximized. To circumvent this problem, we optimize only on
smallest m eigenvalues out of total set of eigenvalues. This can be
done by using only those m eigenvalues that are less than a certain
predetermined threshold. Therefore, the formulation becomes:
argmax
θ
1
k
k
Õ
i=1
vi with {v1, ..., vk } = {vj |vj < min vj |vj + ϵ} (5)
This loss is back-propagated through the network. In our model,
each row in the relevance list specifies a batch. All the movies/shows
that come in the relevance list of a particular show/movie are given
one class and the others which are not in list are assigned a separate
class. Then, we train on the loss using ADAM optimization[18].
4 EVALUATION
4.1 Evaluation Metric
For the challenge, we use two evaluations measures[23] -
(1) Recall@K - For top K predictions the metric recall@K is
defined as -
recall@K =
|or ∩ or |
or (6)
where or is relevance list and and or predicted top K relevant
shows/movies for each item.
(2) Hit@k - If recall@K > 0 for a test case then hit@K has value
1 or else value 0 for that test case.
The average values of recall and hit is taken on all test cases for
evaluation purposes [23].
Conference’17, July 2017, Washington, DC, USA Yaman Kumar et al.
Figure 3: DeepLDA based model’s architecture
Table 1: Track 1 and 2 - Results for the different architectures for TV shows and movies
Data Data Subset Method Hit @ k Recall @ k
k=5 k=10 k=20 k=30 k=50 k=100 k=200 k=300
TV
Shows
Validation Set
Random Forest 0.004 0.016 0.063 0.112 0.017 0.034 0.062 0.084
Regression(Cosine Loss) 0.319 0.388 0.485 0.541 0.171 0.229 0.289 0.323
Regression(Poisson Loss) 0.284 0.353 0.432 0.483 0.145 0.205 0.276 0.322
DeepLDA 0.178 0.234 0.310 0.354 0.065 0.103 0.153 0.196
Testing Set
Random Forest 0.047 0.093 0.178 0.262 0.031 0.056 0.093 0.122
Regression(Cosine Loss) 0.226 0.295 0.38 0.434 0.075 0.104 0.137 0.157
Regression(Poisson Loss) 0.217 0.294 0.381 0.426 0.076 0.106 0.139 0.16
DeepLDA 0.22 0.302 0.404 0.465 0.071 0.117 0.18 0.224
Movies
Validation Set
Random Forest 0.004 0.011 0.037 0.063 0.015 0.025 0.045 0.063
Regression(Cosine Loss) 0.18 0.231 0.313 0.362 0.099 0.139 0.192 0.226
Regression(Poisson Loss) 0.143 0.191 0.268 0.318 0.114 0.166 0.246 0.313
DeepLDA 0.106 0.143 0.192 0.238 0.065 0.091 0.129 0.158
Testing Set
Random Forest 0.055 0.109 0.172 0.216 0.035 0.051 0.077 0.097
Regression(Cosine Loss) 0.108 0.144 0.212 0.256 0.044 0.062 0.084 0.099
Regression(Poisson Loss) 0.107 0.151 0.212 0.25 0.044 0.061 0.082 0.096
DeepLDA 0.161 0.214 0.286 0.33 0.072 0.099 0.137 0.169
These evaluation metrics have been used in several other works
such as [12, 21, 27].
4.2 Results
Table 1 contains the results obtained from all the different archi-
tectural models considered for both TV shows as well as movies
obtained on validation set and training set after training the models.
For various values of K, the results are presented in the table.
The DeepLDA model clearly gives the best results on these met-
rics for both TV shows and movies on the testing set. The competi-
tion used recall@100 and hit@30 for the evaluation of results, thus
using that the analysis of results is presented below.
For TV Shows, considering the hit values on the test data, the
DeepLDA model performs better than the Random Forest model by
77.4%, Regression model with cosine loss by 7.1% and Regression
model with poisson loss by 9.1%. For the metric recall, the DeepLDA
model performs better than the Random Forest model by a full 108%,
Regression model with cosine loss by 12.5% and Regression model
with poisson loss by 10.3%.
For movies, considering the hit values on the test data, the
DeepLDA model performs better than the Random Forest model by
52.7%, Regression model with cosine loss by 28.9% and Regression
model with poisson loss by 32%. Taking into accoutn recall val-
ues, the DeepLDA model performs better than the Random Forest
model by 94.1%, Regression model with cosine loss by 59.6% and
Regression model with poisson loss by 62.2%.
5 CONCLUSIONS AND FUTURE WORK
The paper presented several models on which extensive experi-
ments were done to provide content based recommendations to
users. This solution provided a remarkable solution for the ‘cold-
start’ problem faced by video recommendation engines like Hulu.
The relevance between two videos is computed using different types
of losses, regression techniques and linear discriminant analysis.
In the future, the authors aim to train models using hybrid mod-
els generated after employing explicit as well as implicit features.
These models have the potential of improving the results for video
recommendation significantly. Additionally, they also look forward
to extend the current models for addressing the pernicious cold
start problem holistically by taking into account not just cold videos
but cold-start user profiles too. These approaches could also help
in ameliorating the accuracy of the models considered and make it
more robust.
REFERENCES
[1] Sami Abu-El-Haija, Nisarg Kothari, Joonseok Lee, Apostol Natsev, George
Toderici, Balakrishnan Varadarajan, and Sudheendra Vijayanarasimhan.
2016. YouTube-8M: A Large-Scale Video Classification Benchmark. CoRR
abs/1609.08675 (2016).
[2] Arslan Basharat, Yun Zhai, and Mubarak Shah. 2008. Content based video match-
ing using spatiotemporal volumes. Computer Vision and Image Understanding
110, 3 (2008), 360ś377.
[3] Chumki Basu, Haym Hirsh, William Cohen, et al. 1998. Recommendation as
classification: Using social and content-based information in recommendation.
In AAAI/IAAI. 714ś720.
[4] J Roger Bray and John T Curtis. 1957. An ordination of the upland forest com-
munities of southern Wisconsin. Ecological monographs 27, 4 (1957), 325ś349.
IceBreaker: Solving Cold Start Problem for Video Recommendation Engines Conference’17, July 2017, Washington, DC, USA
[5] Jeff Bullas. 2017. 30 Mind Numbing YouTube Facts,Figures
and Statistics - Infographic. https://www.jeffbullas.com/
35-mind-numbing-youtube-facts-figures-and-statistics-infographic/ Re-
trieved July 7, 2018.
[6] Luca Canini, Sergio Benini, and Riccardo Leonardi. 2013. Affective recommen-
dation of movies based on selected connotative features. IEEE Transactions on
Circuits and Systems for Video Technology 23, 4 (2013), 636ś647.
[7] Meeyoung Cha, Haewoon Kwak, Pablo Rodriguez, Yong-Yeol Ahn, and Sue Moon.
2007. I tube, you tube, everybody tubes: analyzing the world’s largest user gen-
erated content video system. In Proceedings of the 7th ACM SIGCOMM conference
on Internet measurement. ACM, 1ś14.
[8] Sung-Hyuk Cha. 2007. Comprehensive survey on distance/similarity measures
between probability density functions. City 1, 2 (2007), 1.
[9] François Chollet et al. 2015. Keras. https://keras.io.
[10] James Davidson, Benjamin Liebald, Junning Liu, Palash Nandy, Taylor Van Vleet,
Ullas Gargi, Sujoy Gupta, Yu He, Mike Lambert, Blake Livingston, et al. 2010.
The YouTube video recommendation system. In Proceedings of the fourth ACM
conference on Recommender systems. ACM, 293ś296.
[11] Yashar Deldjoo, Mehdi Elahi, Paolo Cremonesi, Franca Garzotto, Pietro Piazzolla,
and Massimo Quadrana. 2016. Content-based video recommendation system
based on stylistic visual features. Journal on Data Semantics 5, 2 (2016), 99ś113.
[12] Yashar Deldjoo, Mehdi Elahi, Massimo Quadrana, and Paolo Cremonesi. 2015. To-
ward building a content-based video recommendation system based on low-level
features. In International Conference on Electronic Commerce and Web Technologies.
Springer, 45ś56.
[13] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009. Imagenet:
A large-scale hierarchical image database. In IEEE Conference on Computer Vision
and Pattern Recognition (CVPR) 2009. Ieee, 248ś255.
[14] Michel-Marie Deza and Elena Deza. 2006. Dictionary of distances. Elsevier.
[15] Matthias Dorfer, Rainer Kelz, and Gerhard Widmer. 2015. Deep Linear Dis-
criminant Analysis. CoRR abs/1511.04707 (2015). arXiv:1511.04707 http:
//arxiv.org/abs/1511.04707
[16] R. A. Fisher. 1936. The Use of Multiple Measurements in Taxonomic Problems.
Annals of Eugenics 7, 7 (1936), 179ś188.
[17] David Goldberg, David Nichols, Brian M Oki, and Douglas Terry. 1992. Using
collaborative filtering to weave an information tapestry. Commun. ACM 35, 12
(1992), 61ś70.
[18] Diederik P. Kingma and Jimmy Ba. 2014. Adam: A Method for Stochastic Optimiza-
tion. CoRR abs/1412.6980 (2014). arXiv:1412.6980 http://arxiv.org/abs/1412.6980
[19] Eugene F Krause. 1975. Taxicab geometry: An adventure in non-Euclidean geometry.
Courier Corporation.
[20] Taras Lehinevych, Nikolaos Kokkinis-Ntrenis, Giorgos Siantikos, A Seza Dogruoz,
Theodoros Giannakopoulos, and Stasinos Konstantopoulos. 2014. Discovering
similarities for content-based recommendation and browsing in multimedia
collections. In 2014 Tenth International Conference on Signal-Image Technology &
Internet-Based Systems (SITIS). IEEE, 237ś243.
[21] Yan Li, Hanjie Wang, Hailong Liu, and Bo Chen. 2017. A study on content-based
video recommendation. In 2017 IEEE International Conference on Image Processing
(ICIP). IEEE, 4581ś4585.
[22] Andy Liaw, Matthew Wiener, et al. 2002. Classification and regression by ran-
domForest. R news 2, 3 (2002), 18ś22.
[23] Mengyi Liu, Xiaohui Xie, and Hanning Zhou. 2018. Content-based Video Rele-
vance Prediction Challenge: Data, Protocol, and Baseline.
[24] Mayank Meghawat, Satyendra Yadav, Debanjan Mahata, Yifang Yin, Rajiv Ratn
Shah, and Roger Zimmermann. 2018. A Multimodal Approach to Predict Social
Media Popularity. In 2018 IEEE Conference on Multimedia Information Processing
and Retrieval (MIPR). IEEE.
[25] Vinod Nair and Geoffrey E. Hinton. 2010. Rectified Linear Units Improve Re-
stricted Boltzmann Machines. In Proceedings of the 27th International Conference
on International Conference on Machine Learning (ICML’10). Omnipress, USA,
807ś814. http://dl.acm.org/citation.cfm?id=3104322.3104425
[26] Neil Rubens, Mehdi Elahi, Masashi Sugiyama, and Dain Kaplan. 2015. Active
learning in recommender systems. In Recommender systems handbook. Springer,
809ś846.
[27] Suvash Sedhain, Scott Sanner, Darius Braziunas, Lexing Xie, and Jordan Chris-
tensen. 2014. Social collaborative filtering for cold-start recommendations. In
Proceedings of the 8th ACM Conference on Recommender systems. ACM, 345ś348.
[28] Rajiv Shah and Roger Zimmermann. 2017. Multimodal analysis of user-generated
multimedia content. Springer. 1ś263 pages.
[29] Rajiv Ratn Shah, Yi Yu, and Roger Zimmermann. 2014. Advisor: Personalized
video soundtrack recommendation by late fusion with heuristic rankings. In
Proceedings of the 22nd ACM international conference on Multimedia. ACM, 607ś
616.
[30] Amit Singhal et al. 2001. Modern information retrieval: A brief overview. IEEE
Data Eng. Bull. 24, 4 (2001), 35ś43.
[31] Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, and Manohar Paluri.
2015. Learning spatiotemporal features with 3d convolutional networks. In
Proceedings of the IEEE international conference on computer vision. 4489ś4497.
[32] Jian Wei, Jianhua He, Kai Chen, Yi Zhou, and Zuoyin Tang. 2017. Collaborative
filtering and deep learning based recommendation system for cold start items.
Expert Systems with Applications 69 (2017), 29ś39.
[33] Zifeng Wu, Chunhua Shen, and Anton van den Hengel. 2016. Wider or deeper:
Revisiting the resnet model for visual recognition. arXiv preprint arXiv:1611.10080
(2016).
[34] Bo Yang, Tao Mei, Xian-Sheng Hua, Linjun Yang, Shi-Qiang Yang, and Mingjing
Li. 2007. Online video recommendation based on multimodal fusion and relevance
feedback. In Proceedings of the 6th ACM international conference on Image and
video retrieval. ACM, 73ś80.
[35] Sergey Zagoruyko and Nikos Komodakis. 2016. Wide residual networks. arXiv
preprint arXiv:1605.07146 (2016).
[36] Herbert Zettl. 2002. Essentials of applied media aesthetics. In Media Computing.
Springer, 11ś38.
[37] Xiaojian Zhao, Guangda Li, Meng Wang, Jin Yuan, Zheng-Jun Zha, Zhoujun Li,
and Tat-Seng Chua. 2011. Integrating rich information for video recommendation
with multi-task rank aggregation. In Proceedings of the 19th ACM international
conference on Multimedia. ACM, 1521ś1524.
[38] Qiusha Zhu, Mei-Ling Shyu, and Haohong Wang. 2013. Videotopic: Content-
based video recommendation using a topic model. In 2013 IEEE International
Symposium on Multimedia (ISM). IEEE, 219ś222.

More Related Content

Similar to IceBreaker Solving Cold Start Problem For Video Recommendation Engines

IMAGE CAPTION GENERATOR USING DEEP LEARNING
IMAGE CAPTION GENERATOR USING DEEP LEARNINGIMAGE CAPTION GENERATOR USING DEEP LEARNING
IMAGE CAPTION GENERATOR USING DEEP LEARNINGIRJET Journal
 
Mtech Second progresspresentation ON VIDEO SUMMARIZATION
Mtech Second progresspresentation ON VIDEO SUMMARIZATIONMtech Second progresspresentation ON VIDEO SUMMARIZATION
Mtech Second progresspresentation ON VIDEO SUMMARIZATIONNEERAJ BAGHEL
 
SVM Based Saliency Map Technique for Reducing Time Complexity in HEVC
SVM Based Saliency Map Technique for Reducing Time Complexity in HEVCSVM Based Saliency Map Technique for Reducing Time Complexity in HEVC
SVM Based Saliency Map Technique for Reducing Time Complexity in HEVCIRJET Journal
 
Perceptually Lossless Compression with Error Concealment for Periscope and So...
Perceptually Lossless Compression with Error Concealment for Periscope and So...Perceptually Lossless Compression with Error Concealment for Periscope and So...
Perceptually Lossless Compression with Error Concealment for Periscope and So...sipij
 
Effective Compression of Digital Video
Effective Compression of Digital VideoEffective Compression of Digital Video
Effective Compression of Digital VideoIRJET Journal
 
Recent advances in content based video copy detection (IEEE)
Recent advances in content based video copy detection (IEEE)Recent advances in content based video copy detection (IEEE)
Recent advances in content based video copy detection (IEEE)PACE 2.0
 
TVSum: Summarizing Web Videos Using Titles
TVSum: Summarizing Web Videos Using TitlesTVSum: Summarizing Web Videos Using Titles
TVSum: Summarizing Web Videos Using TitlesNEERAJ BAGHEL
 
Deep neural networks for Youtube recommendations
Deep neural networks for Youtube recommendationsDeep neural networks for Youtube recommendations
Deep neural networks for Youtube recommendationsAryan Khandal
 
Full reference video quality assessment
Full reference video quality assessmentFull reference video quality assessment
Full reference video quality assessmentHoàng Sơn
 
A Benchmark to Evaluate Mobile Video Upload to Cloud Infrastructures
A Benchmark to Evaluate Mobile Video Upload to Cloud InfrastructuresA Benchmark to Evaluate Mobile Video Upload to Cloud Infrastructures
A Benchmark to Evaluate Mobile Video Upload to Cloud InfrastructuresUniversity of Southern California
 
Resolution Independent 2D Cartoon Video Conversion
Resolution Independent 2D Cartoon Video ConversionResolution Independent 2D Cartoon Video Conversion
Resolution Independent 2D Cartoon Video ConversionEswar Publications
 
System analysis and design for multimedia retrieval systems
System analysis and design for multimedia retrieval systemsSystem analysis and design for multimedia retrieval systems
System analysis and design for multimedia retrieval systemsijma
 
Automatic Visual Concept Detection in Videos: Review
Automatic Visual Concept Detection in Videos: ReviewAutomatic Visual Concept Detection in Videos: Review
Automatic Visual Concept Detection in Videos: ReviewIRJET Journal
 
IRJET- Crowd Density Estimation using Novel Feature Descriptor
IRJET- Crowd Density Estimation using Novel Feature DescriptorIRJET- Crowd Density Estimation using Novel Feature Descriptor
IRJET- Crowd Density Estimation using Novel Feature DescriptorIRJET Journal
 

Similar to IceBreaker Solving Cold Start Problem For Video Recommendation Engines (20)

IMAGE CAPTION GENERATOR USING DEEP LEARNING
IMAGE CAPTION GENERATOR USING DEEP LEARNINGIMAGE CAPTION GENERATOR USING DEEP LEARNING
IMAGE CAPTION GENERATOR USING DEEP LEARNING
 
Mtech Second progresspresentation ON VIDEO SUMMARIZATION
Mtech Second progresspresentation ON VIDEO SUMMARIZATIONMtech Second progresspresentation ON VIDEO SUMMARIZATION
Mtech Second progresspresentation ON VIDEO SUMMARIZATION
 
SVM Based Saliency Map Technique for Reducing Time Complexity in HEVC
SVM Based Saliency Map Technique for Reducing Time Complexity in HEVCSVM Based Saliency Map Technique for Reducing Time Complexity in HEVC
SVM Based Saliency Map Technique for Reducing Time Complexity in HEVC
 
Perceptually Lossless Compression with Error Concealment for Periscope and So...
Perceptually Lossless Compression with Error Concealment for Periscope and So...Perceptually Lossless Compression with Error Concealment for Periscope and So...
Perceptually Lossless Compression with Error Concealment for Periscope and So...
 
Effective Compression of Digital Video
Effective Compression of Digital VideoEffective Compression of Digital Video
Effective Compression of Digital Video
 
Recent advances in content based video copy detection (IEEE)
Recent advances in content based video copy detection (IEEE)Recent advances in content based video copy detection (IEEE)
Recent advances in content based video copy detection (IEEE)
 
TVSum: Summarizing Web Videos Using Titles
TVSum: Summarizing Web Videos Using TitlesTVSum: Summarizing Web Videos Using Titles
TVSum: Summarizing Web Videos Using Titles
 
Video to Video Translation CGAN
Video to Video Translation CGANVideo to Video Translation CGAN
Video to Video Translation CGAN
 
Deep neural networks for Youtube recommendations
Deep neural networks for Youtube recommendationsDeep neural networks for Youtube recommendations
Deep neural networks for Youtube recommendations
 
Full reference video quality assessment
Full reference video quality assessmentFull reference video quality assessment
Full reference video quality assessment
 
C04841417
C04841417C04841417
C04841417
 
gist.doc
gist.docgist.doc
gist.doc
 
A Benchmark to Evaluate Mobile Video Upload to Cloud Infrastructures
A Benchmark to Evaluate Mobile Video Upload to Cloud InfrastructuresA Benchmark to Evaluate Mobile Video Upload to Cloud Infrastructures
A Benchmark to Evaluate Mobile Video Upload to Cloud Infrastructures
 
Resolution Independent 2D Cartoon Video Conversion
Resolution Independent 2D Cartoon Video ConversionResolution Independent 2D Cartoon Video Conversion
Resolution Independent 2D Cartoon Video Conversion
 
System analysis and design for multimedia retrieval systems
System analysis and design for multimedia retrieval systemsSystem analysis and design for multimedia retrieval systems
System analysis and design for multimedia retrieval systems
 
Automatic Visual Concept Detection in Videos: Review
Automatic Visual Concept Detection in Videos: ReviewAutomatic Visual Concept Detection in Videos: Review
Automatic Visual Concept Detection in Videos: Review
 
IRJET- Crowd Density Estimation using Novel Feature Descriptor
IRJET- Crowd Density Estimation using Novel Feature DescriptorIRJET- Crowd Density Estimation using Novel Feature Descriptor
IRJET- Crowd Density Estimation using Novel Feature Descriptor
 
journal in research
journal in research journal in research
journal in research
 
journal published
journal publishedjournal published
journal published
 
A0540106
A0540106A0540106
A0540106
 

More from Jamie Boyd

Free Printable Christmas Writing Paper - Printable
Free Printable Christmas Writing Paper - PrintableFree Printable Christmas Writing Paper - Printable
Free Printable Christmas Writing Paper - PrintableJamie Boyd
 
How Research Paper Writing Service Can Help You Ne
How Research Paper Writing Service Can Help You NeHow Research Paper Writing Service Can Help You Ne
How Research Paper Writing Service Can Help You NeJamie Boyd
 
KS3 Year 8 Spanish Example Paragraphs For Food, Leisur
KS3 Year 8 Spanish Example Paragraphs For Food, LeisurKS3 Year 8 Spanish Example Paragraphs For Food, Leisur
KS3 Year 8 Spanish Example Paragraphs For Food, LeisurJamie Boyd
 
Basics Essay Writing - Bibliographysetup.X.Fc2.Com
Basics Essay Writing - Bibliographysetup.X.Fc2.ComBasics Essay Writing - Bibliographysetup.X.Fc2.Com
Basics Essay Writing - Bibliographysetup.X.Fc2.ComJamie Boyd
 
Self Defense Essay Outline. Presentation Of Self Outline Essay. 2022
Self Defense Essay Outline. Presentation Of Self Outline Essay. 2022Self Defense Essay Outline. Presentation Of Self Outline Essay. 2022
Self Defense Essay Outline. Presentation Of Self Outline Essay. 2022Jamie Boyd
 
Shocking 2015 College Essay Prompts Thatsnotus
Shocking 2015 College Essay Prompts ThatsnotusShocking 2015 College Essay Prompts Thatsnotus
Shocking 2015 College Essay Prompts ThatsnotusJamie Boyd
 
Henderson Vanessar
Henderson VanessarHenderson Vanessar
Henderson VanessarJamie Boyd
 
Uva Admissions Essay
Uva Admissions EssayUva Admissions Essay
Uva Admissions EssayJamie Boyd
 
A Level Economics Essay Writing Guide - Top Tips, U
A Level Economics Essay Writing Guide - Top Tips, UA Level Economics Essay Writing Guide - Top Tips, U
A Level Economics Essay Writing Guide - Top Tips, UJamie Boyd
 
Custom Essay Writing - Master The Art
Custom Essay Writing - Master The ArtCustom Essay Writing - Master The Art
Custom Essay Writing - Master The ArtJamie Boyd
 
Free Gingerbread House Writing Page Free4Class
Free Gingerbread House Writing Page Free4ClassFree Gingerbread House Writing Page Free4Class
Free Gingerbread House Writing Page Free4ClassJamie Boyd
 
Stirring Sample Essay About Yourself For College Tha
Stirring Sample Essay About Yourself For College ThaStirring Sample Essay About Yourself For College Tha
Stirring Sample Essay About Yourself For College ThaJamie Boyd
 
Easy Guide On How To Format A Research Paper CustomEssayMeister.Com
Easy Guide On How To Format A Research Paper CustomEssayMeister.ComEasy Guide On How To Format A Research Paper CustomEssayMeister.Com
Easy Guide On How To Format A Research Paper CustomEssayMeister.ComJamie Boyd
 
Where Do Writers Get Paid For Essays In Blogs And Magazines
Where Do Writers Get Paid For Essays In Blogs And MagazinesWhere Do Writers Get Paid For Essays In Blogs And Magazines
Where Do Writers Get Paid For Essays In Blogs And MagazinesJamie Boyd
 
10 Columbia Supplemental Essay Examples That Worke
10 Columbia Supplemental Essay Examples That Worke10 Columbia Supplemental Essay Examples That Worke
10 Columbia Supplemental Essay Examples That WorkeJamie Boyd
 
Integrating Culture And Diversity In Decision Makin
Integrating Culture And Diversity In Decision MakinIntegrating Culture And Diversity In Decision Makin
Integrating Culture And Diversity In Decision MakinJamie Boyd
 
How To Write A Winning Scholarship Essay In 10 Steps
How To Write A Winning Scholarship Essay In 10 StepsHow To Write A Winning Scholarship Essay In 10 Steps
How To Write A Winning Scholarship Essay In 10 StepsJamie Boyd
 
Persuasive Paragraph Examples Short - Bmp-Get
Persuasive Paragraph Examples Short - Bmp-GetPersuasive Paragraph Examples Short - Bmp-Get
Persuasive Paragraph Examples Short - Bmp-GetJamie Boyd
 
Career Goals 250 Word Essay Scholarships
Career Goals 250 Word Essay ScholarshipsCareer Goals 250 Word Essay Scholarships
Career Goals 250 Word Essay ScholarshipsJamie Boyd
 
INTRODUCTION TO PUBLIC POLICY.Pdf Plag
INTRODUCTION TO PUBLIC POLICY.Pdf  PlagINTRODUCTION TO PUBLIC POLICY.Pdf  Plag
INTRODUCTION TO PUBLIC POLICY.Pdf PlagJamie Boyd
 

More from Jamie Boyd (20)

Free Printable Christmas Writing Paper - Printable
Free Printable Christmas Writing Paper - PrintableFree Printable Christmas Writing Paper - Printable
Free Printable Christmas Writing Paper - Printable
 
How Research Paper Writing Service Can Help You Ne
How Research Paper Writing Service Can Help You NeHow Research Paper Writing Service Can Help You Ne
How Research Paper Writing Service Can Help You Ne
 
KS3 Year 8 Spanish Example Paragraphs For Food, Leisur
KS3 Year 8 Spanish Example Paragraphs For Food, LeisurKS3 Year 8 Spanish Example Paragraphs For Food, Leisur
KS3 Year 8 Spanish Example Paragraphs For Food, Leisur
 
Basics Essay Writing - Bibliographysetup.X.Fc2.Com
Basics Essay Writing - Bibliographysetup.X.Fc2.ComBasics Essay Writing - Bibliographysetup.X.Fc2.Com
Basics Essay Writing - Bibliographysetup.X.Fc2.Com
 
Self Defense Essay Outline. Presentation Of Self Outline Essay. 2022
Self Defense Essay Outline. Presentation Of Self Outline Essay. 2022Self Defense Essay Outline. Presentation Of Self Outline Essay. 2022
Self Defense Essay Outline. Presentation Of Self Outline Essay. 2022
 
Shocking 2015 College Essay Prompts Thatsnotus
Shocking 2015 College Essay Prompts ThatsnotusShocking 2015 College Essay Prompts Thatsnotus
Shocking 2015 College Essay Prompts Thatsnotus
 
Henderson Vanessar
Henderson VanessarHenderson Vanessar
Henderson Vanessar
 
Uva Admissions Essay
Uva Admissions EssayUva Admissions Essay
Uva Admissions Essay
 
A Level Economics Essay Writing Guide - Top Tips, U
A Level Economics Essay Writing Guide - Top Tips, UA Level Economics Essay Writing Guide - Top Tips, U
A Level Economics Essay Writing Guide - Top Tips, U
 
Custom Essay Writing - Master The Art
Custom Essay Writing - Master The ArtCustom Essay Writing - Master The Art
Custom Essay Writing - Master The Art
 
Free Gingerbread House Writing Page Free4Class
Free Gingerbread House Writing Page Free4ClassFree Gingerbread House Writing Page Free4Class
Free Gingerbread House Writing Page Free4Class
 
Stirring Sample Essay About Yourself For College Tha
Stirring Sample Essay About Yourself For College ThaStirring Sample Essay About Yourself For College Tha
Stirring Sample Essay About Yourself For College Tha
 
Easy Guide On How To Format A Research Paper CustomEssayMeister.Com
Easy Guide On How To Format A Research Paper CustomEssayMeister.ComEasy Guide On How To Format A Research Paper CustomEssayMeister.Com
Easy Guide On How To Format A Research Paper CustomEssayMeister.Com
 
Where Do Writers Get Paid For Essays In Blogs And Magazines
Where Do Writers Get Paid For Essays In Blogs And MagazinesWhere Do Writers Get Paid For Essays In Blogs And Magazines
Where Do Writers Get Paid For Essays In Blogs And Magazines
 
10 Columbia Supplemental Essay Examples That Worke
10 Columbia Supplemental Essay Examples That Worke10 Columbia Supplemental Essay Examples That Worke
10 Columbia Supplemental Essay Examples That Worke
 
Integrating Culture And Diversity In Decision Makin
Integrating Culture And Diversity In Decision MakinIntegrating Culture And Diversity In Decision Makin
Integrating Culture And Diversity In Decision Makin
 
How To Write A Winning Scholarship Essay In 10 Steps
How To Write A Winning Scholarship Essay In 10 StepsHow To Write A Winning Scholarship Essay In 10 Steps
How To Write A Winning Scholarship Essay In 10 Steps
 
Persuasive Paragraph Examples Short - Bmp-Get
Persuasive Paragraph Examples Short - Bmp-GetPersuasive Paragraph Examples Short - Bmp-Get
Persuasive Paragraph Examples Short - Bmp-Get
 
Career Goals 250 Word Essay Scholarships
Career Goals 250 Word Essay ScholarshipsCareer Goals 250 Word Essay Scholarships
Career Goals 250 Word Essay Scholarships
 
INTRODUCTION TO PUBLIC POLICY.Pdf Plag
INTRODUCTION TO PUBLIC POLICY.Pdf  PlagINTRODUCTION TO PUBLIC POLICY.Pdf  Plag
INTRODUCTION TO PUBLIC POLICY.Pdf Plag
 

Recently uploaded

How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17Celine George
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxheathfieldcps1
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfsanyamsingh5019
 
Class 11 Legal Studies Ch-1 Concept of State .pdf
Class 11 Legal Studies Ch-1 Concept of State .pdfClass 11 Legal Studies Ch-1 Concept of State .pdf
Class 11 Legal Studies Ch-1 Concept of State .pdfakmcokerachita
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptxVS Mahajan Coaching Centre
 
URLs and Routing in the Odoo 17 Website App
URLs and Routing in the Odoo 17 Website AppURLs and Routing in the Odoo 17 Website App
URLs and Routing in the Odoo 17 Website AppCeline George
 
MENTAL STATUS EXAMINATION format.docx
MENTAL     STATUS EXAMINATION format.docxMENTAL     STATUS EXAMINATION format.docx
MENTAL STATUS EXAMINATION format.docxPoojaSen20
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxiammrhaywood
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsanshu789521
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introductionMaksud Ahmed
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxpboyjonauth
 
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...M56BOOKSTORE PRODUCT/SERVICE
 
Solving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxSolving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxOH TEIK BIN
 
Separation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesSeparation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesFatimaKhan178732
 
How to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxHow to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxmanuelaromero2013
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)eniolaolutunde
 
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfEnzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfSumit Tiwari
 
mini mental status format.docx
mini    mental       status     format.docxmini    mental       status     format.docx
mini mental status format.docxPoojaSen20
 
Concept of Vouching. B.Com(Hons) /B.Compdf
Concept of Vouching. B.Com(Hons) /B.CompdfConcept of Vouching. B.Com(Hons) /B.Compdf
Concept of Vouching. B.Com(Hons) /B.CompdfUmakantAnnand
 

Recently uploaded (20)

How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdf
 
Class 11 Legal Studies Ch-1 Concept of State .pdf
Class 11 Legal Studies Ch-1 Concept of State .pdfClass 11 Legal Studies Ch-1 Concept of State .pdf
Class 11 Legal Studies Ch-1 Concept of State .pdf
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
 
URLs and Routing in the Odoo 17 Website App
URLs and Routing in the Odoo 17 Website AppURLs and Routing in the Odoo 17 Website App
URLs and Routing in the Odoo 17 Website App
 
Staff of Color (SOC) Retention Efforts DDSD
Staff of Color (SOC) Retention Efforts DDSDStaff of Color (SOC) Retention Efforts DDSD
Staff of Color (SOC) Retention Efforts DDSD
 
MENTAL STATUS EXAMINATION format.docx
MENTAL     STATUS EXAMINATION format.docxMENTAL     STATUS EXAMINATION format.docx
MENTAL STATUS EXAMINATION format.docx
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha elections
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptx
 
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
 
Solving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxSolving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptx
 
Separation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesSeparation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and Actinides
 
How to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxHow to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptx
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)
 
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfEnzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
 
mini mental status format.docx
mini    mental       status     format.docxmini    mental       status     format.docx
mini mental status format.docx
 
Concept of Vouching. B.Com(Hons) /B.Compdf
Concept of Vouching. B.Com(Hons) /B.CompdfConcept of Vouching. B.Com(Hons) /B.Compdf
Concept of Vouching. B.Com(Hons) /B.Compdf
 

IceBreaker Solving Cold Start Problem For Video Recommendation Engines

  • 1. IceBreaker: Solving Cold Start Problem for Video Recommendation Engines Agniv Sharma∗ DTU New Delhi agnivsharma96@gmail.com Yaman Kumar∗ Adobe Systems Noida ykumar@adobe.com Abhigyan Khaund∗ IIT Mandi Mandi khaundabhigyan@gmail.com Akash Kumar∗ DTU New Delhi akash_bt2k15@dtu.ac.in Ponnurangam Kumaraguru IIIT-Delhi New Delhi pk@iitd.ac.in Rajiv Ratn Shah IIIT-Delhi New Delhi rajivratn@iiitd.ac.in ABSTRACT Internet has brought about a tremendous increase in content of all forms and, in that, video content constitutes the major backbone of the total content being published as well as watched. Thus it becomes imperative for video recommendation engines such as Hulu to look for novel and innovative ways to recommend the newly added videos to their users. However, the problem with new videos is that they lack any sort of metadata and user interaction so as to be able to rate the videos for the consumers. To this effect, this paper introduces the several techniques we develop for the Content Based Video Relevance Prediction (CBVRP) Challenge being hosted by Hulu for the ACM Multimedia Conference 2018. We employ different architectures on the CBVRP dataset to make use of the provided frame and video level features and generate predictions of videos that are similar to the other videos. We also implement several ensemble strategies to explore complementarity between both the types of provided features. The obtained results are en- couraging and will impel the boundaries of research for multimedia based video recommendation systems. KEYWORDS Video Recommendation System, Implicit Features, Video Relevance Prediction, Content based Recommendation, LDA ACM Reference Format: Agniv Sharma, Yaman Kumar, Abhigyan Khaund, Akash Kumar, Ponnu- rangam Kumaraguru, and Rajiv Ratn Shah. 2018. IceBreaker: Solving Cold Start Problem for Video Recommendation Engines . In Proceedings of ACM Conference (Conference’17). ACM, New York, NY, USA, 5 pages. https://doi. org/10.1145/nnnnnnn.nnnnnnn 1 INTRODUCTION The recent years have seen a huge surge in online content, and, con- tent in the form of videos has led the way. This stupendous wave has been steered by online video recommendation systems like Hulu, Youtube, Netflix, etc. At the same time, the content supplied by traditional media houses and other licensed broadcasters has seen a plunge [7] with respect to amateurs and non-professionals ∗Equal Contribution Conference’17, July 2017, Washington, DC, USA 2018. ACM ISBN 978-x-xxxx-xxxx-x/YY/MM...$15.00 https://doi.org/10.1145/nnnnnnn.nnnnnnn uploading home-made content on various platforms. The popu- larity of video streaming can be judged from the fact that sixty hours of video content is generated every minute on Youtube [5]. A tremendous amount of new content is being generated at a very fast pace and, on top of that, the content metadata itself ranges from incorrect to either none or highly limited information [10, 11]. Additionally, the content thus generated is short (lesser than 10 minutes) and un-reviewed. The user interaction with the horde of these videos is minimal at the very best [10]. Due to these reason, video recommendation systems suffer from a multi-dimensional challenge where they have to deal with un-audited, un-rated and completely new content of which they know nothing about. This leads to the infamous cold start problem where recommendation en- gines have to tackle new content and then process and recommend it to the users in such a manner so as to keep grabbing the eyeballs and dissuade them from shifting to the competitor’s platform. Video recommendation has traditionally been inspired from two types of data: implicit and explicit. While on one hand, video rec- ommendation on the basis of implicit content deals with features like light, mood, color, shape, motion, plot and other aesthetics; on the other hand, explicit content incorporates movie genre, di- rector, actors, description, previous views and likes. Researchers have spent considerable time and effort building recommendation engines based on both the types of data [2, 3, 11, 23, 38]. But as noted in the previous section, for today’s content explosion, the recommendation on the basis of explicit content fails utterly. This leaves one with implicit features. According to Applied Media Aes- thetic [36], media features such as light, color, camera motion etc. serve as crucial elements which render emotional, aesthetic and informative effects. Thus a new artistic work, can be appropriately evaluated and recommended on the basis of intrinsic features. In this paper, we present our extensive experiments in building a video recommendation system which deals with the cold-start problem and uses intrinsic features to recommend videos. Here, we portray our experiments and their results over multiple architec- tures for Content Based Video Relevance Prediction Challenge of ACM Multimedia 2018, sponsored by Hulu. 2 RELATED WORK Building recommendation engines in general, and video recom- mendation engines in particular has always been a hot topic of arXiv:1808.05636v1 [cs.IR] 16 Aug 2018
  • 2. Conference’17, July 2017, Washington, DC, USA Yaman Kumar et al. research. Right from the development of the first recommender sys- tem, Tapestry[17], recommender systems have undergone drastic developmental changes. There have been several works which have employed the low level features, also known as stylistic features for the task of video recommendation[6, 20, 34, 37]. But none of these approaches use only low-level features, thus they cannot solve the cold-start problem in its full glory[11, 26, 28]. However, a few authors have realized this problem and have tried to tackle this problem head-on using various methods [11, 29, 32]. Thus one of the novelties introduced in this work is exploring the recommenda- tion problem using only stylistic and artistic features for solving the daunting cold-start dilemma. 3 METHODOLOGY In this section, we present the dataset and different architectures i.e., random forest regression on video pairs, deep learning based regression and neural network with DeepLDA for the CBCRP chal- lenge. 3.1 Dataset In this paper, we have used the dataset from the CBVRP challenge[23]. It consists of two tracks of data: (1) TV shows and (2) Movies. Track one consists of pre-extracted features from 7536 TV shows trailers and has been divided into three sets - training (3000 shows), vali- dation (864 shows), and testing (3000 shows). Track two consists of pre-extracted features from 10826 movie trailers and has been divided into the following sets - training (4500 movies), validation (1188 movies), and testing (4500 movies). The final results were evaluated on the test set by the organizers. In both the given tracks, the pre-extracted feature vectors were composed of frame-level and video-level features. For frame-level feature vectors, frames of the video at 1 fps were passed into the In- ceptionV3 network[1] trained on ImageNet dataset[13], and output of the ReLU[25] activation of the last hidden layer of dimension 2048 was taken as the frame level feature. For video level features, each video at 8 fps was passed into the C3D network[31] and the activation of the pool5 layer of dimension 512 was taken as the feature vector. Here we use these pre-extracted features from the training dataset as the input to our models. 3.2 Random Forest Regression on Video Pairs To determine the closeness of two given C3D feature vectors, we consider several simplistic distance metrics - Euclidean[14], Cityblock[19], Chebyshev[14], Correlation[8, 14], Square Euclidean, Braycurtis[4, 14] and Cosine Similarity[14, 30] . For a particular video, its positive pairs are formed with the videos that are in its ground truth. Negative pairs are formed by choosing videos randomly from the videos not present in the list of similar videos. To make the model not biased towards any type of pair[14], we keep the number of positive pairs equal to the number of negative pairs. For each such pair, we make a vector X = [x1,x2,x3, ...x7] where each xi is one of the distance types mentioned above calculated between the C3D vectors of each pair. The ground truth value corresponding to each pair is taken as 0 for a positive pair and 1 for a negative pair. We then use Random Forest [22] to regress and choose the best features. 3.3 Deep Learning based Regression 3.3.1 Architecture Overview: This model consists of two parallel networks as shown in Figure 1. The first network consists of time distributed dense layers followed by LSTM layers and the second one contains only dense layers. The final layers concatenate the outputs of the two models and then output the target values for all the TV shows and Movies. The optimal number of layers in each network was found out through experimentation. We used the first network to process the frame-level features from the dataset. In the second network, we used video level features from the dataset. As the experiments panned out, creating a wider network rather than a deeper one proved to be more beneficial [24, 33, 35]. The size of the last layer is equal to the number of videos to be considered for that model. For a given video, it outputs the probability of the similarity of the other video to the given video. Subsequently, the top K predictions are taken for the evaluation of results. In our model, we used the efficient ADAM[18] (lr=0.001) optimization algorithm and ReLU[25] activation function[9]. The same model was trained with losses as cosine proximity loss and poisson loss. The results for both the experiments are presented in Table 1. 3.4 Neural Network with DeepLDA This architecture is composed of a deep neural network which uses the output of a C3D model i.e. the video-level features, as inputs. These are passed through a neural network containing two hidden fully connected layers, followed by a final output layer. A modified version of Linear Discriminant Analysis(LDA), is used as an objective function. Figure 3 shows a high level view of this model. 3.4.1 Model Architecture: In this model, the video level feature set is used as the input to a set of three fully connected layers. The number of fully connected layers was decided empirically. Every fully connected layer involves, linear transformation of the input features, followed by a non-linear activation function. The first two layers, use the ReLU activation. The final layer serves as the output layer. As the C3D embeddings are limited to 512 dimensional feature vector, a smaller model with fewer parameters seem to give better results than a deep model with large number of hidden layers. 3.4.2 Linear Discriminant Analysis Objective Function: Classic Linear Discriminant Analysis: Classic LDA, is a dimen- sionality reduction algorithm, which takes the higher dimensional data, and projects it onto a lower dimensional space such that the separation within data points of the same class is minimized and separation between data points of different classes is maximized[16]. Let x1, ..., xn = X ∈ RN xb be a set of N data points which belong to C different classes. LDA tries to find a transformation matrix A ∈ Rlxb which projects the input to a lower dimensional space L where, l = C-1. The linear combination of features, obtained by multiplying the X with AT has maximum separation in this lower dimensional subspace. The LDA formulation to find A is given as: argmax A |ASb AT | |ASw AT | (1)
  • 3. IceBreaker: Solving Cold Start Problem for Video Recommendation Engines Conference’17, July 2017, Washington, DC, USA Figure 1: Deep Learning based Regression Figure 2: Random Forest based regression on video pairs where, Sb is the between class scatter matrix and Sw is the within class scatter matrix. St denotes the overall scatter matrix. Xc = Xc − mc are the mean-centered observations for each class, the vector mc represents the mean of each class similarly X represents mean centered observation of the complete population. The formula for St , Sw and Sb is given as: Sc = 1 Nc − 1 X T c Xc (2) Sw = 1 C Õ c Sc (3) St = 1 N − 1 X T X (4) The value of transformation matrix A that maximizes Eq. 1, max- imizes the ratio of the ‘between class scatter’ and ‘within class scatter’. Thereby, giving projections with low intraclass and high inter-class variance. To find optimum solution for Eq. 1, we need to solve the eigenvalue problem of Sb e = vSw e, where e and v represent the eigenvector and their corresponding eigenvalues. The matrix A is the set of eigenvectors e. In the paper [15], authors have extrapolated this concept to introduce the DeepLDA objective function. In [15], the LDA function has been modified to be used as an objective function for deep neural networks by maximizing the individual eigenvalues. Each value gives the degree of separation along a particular eigenvector i.e. higher eigenvalues correspond to a greater amount of separation. Maximizing these eigenvalues will help in obtaining an embedding in which classes are well separated. However, if we consider all the eigenvalues in the loss function, it can lead to a trivial solution in which only largest of the eigenvalue is maximized. To circumvent this problem, we optimize only on smallest m eigenvalues out of total set of eigenvalues. This can be done by using only those m eigenvalues that are less than a certain predetermined threshold. Therefore, the formulation becomes: argmax θ 1 k k Õ i=1 vi with {v1, ..., vk } = {vj |vj < min vj |vj + ϵ} (5) This loss is back-propagated through the network. In our model, each row in the relevance list specifies a batch. All the movies/shows that come in the relevance list of a particular show/movie are given one class and the others which are not in list are assigned a separate class. Then, we train on the loss using ADAM optimization[18]. 4 EVALUATION 4.1 Evaluation Metric For the challenge, we use two evaluations measures[23] - (1) Recall@K - For top K predictions the metric recall@K is defined as - recall@K = |or ∩ or | or (6) where or is relevance list and and or predicted top K relevant shows/movies for each item. (2) Hit@k - If recall@K > 0 for a test case then hit@K has value 1 or else value 0 for that test case. The average values of recall and hit is taken on all test cases for evaluation purposes [23].
  • 4. Conference’17, July 2017, Washington, DC, USA Yaman Kumar et al. Figure 3: DeepLDA based model’s architecture Table 1: Track 1 and 2 - Results for the different architectures for TV shows and movies Data Data Subset Method Hit @ k Recall @ k k=5 k=10 k=20 k=30 k=50 k=100 k=200 k=300 TV Shows Validation Set Random Forest 0.004 0.016 0.063 0.112 0.017 0.034 0.062 0.084 Regression(Cosine Loss) 0.319 0.388 0.485 0.541 0.171 0.229 0.289 0.323 Regression(Poisson Loss) 0.284 0.353 0.432 0.483 0.145 0.205 0.276 0.322 DeepLDA 0.178 0.234 0.310 0.354 0.065 0.103 0.153 0.196 Testing Set Random Forest 0.047 0.093 0.178 0.262 0.031 0.056 0.093 0.122 Regression(Cosine Loss) 0.226 0.295 0.38 0.434 0.075 0.104 0.137 0.157 Regression(Poisson Loss) 0.217 0.294 0.381 0.426 0.076 0.106 0.139 0.16 DeepLDA 0.22 0.302 0.404 0.465 0.071 0.117 0.18 0.224 Movies Validation Set Random Forest 0.004 0.011 0.037 0.063 0.015 0.025 0.045 0.063 Regression(Cosine Loss) 0.18 0.231 0.313 0.362 0.099 0.139 0.192 0.226 Regression(Poisson Loss) 0.143 0.191 0.268 0.318 0.114 0.166 0.246 0.313 DeepLDA 0.106 0.143 0.192 0.238 0.065 0.091 0.129 0.158 Testing Set Random Forest 0.055 0.109 0.172 0.216 0.035 0.051 0.077 0.097 Regression(Cosine Loss) 0.108 0.144 0.212 0.256 0.044 0.062 0.084 0.099 Regression(Poisson Loss) 0.107 0.151 0.212 0.25 0.044 0.061 0.082 0.096 DeepLDA 0.161 0.214 0.286 0.33 0.072 0.099 0.137 0.169 These evaluation metrics have been used in several other works such as [12, 21, 27]. 4.2 Results Table 1 contains the results obtained from all the different archi- tectural models considered for both TV shows as well as movies obtained on validation set and training set after training the models. For various values of K, the results are presented in the table. The DeepLDA model clearly gives the best results on these met- rics for both TV shows and movies on the testing set. The competi- tion used recall@100 and hit@30 for the evaluation of results, thus using that the analysis of results is presented below. For TV Shows, considering the hit values on the test data, the DeepLDA model performs better than the Random Forest model by 77.4%, Regression model with cosine loss by 7.1% and Regression model with poisson loss by 9.1%. For the metric recall, the DeepLDA model performs better than the Random Forest model by a full 108%, Regression model with cosine loss by 12.5% and Regression model with poisson loss by 10.3%. For movies, considering the hit values on the test data, the DeepLDA model performs better than the Random Forest model by 52.7%, Regression model with cosine loss by 28.9% and Regression model with poisson loss by 32%. Taking into accoutn recall val- ues, the DeepLDA model performs better than the Random Forest model by 94.1%, Regression model with cosine loss by 59.6% and Regression model with poisson loss by 62.2%. 5 CONCLUSIONS AND FUTURE WORK The paper presented several models on which extensive experi- ments were done to provide content based recommendations to users. This solution provided a remarkable solution for the ‘cold- start’ problem faced by video recommendation engines like Hulu. The relevance between two videos is computed using different types of losses, regression techniques and linear discriminant analysis. In the future, the authors aim to train models using hybrid mod- els generated after employing explicit as well as implicit features. These models have the potential of improving the results for video recommendation significantly. Additionally, they also look forward to extend the current models for addressing the pernicious cold start problem holistically by taking into account not just cold videos but cold-start user profiles too. These approaches could also help in ameliorating the accuracy of the models considered and make it more robust. REFERENCES [1] Sami Abu-El-Haija, Nisarg Kothari, Joonseok Lee, Apostol Natsev, George Toderici, Balakrishnan Varadarajan, and Sudheendra Vijayanarasimhan. 2016. YouTube-8M: A Large-Scale Video Classification Benchmark. CoRR abs/1609.08675 (2016). [2] Arslan Basharat, Yun Zhai, and Mubarak Shah. 2008. Content based video match- ing using spatiotemporal volumes. Computer Vision and Image Understanding 110, 3 (2008), 360ś377. [3] Chumki Basu, Haym Hirsh, William Cohen, et al. 1998. Recommendation as classification: Using social and content-based information in recommendation. In AAAI/IAAI. 714ś720. [4] J Roger Bray and John T Curtis. 1957. An ordination of the upland forest com- munities of southern Wisconsin. Ecological monographs 27, 4 (1957), 325ś349.
  • 5. IceBreaker: Solving Cold Start Problem for Video Recommendation Engines Conference’17, July 2017, Washington, DC, USA [5] Jeff Bullas. 2017. 30 Mind Numbing YouTube Facts,Figures and Statistics - Infographic. https://www.jeffbullas.com/ 35-mind-numbing-youtube-facts-figures-and-statistics-infographic/ Re- trieved July 7, 2018. [6] Luca Canini, Sergio Benini, and Riccardo Leonardi. 2013. Affective recommen- dation of movies based on selected connotative features. IEEE Transactions on Circuits and Systems for Video Technology 23, 4 (2013), 636ś647. [7] Meeyoung Cha, Haewoon Kwak, Pablo Rodriguez, Yong-Yeol Ahn, and Sue Moon. 2007. I tube, you tube, everybody tubes: analyzing the world’s largest user gen- erated content video system. In Proceedings of the 7th ACM SIGCOMM conference on Internet measurement. ACM, 1ś14. [8] Sung-Hyuk Cha. 2007. Comprehensive survey on distance/similarity measures between probability density functions. City 1, 2 (2007), 1. [9] François Chollet et al. 2015. Keras. https://keras.io. [10] James Davidson, Benjamin Liebald, Junning Liu, Palash Nandy, Taylor Van Vleet, Ullas Gargi, Sujoy Gupta, Yu He, Mike Lambert, Blake Livingston, et al. 2010. The YouTube video recommendation system. In Proceedings of the fourth ACM conference on Recommender systems. ACM, 293ś296. [11] Yashar Deldjoo, Mehdi Elahi, Paolo Cremonesi, Franca Garzotto, Pietro Piazzolla, and Massimo Quadrana. 2016. Content-based video recommendation system based on stylistic visual features. Journal on Data Semantics 5, 2 (2016), 99ś113. [12] Yashar Deldjoo, Mehdi Elahi, Massimo Quadrana, and Paolo Cremonesi. 2015. To- ward building a content-based video recommendation system based on low-level features. In International Conference on Electronic Commerce and Web Technologies. Springer, 45ś56. [13] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009. Imagenet: A large-scale hierarchical image database. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2009. Ieee, 248ś255. [14] Michel-Marie Deza and Elena Deza. 2006. Dictionary of distances. Elsevier. [15] Matthias Dorfer, Rainer Kelz, and Gerhard Widmer. 2015. Deep Linear Dis- criminant Analysis. CoRR abs/1511.04707 (2015). arXiv:1511.04707 http: //arxiv.org/abs/1511.04707 [16] R. A. Fisher. 1936. The Use of Multiple Measurements in Taxonomic Problems. Annals of Eugenics 7, 7 (1936), 179ś188. [17] David Goldberg, David Nichols, Brian M Oki, and Douglas Terry. 1992. Using collaborative filtering to weave an information tapestry. Commun. ACM 35, 12 (1992), 61ś70. [18] Diederik P. Kingma and Jimmy Ba. 2014. Adam: A Method for Stochastic Optimiza- tion. CoRR abs/1412.6980 (2014). arXiv:1412.6980 http://arxiv.org/abs/1412.6980 [19] Eugene F Krause. 1975. Taxicab geometry: An adventure in non-Euclidean geometry. Courier Corporation. [20] Taras Lehinevych, Nikolaos Kokkinis-Ntrenis, Giorgos Siantikos, A Seza Dogruoz, Theodoros Giannakopoulos, and Stasinos Konstantopoulos. 2014. Discovering similarities for content-based recommendation and browsing in multimedia collections. In 2014 Tenth International Conference on Signal-Image Technology & Internet-Based Systems (SITIS). IEEE, 237ś243. [21] Yan Li, Hanjie Wang, Hailong Liu, and Bo Chen. 2017. A study on content-based video recommendation. In 2017 IEEE International Conference on Image Processing (ICIP). IEEE, 4581ś4585. [22] Andy Liaw, Matthew Wiener, et al. 2002. Classification and regression by ran- domForest. R news 2, 3 (2002), 18ś22. [23] Mengyi Liu, Xiaohui Xie, and Hanning Zhou. 2018. Content-based Video Rele- vance Prediction Challenge: Data, Protocol, and Baseline. [24] Mayank Meghawat, Satyendra Yadav, Debanjan Mahata, Yifang Yin, Rajiv Ratn Shah, and Roger Zimmermann. 2018. A Multimodal Approach to Predict Social Media Popularity. In 2018 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR). IEEE. [25] Vinod Nair and Geoffrey E. Hinton. 2010. Rectified Linear Units Improve Re- stricted Boltzmann Machines. In Proceedings of the 27th International Conference on International Conference on Machine Learning (ICML’10). Omnipress, USA, 807ś814. http://dl.acm.org/citation.cfm?id=3104322.3104425 [26] Neil Rubens, Mehdi Elahi, Masashi Sugiyama, and Dain Kaplan. 2015. Active learning in recommender systems. In Recommender systems handbook. Springer, 809ś846. [27] Suvash Sedhain, Scott Sanner, Darius Braziunas, Lexing Xie, and Jordan Chris- tensen. 2014. Social collaborative filtering for cold-start recommendations. In Proceedings of the 8th ACM Conference on Recommender systems. ACM, 345ś348. [28] Rajiv Shah and Roger Zimmermann. 2017. Multimodal analysis of user-generated multimedia content. Springer. 1ś263 pages. [29] Rajiv Ratn Shah, Yi Yu, and Roger Zimmermann. 2014. Advisor: Personalized video soundtrack recommendation by late fusion with heuristic rankings. In Proceedings of the 22nd ACM international conference on Multimedia. ACM, 607ś 616. [30] Amit Singhal et al. 2001. Modern information retrieval: A brief overview. IEEE Data Eng. Bull. 24, 4 (2001), 35ś43. [31] Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, and Manohar Paluri. 2015. Learning spatiotemporal features with 3d convolutional networks. In Proceedings of the IEEE international conference on computer vision. 4489ś4497. [32] Jian Wei, Jianhua He, Kai Chen, Yi Zhou, and Zuoyin Tang. 2017. Collaborative filtering and deep learning based recommendation system for cold start items. Expert Systems with Applications 69 (2017), 29ś39. [33] Zifeng Wu, Chunhua Shen, and Anton van den Hengel. 2016. Wider or deeper: Revisiting the resnet model for visual recognition. arXiv preprint arXiv:1611.10080 (2016). [34] Bo Yang, Tao Mei, Xian-Sheng Hua, Linjun Yang, Shi-Qiang Yang, and Mingjing Li. 2007. Online video recommendation based on multimodal fusion and relevance feedback. In Proceedings of the 6th ACM international conference on Image and video retrieval. ACM, 73ś80. [35] Sergey Zagoruyko and Nikos Komodakis. 2016. Wide residual networks. arXiv preprint arXiv:1605.07146 (2016). [36] Herbert Zettl. 2002. Essentials of applied media aesthetics. In Media Computing. Springer, 11ś38. [37] Xiaojian Zhao, Guangda Li, Meng Wang, Jin Yuan, Zheng-Jun Zha, Zhoujun Li, and Tat-Seng Chua. 2011. Integrating rich information for video recommendation with multi-task rank aggregation. In Proceedings of the 19th ACM international conference on Multimedia. ACM, 1521ś1524. [38] Qiusha Zhu, Mei-Ling Shyu, and Haohong Wang. 2013. Videotopic: Content- based video recommendation using a topic model. In 2013 IEEE International Symposium on Multimedia (ISM). IEEE, 219ś222.