Monday, November 23, 2015

Netflix Recommendations

Netflix Recommendations - media opinions

11 Great Moments In Netflix Recommendations
http://www.11points.com/Movies/11_Great_Moments_In_Netflix_Recommendations

collection of posts at Huffington Post
http://www.huffingtonpost.com/news/netflix-recommendations/

http://www.huffingtonpost.com/2012/04/09/netflix-recommendations_n_1413179.html
Good Movies To Watch
http://agoodmovietowatch.com/netflix/

The Netflix Prize Winning Algorithm

http://www.netflixprize.com/assets/GrandPrize2009_BPC_BellKor.pdf

Five Best Movie Recommendation Services

http://lifehacker.com/5884202/five-best-movie-recommendation-services
good sites
ChickFlix
http://chickflix.net/
Rotten Tomatoe
http://www.rottentomatoes.com/
http://www.rottentomatoes.com/about/

WHAT IS THE TOMATOMETER™?

The Tomatometer™ rating - based on the published opinions of hundreds of film and television critics - is a trusted measurement of movie and TV programming quality for millions of moviegoers.
The Tomatometer™ rating represents the percentage of professional critic reviews that are positive for a given film or television show.
Back in the days of the open theaters, when a play was particularly atrocious, the audience expressed their dissatisfaction by not only booing and hissing at the stage, but throwing whatever was at hand -- vegetables and fruits included.
A good review is denoted by a fresh red tomato. In order for a movie or TV show to receive an overall rating of Fresh, the reading on the Tomatometer for that movie must be at least 60%.
A bad review is denoted by a rotten green tomato splat (59% or less).
To receive a Certified Fresh rating a movie must have a steady Tomatometer rating of 75% or better. Movies opening in wide release need at least 80 reviews from Tomatometer Critics (including 5 Top Critics). Movies opening in limited release need at least 40 reviews from Tomatometer Critics (including 5 Top Critics). A TV show must have a Tomatometer Score of 75% or better with 20 or more reviews from Tomatometer Critics (including 5 Top Critics). If the Tomatometer score drops below 70%, then the movie or TV show loses its Certified Fresh status. In some cases, the Certified Fresh designation may be held at the discretion of the Rotten Tomatoes editorial team.

WHAT IS THE AUDIENCE SCORE?

The Audience rating, denoted by a popcorn bucket, is the percentage of all Flixster.com and RottenTomatoes.com users who have rated the movie or TV Show positively.
The full popcorn bucket means the movie received 3.5 stars or higher by Flixster and Rotten Tomatoes users.
The tipped over popcorn bucket means the movie received less than 3.5 stars by Flixster and Rotten Tomatoes users.
The plus sign will appear for movies that do not have audience ratings or reviews. The percentage you see associated with this icon is the percentage of users who added the movie to their Want-to-See list.




mediocre:
Criticker
http://www.criticker.com/
Jinni
http://www.jinni.com/



Netflix Recommendation Algorithm


[PDF] PushTrust: An Efficient Recommendation Algorithm by Leveraging Trust and Distrust Relations

R Forsati, I Barjasteh, F Masrour, AH Esfahanian… - Proceedings of the 9th ACM …, 2015
http://www.egr.msu.edu/waves/people/iman_files/RecSys_15_PushTrust_pres.pdf

Scaling Up Recommendation Services in Many Dimensions

B Németh - Proceedings of the 9th ACM Conference on …, 2015
... After successes in the Netflix Prize competition, the research team transformed into a company
providing recommendation and personalization solutions on a ... Currently he is focused on how
to apply state of the art recommendation algorithms in real world problems to help ...

Recommendation techniques in forensic data analysis: a new approach

M Quintana, S Uribe, F Sánchez, F Álvarez - … Prevention and Detection (ICDP-15), 6th …, 2015
... This technique was widely exploited at the Netflix prize [10], and it is ... provided by past cases, and
they are built over recommendation systems algorithms... methods inspire the “Clues
Recommendation Algorithm” while memory based collaborative filtering techniques are the ...

[PDF] POI Recommendation: Towards Fused Matrix Factorization with Geographical and Temporal Influences

JB Griesner, T Abdessalem, H Naacke - Proceedings of the 9th ACM Conference on …, 2015
... Matrix factorization techniques have demonstrated since the Netflix challenge [8] to be one of the
most accurate recommendation methods many ... Zhang et al. in [16] have proposed Collaborative
Location Activity Filtering (CLAF) algorithm for generic recommendation...

Content recommendation and service costs in swarming systems

D Munaro, C Delgado, DS Menasche - … (ICC), 2015 IEEE International Conference on, 2015
... Netflix alone accounts for up to 34% of North America's downloads during peak hours. ... Some
of the existing P2P systems already have some sort of content recommendation algorithm
embedded in the system [7]–[9], but none of these algorithms was implemented to influence ...




Wednesday, November 18, 2015

Saturday, November 14, 2015

Understanding LSTM RNN Networks


Unsupervised Learning of Video Representations using LSTMs

Long Short Term Memory (LSTM) networks to learn representations of video sequences.
DEEP LEARNING (DL) RECURRENT NEURAL NETWORKS (RNN) GENERATIVE COMPUTER VISION

We use multilayer Long Short Term Memory (LSTM) networks to learn representations of video sequences. Our model uses an encoder LSTM to map an input sequence into a fixed length representation. This representation is decoded using single or multiple decoder LSTMs to perform different tasks, such as reconstructing the input sequence, or predicting the future sequence.
http://www.cs.toronto.edu/~nitish/unsupervised_video/

A test set for evaluating sequence prediction/reconstruction
Moving MNIST [782Mb] contains 10,000 sequences each of length 20 showing 2 digits moving in a 64 x 64 frame.

The results in the updated arxiv paper use this test set to report numbers. For future prediction, the metric is cross entropy loss for predicting the last 10 frames for each sequence conditioned on the first 10 frames.
Code
PapersUnsupervised Learning of Video Representations using LSTMs [pdf]
Nitish Srivastava, Elman Mansimov, Ruslan Salakhutdinov
ICML 2015.
Updated arxiv version with more details -
Unsupervised Learning of Video Representations using LSTMs [arxiv]
Nitish Srivastava, Elman Mansimov, Ruslan Salakhutdinov
unsup_video_lstm.tar.gz [119Kb]

Understanding LSTM Networks

Posted on August 27, 2015

Recurrent Neural Networks

Humans don’t start their thinking from scratch every second. As you read this essay, you understand each word based on your understanding of previous words. You don’t throw everything away and start thinking from scratch again. Your thoughts have persistence.

Traditional neural networks can’t do this, and it seems like a major shortcoming. For example, imagine you want to classify what kind of event is happening at every point in a movie. It’s unclear how a traditional neural network could use its reasoning about previous events in the film to inform later ones.

Recurrent neural networks address this issue. They are networks with loops in them, allowing information to persist.



Deep Learning
Nando de Freitas
26 February 2015
YouTube Oxford CS lectures
https://www.youtube.com/playlist?list=PLE6Wd9FR--EfW8dtjAuPoTuPcqmOV53Fu

Deep Learning Lecture 12: Recurrent Neural Nets and LSTMs
Published on Mar 2, 2015

Slides available at: https://www.cs.ox.ac.uk/people/nando....
Course taught in 2015 at the University of Oxford by Nando de Freitas with great help from Brendan Shillingford.
https://www.cs.ox.ac.uk/people/nando.defreitas/machinelearning/
https://www.youtube.com/watch?v=56TYLaQN4N8

Deep Learning
Alex Graves
5 March 2015
YouTube Oxford CS lectures
https://www.youtube.com/watch?v=-yX1SYeDHbg

Ilya Sutskever's home page

www.cs.toronto.edu/~ilya/

Ilya Sutskever Research Director of OpenAI. I spent three wonderful years as a Research Scientist at the Google Brain Team. Before that, I was a co-founder of ...


Recurrent Batch Normalization
http://arxiv.org/abs/1603.09025

Tim Cooijmans, Nicolas Ballas, César Laurent, Çağlar Gülçehre, Aaron Courville
(Submitted on 30 Mar 2016 (v1), last revised 4 Apr 2016 (this version, v3))
We propose a reparameterization of LSTM that brings the benefits of batch normalization to recurrent neural networks. Whereas previous works only apply batch normalization to the input-to-hidden transformation of RNNs, we demonstrate that it is both possible and beneficial to batch-normalize the hidden-to-hidden transition, thereby reducing internal covariate shift between time steps. We evaluate our proposal on various sequential problems such as sequence classification, language modeling and question answering. Our empirical results show that our batch-normalized LSTM consistently leads to faster convergence and improved generalization.

Ross Goodwin
http://rossgoodwin.com/
https://medium.com/@rossgoodwin/3505ae7a17e7

Adventures in Narrated Reality, Part II
Ongoing experiments in writing & machine intelligence


By Ross Goodwin

[DRAFT]

Due to the popularity of Adventures in Narrated Reality, Part I, I’ve decided to continue narrating my research concerning the creative potential of LSTM recurrent neural networks here on Medium. In this installment, I’ll begin by introducing a new short film: Sunspring, an End Cue film, directed by Oscar Sharp and starring Thomas Middleditch, created for the 2016 Sci-Fi London 48 Hour Film Challenge from a screenplay generated with an LSTM trained on science fiction screenplays.