Saturday, August 27, 2016

Ben-Hur: Social Media Monitoring Report

Source: Official Trailer



Ben-Hur: Social Media Monitoring Report
I was recently reading an interesting article The straight-washing of “Ben-Hur”: Remake of the ’59 epic drops gaysubtext—and beefs up religious themes. This inspired me to consider exploring audience feedback about this film on social media during its opening weekend.  The film’s opening weekend was just a week ago (Aug 19-21, 2016) and resulted in a low box office turnout of $11.4million domestic--quite a poor outcome for a film with an approximate $100 million production budget.


Considering Approaches to Collecting Data
In considering ways to explore how the film was being discussed on social media, I initially considered looking at its Facebook page (Ben-Hur Film) or its Twitter feed.  As of this posting, approximately 226,000 people have liked this Facebook page.  I had considered exploring the feedback of audiences to the posted video clips or to the film in general, with a focus on posts during the opening weekend, but this resulted in too large a sample for the purposes of this dialogue.

I also considered looking to Twitter and found there were several hashtags I could explore, including #BenHur, #BenHurMovies, and #BenHur2016.  However, looking at some of these, I was finding a mix of audience response, celebrity encouragements to go see the film, and tweets from film producers, such as Roma Downey (@RealRomaDowney).  This mix of dialogues was going to be challenging to explore—particularly as it might require my further investigation into the tweeter to gauge if they were connected to the film.

Rotten Tomatoes: Site of Examination
Ultimately, I decided to use Rotten Tomatoes,  a film review website that has an active audience participation pool that is separate from their “critic” or reviewers.  Looking at the page for Ben-Hur (2016) we are exposed to preliminary feedback on the film’s response.  As seen on the “Tomatometer” the film was rated at a 28%, which indicated the percentage of approved critics who have given the film a positive review, which is assessed at a 3.5 or higher out of 5 stars.  Meanwhile, we see that 66% of the audience respondents rated the film positively.  While these initial numbers are somewhat informative, presenting information about critical and audience response it does not dig into the opening weekend feedback.  These numbers speak to feedback up to the current time.

Source: Rottentomatoes.com

To explore the audience feedback during opening weekend I looked at the audience reviews posted on Aug 19, Aug 20, and Aug 21, 2016.  It should be noted that these reviews are fairly immediate and does not explore the reviews people posted in the days that followed, who may have seen the film during opening weekend.  The collected reviews from these three days resulted in a sample of 208 reviews (n=208).  Given the variety of review lengths and assessments that could be done, I chose to centrally focus on a quassi-measure of reviewer legitimacy/social engagement, who was posting a review, and their star rating.  

Findings
Firstly, to assess the legitimacy and social media engagement of the reviewers I examined their “profiles.”  “Profiles” are readily visible on the page of reviews in the form of an image and a screen name. Unlike some social media platforms, these profiles are not full profiles in the traditional sense, by in large.  One is unable to click into a person to learn more about them, we are constrained to examinations of their identity based on image and screen name.  As a quassi-measure of likely digital media engagement I collected data on the number of reviewers who included a photo in their profile and those who did not.  I found that nearly ¾ of the audience reviewers included an image, which may suggest a certain familiarity and engagement with both digital technologies and profile construction/meaning.

Secondly, informed by the profile information, I explored the likely gender of the reviewer.  Given the limitations of this project, I centrally focused upon categories of men, women, and unknown.  Reviewers were categorized primarily based upon their screen name, using personal knowledge of name gendering based on a western context.  For a number of profiles it was impossible to determine gender on name alone—in such a case, if a picture was present, it would also be used in the gender determination when appropriate.  This did leave a number of reviewers as unknown.  This demonstrated that among those posting reviews on opening weekend on this website, approximately 60% were men, 25% were women, and 15% were gender unknown (see image for precise statistics).




Finally, I explored the star-rating these reviewers gave to the film.  The ratings ranged from zero to five, with five being the strongest possible review.  Below you may see the breakdown of ratings.  The average number of stars given was 3.5, with a median of 4 stars. Based on this sample and the Rotten Tomatoes rating system, this demonstrated 64% of the audience who wrote a review on this website on the opening weekend liked the film.  This is highly consistent with the current rating, a week later, of 66%.  This said, the numbers are skewed in part due to at least two reviews: one reviewer gave the film zero stars, but had evidently not seen the film based on their written commentary “This won’t be very good. I am not going to watch it.” and another gave the film zero stars, but has extremely positive things to say “I thought the film was awesome!” These two reviews demonstrate some of the problems with this sort of data – some may post feedback/reviews without being personally informed and others may not fully understand the correct manner to assess or evaluate something in a system such as this.




Looking Elsewhere
Another obvious place one might look for film feedback is the Internet Media Database, largely known as IMDb.  On the IMDb website when we look at Ben-Hur we see a 5.5 star out of 10 star score, based on just over 3,800 current reviews. It is interesting that one can actually dig into the reviewer statistics specifically.  Herein, we can find information on reviewer characteristics such as sex, age, age, and U.S./non-U.S. users. For example we can see, that among reviewers whose sex (male or female) is determined, that approximately 80% are male and 20% are female. 


Summarizing Thoughts
Looking at the audience response to this film on this social media site and others offers perspective for various consumers and producers of media, particularly film in this case.  As a consumer, before paying substantial money to see a film in the theater, many may consult reviews from experts or from online reviews in general. Certainly, we may weight expert reviews more heavily on average, but the word of mouth of the audience can make a real difference.  There sure have been films that opened to poor reviews, but became hits with the audiences -- you've probably seen at least some of them!

 
Looking to the reviews for this film on Rotten Tomatoes offers a more positive response to the film than IMDb.  However, the current general reviews on both arguably align with an IMDb overall rating of 5.5 out of 10 stars and a Rotten Tomatoes rating of 66% liking the film.  Indeed, if the Rotten Tomatoes scale is adjusted to a 10 points scale, looking at the opening weekend reviews, it would demonstrate a 7 out of 10 stars.  When examining these numbers we do need to consider the longevity and internet traffic of these respective sites.  IMDb was launched in 1990 and has an Alexa rank of 53.  Meanwhile,  Rotten Tomatoes was launched in 1998, but has a Alexa rank of 552.  Clearly, IMDb is more trafficked and thus likely more respected by the average consumer.  This said, there may different types of audiences using these websites – perhaps Rotten Tomatoes is the “underground” reviewer locale?


Regardless of source comparisons, Rotten Tomatoes demonstrates that the opening weekend’s audience reviews were largely consistent with ratings throughout the week. Despite the various media word of mouth, this film is clearly appealing to a significant number of reviewers.  While not specifically assessed, anecdotally, it appeared that those offering particularly negative reviews were often making comparisons to the classic Ben-Hur (1959). Those with very strong reviews appeared to reference the action of the film as well as its strong religious messages.


Beyond the informing of audiences, these reviews offer insight to film producers and companies as they consider remaking films—especially critically acclaimed films that persist in their reputation.  However, an exploration of domestic versus international reception can also be part of the equation as they consider the financial implications for a film remake such as this – while many Americans may be familiar with the 1959 version, foreign audiences may not be and may receive such a film more openly, without baggage of prior perspective.