YouTube is an American online video-sharing platform headquartered in San Bruno, California. The service, created in February 2005 by three former PayPal employees—Chad Hurley, Steve Chen, and Jawed Karim—was bought by Google in November 2006 for US$1.65 billion and now operates as one of the company’s subsidiaries. YouTube is the second most-visited website in the world after Google Search, according to Alexa Internet rankings.
YouTube allows users to upload, view, rate, share, add to playlists, report, comment on videos, and subscribe to other users. Available content includes video clips, TV show clips, music videos, short and documentary films, audio recordings, movie trailers, live streams, video blogging, short original videos, and educational videos. Most content is generated and uploaded by individuals, but media corporations including CBS, the BBC, Vevo, and Hulu offer some of their material via YouTube as part of the YouTube partnership program. Unregistered users can watch, but not upload, videos on the site, while registered users can upload an unlimited number of videos and add comments. Age-restricted videos are available only to registered users affirming themselves to be at least 18 years old.
As of May 2019, there were more than 500 hours of content uploaded to YouTube each minute and one billion hours of content being watched on YouTube every day. YouTube and selected creators earn advertising revenue from Google AdSense, a program that targets ads according to site content and audience. The vast majority of videos are free to view, but there are exceptions, including subscription-based premium channels, film rentals, as well as YouTube Music and YouTube Premium, subscription services respectively offering premium and ad-free music streaming, and ad-free access to all content, including exclusive content commissioned from notable personalities. Based on reported quarterly advertising revenue, YouTube is estimated to have US$15 billion in annual revenues.
YouTube has faced criticism over aspects of its operations, including its handling of copyrighted content contained within uploaded videos, its recommendation algorithms perpetuating videos that promote conspiracy theories and falsehoods, hosting videos ostensibly targeting children but containing violent or sexually suggestive content involving popular characters, videos of minors attracting pedophilic activities in their comment sections, and fluctuating policies on the types of content that is eligible to be monetized with advertising.
History
Founding and initial growth (2005–2006)
YouTube was founded by Steve Chen, Chad Hurley, and Jawed Karim, who were all early employees of PayPal. Hurley had studied design at Indiana University of Pennsylvania, and Chen and Karim studied computer science together at the University of Illinois at Urbana–Champaign.
Karim said the inspiration for YouTube first came from Janet Jackson’s role in the 2004 Super Bowl incident when her breast was exposed during her performance, and later from the 2004 Indian Ocean tsunami. Karim could not easily find video clips of either event online, which led to the idea of a video sharing site. Hurley and Chen said that the original idea for YouTube was a video version of an online dating service, and had been influenced by the website Hot or Not. They created posts on Craigslist asking attractive women to upload videos of themselves to YouTube in exchange for a $100 reward. Difficulty in finding enough dating videos led to a change of plans, with the site’s founders deciding to accept uploads of any type of video.
According to a story that has often been repeated in the media, Hurley and Chen developed the idea for YouTube during the early months of 2005, after they had experienced difficulty sharing videos that had been shot at a dinner party at Chen’s apartment in San Francisco. Karim did not attend the party and denied that it had occurred, but Chen commented that the idea that YouTube was founded after a dinner party “was probably very strengthened by marketing ideas around creating a story that was very digestible”.
YouTube began as a venture capital–funded technology startup, primarily from an $11.5 million investment by Sequoia Capital and an $8 million investment from Artis Capital Management between November 2005 and April 2006. YouTube’s early headquarters were situated above a pizzeria and Japanese restaurant in San Mateo, California. The domain name www.youtube.com
was activated on February 14, 2005, and the website was developed over the subsequent months. The first YouTube video, titled Me at the zoo, shows co-founder Jawed Karim at the San Diego Zoo. The video was uploaded on April 23, 2005, and can still be viewed on the site. YouTube offered the public a beta test of the site in May 2005. The first video to reach one million views was a Nike advertisement featuring Ronaldinho in November 2005. Following a $3.5 million investment from Sequoia Capital in November, the site launched officially on December 15, 2005, by which time the site was receiving 8 million views a day.
At the time of the official launch, YouTube did not have much market recognition. It was not the first video-sharing site on the Internet, as Vimeo was launched in November 2004, though that site remained a side project of its developers from CollegeHumor at the time and did not grow much either. The week of YouTube’s launch, NBC-Universal’s Saturday Night Live ran a skit “Lazy Sunday” by The Lonely Island. Besides helping to bolster ratings and long-term viewership for Saturday Night Live, “Lazy Sunday”‘s status as an early viral video helped established YouTube as an important website. Unofficial uploads of the skit to YouTube drew in more than five million collective views by February 2006 before they were removed at request of NBC-Universal about two months later, raising questions of copyright related to viral content. Despite eventually being taken down, these duplicate uploads of the skit helped popularize YouTube’s reach and led to the upload of further third-party content. The site grew rapidly and, in July 2006, the company announced that more than 65,000 new videos were being uploaded every day, and that the site was receiving 100 million video views per day.
The choice of the name www.youtube.com
led to problems for a similarly named website, www.utube.com
. The site’s owner, Universal Tube & Rollform Equipment, filed a lawsuit against YouTube in November 2006 after being regularly overloaded by people looking for YouTube. Universal Tube has since changed the name of its website to www.utubeonline.com
.
Acquisition by Google (2006–2013)
On October 9, 2006, Google announced that it had acquired YouTube for $1.65 billion in Google stock. The deal was finalized on November 13, 2006. Google’s acquisition launched new newfound interest in video-sharing sites; IAC, which now owned Vimeo after acquiring CollegeHumor, used its asset to develop a competing site to YouTube, focusing on supporting the content creator to distinguish itself from YouTube.
In March 2010, YouTube began free streaming of certain content, including 60 cricket matches of the Indian Premier League. According to YouTube, this was the first worldwide free online broadcast of a major sporting event. On March 31, 2010, the YouTube website launched a new design, with the aim of simplifying the interface and increasing the time users spend on the site. In May 2010, YouTube videos were watched more than two billion times per day. This increased to three billion in May 2011, and four billion in January 2012. In February 2017, one billion hours of YouTube were watched every day.
According to data published by market research company comScore, YouTube is the dominant provider of online video in the United States, with a market share of around 43% and more than 14 billion views of videos in May 2010.
In October 2010, Hurley announced that he would be stepping down as chief executive officer of YouTube to take an advisory role, and that Salar Kamangar would take over as head of the company. In April 2011, James Zern, a YouTube software engineer, revealed that 30% of videos accounted for 99% of views on the site. In November 2011, the Google+ social networking site was integrated directly with YouTube and the Chrome web browser, allowing YouTube videos to be viewed from within the Google+ interface.
In May 2011, 48 hours of new videos were uploaded to the site every minute, which increased to 60 hours every minute in January 2012, 100 hours every minute in May 2013, 300 hours every minute in November 2014, and 400 hours every minute in February 2017. As of January 2012, the site had 800 million unique users a month. It has been claimed, by The Daily Telegraph in 2008, that in 2007, YouTube consumed as much bandwidth as the entire Internet in 2000. According to third-party web analytics providers, Alexa and SimilarWeb, YouTube is the second-most visited website in the world, as of December 2016; SimilarWeb also lists YouTube as the top TV and video website globally, attracting more than 15 billion visitors per month. In October 2006, YouTube moved to a new office in San Bruno, California.
In December 2011, YouTube launched a new version of the site interface, with the video channels displayed in a central column on the home page, similar to the news feeds of social networking sites. At the same time, a new version of the YouTube logo was introduced with a darker shade of red, the first change in design since October 2006.
In early March 2013, YouTube finalized the transition for all channels to the previously optional “One Channel Layout,” which removed many customization options and custom background images for consistency, and split up the channel information to different tabs (Home/Feed, Videos Playlists, Discussion, About) rather than one unified page.
New revenue streams (2013–ongoing)
In May 2013, YouTube launched a pilot program for content providers to offer premium, subscription-based channels within the platform. In February 2014, Susan Wojcicki was appointed CEO of YouTube. In November 2014, YouTube announced a subscription service known as “Music Key,” which bundled ad-free streaming of music content on YouTube with the existing Google Play Music service.
In February 2015, YouTube released a secondary mobile app known as YouTube Kids. The app is designed to provide an experience optimized for children. It features a simplified user interface, curated selections of channels featuring age-appropriate content, and parental control features. Later on August 26, 2015, YouTube launched YouTube Gaming—a video gaming-oriented vertical and app for videos and live streaming, intended to compete with the Amazon.com-owned Twitch.
In October 2015, YouTube announced YouTube Red (now YouTube Premium), a new premium service that would offer ad-free access to all content on the platform (succeeding the Music Key service released the previous year), premium original series, and films produced by YouTube personalities, as well as background playback of content on mobile devices. YouTube also released YouTube Music, a third app oriented towards streaming and discovering the music content hosted on the YouTube platform.
In January 2016, YouTube expanded its headquarters in San Bruno by purchasing an office park for $215 million. The complex has 51,468 square metres (554,000 square feet) of space and can house up to 2,800 employees.
On August 29, 2017, YouTube officially launched the “polymer” redesign of its user interfaces based on Material Design language as its default, as well a redesigned logo that is built around the service’s play button emblem.
On April 3, 2018, a shooting took place at YouTube’s headquarters in San Bruno, California.
On May 17, 2018, YouTube announced the re-branding of YouTube Red as YouTube Premium (accompanied by a major expansion of the service into Canada and 13 European markets), as well as the upcoming launch of a separate YouTube Music subscription.
In September 2018, YouTube began to phase out the separate YouTube Gaming website and app and introduced a new Gaming portal within the main service. YouTube staff argued that the separate platform was causing confusion and that the integration would allow the features developed for the service (including game-based portals and enhanced discoverability of gaming-related videos and live streaming) to reach a broader audience through the main YouTube website.
In July 2019, It was announced that YouTube will discontinue support for Nintendo 3DS systems on September 3, 2019. However, owners of New Nintendo 3DS, or New Nintendo 3DS XL, can still access YouTube on the Internet browser.
In November 2019, it was announced that YouTube was gradually phasing out the classic version of its Creator Studio across all users by the spring of 2020. As of August 2020, the classic studio is no longer available.
During the COVID-19 pandemic, when most of the world was under stay-at-home orders, usage of services such as YouTube grew greatly. In response to EU officials requesting that such services reduce bandwidth as to make sure medical entities had sufficient bandwidth to share information, YouTube along with Netflix stated they would reduce streaming quality for at least thirty days as to cut bandwidth use of their services by 25% to comply with the EU’s request. YouTube later announced that they will continue with this move worldwide, “We continue to work closely with governments and network operators around the globe to do our part to minimize stress on the system during this unprecedented situation”.
In June 2020, the ability to use categories was phased out.
Features
Video technology
YouTube primarily uses the VP9 and H.264/MPEG-4 AVC video formats, and the Dynamic Adaptive Streaming over HTTP protocol. By January 2019, YouTube had begun rolling out videos in AV1 format.
Playback
Previously, viewing YouTube videos on a personal computer required the Adobe Flash Player plug-in to be installed in the browser. In January 2010, YouTube launched an experimental version of the site that used the built-in multimedia capabilities of web browsers supporting the HTML5 standard. This allowed videos to be viewed without requiring Adobe Flash Player or any other plug-in to be installed. The YouTube site had a page that allowed supported browsers to opt into the HTML5 trial. Only browsers that supported HTML5 Video using the MP4 (with H.264 video) or WebM (with VP8 video) formats could play the videos, and not all videos on the site were available.
On January 27, 2015, YouTube announced that HTML5 would be the default playback method on supported browsers. YouTube used to employ Adobe Dynamic Streaming for Flash, but with the switch to HTML5 video now streams video using Dynamic Adaptive Streaming over HTTP (MPEG-DASH), an adaptive bit-rate HTTP-based streaming solution optimizing the bitrate and quality for the available network.
Uploading
All YouTube users can upload videos up to 15 minutes each in duration. Users can verify their account, normally through a mobile phone, to gain the ability to upload videos up to 12 hours in length, as well as produce live streams. When YouTube was launched in 2005, it was possible to upload longer videos, but a ten-minute limit was introduced in March 2006 after YouTube found that the majority of videos exceeding this length were unauthorized uploads of television shows and films. The 10-minute limit was increased to 15 minutes in July 2010. In the past, it was possible to upload videos longer than 12 hours. Videos can be at most 128 GB in size. Video captions are made using speech recognition technology when uploaded. Such captioning is usually not perfectly accurate, so YouTube provides several options for manually entering the captions for greater accuracy. YouTube formerly offered a ‘Community Captions’ feature, where viewers could write and submit captions for public display upon approval by the video uploader, but this was deprecated on September 28, 2020. (Captions that have already been added using the feature will be retained.)
YouTube accepts nearly all consumer video and audio encodings, including but not limited to MPEG-1, MPEG-2, H.264 AVC, H.265 HEVC, VC-1, VP8, VP9, and AV1 for video and MP3, linear PCM, AAC, FLAC, Vorbis, Opus, and Dolby Digital (also known as AC-3) for audio. Nearly any container format can be used, including but not limited to MP4, Matroska, FLV, AVI, WebM, 3GP, MPEG-PS, and the QuickTime File Format. Some intermediate video formats (i.e., primarily used for professional video editing, not for final delivery or storage) are also accepted, such as ProRes. YouTube does stipulate a “recommended” upload format to maximize the quality of the transcoded video that will be shown to viewers; as of late 2020 this is H.264 AVC video and AAC audio in an MP4 container.
Videos with progressive scanning or interlaced scanning can be uploaded, but for the best video quality, YouTube suggests interlaced videos be deinterlaced before uploading. All the video formats on YouTube use progressive scanning. YouTube’s statistics shows that interlaced videos are still being uploaded to YouTube, and there is no sign of that dwindling. YouTube attributes this to uploading of made-for-TV content.
Quality and formats
YouTube originally offered videos at only one quality level, displayed at a resolution of 320×240 pixels using the Sorenson Spark codec (a variant of H.263), with mono MP3 audio. In June 2007, YouTube added an option to watch videos in 3GP format on mobile phones. In March 2008, a high-quality mode was added, which increased the resolution to 480×360 pixels. In December 2008, 720p HD support was added. At the time of the 720p launch, the YouTube player was changed from a 4:3 aspect ratio to a widescreen 16:9. With this new feature, YouTube began a switchover to H.264/MPEG-4 AVC as its default video compression format. In November 2009, 1080p HD support was added. In July 2010, YouTube announced that it had launched a range of videos in 4K format, which allows a resolution of up to 4096×3072 pixels. In March 2015, support for 4K resolution was added, with the videos playing at 3840 × 2160 pixels. In June 2015, support for 8K resolution was added, with the videos playing at 7680×4320 pixels. In November 2016, support for HDR video was added which can be encoded with Hybrid Log-Gamma (HLG) or Perceptual Quantizer (PQ). HDR video can be encoded with the Rec. 2020 color space.
In June 2014, YouTube began to deploy support for high frame rate videos up to 60 frames per second (as opposed to 30 before), becoming available for user uploads in October. YouTube stated that this would enhance “motion-intensive” videos, such as video game footage.
YouTube videos are available in a range of quality levels. The former names of standard quality (SQ), high quality (HQ), and high definition (HD) have been replaced by numerical values representing the vertical resolution of the video. The default video stream is encoded in the VP9 format with stereo Opus audio; if VP9/WebM is not supported in the browser/device or the browser’s user agent reports Windows XP, then H.264/MPEG-4 AVC video with stereo AAC audio is used instead.
2020 picture quality cut
On March 18, 2020, Thierry Breton, a European commissioner in charge of digital policy of the European Union urged streaming services including YouTube to limit their services. The request came as a result of the prevention of Europe’s broadband networks from crashing as tens of millions of people started telecommuting, due to the COVID-19 pandemic in Europe. According to the EU, the streaming platforms should consider offering only standard definition, rather than high-definition, programs and users should be responsible for their data consumption. On March 20, YouTube responded by temporarily downgrading the videos in standard definition across the EU including the traffic in the UK as well.
Annotations
From 2008 to 2017, users could add “annotations” to their videos—such as pop-up text messages and hyperlinks. These functions were notably used as the basis for interactive videos, which used hyperlinks to other videos to achieve branching elements. In March 2017, it was announced that the annotations editor had been discontinued and the feature would be sunset because their use had fallen rapidly, users had found them to be an annoyance, and because they were incompatible with mobile versions of the service. Annotations were removed entirely from all videos on January 15, 2019. YouTube had introduced standardized widgets intended to replace annotations in a cross-platform manner, including “end screens” (a customizable array of thumbnails for specified videos displayed near the end of the video) and “cards”, but they are not backwards compatible with existing annotations, while the removal of annotations will also break all interactive experiences which depended on them.
Live streaming
YouTube carried out early experiments with live streaming, including a concert by U2 in 2009, and a question-and-answer session with US President Barack Obama in February 2010. These tests had relied on technology from 3rd-party partners, but in September 2010, YouTube began testing its own live streaming infrastructure. In April 2011, YouTube announced the rollout of YouTube Live, with a portal page at the URL “www.youtube.com/live”. The creation of live streams was initially limited to select partners. It was used for real-time broadcasting of events such as the 2012 Olympics in London. In October 2012, more than 8 million people watched Felix Baumgartner’s jump from the edge of space as a live stream on YouTube.
In May 2013, creation of live streams was opened to verified users with at least 1,000 subscribers; in August of that year the number was reduced to 100 subscribers, and in December the limit was removed. In February 2017, live streaming was introduced to the official YouTube mobile app. Live streaming via mobile was initially restricted to users with at least 10,000 subscribers, but as of mid-2017 it has been reduced to 100 subscribers. Live streams can be up to 4K resolution at 60 fps, and also support 360° video. In February 2017, a live streaming feature called Super Chat was introduced, which allows viewers to donate between $1 and $500 to have their comment highlighted.
3D videos
In a video posted on July 21, 2009, YouTube software engineer Peter Bradshaw announced that YouTube users could now upload 3D videos. The videos can be viewed in several different ways, including the common anaglyph (cyan/red lens) method which utilizes glasses worn by the viewer to achieve the 3D effect. The YouTube Flash player can display stereoscopic content interleaved in rows, columns or a checkerboard pattern, side-by-side or anaglyph using a red/cyan, green/magenta or blue/yellow combination. In May 2011, an HTML5 version of the YouTube player began supporting side-by-side 3D footage that is compatible with Nvidia 3D Vision. The feature set has since been reduced, and the 3D feature currently only supports red/cyan anaglyph with no side-by-side support.
360-degree videos
In January 2015, Google announced that 360-degree video would be natively supported on YouTube. On March 13, 2015, YouTube enabled 360° videos which can be viewed from Google Cardboard, a virtual reality system. YouTube 360 can also be viewed from all other virtual reality headsets. Live streaming of 360° video at up to 4K resolution is also supported.
In 2017, YouTube began to promote an alternative stereoscopic video format known as VR180, which is limited to a 180-degree field of view but is promoted as being easier to produce than 360-degree video and allowing more depth to be maintained by not subjecting the video to equirectangular projection.
User features
Community
On September 13, 2016, YouTube launched a public beta of Community, a social media-based feature that allows users to post text, images (including GIFs), live videos and others in a separate “Community” tab on their channel. Prior to the release, several creators had been consulted to suggest tools Community could incorporate that they would find useful; these YouTubers included Vlogbrothers, AsapScience, Lilly Singh, The Game Theorists, Karmin, The Key of Awesome, The Kloons, Peter Hollens, Rosianna Halse Rojas, Sam Tsui, Threadbanger and Vsauce3.
After the feature has been officially released, the community post feature gets activated automatically for every channel that passes a specific threshold of subscriber counts or already has more subscribers. This threshold was lowered over time, from 10000 subscribers to 1500 subscribers, to 1000 subscribers, which is the current threshold as of September 2019.
Channels that the community tab becomes enabled for, get their channel discussions (the name before March 2013 “One channel layout” redesign finalization: “channel comments”) permanently erased, instead of co-existing or migrating.
Content accessibility
YouTube offers users the ability to view its videos on web pages outside their website. Each YouTube video is accompanied by a piece of HTML that can be used to embed it on any page on the Web. This functionality is often used to embed YouTube videos in social networking pages and blogs. Users wishing to post a video discussing, inspired by, or related to another user’s video can make a “video response”.
On August 27, 2013, YouTube announced that it would remove video responses for being an underused feature. Embedding, rating, commenting and response posting can be disabled by the video owner. YouTube does not usually offer a download link for its videos, and intends for them to be viewed through its website interface. A small number of videos can be downloaded as MP4 files. Numerous third-party web sites, applications and browser plug-ins allow users to download YouTube videos.
In February 2009, YouTube announced a test service, allowing some partners to offer video downloads for free or for a fee paid through Google Checkout. In June 2012, Google sent cease and desist letters threatening legal action against several websites offering online download and conversion of YouTube videos. In response, Zamzar removed the ability to download YouTube videos from its site. Users retain copyright of their own work under the default Standard YouTube License, but have the option to grant certain usage rights under any public copyright license they choose.
Since July 2012, it has been possible to select a Creative Commons attribution license as the default, allowing other users to reuse and remix the material.
Platforms
Most modern smartphones are capable of accessing YouTube videos, either within an application or through an optimized website. YouTube Mobile was launched in June 2007, using RTSP streaming for the video. Not all of YouTube’s videos are available on the mobile version of the site.
Since June 2007, YouTube’s videos have been available for viewing on a range of Apple products. This required YouTube’s content to be transcoded into Apple’s preferred video standard, H.264, a process that took several months. YouTube videos can be viewed on devices including Apple TV, iPod Touch and the iPhone.
In July 2010, the mobile version of the site was relaunched based on HTML5, avoiding the need to use Adobe Flash Player and optimized for use with touch screen controls. The mobile version is also available as an app for the Android platform.
In September 2012, YouTube launched its first app for the iPhone, following the decision to drop YouTube as one of the preloaded apps in the iPhone 5 and iOS 6 operating system. According to GlobalWebIndex, YouTube was used by 35% of smartphone users between April and June 2013, making it the third-most used app.
A TiVo service update in July 2008 allowed the system to search and play YouTube videos.
In January 2009, YouTube launched “YouTube for TV”, a version of the website tailored for set-top boxes and other TV-based media devices with web browsers, initially allowing its videos to be viewed on the PlayStation 3 and Wii video game consoles.
In June 2009, YouTube XL was introduced, which has a simplified interface designed for viewing on a standard television screen. YouTube is also available as an app on Xbox Live.
On November 15, 2012, Google launched an official app for the Wii, allowing users to watch YouTube videos from the Wii channel. An app was available for Wii U and Nintendo 3DS, but was discontinued in August 2019. Videos can also be viewed on the Wii U Internet Browser using HTML5. Google made YouTube available on the Roku player on December 17, 2013, and, in October 2014, the Sony PlayStation 4.
In November 2018, YouTube launched as a downloadable app for the Nintendo Switch.
Localization
On June 19, 2007, Google CEO Eric Schmidt appeared to Paris to launch the new localization system. The interface of the website is available with localized versions in 104 countries, one territory (Hong Kong) and a worldwide version.
The YouTube interface suggests which local version should be chosen based on the IP address of the user. In some cases, the message “This video is not available in your country” may appear because of copyright restrictions or inappropriate content. The interface of the YouTube website is available in 76 language versions, including Amharic, Albanian, Armenian, Burmese, Khmer, Kyrgyz, Laotian, Mongolian, Persian and Uzbek, which do not have local channel versions. Access to YouTube was blocked in Turkey between 2008 and 2010, following controversy over the posting of videos deemed insulting to Mustafa Kemal Atatürk and some material offensive to Muslims. In October 2012, a local version of YouTube was launched in Turkey, with the domain youtube.com.tr
. The local version is subject to the content regulations found in Turkish law. In March 2009, a dispute between YouTube and the British royalty collection agency PRS for Music led to premium music videos being blocked for YouTube users in the United Kingdom. The removal of videos posted by the major record companies occurred after failure to reach an agreement on a licensing deal. The dispute was resolved in September 2009. In April 2009, a similar dispute led to the removal of premium music videos for users in Germany.
YouTube Premium
YouTube Premium (formerly YouTube Red) is YouTube’s premium subscription service. It offers advertising-free streaming, access to exclusive content, background and offline video playback on mobile devices, and access to the Google Play Music “All Access” service. YouTube Premium was originally announced on November 12, 2014, as “Music Key”, a subscription music streaming service, and was intended to integrate with and replace the existing Google Play Music “All Access” service. On October 28, 2015, the service was relaunched as YouTube Red, offering ad-free streaming of all videos, as well as access to exclusive original content. As of November 2016, the service has 1.5 million subscribers, with a further million on a free-trial basis. As of June 2017, the first season of YouTube Originals had gotten 250 million views in total.
In May 2014, before Music Key service was launched, the independent music trade organization Worldwide Independent Network alleged that YouTube was using non-negotiable contracts with independent labels that were “undervalued” in comparison to other streaming services and that YouTube would block all music content from labels who do not reach a deal to be included on the paid service. In a statement to the Financial Times in June 2014, Robert Kyncl confirmed that YouTube would block the content of labels who do not negotiate deals to be included in the paid service “to ensure that all content on the platform is governed by its new contractual terms.” Stating that 90% of labels had reached deals, he went on to say that “while we wish that we had 100% success rate, we understand that is not likely an achievable goal and therefore it is our responsibility to our users and the industry to launch the enhanced music experience.” The Financial Times later reported that YouTube had reached an aggregate deal with Merlin Network—a trade group representing over 20,000 independent labels, for their inclusion in the service. However, YouTube itself has not confirmed the deal.
On September 28, 2016, YouTube named Lyor Cohen, the co-founder of 300 Entertainment and former Warner Music Group executive, the Global Head of Music.
YouTube TV
On February 28, 2017, in a press announcement held at YouTube Space Los Angeles, YouTube announced the launch of YouTube TV, an over-the-top MVPD-style subscription service that would be available for United States customers at a price of US$35 per month. Initially launching in five major markets (New York City, Los Angeles, Chicago, Philadelphia and San Francisco) on April 5, 2017, the service offers live streams of programming from the five major broadcast networks (ABC, CBS, The CW, Fox and NBC), as well as approximately 40 cable channels owned by the corporate parents of those networks, The Walt Disney Company, CBS Corporation, 21st Century Fox, NBCUniversal and Turner Broadcasting System (including among others Bravo, USA Network, Syfy, Disney Channel, CNN, Cartoon Network, E!, Fox Sports 1, Freeform, FX and ESPN). Subscribers can also receive Showtime and Fox Soccer Plus as optional add-ons for an extra fee, and can access YouTube Premium original content (YouTube TV does not include a YouTube Premium subscription).
During the 2017 World Series (in which it was the presenting sponsor), YouTube TV ads were placed behind the home plate. The trademarked red play button logo appeared at the center of the screen, mimicking YouTube’s interface.
YouTube Go
YouTube Go is an Android app aimed at making YouTube easier to access on mobile devices in emerging markets. It is distinct from the company’s main Android app and allows videos to be downloaded and shared with other users. It also allows users to preview videos, share downloaded videos through Bluetooth, and offers more options for mobile data control and video resolution.
YouTube announced the project in September 2016 at an event in India. It was launched in India in February 2017, and expanded in November 2017 to 14 other countries, including Nigeria, Indonesia, Thailand, Malaysia, Vietnam, the Philippines, Kenya, and South Africa. It was rolled out in 130 countries worldwide, including Brazil, Mexico, Turkey, and Iraq on February 1, 2018. The app is available to around 60% of the world’s population.
YouTube Music
In early 2018, Cohen began hinting at the possible launch of YouTube’s new subscription music streaming service, a platform that would compete with other services such as Spotify and Apple Music. On May 22, 2018, the music streaming platform named “YouTube Music” was launched.
YouTube Shorts
In September 2020, YouTube announced that it would be launching a beta version of a new platform of 15–second videos, similar to TikTok, called YouTube Shorts. The platform was first tested in India. The platform is not at first be a standalone app, but is integrated into the main YouTube app. Like TikTok, it gives users access to built-in creative tools, including the possibility of adding licensed music to their videos.
YouTube Stories
In 2018, YouTube started testing a new feature initially called “YouTube Reels”. The feature is nearly identical to Instagram Stories and Snapchat Stories. YouTube later renamed the feature “YouTube Stories”. It is only available to creators who have more than 10,000 subscribers and can only be posted/seen in the YouTube mobile app.
ISNI
In 2018, YouTube became an ISNI registry, and announced its intention to begin creating ISNI identifiers to uniquely identify the musicians whose videos it features. ISNI anticipate the number of ISNI IDs “going up by perhaps 3–5 million over the next couple of years” as a result.
April Fools
YouTube featured an April Fools prank on the site on April 1 of every year from 2008 to 2016. In 2008, all links to videos on the main page were redirected to Rick Astley’s music video “Never Gonna Give You Up”, a prank known as “rickrolling”. The next year, when clicking on a video on the main page, the whole page turned upside down, which YouTube claimed was a “new layout”. In 2010, YouTube temporarily released a “TEXTp” mode which rendered video imagery into ASCII art letters “in order to reduce bandwidth costs by $1 per second.”
The next year, the site celebrated its “100th anniversary” with a range of sepia-toned silent, early 1900s-style films, including a parody of Keyboard Cat. In 2012, clicking on the image of a DVD next to the site logo led to a video about a purported option to order every YouTube video for home delivery on DVD.
In 2013, YouTube teamed up with satirical newspaper company The Onion to claim in an uploaded video that the video-sharing website was launched as a contest which had finally come to an end, and would shut down for ten years before being re-launched in 2023, featuring only the winning video. The video starred several YouTube celebrities, including Antoine Dodson. A video of two presenters announcing the nominated videos streamed live for 12 hours.
In 2014, YouTube announced that it was responsible for the creation of all viral video trends, and revealed previews of upcoming trends, such as “Clocking”, “Kissing Dad”, and “Glub Glub Water Dance”. The next year, YouTube added a music button to the video bar that played samples from “Sandstorm” by Darude. In 2016, YouTube introduced an option to watch every video on the platform in 360-degree mode with Snoop Dogg.
Content partnerships
In 2016, YouTube introduced a global program to develop creators whose videos produce a positive social impact. Google dedicated $1 million to this Creators for Change program. The first three videos from the program premiered at the 2017 Tribeca TV Festival. YouTube expanded the program in 2018. YouTube also launched YouTube Space in 2012, and has currently expanded to 10 global locations. The Space gives content creators a physical location to learn about producing content as well as providing them with facilities to create content for their YouTube channels.
Experimental (TestTube)
Experimental features of YouTube can be accessed in an area of the site named TestTube.
For example, in October 2009, a comment search feature accessible under /comment_search
was implemented as part of the this program. The feature was removed later.
Later the same year, YouTube Feather was introduced as a lightweight alternative website for countries with limited internet speeds.
Social impact
Both private individuals and large production companies have used YouTube to grow audiences. Independent content creators have built grassroots followings numbering in the thousands at very little cost or effort, while mass retail and radio promotion proved problematic. Concurrently, old media celebrities moved into the website at the invitation of a YouTube management that witnessed early content creators accruing substantial followings and perceived audience sizes potentially larger than that attainable by television. While YouTube’s revenue-sharing “Partner Program” made it possible to earn a substantial living as a video producer—its top five hundred partners each earning more than $100,000 annually and its ten highest-earning channels grossing from $2.5 million to $12 million—in 2012 CMU business editor characterized YouTube as “a free-to-use … promotional platform for the music labels.” In 2013 Forbes‘ Katheryn Thayer asserted that digital-era artists’ work must not only be of high quality, but must elicit reactions on the YouTube platform and social media. Videos of the 2.5% of artists categorized as “mega”, “mainstream” and “mid-sized” received 90.3% of the relevant views on YouTube and Vevo in that year. By early 2013 Billboard had announced that it was factoring YouTube streaming data into calculation of the Billboard Hot 100 and related genre charts.
Observing that face-to-face communication of the type that online videos convey has been “fine-tuned by millions of years of evolution,” TED curator Chris Anderson referred to several YouTube contributors and asserted that “what Gutenberg did for writing, online video can now do for face-to-face communication.” Anderson asserted that it is not far-fetched to say that online video will dramatically accelerate scientific advance, and that video contributors may be about to launch “the biggest learning cycle in human history.” In education, for example, the Khan Academy grew from YouTube video tutoring sessions for founder Salman Khan’s cousin into what Forbes‘ Michael Noer called “the largest school in the world,” with technology poised to disrupt how people learn. YouTube was awarded a 2008 George Foster Peabody Award, the website being described as a Speakers’ Corner that “both embodies and promotes democracy.” The Washington Post reported that a disproportionate share of YouTube’s most subscribed channels feature minorities, contrasting with mainstream television in which the stars are largely white. A Pew Research Center study reported the development of “visual journalism,” in which citizen eyewitnesses and established news organizations share in content creation. The study also concluded that YouTube was becoming an important platform by which people acquire news.
YouTube has enabled people to more directly engage with government, such as in the CNN/YouTube presidential debates (2007) in which ordinary people submitted questions to U.S. presidential candidates via YouTube video, with a techPresident co-founder saying that Internet video was changing the political landscape. Describing the Arab Spring (2010–2012), sociologist Philip N. Howard quoted an activist’s succinct description that organizing the political unrest involved using “Facebook to schedule the protests, Twitter to coordinate, and YouTube to tell the world.” In 2012, more than a third of the U.S. Senate introduced a resolution condemning Joseph Kony 16 days after the “Kony 2012” video was posted to YouTube, with resolution co-sponsor Senator Lindsey Graham remarking that the video “will do more to lead to (Kony’s) demise than all other action combined.”
Conversely, YouTube has also allowed government to more easily engage with citizens, the White House’s official YouTube channel being the seventh top news organization producer on YouTube in 2012 and in 2013 a healthcare exchange commissioned Obama impersonator Iman Crosson’s YouTube music video spoof to encourage young Americans to enroll in the Affordable Care Act (Obamacare)-compliant health insurance. In February 2014, U.S. President Obama held a meeting at the White House with leading YouTube content creators to not only promote awareness of Obamacare but more generally to develop ways for government to better connect with the “YouTube Generation.” Whereas YouTube’s inherent ability to allow presidents to directly connect with average citizens was noted, the YouTube content creators’ new media savvy was perceived necessary to better cope with the website’s distracting content and fickle audience.
Some YouTube videos have themselves had a direct effect on world events, such as Innocence of Muslims (2012) which spurred protests and related anti-American violence internationally. TED curator Chris Anderson described a phenomenon by which geographically distributed individuals in a certain field share their independently developed skills in YouTube videos, thus challenging others to improve their own skills, and spurring invention and evolution in that field. Journalist Virginia Heffernan stated in The New York Times that such videos have “surprising implications” for the dissemination of culture and even the future of classical music.
A 2017 New York Times Magazine article posited that YouTube had become “the new talk radio” for the far right. Almost a year before YouTube’s January 2019 announcement that it would begin a “gradual change” of “reducing recommendations of borderline content and content that could misinform users in harmful ways”, Zeynep Tufekci had written in The New York Times that, “(g)iven its billion or so users, YouTube may be one of the most powerful radicalizing instruments of the 21st century”. Under YouTube’s changes to its recommendation engine, the most recommended channel evolved from conspiracy theorist Alex Jones (2016) to Fox News (2019).
The Legion of Extraordinary Dancers and the YouTube Symphony Orchestra selected their membership based on individual video performances. Further, the cybercollaboration charity video “We Are the World 25 for Haiti (YouTube edition)” was formed by mixing performances of 57 globally distributed singers into a single musical work, with The Tokyo Times noting the “We Pray for You” YouTube cyber-collaboration video as an example of a trend to use crowdsourcing for charitable purposes. The anti-bullying It Gets Better Project expanded from a single YouTube video directed to discouraged or suicidal LGBT teens, that within two months drew video responses from hundreds including U.S. President Barack Obama, Vice President Biden, White House staff, and several cabinet secretaries. Similarly, in response to fifteen-year-old Amanda Todd’s video “My story: Struggling, bullying, suicide, self-harm,” legislative action was undertaken almost immediately after her suicide to study the prevalence of bullying and form a national anti-bullying strategy. In May 2018, London Metropolitan Police claimed that the drill videos that talk about violence give rise to the gang-related violence. YouTube deleted 30 music videos after the complaint.
Viewership
In January 2012, it was estimated that visitors to YouTube spent an average of 15 minutes a day on the site, in contrast to the four or five hours a day spent by a typical US citizen watching television. In 2017, viewers on average watched YouTube on mobile devices for more than an hour every day.
Revenue
Prior to 2020, Google did not provide detailed figures for YouTube’s running costs, and YouTube’s revenues in 2007 were noted as “not material” in a regulatory filing. In June 2008, a Forbes magazine article projected the 2008 revenue at $200 million, noting progress in advertising sales. In 2012, YouTube’s revenue from its ads program was estimated at $3.7 billion. In 2013 it nearly doubled and estimated to hit $5.6 billion according to eMarketer, while others estimated $4.7 billion. The vast majority of videos on YouTube are free to view and supported by advertising. In May 2013, YouTube introduced a trial scheme of 53 subscription channels with prices ranging from $0.99 to $6.99 a month. The move was seen as an attempt to compete with other providers of online subscription services such as Netflix and Hulu.
Google first published exact revenue numbers for YouTube in February 2020 as part of Alphabet’s 2019 financial report. According to Google, YouTube had made US$15.1 billion in ad revenue in 2019, in contrast to US$8.1 billion in 2017 and US$11.1 billion in 2018. YouTube’s revenues made up nearly 10% of the total Alphabet revenue in 2019. These revenues accounted for approximately 20 million subscribers combined between YouTube Premium and YouTube Music subscriptions, and 2 million subscribers to YouTube TV.
Advertisement partnerships
YouTube entered into a marketing and advertising partnership with NBC in June 2006. In March 2007, it struck a deal with BBC for three channels with BBC content, one for news and two for entertainment. In November 2008, YouTube reached an agreement with MGM, Lions Gate Entertainment, and CBS, allowing the companies to post full-length films and television episodes on the site, accompanied by advertisements in a section for U.S. viewers called “Shows”. The move was intended to create competition with websites such as Hulu, which features material from NBC, Fox, and Disney. In November 2009, YouTube launched a version of “Shows” available to UK viewers, offering around 4,000 full-length shows from more than 60 partners. In January 2010, YouTube introduced an online film rentals service, which is only available to users in the United States, Canada, and the UK as of 2010. The service offers over 6,000 films.
Partnership with video creators
In May 2007, YouTube launched its Partner Program (YPP), a system based on AdSense which allows the uploader of the video to share the revenue produced by advertising on the site. YouTube typically takes 45 percent of the advertising revenue from videos in the Partner Program, with 55 percent going to the uploader.
There are over a million members of the YouTube Partner Program. According to TubeMogul, in 2013 a pre-roll advertisement on YouTube (one that is shown before the video starts) cost advertisers on average $7.60 per 1000 views. Usually no more than half of the eligible videos have a pre-roll advertisement, due to a lack of interested advertisers.
YouTube policies restrict certain forms of content from being included in videos being monetized with advertising, including videos containing violence, strong language, sexual content, “controversial or sensitive subjects and events, including subjects related to war, political conflicts, natural disasters and tragedies, even if graphic imagery is not shown” (unless the content is “usually newsworthy or comedic and the creator’s intent is to inform or entertain”), and videos whose user comments contain “inappropriate” content.
In 2013, YouTube introduced an option for channels with at least a thousand subscribers to require a paid subscription in order for viewers to watch videos. In April 2017, YouTube set an eligibility requirement of 10,000 lifetime views for a paid subscription. On January 16, 2018, the eligibility requirement for monetization was changed to 4,000 hours of watchtime within the past 12 months and 1,000 subscribers. The move was seen as an attempt to ensure that videos being monetized did not lead to controversy, but was criticized for penalizing smaller YouTube channels.
YouTube Play Buttons, a part of the YouTube Creator Rewards, are a recognition by YouTube of its most popular channels. The trophies made of nickel plated copper-nickel alloy, golden plated brass, silver plated metal and ruby are given to channels with at least one hundred thousand, a million, ten million and fifty million subscribers, respectively.
Revenue to copyright holders
The majority of YouTube’s advertising revenue goes to the publishers and video producers who hold the rights to their videos; the company retains 45% of the ad revenue. In 2010, it was reported that nearly a third of the videos with advertisements were uploaded without permission of the copyright holders. YouTube gives an option for copyright holders to locate and remove their videos or to have them continue running for revenue. In May 2013, Nintendo began enforcing its copyright ownership and claiming the advertising revenue from video creators who posted screenshots of its games. In February 2015, Nintendo agreed to share the revenue with the video creators.
Controversies
YouTube has a set of community guidelines aimed to reduce abuse of the site’s features. Generally prohibited material includes sexually explicit content, videos of animal abuse, shock videos, content uploaded without the copyright holder’s consent, hate speech, spam, and predatory behavior. Despite the guidelines, YouTube has faced criticism from news sources for retaining content in violation of these guidelines.
Copyrighted material
At the time of uploading a video, YouTube users are shown a message asking them not to violate copyright laws. Despite this advice, many unauthorized clips of copyrighted material remain on YouTube. YouTube does not view videos before they are posted online, and it is left to copyright holders to issue a DMCA takedown notice pursuant to the terms of the Online Copyright Infringement Liability Limitation Act. Any successful complaint about copyright infringement results in a YouTube copyright strike. Three successful complaints for copyright infringement against a user account will result in the account and all of its uploaded videos being deleted. Organizations including Viacom, Mediaset, and the English Premier League have filed lawsuits against YouTube, claiming that it has done too little to prevent the uploading of copyrighted material. Viacom, demanding $1 billion in damages, said that it had found more than 150,000 unauthorized clips of its material on YouTube that had been viewed “an astounding 1.5 billion times”. YouTube responded by stating that it “goes far beyond its legal obligations in assisting content owners to protect their works”.
During the same court battle, Viacom won a court ruling requiring YouTube to hand over 12 terabytes of data detailing the viewing habits of every user who has watched videos on the site. The decision was criticized by the Electronic Frontier Foundation, which called the court ruling “a setback to privacy rights”. In June 2010, Viacom’s lawsuit against Google was rejected in a summary judgment, with U.S. federal Judge Louis L. Stanton stating that Google was protected by provisions of the Digital Millennium Copyright Act. Viacom announced its intention to appeal the ruling. On April 5, 2012, the United States Court of Appeals for the Second Circuit reinstated the case, allowing Viacom’s lawsuit against Google to be heard in court again. On March 18, 2014, the lawsuit was settled after seven years with an undisclosed agreement.
In August 2008, a US court ruled in Lenz v. Universal Music Corp. that copyright holders cannot order the removal of an online file without first determining whether the posting reflected fair use of the material. The case involved Stephanie Lenz from Gallitzin, Pennsylvania, who had made a home video of her 13-month-old son dancing to Prince’s song “Let’s Go Crazy”, and posted the 29-second video on YouTube. In the case of Smith v. Summit Entertainment LLC, professional singer Matt Smith sued Summit Entertainment for the wrongful use of copyright takedown notices on YouTube. He asserted seven causes of action, and four were ruled in Smith’s favor.
In April 2012, a court in Hamburg ruled that YouTube could be held responsible for copyrighted material posted by its users. The performance rights organization GEMA argued that YouTube had not done enough to prevent the uploading of German copyrighted music. YouTube responded by stating:
We remain committed to finding a solution to the music licensing issue in Germany that will benefit artists, composers, authors, publishers, and record labels, as well as the wider YouTube community.
On November 1, 2016, the dispute with GEMA was resolved, with Google content ID being used to allow advertisements to be added to videos with content protected by GEMA.
In April 2013, it was reported that Universal Music Group and YouTube have a contractual agreement that prevents content blocked on YouTube by a request from UMG from being restored, even if the uploader of the video files a DMCA counter-notice. When a dispute occurs, the uploader of the video has to contact UMG. YouTube’s owner Google announced in November 2015 that they would help cover the legal cost in select cases where they believe fair use defenses apply.
Content ID
In June 2007, YouTube began trials of a system for automatic detection of uploaded videos that infringe copyright. Google CEO Eric Schmidt regarded this system as necessary for resolving lawsuits such as the one from Viacom, which alleged that YouTube profited from content that it did not have the right to distribute. The system, which was initially called “Video Identification” and later became known as Content ID, creates an ID File for copyrighted audio and video material, and stores it in a database. When a video is uploaded, it is checked against the database, and flags the video as a copyright violation if a match is found. When this occurs, the content owner has the choice of blocking the video to make it unviewable, tracking the viewing statistics of the video, or adding advertisements to the video.
By 2010, YouTube had “already invested tens of millions of dollars in this technology”.
In 2011, YouTube described Content ID as “very accurate in finding uploads that look similar to reference files that are of sufficient length and quality to generate an effective ID File”.
By 2012, Content ID accounted for over a third of the monetized views on YouTube.
An independent test in 2009 uploaded multiple versions of the same song to YouTube and concluded that while the system was “surprisingly resilient” in finding copyright violations in the audio tracks of videos, it was not infallible. The use of Content ID to remove material automatically has led to controversy in some cases, as the videos have not been checked by a human for fair use. If a YouTube user disagrees with a decision by Content ID, it is possible to fill in a form disputing the decision.
Before 2016, videos were not monetized until the dispute was resolved. Since April 2016, videos continue to be monetized while the dispute is in progress, and the money goes to whoever won the dispute. Should the uploader want to monetize the video again, they may remove the disputed audio in the “Video Manager”. YouTube has cited the effectiveness of Content ID as one of the reasons why the site’s rules were modified in December 2010 to allow some users to upload videos of unlimited length.
Controversial videos
YouTube has also faced criticism over the handling of offensive content in some of its videos. The uploading of videos containing defamation, pornography, and material encouraging criminal conduct is forbidden by YouTube’s “Community Guidelines”. YouTube relies on its users to flag the content of videos as inappropriate, and a YouTube employee will view a flagged video to determine whether it violates the site’s guidelines.
To limit the spread of misinformation and fake news via YouTube, it has rolled out a comprehensive policy regarding how to planned to deal with technically manipulated videos.
YouTube has also been criticized for suppressing opinions dissenting from governments’ positions, especially related to the COVID-19 pandemic.
Public disasters
Controversial content has included material relating to Holocaust denial and the Hillsborough disaster, in which 96 football fans from Liverpool were crushed to death in 1989. In July 2008, the Culture and Media Committee of the House of Commons of the United Kingdom stated that it was “unimpressed” with YouTube’s system for policing its videos, and argued that “proactive review of content should be standard practice for sites hosting user-generated content”. YouTube responded by stating:
We have strict rules on what’s allowed, and a system that enables anyone who sees inappropriate content to report it to our 24/7 review team and have it dealt with promptly. We educate our community on the rules and include a direct link from every YouTube page to make this process as easy as possible for our users. Given the volume of content uploaded on our site, we think this is by far the most effective way to make sure that the tiny minority of videos that break the rules come down quickly. (July 2008)
Imam Anwar al-Awlaki
In October 2010, U.S. Congressman Anthony Weiner urged YouTube to remove from its website videos of imam Anwar al-Awlaki. YouTube pulled some of the videos in November 2010, stating they violated the site’s guidelines. In December 2010, YouTube added the ability to flag videos for containing terrorism content.
PRISM
Following media reports about PRISM, NSA’s massive electronic surveillance program, in June 2013, several technology companies were identified as participants, including YouTube. According to leaks from the program, YouTube joined the PRISM program in 2010.
Restricted monetization policies
YouTube’s policies on “advertiser-friendly content” restrict what may be incorporated into videos being monetized; this includes strong violence, language, sexual content, and “controversial or sensitive subjects and events, including subjects related to war, political conflicts, natural disasters and tragedies, even if graphic imagery is not shown”, unless the content is “usually newsworthy or comedic and the creator’s intent is to inform or entertain”. In September 2016, after introducing an enhanced notification system to inform users of these violations, YouTube’s policies were criticized by prominent users, including Phillip DeFranco and Vlogbrothers. DeFranco argued that not being able to earn advertising revenue on such videos was “censorship by a different name”. A YouTube spokesperson stated that while the policy itself was not new, the service had “improved the notification and appeal process to ensure better communication to our creators”. Boing Boing reported in 2019 that LGBT keywords resulted in demonetization.
Advertiser mass boycott
In March 2017, the government of the United Kingdom pulled its advertising campaigns from YouTube, after reports that its ads had appeared on videos containing extremist content. The government demanded assurances that its advertising would “be delivered safely and appropriately”. The Guardian newspaper, as well as other major British and U.S. brands, similarly suspended their advertising on YouTube in response to their advertising appearing near offensive content. Google stated that it had “begun an extensive review of our advertising policies and have made a public commitment to put in place changes that give brands more control over where their ads appear”. In early April 2017, the YouTube channel h3h3Productions presented evidence claiming that a Wall Street Journal article had fabricated screenshots showing major brand advertising on an offensive video containing Johnny Rebel music overlaid on a Chief Keef music video, citing that the video itself had not earned any ad revenue for the uploader. The video was retracted after it was found that the ads had been triggered by the use of copyrighted content in the video.
On April 6, 2017, YouTube announced that to “ensure revenue only flows to creators who are playing by the rules”, it would change its practices to require that a channel undergo a policy compliance review, and have at least 10,000-lifetime views, before they may join the Partner Program.
Logan Paul corpse scandal
In January 2018, YouTube creator Logan Paul faced criticism for a video he had uploaded from a trip to Japan, where he encountered a body of a suicide death in the Aokigahara forest. The corpse was visible in the video, although its face was censored. The video proved controversial due to its content, with its handling of the subject matter being deemed insensitive by critics. On January 10—eleven days after the video was published—YouTube announced that it would cut Paul from the Google Preferred advertising program. Six days later, YouTube announced tighter thresholds for the partner program to “significantly improve our ability to identify creators who contribute positively to the community”, under which channels must have at least 4,000 hours of watch time within the past 12 months and at least 1,000 subscribers. YouTube also announced that videos approved for the Google Preferred program would become subject to manual review and that videos would be rated based on suitability (with advertisers allowed to choose).
These changes led to further criticism of YouTube from independent channels, who alleged that the service had been changing its algorithms to give higher prominence to professionally produced content (such as celebrities, music videos, and clips from late night talk shows), that attract wide viewership and has a lower risk of alienating mainstream advertisers, at the expense of the creators that had bolstered the service’s popularity.
Conspiracy theories and fringe discourse
YouTube has been criticized for using an algorithm that gives great prominence to videos that promote conspiracy theories, falsehoods and incendiary fringe discourse. According to an investigation by The Wall Street Journal, “YouTube’s recommendations often lead users to channels that feature conspiracy theories, partisan viewpoints and misleading videos, even when those users haven’t shown interest in such content. When users show a political bias in what they choose to view, YouTube typically recommends videos that echo those biases, often with more-extreme viewpoints.” When users search for political or scientific terms, YouTube’s search algorithms often give prominence to hoaxes and conspiracy theories. After YouTube drew controversy for giving top billing to videos promoting falsehoods and conspiracy when people made breaking-news queries during the 2017 Las Vegas shooting, YouTube changed its algorithm to give greater prominence to mainstream media sources. In 2018, it was reported that YouTube was again promoting fringe content about breaking news, giving great prominence to conspiracy videos about Anthony Bourdain’s death.
In 2017, it was revealed that advertisements were being placed on extremist videos, including videos by rape apologists, anti-Semites, and hate preachers who received ad payouts. After firms started to stop advertising on YouTube in the wake of this reporting, YouTube apologized and said that it would give firms greater control over where ads got placed.
Alex Jones, known for right-wing conspiracy theories, had built a massive audience on YouTube. YouTube drew criticism in 2018 when it removed a video from Media Matters compiling offensive statements made by Jones, stating that it violated its policies on “harassment and bullying”. On August 6, 2018, however, YouTube removed Alex Jones’ YouTube page following a content violation.
University of North Carolina professor Zeynep Tufekci has referred to YouTube as “The Great Radicalizer”, saying “YouTube may be one of the most powerful radicalizing instruments of the 21st century.” Jonathan Albright of the Tow Center for Digital Journalism at Columbia University described YouTube as a “conspiracy ecosystem”.
In January 2019, YouTube said that it had introduced a new policy starting in the United States intended to stop recommending videos containing “content that could misinform users in harmful ways.” YouTube gave flat earth theories, miracle cures, and 9/11 trutherism as examples. Efforts within YouTube engineering to stop recommending borderline extremist videos falling just short of forbidden hate speech, and track their popularity were originally rejected because they could interfere with viewer engagement. In late 2019, the site began implementing measures directed towards “raising authoritative content and reducing borderline content and harmful misinformation.”
A July 2019 study based on ten YouTube searches using the Tor Browser related to the climate and climate change, the majority of videos were videos that communicated views contrary to the scientific consensus on climate change.
A 2019 BBC investigation of YouTube searches in ten different languages found that YouTube’s algorithm promoted health misinformation, including fake cancer cures. In Brazil, YouTube has been linked to pushing pseudoscientific misinformation on health matters, as well as elevated far-right fringe discourse and conspiracy theories.
Following the dissemination via YouTube of misinformation related to the COVID-19 pandemic that 5G communications technology was responsible for the spread of coronavirus disease 2019 which led to numerous 5G towers in the United Kingdom to be destroyed, YouTube removed all such videos linking 5G and the coronavirus in this manner.
Hateful content
Before 2019, YouTube has taken steps to remove specific videos or channels related to supremacist content that had violated its acceptable use policies but otherwise did not have site-wide policies against hate speech.
In the wake of the March 2019 Christchurch mosque attacks, YouTube and other sites like Facebook and Twitter that allowed user-submitted content drew criticism for doing little to moderate and control the spread of hate speech, which was considered to be a factor in the rationale for the attacks. These platforms were pressured to remove such content, but in an interview with The New York Times, YouTube’s chief product officer Neal Mohan said that unlike content such as ISIS videos which take a particular format and thus easy to detect through computer-aided algorithms, general hate speech was more difficult to recognize and handle, and thus could not readily take action to remove without human interaction.
YouTube joined an initiative led by France and New Zealand with other countries and tech companies in May 2019 to develop tools to be used to block online hate speech and to develop regulations, to be implemented at the national level, to be levied against technology firms that failed to take steps to remove such speech, though the United States declined to participate. Subsequently, on June 5, 2019, YouTube announced a major change to its terms of service, “specifically prohibiting videos alleging that a group is superior in order to justify discrimination, segregation or exclusion based on qualities like age, gender, race, caste, religion, sexual orientation or veteran status.” YouTube identified specific examples of such videos as those that “promote or glorify Nazi ideology, which is inherently discriminatory”. YouTube further stated it would “remove content denying that well-documented violent events, like the Holocaust or the shooting at Sandy Hook Elementary, took place.”
In June 2020, YouTube banned several channels associated with white supremacy, including those of Stefan Molyneux, David Duke, and Richard B. Spencer, asserting these channels violated their policies on hate speech. The ban occurred the same day that Reddit announced the ban on several hate speech sub-forums including r/The_Donald.
Child protection
Leading into 2017, there was a significant increase in the number of videos related to children, coupled between the popularity of parents vlogging their family’s activities, and previous content creators moving away from content that often was criticized or demonetized into family-friendly material. In 2017, YouTube reported that time watching family vloggers had increased by 90%. However, with the increase in videos featuring children, the site began to face several controversies related to child safety. During Q2 2017, the owners of popular channel FamilyOFive, which featured themselves playing “pranks” on their children, were accused of child abuse. Their videos were eventually deleted, and two of their children were removed from their custody. A similar case happened in 2019 when the owners of the channel Fantastic Adventures was accused of abusing her adopted children. Her videos would later be deleted.
Later that year, YouTube came under criticism for showing inappropriate videos targeted at children and often featuring popular characters in violent, sexual or otherwise disturbing situations, many of which appeared on YouTube Kids and attracted millions of views. The term “Elsagate” was coined on the Internet and then used by various news outlets to refer to this controversy. On November 11, 2017, YouTube announced it was strengthening site security to protect children from unsuitable content. Later that month, the company started to mass delete videos and channels that made improper use of family friendly characters. As part as a broader concern regarding child safety on YouTube, the wave of deletions also targeted channels which showed children taking part in inappropriate or dangerous activities under the guidance of adults. Most notably, the company removed Toy Freaks, a channel with over 8.5 million subscribers, that featured a father and his two daughters in odd and upsetting situations. According to analytics specialist SocialBlade, it earned up to £8.7 million annually prior to its deletion.
Even for content that appears to be aimed at children and appears to contain only child-friendly content, YouTube’s system allows for anonymity of who uploads these videos. These questions have been raised in the past, as YouTube has had to remove channels with children’s content which, after becoming popular, then suddenly include inappropriate content masked as children’s content. Alternative, some of the most-watched children’s programming on YouTube comes from channels who have no identifiable owners, raising concerns of intent and purpose. One channel that had been of concern was “Cocomelon” which provided numerous mass-produced animated videos aimed at children. Up through 2019, it had drawn up to US$10 million a month in ad revenue, and was one of the largest kid-friendly channels on YouTube before 2020. Ownership of Cocomelon was unclear outside of its ties to “Treasure Studio”, itself an unknown entity, raising questions as to the channel’s purpose, but Bloomberg News had been able to confirm and interview the small team of American owners in February 2020 regarding “Cocomelon”, who stated their goal for the channel was to simply entertain children, wanting to keep to themselves to avoid attention from outside investors. The anonymity of such channel raise concerns because of the lack of knowledge of what purpose they are trying to serve. The difficulty to identify who operates these channels “adds to the lack of accountability”, according to Josh Golin of the Campaign for a Commercial-Free Childhood, and educational consultant Renée Chernow-O’Leary found the videos were designed to entertain with no intent to educate, all leading to both critics and parents to be concerns for their children becoming too enraptured by the content from these channels. Content creators that earnestly make kid-friendly videos have found it difficult to compete with larger channels like ChuChu TV, unable to produce content at the same rate as these large channels, and lack the same means of being promoted through YouTube’s recommendation algorithms that the larger animated channel networks have shared.
In January 2019, YouTube officially banned videos containing “challenges that encourage acts that have an inherent risk of severe physical harm” (such as, for example, the Tide Pod Challenge), and videos featuring pranks that “make victims believe they’re in physical danger” or cause emotional distress in children.
Sexualization of children
Also in November 2017, it was revealed in the media that many videos featuring children—often uploaded by the minors themselves, and showing innocent content such as the children playing with toys or performing gymnastics—were attracting comments from pedophiles with predators finding the videos through private YouTube playlists or typing in certain keywords in Russian. Other child-centric videos originally uploaded to YouTube began propagating on the dark web, and uploaded or embedded onto forums known to be used by pedophiles.
As a result of the controversy, which added to the concern about “Elsagate”, several major advertisers whose ads had been running against such videos froze spending on YouTube. In December 2018, The Times found more than 100 grooming cases in which children were manipulated into sexually implicit behavior (such as taking off clothes, adopting sexualised poses and touching other children inappropriately) by strangers. After a reporter flagged the videos in question, half of them were removed, and the rest were removed after The Times contacted YouTube’s PR department.
In February 2019, YouTube vlogger Matt Watson identified a “wormhole” that would cause the YouTube recommendation algorithm to draw users into this type of video content, and make all of that user’s recommended content feature only these types of videos. Most of these videos had comments from sexual predators commenting with timestamps of when the children were shown in compromising positions, or otherwise making indecent remarks. In some cases, other users had reuploaded the video in unlisted form but with incoming links from other videos, and then monetized these, propagating this network. In the wake of the controversy, the service reported that they had deleted over 400 channels and tens of millions of comments, and reported the offending users to law enforcement and the National Center for Missing and Exploited Children. A spokesperson explained that “any content—including comments—that endangers minors is abhorrent and we have clear policies prohibiting this on YouTube. There’s more to be done, and we continue to work to improve and catch abuse more quickly.” Despite these measures, AT&T, Disney, Dr. Oetker, Epic Games, and Nestlé all pulled their advertising from YouTube.
Subsequently, YouTube began to demonetize and block advertising on the types of videos that have drawn these predatory comments. The service explained that this was a temporary measure while they explore other methods to eliminate the problem. YouTube also began to flag channels that predominantly feature children, and preemptively disable their comments sections. “Trusted partners” can request that comments be re-enabled, but the channel will then become responsible for moderating comments. These actions mainly target videos of toddlers, but videos of older children and teenagers may be protected as well if they contain actions that can be interpreted as sexual, such as gymnastics. YouTube stated it was also working on a better system to remove comments on other channels that matched the style of child predators.
A related attempt to algorithmically flag videos containing references to the string “CP” (an abbreviation of child pornography) resulted in a number of prominent false positives involving unrelated topics using the same abbreviation, including videos related to the mobile video game Pokémon Go (which uses “CP” as an abbreviation of the statistic “Combat Power”), and Club Penguin. YouTube apologized for the errors, and reinstated the affected videos. Separately, online trolls have attempted to have videos flagged for takedown or removal by commenting with statements similar to what the child predators had said; this activity became an issue during the PewDiePie vs T-Series rivalry in early 2019. YouTube stated they do not take action on any video with these comments but those that they have flagged that are likely to draw child predator activity.
In June 2019, The New York Times cited researchers who found that users who watched erotic videos could be recommended seemingly innocuous videos of children. As a result, Senator Josh Hawley stated plans to introduce federal legislation that would ban YouTube and other video sharing sites from including videos that predominantly feature minors as “recommended” videos, excluding those that were “professionally produced”, such as videos of televised talent shows. YouTube has suggested potential plans to remove all videos featuring children from the main YouTube site and transferring them to the YouTube Kids site where they would have stronger controls over the recommendation system, as well as other major changes on the main YouTube site to the recommended feature and autoplay system.
COPPA settlement
In April 2018, a coalition of 23 groups (including the CCFC, CDD, as well as Common Sense Media) filed a complaint with the Federal Trade Commission, alleging that YouTube collected information from users under the age of 13 without parental consent, in violation of the Children’s Online Privacy Protection Act (COPPA).
In September 2019, YouTube was fined $170 million by the FTC for collecting personal information from minors under the age of 13 (in particular, viewing history) without parental consent, in order to allow channel operators to serve targeted advertising on their videos. In particular, the FTC ruled that YouTube was partly liable under COPPA, as the service’s rating and curation of content as being suitable for children constituted the targeting of the website towards children. In order to comply with the settlement, YouTube was ordered to “develop, implement, and maintain a system for Channel Owners to designate whether their Content on the YouTube Service is directed to Children.” YouTube also announced that it would invest $100 million over the next three years to support the creation of “thoughtful, original children’s content”.
YouTube began to implement its compliance approach in November 2019, in line with its settlement with the FTC. All channels must declare if their content is “made for kids”, either as a blanket claim for their entire channel, or on a per-video basis. The company states that a video is considered “made for kids” if its primary audience is children, or is “directed” to children based on various factors as guidelines (even if they are not the primary audience), including use of child actors, “characters, celebrities, or toys that appeal to children”, depictions of “activities that appeal to children, such as play-acting, simple songs or games, or early education”, and poems, songs, and stories intended for children, among others. YouTube will employ machine learning to find videos that they believe are clearly “made for kids” and automatically mark them as such, but will not help or advise content creators for videos that fall into unclear categories, as this constitutes legal advice. In order to prevent data to be collected from minors without consent, videos marked as being “made for kids” were automatically reduced in functionality beginning on January 6, 2020. As a result, social and community features such as end screens and other widgets, notification functions, and comments are disabled, and videos can only be monetized with contextual advertising based on the video’s metadata. Further, liability for failing proper marking channels or videos as “made for kids” would fall onto the channel owners, with the FTC able to issue up to $42,000 fines per infringing video, though the FTC clarified that the amount would be based on “a company’s financial condition and the impact a penalty could have on its ability to stay in business”.
The new policies have faced criticism, with some channel owners having considered YouTube and the FTC’s guidance to be unclear in certain edge cases, such as video gaming (where content may typically be directed towards teens and young adults, but may still contain characters that appeal to children). They also noted that according to YouTube, a lack of targeted ads can reduce a video’s revenue, and that the lack of social features may impact the ability for their videos to receive promotion. Videos marked as “made for kids” are also excluded from Google search engine results, further lowering revenue for content creators. Content creators who were unsure of whether their content was “made for kids” argued they would either need to preemptively mark their videos as such or make their videos private, or otherwise would be at risk of being fined by the FTC.
The legal language of COPPA offered the ability for content to be marked for “mixed audience”, which would allow for data collection from the viewers once the viewer had identified themselves of being 13 years or older. YouTube’s decision not to include a “mixed audience” as a third option has been criticized, since this option would alleviate content creator concerns. YouTube had stated in their information page related to the COPPA requirements that “there are some complexities with the mixed audience category” which they have submitted to the FTC during public comment periods, and in the interim “decided to streamline the options for creators by creating a single ‘Made for kids’ category to avoid further confusion in an already unclear space.”
FTC commissioner Rebecca Kelly Slaughter noted in dissenting remarks that many of the child-directed channels on YouTube were run by owners outside of the U.S., which may fall outside the jurisdiction of COPPA and the FTC’s “practical reach”. The FTC issued a blog post on November 22, 2019, to clarify what it considered “made for kids” and that several factors will be used to make this determination, and that it recognized that some types of content, like animated programming, can appeal to all ages and would not be immediately become subject to COPPA’s requirements. The FTC also directed creators to its original complaint against YouTube, identifying channels and video content that they considered to be under COPPA that was the basis of their case.
On December 10, 2019, citing the aforementioned criticism, YouTube filed comments with the FTC requesting clarity on its rules, citing the aforementioned edge cases. YouTube started treating all videos designated as “made for kids” as liable under COPPA on January 6, 2020.
User comments
Most videos enable users to leave comments, and these have attracted attention for the negative aspects of both their form and content. In 2006, Time praised Web 2.0 for enabling “community and collaboration on a scale never seen before”, and added that YouTube “harnesses the stupidity of crowds as well as its wisdom. Some of the comments on YouTube make you weep for the future of humanity just for the spelling alone, never mind the obscenity and the naked hatred”. The Guardian in 2009 described users’ comments on YouTube as:
Juvenile, aggressive, misspelled, sexist, homophobic, swinging from raging at the contents of a video to providing a pointlessly detailed description followed by a LOL, YouTube comments are a hotbed of infantile debate and unashamed ignorance—with the occasional burst of wit shining through.
In September 2008, The Daily Telegraph commented that YouTube was “notorious” for “some of the most confrontational and ill-formed comment exchanges on the internet”, and reported on YouTube Comment Snob, “a new piece of software that blocks rude and illiterate posts”. The Huffington Post noted in April 2012 that finding comments on YouTube that appear “offensive, stupid and crass” to the “vast majority” of the people is hardly difficult.
On November 6, 2013, Google implemented a comment system oriented on Google+ that required all YouTube users to use a Google+ account in order to comment on videos. The stated motivation for the change was giving creators more power to moderate and block comments, thereby addressing frequent criticisms of their quality and tone. The new system restored the ability to include URLs in comments, which had previously been removed due to problems with abuse. In response, YouTube co-founder Jawed Karim posted the question “why the fuck do I need a google+ account to comment on a video?” on his YouTube channel to express his negative opinion of the change. The official YouTube announcement received 20,097 “thumbs down” votes and generated more than 32,000 comments in two days. Writing in the Newsday blog Silicon Island, Chase Melvin noted that “Google+ is nowhere near as popular a social media network as Facebook, but it’s essentially being forced upon millions of YouTube users who don’t want to lose their ability to comment on videos” and “Discussion forums across the Internet are already bursting with outcry against the new comment system”. In the same article Melvin goes on to say:
Perhaps user complaints are justified, but the idea of revamping the old system isn’t so bad. Think of the crude, misogynistic and racially-charged mudslinging that has transpired over the last eight years on YouTube without any discernible moderation. Isn’t any attempt to curb unidentified libelers worth a shot? The system is far from perfect, but Google should be lauded for trying to alleviate some of the damage caused by irate YouTubers hiding behind animosity and anonymity.
On July 27, 2015, Google announced in a blog post that it would be removing the requirement to sign up to a Google+ account to post comments to YouTube.
On November 3, 2016, YouTube announced a trial scheme which allows the creators of videos to decide whether to approve, hide or report the comments posted on videos based on an algorithm that detects potentially offensive comments. Creators may also choose to keep or delete comments with links or hashtags in order to combat spam. They can also allow other users to moderate their comments.
In May 2020, it was noted that comments in videos and live streams containing the mandarin phrases “共匪” (“communist bandit”) or “五毛” (“50-cent party”) were automatically deleted within 15 seconds. The censorship was considered odd as YouTube is blocked in China.
In December 2020 it was reported that YouTube is launching a new feature that will warn users who post a comment that “may be offensive to others.”
View counts
In December 2012, two billion views were removed from the view counts of Universal and Sony music videos on YouTube, prompting a claim by The Daily Dot that the views had been deleted due to a violation of the site’s terms of service, which ban the use of automated processes to inflate view counts. This was disputed by Billboard, which said that the two billion views had been moved to Vevo, since the videos were no longer active on YouTube. On August 5, 2015, YouTube patched the formerly notorious behaviour which caused a video’s view count to freeze at “301” (later “301+”) until the actual count was verified to prevent view count fraud. YouTube view counts once again updated in real time.
Other algorithm design or implementation issues
In 2018, YouTube introduced a system that would automatically add information boxes to videos that its algorithms determined may present conspiracy theories and other fake news, filling the infobox with content from Encyclopedia Britannica and Wikipedia as a means to inform users to minimize misinformation propagation without impacting freedom of speech. In the wake of the Notre-Dame de Paris fire on April 15, 2019, several user-uploaded videos of the landmark fire were flagged by YouTube’ system automatically with an Encyclopedia Britannica article on the false conspiracy theories around the September 11 attacks. Several users complained to YouTube about this inappropriate connection. YouTube officials apologized for this, stating that their algorithms had misidentified the fire videos and added the information block automatically, and were taking steps to remedy this.
Five leading content creators whose channels were based on LGBTQ+ materials filed a federal lawsuit against YouTube in August 2019, alleging that YouTube’s algorithms diverts discovery away from their channels, impacting their revenue. The plaintiffs claimed that the algorithms discourage content with words like “lesbian” or “gay”, which would be predominant in their channels’ content, and because of YouTube’s near-monopolization of online video services, they are abusing that position.
Content moderation
YouTube contracts companies to hire content moderators, who view content flagged as potentially violating YouTube’s content policies and determines if they should be removed. In September 2020, a class-action suit was filed by a former content moderator who reported developing post-traumatic stress disorder (PTSD) after an 18-month period on the job. The former content moderator said that she was regularly made to exceed YouTube’s stated limited of four hours per day of viewing graphic content. The lawsuit alleges that YouTube’s contractors gave little to no training or support for its moderator’s mental health, made prospective employees sign NDAs before showing them any examples of content they would see while reviewing, and censored all mention of trauma from its internal forums. It also purports that requests for extremely graphic content to be blurred, reduced in size or made monochrome, per recommendations from the National Center for Missing and Exploited Children, were rejected by YouTube as not a high priority for the company.
Addy A-Game and Street Attraction channels design
In October 2019, YouTube deactivated two channels run by “pick-up artists” after a BBC investigation into the online industry. The company added: “YouTube strictly prohibits explicit sexual, graphic or harassing content. Nothing is more important than protecting the safety of our community, and we will continue to review and refine our policies in this area.”
Censorship and bans
YouTube has been censored, filtered, or banned for a variety of reasons, including:
- Limiting public access and exposure to content that may ignite social or political unrest.
- Preventing criticism of a ruler (e.g. in North Korea), government (e.g. in China) or its actions (e.g. in Morocco), government officials (e.g. in Turkey and Libya), or religion (e.g. in Pakistan).
- Morality-based laws, e.g. in Iran.
Access to specific videos is sometimes prevented due to copyright and intellectual property protection laws (e.g. in Germany), violations of hate speech, and preventing access to videos judged inappropriate for youth, which is also done by YouTube with the YouTube Kids app and with “restricted mode”. Businesses, schools, government agencies, and other private institutions often block social media sites, including YouTube, due to its bandwidth limitations and the site’s potential for distraction.
As of 2018, public access to YouTube is blocked in many countries, including China, North Korea, Iran, Syria, Turkmenistan, Uzbekistan, Tajikistan, Eritrea, Sudan and South Sudan. In some countries, YouTube is blocked for more limited periods of time such as during periods of unrest, the run-up to an election, or in response to upcoming political anniversaries. In cases where the entire site is banned due to one particular video, YouTube will often agree to remove or limit access to that video in order to restore service.
Reports emerged that since October 2019, comments posted with Chinese characters insulting the Communist Party of China (共匪 or “communist bandit”) or (五毛 or “50 Cent Party”, referring to state-sponsored commentators) were being automatically deleted within 15 seconds.
Specific incidents where YouTube has been blocked include:
- Thailand blocked access in April 2007 over a video said to be insulting the Thai king.
- Morocco blocked access in May 2007, possibly as a result of videos critical of Morocco’s occupation of Western Sahara. YouTube became accessible again on May 30, 2007, after Maroc Telecom unofficially announced that the denied access to the website was a mere “technical glitch”.
- Turkey blocked access between 2008 and 2010 after controversy over videos deemed insulting to Mustafa Kemal Atatürk. In November 2010, a video of the Turkish politician Deniz Baykal caused the site to be blocked again briefly, and the site was threatened with a new shutdown if it did not remove the video. During the two and a half-year block of YouTube, the video-sharing website remained the eighth-most-accessed site in Turkey. In 2014, Turkey blocked the access for the second time, after “a high-level intelligence leak.”
- Pakistan blocked access on February 23, 2008, because of “offensive material” towards the Islamic faith, including display of the Danish cartoons of Muhammad. This led to a near global blackout of the YouTube site for around two hours, as the Pakistani block was inadvertently transferred to other countries. On February 26, 2008, the ban was lifted after the website had removed the objectionable content from its servers at the request of the government. Many Pakistanis circumvented the three-day block by using virtual private network software. In May 2010, following the Everybody Draw Mohammed Day, Pakistan again blocked access to YouTube, citing “growing sacrilegious content”. The ban was lifted on May 27, 2010, after the website removed the objectionable content from its servers at the request of the government. However, individual videos deemed offensive to Muslims posted on YouTube will continue to be blocked. Pakistan again placed a ban on YouTube in September 2012, after the site refused to remove the film Innocence of Muslims, with the ban still in operation as of September 2013. The ban was lifted in January 2016 after YouTube launched a Pakistan-specific version.
- Libya blocked access on January 24, 2010, because of videos that featured demonstrations in the city of Benghazi by families of detainees who were killed in Abu Salim prison in 1996, and videos of family members of Libyan leader Muammar Gaddafi at parties. The blocking was criticized by Human Rights Watch. In November 2011, after the Libyan Civil War, YouTube was once again allowed in Libya.
- Afghanistan, Bangladesh, Pakistan, and Sudan blocked access in September 2012 following controversy over a 14-minute trailer for the film Innocence of Muslims which had been posted on the site. A court in the southern Russian Republic of Chechnya ruled Innocence of Muslims should be banned. In Libya and Egypt, it was blamed for violent protests. YouTube stated that “This video—which is widely available on the Web—is clearly within our guidelines and so will stay on YouTube. However, given the very difficult situation in Libya and Egypt we have temporarily restricted access in both countries.”