Production & Exchange Formats for 3DTV Programmes
The purpose of this EBU recommendation is to give technical aid to broadcasters who intend to use current (or future) 2D HDTV infrastructures to produce 3DTV programmes.
A curation about new media technologies
The purpose of this EBU recommendation is to give technical aid to broadcasters who intend to use current (or future) 2D HDTV infrastructures to produce 3DTV programmes.
Online and mobile viewing of widely-available, high-quality video content including TV programming, movies, sports events, and news is now poised to go mainstream. Driven by the recent availability of low-cost, high-resolution desktop/laptop/tablet PCs, smart phones, set-top boxes and now Ethernet-enabled TV sets, consumers have rapidly moved through the ‘novelty’ phase of acceptance into expectation that any media should be available essentially on any device over any network connection. Whether regarded as a disruption for cable TV, telco or satellite TV providers, or an opportunity for service providers to extend TV services onto the web for on-demand, time-shifted and place-shifted programming environments – often referred to as ‘three screen delivery’ or ‘TV Anywhere’ – this new video delivery model is here to stay.
While tremendous advancements in core and last mile bandwidth have been achieved in the last decade around the world – primarily driven by web-based data consumption – video traffic represents a quantum leap in bandwidth requirements. Coupled with the fact that the Internet at large is not a managed quality-of-service environment, requires that new methods of video transport be considered to provide the quality of video experience across any device and network that we have come to expect from managed TV-delivery networks.
The evolution of video delivery transport has led to a new set of de facto standard adaptive delivery protocols from Apple, Microsoft, Adobe that are now positioned for broad adoption. Consequently, networks must now be equipped with servers that can take high-quality video content from its source live or file format and ‘package’ it for transport to devices ready to accept these new delivery protocols.
Video Delivery Background
The Era of Stateful Protocols
For many years, stateful protocols including Real Time Streaming Protocol (RTSP), Adobe’s Real Time Messaging Protocol (RTMP), and Real Networks' RTSP over Real Data Transport (RDT) protocol were utilized to stream video content to desktop and mobile clients. Stateful protocols require that from the time a client connects to a streaming server until the time it disconnects, the server tracks client state. If the client needs to perform any video session control commands like start, stop, pause or fast-forward it must do so by communicating state information back to the streaming server.
Once a session between the client and the server has been established, the server sends media as a stream of small packets typically representing a few milliseconds of video. These packets can be transmitted over UDP or TCP. TCP overcomes firewall blocking of UDP packets, but may also incur increased latency as packets are sent, and resent if not acknowledged, until received at the far end.
These protocols served the market well, particularly during the era where desktop and mobile device experiences were limited by frequency, quality, duration, screen/window size/resolution, constrained processor, memory and storage capabilities of mobile devices, etc.
However, the above experience factors have all changed dramatically in the last few years. And that has exposed a number of stateful protocol implementation weaknesses:
3ality Digital, considered to be the leading innovator of the most sophisticated S3D production technology in the industry, announced it has acquired Element Technica, a company long-known for its manufacturing expertise, accessories, and mechanical engineering of motorized S3D camera rigs.
3ality Digital is now 3ality Technica, and with its acquisition of Element Technica, 3ality Technica now provides all of the control, accuracy, breadth, automation, modularity, accessories, and design of both existing product lines. Physically, the companies will combine in an expansion to the 3ality Digital headquarters in Burbank, CA.
Source: 3ality Digital
TvTak is the fastest way to tell your friends what you’re watching on Facebook, Twitter or Email. All you need is an iPhone and regular TV. The TvTak advanced recognition technology does the rest.
When you’re watching TV and see something interesting, just point your iPhone to the screen. TvTak instantly detects what show or commercial you’re watching. It only takes one second. No need for typing the name of the show, or for cumbersome check-ins among long lists of programs.
Walt Disney Studios' vice president of production technology believes that 3D rigs have a limited lifespan and that a more post-production oriented approach to filming stereo feature film is required.
Disney’s Howard Lukk argues for a hybrid approach to stereoscopic filmmaking which would supplement a 2D camera with smaller ‘witness’ cameras to pick up the 3D volumes, then apply algorithms at a visual effects house or a conversion company to create the 3D and free the filmmaker from cumbersome on-set equipment.
For Lukk, the problem is that 3D rigs are complicated to build and harder still to calibrate for true accuracy.
“There are enough things for the DOP, director and camera operators to try to track on the set as it is, without having to track interaxial and convergence,” he said. “We are making it more complicated on the set, where I think it needs to be less complicated.”
There should not be a battle between natively captured 3D and conversion, he contends. “We should start to ask if there is another method – even combine the two? We really should be looking at all methods to make better quality 3D. In the end, if 3D is not good quality, we are going to kill this stereoscopic industry just as it is re-emerging.”
If 3D camera rigs are not the future of the industry, Lukk suggests that a hybrid approach will develop which will be a combination of capturing volumetric space on set and being able to produce the 3D in a post-production environment at the back end.
“This will give you much more versatility in manipulating the images. This idea feeds on the idea of computational cinematography conceived by Marc Levoy (a computer graphics research at Stanford University) a few years ago. Basically this says that if we capture things in a certain way, we can compute things that we really need in the back end.
“You can be less accurate on the front end. Adobe has been doing a lot of work in this area, where you can refocus the image after the event. You can apply this concept to high dynamic range and higher frame rates.”
Disney is currently researching this method at Disney Research in Zurich, Lukk added. In addition Lukk says that research is also being conducted at the Fraunhofer Institute in Germany.
“I think eventually we’ll get back to capturing the volumetric space and allowing cinematographers and directors to do what they do best - that is, capturing the performance,” he said.
Source: TVB Europe
Light field displays can provide glasses-free 3D images, under the right circumstances. One of their big advantages is they are independent of "sweet spots" and the 3D image can be seen from any location in the viewing area. With proper design, they can show both horizontal and vertical parallax and they have "look around" capability, allowing you to see what is behind a foreground object by moving sideways in the viewing region.
Unfortunately, they are not problem free. One solution involves multiple projectors, up to 40 or more, in a rear projection configuration, and special processing software/hardware to drive the projectors. That is not a hand-held system by any means. Another issue has been their modest image quality, good enough for some digital signage applications, perhaps, but not good enough for TV.
In a paper at SIGGRAPH 2011, which is continuing through tomorrow in Vancouver, British Columbia, Gordon Wetzstein from the University of British Columbia and 3 co-authors from UBC and the MIT Media Lab, presented a paper that may revolutionize 3D displays and bring light field technology into hand-held systems or flat panel TV. Someday-don’t hold your breath, though.
According to the authors, the approach is a "multi-layer generalization of conventional parallax barriers." Instead of being two-layer, the LCD and the parallax barrier, their demonstration system is 5 layers plus a backlight. These layers, instead of being black and clear like the parallax barrier, are "attenuation layers" that selectively reduce the intensity of the light and produce gray scale from full black to fully transparent.
Multi-layer 3D displays in the past have normally produced the depth directly, with each layer in display producing a different depth plane. The depth volume of the system is then limited to the thickness of the display, which would be a severe limitation for a handheld display or a flat-panel 3D TV. Such an approach is more generally considered a volumetric display.
The UBC/MIT approach doesn’t work this way. Instead, the different attenuation layers interact with each other to control the intensity and color of the light in each different direction. This reproduces the light field produced by light reflecting off of the original object, hence the name light field reconstruction. This approach, in theory at least, can produce both out-of-screen and behind-the-screen 3D effects.
Skype has adopted Google’s open source video codec VP8 as its default solution for video conferencing, according to a blog post from Google Product Manager John Luther. The new Skype for Windows client 5.5 will automatically use VP8 both for one-on-one and group video calls as long as other participants are using the same version.
Skype has been using VP8 for group video calls since late last year, but the adoption of the codec for one-on-one calls as well is definitely a boost for Google’s open video ambitions. Google open sourced VP8 in May of 2010 as part of its WebM video format, but many end users likely haven’t seen VP8 in action just yet. WebM is supported by Chrome, Opera and Firefox, and YouTube has been converting its entire catalog to the format. However, the site still serves up H.264-encoded videos in Flash by default, and users have to opt in to a special trial to get to see the WebM versions of YouTube videos.
The codec has also been targeted by patent pool entity MPEG LA, which is threatening to form a patent pool for VP8. Google has maintained that companies adopting WebM or VP8 have nothing to fear, and the fact that a company that’s being acquired by Microsoft is willing to put its eggs in the open codec basket definitely should quell some fears and possibly encourage other video sites as well as video conferencing providers to switch to embrace the format. One should note, however, that Microsoft has so far shied away from adopting WebM for its Internet Explorer browser.
WebM developers have long been saying that it is well-suited for real-time applications, and Google itself is working on making VP8 the default video codec for both Google Talk and its new group video chat platform Google+ Hangouts.
By Janko Roettgers, GigaOM
I have seen the future of TV and it is called Zeebox. The next project from Anthony Rose, the technologist who built KaZaA and BBC iPlayer in to some of the most disruptive digital media plays, is due to go live in October. Topped by Peoplesound founder and ex EMI SVP Ernesto Schmitt as CEO, the pair’s startup raised $5 million in seed funds from unidentified investors in June and has been operating in stealth as “tBone”. But it has been renamed and has just located at offices at London’s Covent Garden, where it has a staff of almost 30 (including former iPlayer engineers) and where the pair showed me an exclusive demo…
What is Zeebox?
Attempting to ride both the multi-screen TV engagement trend and the increasing adoption of internet TVs, Zeebox is a real-time system for social TV viewing and for engaging deeply around those shows that depends on recognising sofa-based second screens as the place for innovation.
The free Zeebox app for iPad (and, later, iPhone and Android) is a TV guide that displays what shows Facebook and Twitter friends are watching. Owners of compatible connected TVs can flip channel straight from the app, as though it were a remote control. Although the command takes place over the internet, the change happens as quick as or quicker than some standard infrared remotes.
Viewing Together
Notifications appear on-screen to indicate friends’ presence in channels. Users can chat in the iPad app and send invites to join one another for simultaneous viewing - accepting an invite results in the channel changing. “Jack, come and watch The Apprentice with me,” Schmitt tells me, by means of example.
As well as these personal connections, Zeebox users’ collective actions can shape the experience. The app displays real-time data for which shows are “trending” up or down. In a scenario Rose presents, a notification appears to say Top Gear is currently “hot” (perhaps Jeremy Clarkson has said something particularly egregious). The opportunity to surface breaking news in this serendipitous way is clear, along with the prospect of improving TV ratings measurement with actual real-time data.
Making TV Hyperlinked
But this “second-by-second” approach is the fabric of more than just Zeebox’s social interaction. Using both commercially licensable broadcast metadata and frame-by-frame analysis of live TV pictures and audio, Zeebox will apparently understand exactly what is on the TV screen at any given moment (“just as Google spiders the web”), in order to serve up all manner of related material on the handheld app.
As example, Schmitt shows how, whilst Tom Cruise is interviewed on Top Gear, the app will auto-display “infotags” for spoken topics (say, “Ferrari 458”, “Abu Dhabi”, “Sebastian Vettel” and “Tom Cruise” himself), as Cruise is speaking. Each topic becomes an in-app link to a corresponding piece of online content, on Wikipedia, IMDB, iTunes Store or whatever.
The method involved is Zeebox’s “secret sauce”, the subject of a pending patent application, but it’s called automated content recognition (ACR), a field with several vendors including Civolution.
“As context emerges on TV, these infotags just keep ticking up,” Schmitt says. “I find this so unbelievably exciting. Anything being discussed on TV is right there for you.” Or, as Rose puts it: “It’s like crack - you just keep wanting stuff, and getting it second-by-second. TV just becomes better.”
One of the intended uses of “infotags” is commerce. Schmitt wants viewers to be able to buy things relating to what they see on screen. As I flip channel to QVC, he assures me Zeebox will know what’s on-screen is a cubic zirconia ring - and offer me more information, as well as ways to buy that ring.
Programme Context
Rose wants in-app TV show pages to display live tweet streams as well as broadcaster-owned HTML “widgets” for custom show engagement. “BBC Red Button’s non-interactive, a bugger to author for and a bugger to use,” he says. “Imagine a next-generation Red Button toolset that allows people to author things for an IP age.”
From these show pages, Zeebox will also offer links to available on-demand episodes. They could be played on the tablet or smartphone, but Rose tells me users may eventually be able to use those handhelds to invoke playback on the TV. “I don’t have a full answer to these things yet, but we’ll experiment with the full infrastructure,” he says.
How it Works
At its most basic, an iPad user can “check-in” to shows manually (though Rose and Schmitt hate the GetGlue- and Foursquare-style gamification concept). To automate that process, the app can listen for shows’ audio fingerprints, Shazam-style. Connected TV owners get the full automatic experience because those TVs already know what shows are on.
“The browser in connected TVs lets you create HTML overlays and widgets,” Rose says. “We’ve created a lightweight, Javascript-based plugin that, on many 2011 and 2010 TVs, can be software-updated and user-installed.”
Zeebox is currently in demo on Samsung Smart TV and Rose, the former CTO of the YouView connected TV consortium, says: “YouView’s got a nice underlying architecture that will allow the Zeebox plugin to run on it, so we look forward to those discussions in the fullness of time.”
A Zeebox open API is also proposed to empower developers to build similar functionality in to their own apps. “There’s a shitload of technical work that needs to be done,” Rose says. “Getting there is non-trivial. We want to go to the moon.”
By By Robert Andrews, PaidContent
Tuesday, August 02, 2011
Labels: Automatic Content Recognition, Connected TV, Fingerprinting, Second Screen, Social TV
If connected TV services are to carry premium content that people are willing to pay for, security must be highly renewable with the whole ecosystem under constant surveillance for threats. It must be possible to authenticate Consumer Electronic (CE) devices approved for access to premium content and cut them off instantly in the event of a breach, with the ability to swap out to a new security system just as happens with smartcard based CA (Conditional Access). This is the view of traditional Conditional Access (CA) vendors that hope to be custodians of security in the OTT world as well.
In this connected TV world, operators will no longer have direct control over end devices but must still maintain a one-to-one relationship with all devices allowed to access premium content, according to Fred Ellis, Director of Operations and General Manager at SecureMedia, part of the Motorola Mobility Group. Ellis recommends two types of embedded security to achieve this, either using digital certificates signed and inserted in the device by the manufacturer, or a secure software client that exploits some unique feature of the box, such as its native HLS (HTTP Live Streaming ) player. Both of these will allow the operator to authenticate the device each time a session is opened, for example by verifying the digital certificate.
Such client security methods must blend with OTT delivery methods such as Adaptive Bit Rate Streaming (ABRS), but above all they must support continuous monitoring. This is even more essential for OTT than in a traditional walled garden Pay TV broadcast infrastructure, according to Christopher Schouten, Senior Marketing Director for online services at CA vendor Irdeto, whose ActiveCloak platform is used by US OTT provider Netflix. Successful defence against OTT piracy involves a combination of continuous surveillance with dynamic security that can apply changes very quickly, Schouten argues. “The key to proactive hacking prevention is the ability to monitor hacker chatter on the web to analyse patterns and isolate instances of real threats that require expert attention,” he argues.
SecureMedia is equally convinced that sophisticated monitoring will be an essential component for OTT success. “Our SecureMedia’s iDetect tamper detection technology immediately notifies operators of any attempted breach in the delivery system and they can promptly terminate content delivery to the device,” says Ellis.
There is also the higher level issue of rights enforcement, and here too OTT extends the challenge because the CE devices may not all support the same DRMs. So operators cannot apply a single security system end-to-end as they can within their walled gardens, according to Tore Gimse, CTO at Norwegian CA vendor Conax. “There is no definite reason why there should be the same security end-to-end," says Tore. “We already offer bridges between the Conax CA and other DRMs, securely handing the protected content from one system over to another. Such solutions could, in principle, be extended to other forms of security transfer.”
In a sense, OTT creates two conflicting security demands. On the one hand, the infrastructure must be even more closely monitored and managed to combat the new piracy threats that will arise in a wide open infrastructure. But on the other hand, no single CA vendor can rule, requiring interoperability between multiple systems, and this can create points of weakness. One implication is that CE vendors may mount a challenge to Pay TV operators by arguing that the network in the middle cannot be secured anyway so you might as well put all the effort into the device. Then content owners would bypass the operators and go straight to consumers.
By Philip Hunter, Videonet