• Lab
  • AndroidForMobile Foundation at
    HOME
              
    LATEST STORY
    By running unwitting PR for Jeffrey Epstein, Forbes shows the risks of a news outlet thinking like a tech platform
    ABOUT                    SUBSCRIBE
    May 13, 2009, 10:48 a.m.

    The New York Times would like to join you in the living room

    In a corner of the at The New York Times Co., they’ve prototyped a living room of the future. It’s not as whizbang awesome as you might hope — a glows red or green depending on how the markets are doing — but it does feel like a reasonable conception of Living Room 2.0. Their major bet: as become more common, people will increasingly choose to consume web material on those huge, high-definition screens.

    That wouldn’t, on its face, be an advantageous development for the Times, which produces the vast majority of its content in longform text you’d never consider reading on TV. But as , a creative technologist in the R&D group, explains in today’s video, it may be possible to shift gears in the living room and emphasize the newspaper’s multimedia content. She demonstrates the concept with “,” a major series on environmental damage in China from 2007.

    This is the third in our weeklong series of videos from the R&D group, and it may be the one that’s easiest to imagine coming to pass. Laptop and desktop computers are already commonplace in the living room, is a , and Apple keeps at converging TV and the Internet. (On Oxygen’s , the cast members check their email on a television in the living room. QED.)

    Still, reimagining The New York Times in HDTV is a challenging leap. (You might recall the Times Co. made an unsuccessful into television with the Discovery Channel earlier this decade.) The newspaper produces a ton of multimedia content — certainly more than its competitors — but a satisfactory living-room experience would require video on a scale the Times isn’t yet producing. That’s why they call it the future.

    You’ll see more of the R&D group’s living room in tomorrow’s video (yesterday‘s was also shot in there). After the jump, you’ll find a mock-up by design integration editor , which adds a projector but is otherwise pretty faithful to the actual room. And below that, there a transcript of today’s video.

    Alexis Lloyd: The main problem we see with content from The New York Times in the living room is that our primary form of storytelling is still long-form text, which works really well on paper, still works well on the web — but once you’re sitting ten feet from a television in your living room, that pretty much breaks down. But we do produce all this great multimedia content. It’s just usually pushed off to the side a little bit. So in this demo we are asking the question: Can we flip that paradigm around and use the media that works really well in the living room — the video and the images — and make that the spine of the story, but still pull in some of the text and pull in some degree of interactivity that you might want when you’re in the living room?

    So I’ll show you this. In this case, I’m using a standard mouse to navigate this, but we’re also looking at a lot of devices like these  that I could sit and navigate from my sofa, as well as doing custom remote controls and interfaces like the kind that Mike showed you on CustomTimes.

    So this is just a that I’m going to start playing, and as the video plays there are these panels that appear that I can open up to show you contextual information about what’s being discussed in the video. So in this case, I can get background information about the turtle that’s being mentioned. It’s text, but it’s short, it’s big, and furthermore, it’s optional. So I can just open it up, read it, and then I’m back in the video. So it doesn’t take me out of that central experience of sitting back and being told a story, which is my primary kind of mode when I’m in the living room.

    And we can do this with all kinds of content. So in that case, that was some, a piece of text that was related to what they’re talking about in the video. In this case, there’s who is being interviewed. She’s written . I can open this up to read an excerpt from that book. Furthermore, it knows it’s a book, so there’s an e-commerce component that’s integrated into this. And I can just choose from this interface to buy the book. It goes into my Amazon shopping process, and I’m back in the video. So I’ve done all this, but I haven’t been taken out of that basic experience.

    And this is really pointing to the idea of creating more granular levels of about content. So we have metadata about our videos as a whole, but now we can begin to say, at this particular point in time in the video, we have a related map or at this particular point in time, they’re talking about this lake. And we have a slide show about that. So I’m going to open that up.

    And then you can see our photojournalism really has a place on the big screen because the photos are stunning at this size. And furthermore, the photos themselves have this more granular level of metadata where there are these hot spots that I can use to get deeper information about objects or people in the video — or in the photo, rather. So I can find out all about this toxic algae that’s growing on this lake as a result of .

    And at the end of the video, we’ve also integrated some social functionality, so I can choose to share this video with other friends, and it pulls in the people I most frequently share with. I can say, I want to share this with , and then it will go into any number of his social feeds and into our lifestream app, which Ted will show you in a moment. There’s also some in here, so you might be inspired to give to ‘s clean water campaign after watching that video.

    And furthermore, that’s just a one-minute video piece, but this was a yearlong series. There is a huge collection of multimedia content that was created for it. So we started asking the question of can we use the metadata that we’re already creating for our content to allow readers and users different lenses into these large collection of media that might be overwhelming to them?

    So in this case I have four different views that I can go into that are dynamically created from the metadata associated with it. So there’s an editors’ choice where I can say I just want to know what the New York Times editors think are the highlights of this collection or this package. But I can take that same content and I can sort it geographically, or I can sort it over time and see a timeline. Or I can sort it thematically and start to see relationships between different themes in the collection of media.

    So those are some of the different ideas we’re looking at around how our content could be produced and packaged and repurposed for exploration in the living room.

    POSTED     May 13, 2009, 10:48 a.m.
    PART OF A SERIES     The New York Times R&D Lab
    SHARE THIS STORY
       
     
    Join the 50,000 who get the freshest future-of-journalism news in our daily email.
    By running unwitting PR for Jeffrey Epstein, Forbes shows the risks of a news outlet thinking like a tech platform
    If journalists want to criticize the anything-goes ethos of Facebook, it’s only fair to note when news organizations’ hunger for scale leads them down the same problematic path.
    Can’t read just one: Slate’s daily advice columns are strange, funny, deep, and increasingly a major traffic driver for the site
    “We probably won’t do twincest again.”
    O, a meaning!
    Fifty years ago, with humanity about to reach the moon, The New York Times gave a poet a corner of the front page.
    viagra générique pas cher

    www.np.com.ua/elektromehanicheskoe-oborudovanie/miksery-planetarnye.html

    У нашей фирмы важный интернет-сайт со статьями про www.progressive.ua.