About six months ago, Microsoft released an alternative proposal to the W3C WebRTC 1.0 Working Draft, dubbed CU-RTC-Web. Like all W3C groups, the WebRTC Working Group enlists membership from a majority of the industry, including names like Nokia, Cisco, Google, and Mozilla. The most important question raised by the Microsoft proposal is how the Working Group would react to criticism of its draft proposal, and whether Microsoft would accept the published APIs of the Working Group, even if CU-RTC-Web is not adopted. So what exactly does this mean for the development community?
The Microsoft draft outlines a low-level API that allows developers more direct access to the underlying network and media delivery components. It exposes objects representing network sockets and gives explicit application control over the media transport. In contrast, the WebRTC API abstracts these details with a text-based interface that passes encoded strings between the two participants in the call. With the WebRTC draft, developers are responsible for passing the strings between communicating browsers, but not explicitly configuring media transport for a video chat.
PennApps hackathon was the largest college hackathon in the world and it took place this past weekend. It produced some of the best/most entertaining hacks that I’ve seen at any hackathon: Remote controlled battle bots, Automatic Wifi Authentication for facebook friends, enlarging media seamlessly from one to multiple mobile screens, app that messages you if you forget to put required items in your backpack, exploring neighborhoods from the comfort of your couch with augmented reality, just to name a few.
Looking back, I would say that this hackathon was a smashing success, and I’m sure the other sponsors would say the same. From my perspective as a developer evangelist, here’s why PennApps turned out to be a legendary hackathon and what we can learn from it:
Last weekend we had the pleasure of sponsoring University Hacker Olympics. Unlike your typical hackathons, this one emphasized connecting University students with industry professionals.
Personally, I thought the event was innovative in the field of recruiting. In the traditional interview process, sometimes great candidates were dismissed because their shyness or nervousness inhibited them from performing. 1-1 interviews can be intimidating, we’ve all been there. From the interviewer’s perspective, asking candidates to solve problems does not provide any valuable insight into how pleasant it would be to work with them in a working environment.
A number of developers have asked for ways to help end users diagnose potential problems that disallow them from being able to successfully video chat using OpenTok. A while ago we introduced a troubleshooter page that test a user’s network, hardware, and software to check things like ports, camera, and Flash version to be sure that they have what they need in order to use OpenTok.
While this tool has been useful for diagnosing problems, one obvious pitfall is that you have to send your users to our website—away from your experience.
Developing an iOS App itself is a huge undertaking: you want your product to be beautiful, interactive, and functional. That’s why Parse makes so much sense, it helps you avoid writing a backend server to power your App by giving you a data store and providing the most basic web services. These days many web services are incredibly powerful and help developers do really amazing things, like OpenTok, but they are targeted at having a backend. That’s where Parse Cloud Code comes in: it gives developers the ability to leverage the best of a back-end server in the path of least resistance.
Yesterday, the talented folks at Double Robotics rocked the stage at LeWeb in Paris. David Cann, Co-founder and CEO, demoed their telepresence robot, Double. The sleek Double combines Segway-style movement with video presence delivered through an iPad.
Leveraging OpenTok’s iOS SDK on WebRTC, Double enables owners to broadcast their video stream on the robot’s iPad from any location. Not only that, but the operator can stream and control its movements from a smartphone.
Cann demoed the Double’s original inspiration, which was to connect remote workers with their home offices. Calling his CTO back in Sunnyvale, Cann took their Double for a three a.m. spin around the office from Paris. Customers are also finding uses for Double for remote tours of factories, museums, schools, and retail outlets.
Here at TokBox, we’ve been supporting AngelHack since its infancy. This year it’s no different, and this time we sponsored at a city bursting with startup energy, Los Angeles.
The event is hosted at a spacious and comfortable co-working space called Cross Campus, a place to inspire “creative collisions through space design, learning platforms, and extraordinary events.” If you are an entrepreneur you might want to check it out!
The event started off with sponsor pitches and API talks. Singly provides SDK for developers to get their app connected quickly and easily with services like Facebook, Twitter, Google, just to name a few. Gimbal provides a mobile context awareness platform that includes image recognition and geofencing. TokBox provides a video chat API (called OpenTok), and for demo I live coded a web and iOS app that video chatted with each other.
WebRTC Demo Day! Unless you have been living under a rock, you have heard of WebRTC. A few weeks ago, Google unleashed Chrome 23 which has WebRTC and PeerConnection support. This is really exciting because everyone on chrome (and IE users with Chrome Frame plugin) can now experience the next generation of communication via live video. Today is the day to experience it. Simply go to https://OpenTokRTC.com and join a room! If you’re lucky, you might catch a few WebRTC Streams from iOS devices.
There’s irony in being a cloud provider when a major piece of your infrastructure lies below sea-level. Of course, the premise behind the cloud is that you never have to think about the underlying physical location of your servers. If a flood were to happen in a critical network hub, say New York City, it should in theory have no impact on a cloud service (because, you know—clouds are in the sky).
As Hurricane Sandy forecasts came in, our ops team began to consider what impact it might have on our service. We had two options: A) keep our NYC servers as they were and hope that power and connectivity would remain intact, or B) take our NYC servers out of rotation and direct that traffic elsewhere.
Nearly 7 months ago, we publicly announced that the OpenTok API would extend its reach to native mobile application developers by publishing the OpenTok iOS SDK. In the time since, we have tightened the performance of the SDK runtime for iOS devices and spent a good deal of time learning about how best to deliver video to the mobile platform. While iOS commands a large portion of the mobile app market, it is intuitive that we should build similar SDKs for other popular platforms outside of the browser. It is a pleasure to announce that we are developing the OpenTok Android SDK, to allow native Android developers to bring live video chat to their apps.