Architectural advice - web camera remote access

Posted by Alan Hollis on Server Fault See other posts from Server Fault or by Alan Hollis
Published on 2012-12-17T20:54:03Z Indexed on 2012/12/17 23:04 UTC
Read the original article Hit count: 192

Filed under:
|
|
|
|

I'm looking for architectural advice. I have a client who I've built a website for which essentially allows users to view their web cameras remotely.

The current flow of data is as follows:

User opens page to view web camera image. Javascript script polls url on server ( appended with unique timestamp ) every 1000ms

Ftp connection is enabled for the cameras ftp user. Web camera opens ftp connection to server. Web camera begins taking photos. Web camera sends photo to ftp server.

On image url request: Server reads latest image on hard drive uploaded via ftp for camera. Server deleted any older images from the server.

This is working okay at the moment for a small amount of users/cameras ( about 10 users and around the same amount of cameras), but we're starting to worrying about the scalability of this approach.

My original plan was instead of having the files read from the server, the web server would open up an ftp connection to the web server and read the latest images directly from there meaning we should have been able to scale horizontally fairly easily. But ftp connection establishment times were too slow ( mainly due to the fact that PHP out of the ox is unable to persist ftp connections ) and so we abandoned this approach and went straight for reading from the hard drive.

The firmware provider for the cameras state they're able to build a http client which instead of using ftp to upload the image could post the image to a web server. This seems plausible enough to me, but I'm looking for some architectural advice.

My current thought is a simple Nginx/PHP/Redis stack.

Web camera issues post requests of latest image to Nginx/PHP and the latest image for that camera is stored in Redis.

The clients can then pull the latest image from Redis which should be extremely quick as the images will always be stored in memory.

The data flow would then become:

User opens page to view web camera image. Javascript script polls url on server ( appended with unique timestamp ) every 1000ms

Camera is sent an http request to start posting images to a provided url Web camera begins taking photos. Web camera sends post requests to server as fast as it can

On image url request: Server reads latest image from redis Server tells redis to delete later image

My questions are:

  1. Are there any greater overheads of transferring images via HTTP instead of FTP?
  2. Is there a simple way to calculate how many potential cameras we could have streaming at once?
  3. Is there any way to prevent potentially DOS'ing our own servers due to web camera requests?
  4. Is Redis a good solution to this problem?
  5. Should I abandon PHP/Ngix combination and go for something else?
  6. Is this proposed solution actually any good?
  7. Will adding HTTPs to the mix cause posting the image to become too slow?

Thanks in advance

Alan

© Server Fault or respective owner

Related posts about http

Related posts about Web