Search Results

Search found 11573 results on 463 pages for 'store'.

Page 231/463 | < Previous Page | 227 228 229 230 231 232 233 234 235 236 237 238  | Next Page >

  • VIDEO Streaming - How to output the video timestamp?

    - by Emmanuel Brunet
    I would like to backup an ASF video stream with the time stamp (I mean the original recording date/time) on the output stream ? Usage I convert / store my video using the mkv format (Matroska) with libx264 (video) and aac (audio) codecs. Assume the IP camera webcam user account is account, the password password $ ffmpeg -i http://admin:alpha1237@webcam/videostream.asf -c:v libx264 -s 768X432 -crf 13 -b:v 2500K -pix_fmt yuv420p -c:a libfdk_aac output.mkv This works fine on a tenvis JPT3815W camera How to I need to get the video timestamp available for display as a subtitle or other meta data field managed by standard video players, and ideally to be able to hide it or not during video reading. Does anybody knows how to achieve that ? Thanks in advance for your help.

    Read the article

  • development and music recording machine suggestions?

    - by dean nolan
    I wasn't sure if this belongs on SuperUser so flag if so. I am looking to build, primarily, a windows development machine that is also good for recording using Cubase. I know I should use seperate machines but I'm on a budget this time of year. I also havn't kept up with hardware for quite a few years. Basically I know I want quad core, multiple monitor support (no gaming requirements). A lot of RAM, very quiet case and super fast HDD (SSD OR 10,000RPM)for compiling and latency. I will store libraries and other data on a USB drive. Sound card is not needed as I will be using an audio interface, all other music recording equipment is taken care of also. I could do with some decent monitor recomendations also. All suggestions welcome, thanks.

    Read the article

  • Can I get "disk utilization" from a NetApp filer via SNMP?

    - by Andrew
    On a NetApp filer's command line I'm running "sysstat -u" to show disk utilization, (actually the utilization of the single busiest disk). By disk utilization, I mean "percent of time the disk is busy", not "how much space on the disk is being used to store data/metadata". Is there a way to get disk utilization info through SNMP? The netapp.mib file doesn't appear to expose this. It does have CPU utilization, disk usage & capacity information, etc, but not disk utilization. The MIB-II (rfc1213) seems to be the only other information exposed by the filer through SNMP. I hope I am missing something. The "CP (consistency point) time" metric is exposed through the NETAPP-MIB in SNMP, but this seems to only partially correlate with disk utilization under write load, and not really at all under read load.

    Read the article

  • The bcdedit export operation fails. windows vista not booting.

    - by ocurro
    I'm fixing my boot record with bootrec.exe, the first command goes: C:\>bcdedit /export c:\BCD_Backup Immediately I receive this error here: The store export operation has failed. The requested system device cannot be found. Likewise, the next command fails, because bcd file is missing: C:\>attrib bcd -s -h -r Lastly, when trying the next command, all it returns is "The requested system device cannot be found". C:\>bootrec /RebuildBcd I'm using Windows 7 setup image to operate. Have Vista installed atm. Has it happened to you? I'd appreciate some orientation.

    Read the article

  • What is the best file system and allocation size for a USB flash drive?

    - by e-t172
    I'm considering using my 4 GB Kingston DataTraveler USB stick to store my Firefox and Thunderbird profiles for my laptop and desktop PCs. I want to maximize performance when using Firefox. The question is: what is the best file system and allocation size for the fastest Firefox profile operation on a USB flash drive? I'm using Windows 7 on both machines and I don't care about compatibility or the drive's lifetime. I just want to maximize performance. I could even use ext2 with the Ext2 IFS driver if that means it'll be faster. I'm assuming (perhaps I'm wrong) that putting a Firefox profile on a USB stick would be a "lots of small files" usage. In that case, it seems that NTFS would perform best, but I'm not sure. Besides I found nothing regarding the best allocation size to use. Considering that the default allocation size is designed for hard drives (which have different characteristics), I'm assuming that the default allocation size is not the best.

    Read the article

  • Saving music wisely: Why save 'Queen - Bohemian Rhapsody.mp3' millions of times?

    - by hsmit
    As far as I'm concerned, Queen's song 'bohemian rhapsody' is one of the most popular songs all time. But for the purpose of this message you may replace this with another track. At the same time I think 60% of the digital-music listeners have this track. Sometimes we have multiple copies: different versions of the track, different devices, unwanted duplicates in download folders, itunes folders etc.. Wouldn't it be much smarter to store these songs only once? You can imagine various solutions for this. How would you accomplish this? Some criteria that may help you find an answer: It must reduce disk space It must remember which music belongs to you (DRM) It must use network traffic efficiently

    Read the article

  • Opera: closed window w/many tabs, window still open w/one tab, how to retrieve the many tabs

    - by Abel
    This is a little recipe for a little disaster: accidentally close a window with many important tabs open, only to find out that another window is still active with one tab open. Closing it will overwrite the saved tabs of the other window, which I want to recover. I need to recover the window with "many tabs" if possible,I didn't yet close the window with the one tab open, hopefully that prevents overwriting the saved settings. How do I retrieve, now that Opera is still active, the tabs of the closed window? I.e., where would Opera normally store these tabs when it closes a window? Can I prevent it from overwriting the saved state of "many tabs" with the current "one tab"?

    Read the article

  • Automatically reconnect to ODBC sources?

    - by stefan.at.wpf
    I am using Asterisk 1.8.10.1 and a MySQL database connected via ODBC to store CDRs. When my MySQL database isn't available when Asterisk starts or has an outage while Asterisk is running, I would expect Asterisk to retry to connect to the database, but this doesn't happen! Anyone knows where I can enable some kidn of automatic reconnect to databases in Asterisk? My res_odbc.conf looks like this: [asterisk] enabled => yes dsn => asterisk-connector username => user password => pass pre-connect => yes pooling => no limit => 1 idlecheck => 1 negative_connection_cache => 1

    Read the article

  • Implementing Variable Envelope Return Path (VERP) using Exchange

    - by iammichael
    We're looking into implementing Variable Envelope Return Path (VERP) for improved bounce processing for our application. Our current mail infrastructure is MS Exchange 2007 but are in the process of upgrading to 2010. We're also implementing Postini for spam filtering. Exchange doesn't support sub-addressing (see also this question on disposable addresses) -- and VERP is somewhat of a specialized application of sub-addressing. Are there any options for implementing VERP in Exchange without putting another non-Exchange SMTP relay in front of Exchange to pre-process incoming messages? Specifically could a transport rule be created that could match against the target (non-existing) recipient, store that recipient address in a special header added to the message, and redirect the message to a pre-created mailbox? Note: we have developer resources available if custom code could be used somehow.

    Read the article

  • How to maintain one file across many production servers (Windows and Linux)

    - by Brien
    My organization wants to centrally manage an Oracle TNSnames file for all of their production servers. When that file changes, they want to be able to push out the changes to all servers that use it with a minimal effort. Approaches that have been considered: Centralized file server (drawback: if the file server or the network connection to the file server goes down, the servers have no access to the critical file) Subversion client on each server (drawback: using a source control tool in production, added complexity) Store an individual copy of the file on each server (drawback: changing the file contents involves making changes on many different servers) Update Can I use DFS to do this?

    Read the article

  • Anyone love/hate the PowerConnect line of switches from Dell?

    - by Rob Bergin
    I am looking at replacing some unmanaged 16 port store bought GB switches and wanted to go with Cisco but it may be cost prohibitive. Instead I am looking at ProCurve or Dell's PowerConnect line up. I am looking for SNMP, Management, VLANs, and SFLOW would icing on the switch cupcake. I would get the 6224 or the 6248 and then maybe add the RPS-600 to it for redundant power. I think the RPS-600 supports multiple switches. Rackspace is also a little challenge so I am trying to do it with as little Rack Units as possible. Ideally I would go with two 6224's or a single 6248 and then do two VLANs. Thanks for any feedback. Rob

    Read the article

  • How to create a Linux Media Server using Ubuntu?

    - by Thomas
    Hello all: I'm an intermediate Linux user and a relative beginner to servers. I would like some help finding resources on setting up a basic server. I have Googled, and am a member of the Ubuntu Forums, but just figure it can't hurt to ask the Stack Overflow community for help as well. I plan on installing on an old laptop (Lenovo Thinkpad R61i or Toshiba Satellite A105). I have downloaded the latest Ubuntu (9.10) but don't know how to do any of the configurations. I just want a server to store my files where I can access (download and/ or stream) from a browser. Any help you can give is greatly appreciated. Thanks! Thomas

    Read the article

  • Postfix tutorial inconsistency

    - by Desmond Hume
    I'm following this tutorial to setup a Postfix/Dovecot mail server with Postfix Admin as a web front end. As regards directory structure for virtual mail users, the author of the tutorial writes: Virtual mail users are those that do not exist as Unix system users. They thus don't use the standard Unix methods of authentication or mail delivery and don't have home directories. That is how we are managing things here: mail users are defined in the database created by Postfix Admin rather than existing as system users. Mail will be kept in subfolders per domain and account under /var/vmail - e.g. [email protected] will have a mail directory of /var/vmail/example.com/me. But when he gives instructions about configuring Postfix Admin, he suggests this to be contained by Postfix Admin's config.inc.php: // Mailboxes // If you want to store the mailboxes per domain set this to 'YES'. // Examples: // YES: /usr/local/virtual/domain.tld/[email protected] // NO: /usr/local/virtual/[email protected] $CONF['domain_path'] = 'NO'; Is there an inconsistency?

    Read the article

  • Amazon EBS root volume persistence

    - by hipplar
    When I launch a new Windows EC2 instance I am given a 30 gig root EBS volume. I'm trying to make sure I understand the EBS terminology and want to make sure I understand this correctly: Q: What happens to my data when a system terminates? The data stored on a local instance store will persist only as long as that instance is alive. However, data that is stored on an Amazon EBS volume will persist independently of the life of the instance. What exactly does "instance is alive" mean? If I write files to the root volume and reboot the instance will the files remain? Or do I physically have to terminate (delete) the instance for the root volume to go away? Thanks

    Read the article

  • Network connectivity logs on OSX

    - by Stephen
    I'm trying to determine whether or not Macs store a log of network connectivity. Specifically, I'm trying to find out if one of my machines (Mac Tower, OSX 10.7.4, 2-2.26 GHz, 16 GB DDR3) lost connection to pinpoint an issue. This machine is on an external network where and the IT staff is not very helpful, so I'm trying to ensure that there were network issues rather than some other local issue on that particular machine. Can you pull network logs thru Console that show when a network connection was established and was disconnected?

    Read the article

  • Sharing external storage between different operating systems?

    - by CT
    I just received a Lacie 1TB external usb/firewire/esata drive. I have 2 machines. A macbook running osx 10.6 and a desktop running windows 7. I would like to rip my dvd collection to iso and store on my external. Right now I use my macbook's disk utility to rip dvds to iso. However my desktop is what is connected to my hdtv. I mainly just use the desktop for media. I'd like to format the external with a 200 GB partition for time machine backups and have the rest for storage. DVD iso are often above 8GB so that sort of eliminates FAT file systems correct? Do I have any options to be able to have my mac and pc both see the drive?

    Read the article

  • Simple, manageable DNS on EC2?

    - by dkulchenko
    I'm working on a large network of servers sitting on EC2, and need a way for the servers to know about each other's locations in the cloud. I thought the simplest way would be to use DNS, because if I replace the EC2 instance, I simply update the DNS record, and the rest of the servers will know about it (with names like users.db.mysoft.com, routing.mysoft.com, cluster1.memcached.mysoft.com). I'm considering setting up a master DNS server on a micro/small instance to accommodate this. I'd preferably need something that's as simple as a key-value store (hostname - IP) into which the platform could remotely add/remove entries. Can I do this with BIND? Or is there a better solution?

    Read the article

  • Fast or automatic way to organize iphone/ipod apps

    - by Notitze
    Is there a tool or application that can help me quickly organize my iphone applications? The standard way of dragging and dropping 1 application at a time in folder is too frustrating for me. I gave up after 1 hour of trying... I found it to be such a waste of time ... I am looking for an application to automatically organize my apps in the standard folder categories at least (Productivity, Navigation, Photography, etc). I was not able to find one in the iTunes store. Any suggestions?

    Read the article

  • How to clone and deploy multiple machines?

    - by Mimi
    Hi, See what I wanna do is basically make bootable clones or disc images or whatever(let's call it "stuff" for now ) and store many of them in a single big HDD. Then, I wanna use that big HDD and the "stuff" in it to be able to deploy new machines without the hassle of reinstalling Windows and whatnot(i.e. I do not want to restore nor simply clone a same machine, I literally want to use clones to save time during machine set up). Obviously I'd deploy the "stuff" onto an HDD going into a machine that I know the hardware of to avoid driver issues and whatnot. I alos have enough Windows keys to deal with the multiple activations. I tried using Acronis True Image but it doesn't seem to do what I want(unless idk how to use it properly). Any advice is welcome thanks :)

    Read the article

  • Windows 2008 File Share

    - by user36540
    Hi, I have 3 Windows 2008 Standard servers in my system with no domain controller. Two of the servers are running a NLB cluster and the third server is a file server that the web servers connect to. I want to store my source code on the file server and point the IIS config to the network file share. The web sites also need access to a file share on the file server. I was able to share the network drive and access while logged into either of the web servers but my web apps are unable to access the file share - I assume due to permissions. Does anybody know the correct way to do this? Thanks, Chris

    Read the article

  • Where is the best location to keep shared-developer website files in the linux hierarchy?

    - by Tchalvak
    I just started hosting files for a website on my server, and I'm not sure where is an appropriate place to keep them. At the moment, I have them in /var/www/name.of.virtualhost.site/www/. That's obviously not secure because anything below the final public /www/ folder is also available since the /var/www/ contents are already being served up. For example, /var/www/name.of.virtualhost.site/docs/site_policies.txt is accessible via something like defaultsite.com/name.of.virtualhost.site/docs/site_policies.txt. So where is a good place to store the files that make up a website? (when it's a site that only I'm developing, I can obviously just stick them in /home/my_username/sites/name.of.virtualhost.site/, but that doesn't work well when I want other developers to be working on the site's files as well) I'm running a LAMP stack, not that I expect it to matter.

    Read the article

  • Setting up an Active-Active IIS Cluster with ARR - is it possible?

    - by Ahmed Zubair
    I would like to know if we can setup an Active-Active IIS Cluster using Windows Cluster services that shares a common storage to store web content and WITHOUT the use of Windows NLB. I'm aware that this may not be a best practice or not a recommended setup, however, the setup is to be configured as below: Two web servers running IIS 7.5 (needs a common storage for web content) for HA and another set of two servers for sql cluster in active-passive mode for HA. Also is it possible to enable ARR on 2 node active-active IIS cluster for load balancing http requests? Appreciate if someone replies with both pros & cons of the setup.

    Read the article

  • LDAP (slapd) creating users with access to specific trees

    - by Josh
    I am setting up a CentOS server with Virtualmin and Postfix, and I am trying to use LDAP to store unix users, groups, Postfix aliases and virtual domains. I am following the instructions from Webmin's site. I have created an LDAP domain and configured Postfix to fetch Aliases and Virtual Domains from LDAP, but in order to do so I had to configure postfix to authenticate with the master LDAP account, cn=Manager,dc=mydomain,dc=com. This seems like a terrible idea because that account has access to the Users and Groups, which postfix does not need access to. How can I create a new LDAP account for Postfix which only has access to the LDAP trees Postfix needs?

    Read the article

  • Are Plesk server backups useful?

    - by Michael T. Smith
    I'm working for a startup now, and I'm the programmer. Because of our small team size, I'm also handling the server management for now (until we get a dedicated server administrator.) I've never used Plesk before, and the server we're using (a Media Temple Dedicated Virtual server) had it installed when I got here. One of my first jobs was to set up backups: Plesk was already running it's nightly server-wide backups. I created a small script to dump the web app, it's DBs and any assets, tar them, store them, and then copy them to another small server we have (to backup the backups.) But, we're constantly running into hard drive space issues because of the Plesk backups. And I'm wondering, are they useful? If I have the web app and all of it's assets, I could easily enough get another server up and running. Do we need to keep running Plesk's backups? Thoughts?

    Read the article

  • Backing Up Transaction Logs to Tape?

    - by David Stein
    I'm about to put my database in Full Recovery Model and start taking transaction log backups. I am taking a full nightly backup to another server and later in the evening this file and many others are backed up to tape. My question is this. I will take hourly (or more if necessary) t-log backups and store them on the other server as well. However, if my full backups are passing DBCC and integrity checks, do I need to put my T-Logs on tape? If someone wants point in time recovery to yesterday at 2pm, I would need the previous full backup and the transaction logs. However, other than that case, if I know my full back ups are good, is there value in keeping the previous day's transaction log backups?

    Read the article

< Previous Page | 227 228 229 230 231 232 233 234 235 236 237 238  | Next Page >