Search Results

Search found 6169 results on 247 pages for 'future proof'.

Page 115/247 | < Previous Page | 111 112 113 114 115 116 117 118 119 120 121 122  | Next Page >

  • Sync data between a windows desktop app and windows mobile client app

    - by Chris W
    I need to knock up a very quick prototype/proof of concept application to demo to someone within the next couple of days so I've minimal time to research this as fully as I normally would. The set-up is a very simple database application running on a laptop - will only ever be a single user updating a couple of tables so I was thinking of knocking up a basic Win Forms app against SQL Compact. Visual Studio's auto generated data grid edit screens will be fine with a little customisation. The second aspect is to then add a windows mobile client application that can pull data from both tables stored on the laptop, edit some data and insert some extra rows before sending the changes back to the laptop copy of the database. I've not done any WinMo development so what's the best approach for me to look at. Is it easy enough to sync data between the two databases when the WinMo device is connected to the laptop with USB? Most of the samples I've looked at so far seem to be syncing SQL Compact with SQL Standard using IIS which seems a bit overkill. The volumes of data to be synced are so small that I can easily write some manual sync code if it's easy for me to query/update the Compact DB from the laptop application when the device is connected.

    Read the article

  • How to backup data from linux servers to linux server (incremental + snapshot)?

    - by wag2639
    We have a handful of hosted servers running RHEL4 and RHEL5 and would like to backup some key folders (I'm thinking /var /srv and /etc) to a local server we have in house. The local server is running Ubuntu 9.10 Server edition. I'm looking for a free (preferably OSS) way to grab (or push) incremental backups to my local server and once a month or so, make a new snapshot for incremental updates in between snapshots. Also, while I'm comfortable with using a command line, others may need to use the system in the future, and I would like some kind of graphical or web interface to browse the backup repository. Suggestions?

    Read the article

  • How to backup data from linux servers to linux server (incremental + snapshot)?

    - by wag2639
    We have a handful of hosted servers running RHEL4 and RHEL5 and would like to backup some key folders (I'm thinking /var /srv and /etc) to a local server we have in house. The local server is running Ubuntu 9.10 Server edition. I'm looking for a free (preferably OSS) way to grab (or push) incremental backups to my local server and once a month or so, make a new snapshot for incremental updates in between snapshots. Also, while I'm comfortable with using a command line, others may need to use the system in the future, and I would like some kind of graphical or web interface to browse the backup repository. Suggestions?

    Read the article

  • Ensure that my C# desktop application is making requests to my ASP .NET MVC action?

    - by Mathias Lykkegaard Lorenzen
    I've seen questions that are almost identical to this one, except minor but important differences that I would like to get detailed. Let's say that I have a controller and an action method in MVC which therefore accepts requests on the following URL: http://example.com/api/myapimethod?data=some-data-here. This URL is then being called regularly by 1000 clients or more spread out in the public. The reason for this is crowdsourcing. The clients around the globe help feed a global cache on my server, which makes it faster for the rest of the clients to fetch the data. Now, if I'm sneaky (and I am), I can go into Fiddler, Ethereal, Wireshark or any other packet sniffing tool and figure out which requests the program is making. By figuring that out, I can also replicate them, and fill the service with false corrupted data. What is the best approach to ensuring that the data received in my ASP .NET MVC action method is actually from the desktop client application, and not some falsely generated data that the user invented? Since it is all based on crowdsourcing, would it be a good idea for my users to be able to "vote" if some data is falsified, and then let an automatic cleanup commence if there are enough votes? I do not have access to a tool like SmartAssembly, so unfortunately my .NET program is fully decompilable. I realize this might be impossible to accomplish in an error-proof manner, but I would like to know where my best chances are.

    Read the article

  • Nginx as a reverse proxy + Apache or completely cut out Apache? (WordPress Multisite)

    - by user715564
    I am starting a WordPress multisite network with domain mapping and I am trying to think through my server set up. Right now, I only have one medium sized VPS but hopefully I will need to add more servers later so ideally the solution will also accommodate future growth. My question is, would it be better to set up Nginx as a reverse proxy with Apache or use only Nginx? It seems like setting up Nginx as a reverse proxy would be easier and offer less of a possibility of problems but, on the other hand, would using only Nginx add substantial benefits?

    Read the article

  • Public key repository - does this exist?

    - by allegroconmolto
    This is one of those "Surely this already exists, and if not, I damn well better build it!" moments. Here's my problem: I run a devteam, of around 10 individuals. Each of us has our own private key(s). When I go set up a new server for us, I have to import all of their keys to it. And if we start working with a contractor, I have to get his or her key too, and then import it and place it in .ssh/authorized_keys on the server. Ideally I'd like to be able to do something along the lines of: import_key allegroconmolto And it would contact a public repository of public keys, see if there was a user named allegroconmolto, and if so, import their key and add it to my authorized_keys. Adding future users would then be as simple as running import_key for them. I could then scp the authorized_keys file to all my servers and be done! This seems so obvious that I feel like it surely must exist somewhere.

    Read the article

  • Using prf file to configure outlook clients

    - by chrissie1
    I have a prf file to outlook clients. For the moment the new exchange will be in another domain, that will change in the future. But the usernames on both domains are the same. I see that the example prf just takes the username ;*************************************************************** ; Section 4 - Default values for each service. ;*************************************************************** [ServiceEGS] MailboxName=%UserName% HomeServer=exchange_server_name [Service1] OverwriteExistingService=No UniqueService=Yes MailboxName=%UserName% HomeServer=exchange_server_name AccountName=Microsoft Exchange Server Can I add the domainname to that like so. MailboxName=\\domain\%UserName´% or is there a better way of doing this.

    Read the article

  • Can I install fresh Linux accross partitions (LUKS & LVM) and preserve/use existing home user?

    - by xtian
    With an existing LUKS encrypted logical volume partitioned hard disk dual boot to Windoz and Linux (Fedora 15), is it necessary to "start over" with the LUKS setup when upgrading the system? I recall some note about dividing the Linux installation over different partitions would help to preserve the home data in future update (I can't find this now) Before I try it, is this possible and intended use case for partitioning a Linux installation? # lsblk -fa NAME FSTYPE LABEL MOUNTPOINT sda [80G] +-sda1 [system W95 FAT 32] vfat +-sda2 ext4 /boot +-sda3 [52.4G] crypto_LUKS +-luks-de25ac97-6a32-4b79-a6a0-296a39376b3b (dm-0) LVM2_member +-cryptVG-root (dm-1) [21.5G] ext4 / +-cryptVG-swap (dm-2) [5.4MB] swap [SWAP] +-cryptVG-data (dm-3) [25.6G] ext4 /home

    Read the article

  • migration of physical server to a virtual solution, what i have to do?

    - by bibarse
    Hello I'm new in this forum, so i would like that you forgive me for my blissfully and my low English level. I'm a trainee in company one month ago, and my mission is to migrate 3 physicals servers to a virtualization technology. The company edit softwares for E-learning so there are lots of data like videos, flash and compressed (zip). This is some inventory of the servers: OS: Debian, 2 redhat, apache, php/mysql, sendMail/Dovecot, webmin with virtualmin template to create dynamically the web sites because there is no sysadmin ... The future provider will be responsible of to secure, update and create the virtual machines (outsourcing) and with a RedHat OS's. So i want that you help me to choose a virtualisation technologie (for the i prefer KVM of Redhat RHEV, VMWare is expensive), how evaluate the hardware needs (this for evolution of 4 or 5 years) and to elaborate a good planing to don't forget any think. Thank you for your responses.

    Read the article

  • Opening notebooks

    - by Plastkort
    I have a Compaq CQ40 which has a defective CPU fan, it seems unable to cool the notebook enough at this time, it is constantly running, and if you move the pc up/down to look under it, it makes some sort of chainsaw sound. I have tried to take out all the screws underneath, but I still have a problem with the plastic cover on the above; I am afraid I might break it since it seem to be connected with plastic clips underneath. I have fixed notebooks before, but this one seem new, how can I open the plastic cover above without breaking it? Any advice, tips and tricks would be so nice to have, not only for this notebook, but any future notebook also I might need to repair. It would be nice if anyone here knows a place with video tutorials on this subject as well.

    Read the article

  • two shops network

    - by edward
    okay so, I just opened up two shops in my hometown. The two stores is about 6 blocks apart, connecting them by wire is not really feasible in cost wise. What kind of network topology should I use for my small shops, there will be 5 computers, one is the sales computer ,another 4 as mentioned is the guest computer. I want the sales and guest computers network to be seperated. Both shops have same computers. The guest computers serves up simple website that has my shop catalog on it, I'm thinking of using a web server. So, how am i suppose to setup these networks, im planning to add in more computers in the future. Is it I need to station a single server at a shop, and all the computer connected to it? or is there any more effective methods? I'm no networking expert, would love to hear some advice.

    Read the article

  • Running an application from an USB device...

    - by Workshop Alex
    I'm working on a proof-of-concept application, containing a WCF service with console host and client, both on a single USB device. On the same device I will also have the client application which will connect to this service. The service uses the entity framework to connect to the database, which in this POC will just return a list of names. If it works, it will be used for a larger project. Creating the client and service was easy and this works well from USB. But getting the service to connect to the database isn't. I've found this site, suggesting that I should modify machine.config but that stops the XCopy deployment. This project cannot change any setting of the PC, so this suggestion is bad. I cannot create a deployment setup either. The whole thing just needs to run from USB disk. So, how do I get it to run? (The service just selects a list of names from the database, which it returns to the client. If this POC works, it will do far more complex things!)

    Read the article

  • Can I transform this asynchronous java network API into a monadic representation (or something else

    - by AlecZorab
    I've been given a java api for connecting to and communicating over a proprietary bus using a callback based style. I'm currently implementing a proof-of-concept application in scala, and I'm trying to work out how I might produce a slightly more idiomatic scala interface. A typical (simplified) application might look something like this in Java: DataType type = new DataType(); BusConnector con = new BusConnector(); con.waitForData(type.getClass()).addListener(new IListener<DataType>() { public void onEvent(DataType t) { //some stuff happens in here, and then we need some more data con.waitForData(anotherType.getClass()).addListener(new IListener<anotherType>() { public void onEvent(anotherType t) { //we do more stuff in here, and so on } }); } }); //now we've got the behaviours set up we call con.start(); In scala I can obviously define an implicit conversion from (T = Unit) into an IListener, which certainly makes things a bit simpler to read: implicit def func2Ilistener[T](f: (T => Unit)) : IListener[T] = new IListener[T]{ def onEvent(t:T) = f } val con = new BusConnector con.waitForData(DataType.getClass).addListener( (d:DataType) => { //some stuff, then another wait for stuff con.waitForData(OtherType.getClass).addListener( (o:OtherType) => { //etc }) }) Looking at this reminded me of both scalaz promises and f# async workflows. My question is this: Can I convert this into either a for comprehension or something similarly idiomatic (I feel like this should map to actors reasonably well too) Ideally I'd like to see something like: for( d <- con.waitForData(DataType.getClass); val _ = doSomethingWith(d); o <- con.waitForData(OtherType.getClass) //etc )

    Read the article

  • Hosing website on own server. What is Hardware requirement and Cost [closed]

    - by KuKu
    As i am about to finish my dream website, i need to host it on some server. I checked Amazon AWS Free Tier, i found it pretty complex. As i made full website in java(JSP + Servlet + mysql + Node.js), its been expensive to host. So i thought that why not to host on own server so that i will be fully dependent on my resources, not on any one else. And i know, in future i need to pay more and more to the other hosting company(because of uploading images and number of incoming users). So my question is, On initial stage what kind of hardware i will required. What can be the cost of that hardware? I already have 12MBPS broadband connection, will it be sufficient? It has static IP Address as well.

    Read the article

  • Perfmon quick rundown

    - by anon
    I've known quite along while about performance monitor on Windows. I have decided now to create a scheduled performance monitoring of my entire system so I can find bottlenecks for future improvements. So as you can imagine this is going to run 24/7 so I can identify peak utilization. With performance monitor on Windows 7 for example where are the logs stored (c:\perfmon)? Is there a log size? Even better is the a website that can get me up to speed with scheduling and best practices of perfmon? (I don't need an explanation of what I can monitor)

    Read the article

  • Uptime concerns in case of AWS outage

    - by Aditya Patawari
    I am running an Elastic Load Balancer backup by 2 instances in different Availability Zones in US East. I am using Multi-AZ RDS as well. Ideally this should ensure that if one AZ goes down, it should not effect the app because everything is spread across multiple AZs. But the recent AWS outage took the app down for a long time. I am not sure how this can happen. It would be great if someone can point out what went wrong. Major question here I have is how can I avoid this in future? I can setup app servers across different regions or even providers and use DNS for load balancing but what do I do with MySQL? Read Replicas will introduce some lag which I would want to avoid.

    Read the article

  • NHibernate lazy properties behavior?

    - by GeReV
    I've been trying to get NHibernate into development for a project I'm working on at my workplace. Since I have to put a strong emphasis on performance, I've been running a proof-of-concept stress test on an existing project's table with thousands of records, all of which contain a large text column. However, when selecting a collection of these records, the select statement takes a relatively long time to execute; apparently due to the aforementioned column. The first solution that comes to mind is setting this property as lazy: <property name="Content" lazy="true"/> But there seems to be no difference in the SQL generated by NHibernate. My question is, how do lazy properties behave in NHibernate? Is there some kind of type limitations I could be missing? Should I take a different approach altogether? Using HQL's new Class(column1, column2) approach works, but lazy properties sounds like a simpler solution. It's perhaps worth mentioning I'm using NHibernate 2.1.2GA with the Castle DynamicProxy. Thanks!

    Read the article

  • What are my options for a disk with what seems to be a corrupted filesystem?

    - by CT
    I have a friend with an old Dell that will not boot into Windows. It has an IDE drive. It spins up. I have an IDE to USB device. I've attached the drive via that device to a working laptop. The drive does not mount. If I go into Disk Management I can see the drive but it will not initalize, says "Drive not ready." I've also booted into a linux live cd to see if the drive mounts, it does not. I am just trying to recover some pictures from the drive. The data is not important enough to send to a professional. The issue is more of a curosity on how to recover data if and when these situations would occur in the future.

    Read the article

  • .NET Best Way to move many files to and from various directories??

    - by Dan
    I've created a program that moves files to and from various directories. An issue I've come across is when you're trying to move a file and some other program is still using it. And you get an error. Leaving it there isn't an option, so I can only think of having to keep trying to move it over and over again. This though slows the entire program down, so I create a new thread and let it deal with the problem file and move on to the next. The bigger problem is when you have too many of these problem files and the program now has so many threads trying to move these files, that it just crashes with some kernel.dll error. Here's a sample of the code I use to move the files: Public Sub MoveIt() Try File.Move(_FileName, _CopyToFileName) Catch ex As Exception Threading.Thread.Sleep(5000) MoveIt() End Try End Sub As you can see.. I try to move the file, and if it errors, I wait and move it again.. over and over again.. I've tried using FileInfo as well, but that crashes WAY sooner than just using the File object. So has anyone found a fool proof way of moving files without it ever erroring? Note: it takes a lot of files to make it crash. It'll be fine on the weekend, but by the end of the day on monday, it's done.

    Read the article

  • Opinion choosing Switch

    - by mastercode
    ) i have to reestruct a LAN network, with (currently) +/- 60hosts connected ... i have File Servers hosted, VoIP Phones,wireless AP's,printers, scanners, plotters,biometric dispositive,and 2 QNAP TS412 as FileServer and BackupServer, a Mac Mini as main Server of almost all services that need server ... and, a HP V1910-24 (L2+) and another two switches,but, only L2. which switch in your opinion, could fit better this reestruct, to ensure a VLAN division- and have to support Inter VLAN routing also - provide better performance, and also, allow a Future expansion. the budget, is low xD hehe!!

    Read the article

  • How to backup old emails locally in Thunderbird and then remove them from IMAP server?

    - by saicode
    I am using Godaddy IMAP email with Thunderbird as my desktop client on Windows 7. The email service has unlimited mailbox size but the local Thunderbird is having troubles due to the large size of the inbox/outbox. I would like to take out old emails from IMAP server and backup them locally. After backing up the old email I would like to delete old email (older than let's say, 2012) from the server. Also I'd like to have them accessible from the local backup if ever needed in the future. This way I might be able to make Thunderbird fast and problem free. Problem is, I am not able to find any instruction to do this in an automated way based on dates etc. I can find some links for Archiving, Compacting and Backup. But unable to find any tutorial about how to backup and archive it locally and delete the original emails from the server.

    Read the article

  • Is ceph usable with only 100Mbps bandwidth between nodes?

    - by vaab
    I haven't great hardware, but my requirements are low, I would like to start using ceph so as to abstract filesystem location and allow potential easy scaling to bigger hardware in an hypothetical future. My actual hardware meets ceph hardware requirements except the ethernet bandwidth part between the hosts. Mine is 100 Mbit/s which is much lower than the 1Gbps expected in ceph, even from the minimal requirement. Will I be able to use ceph in a very small smili-prod environnement (with limited number of clients) ? FYI: My hardware is 2 or maybe 3 hosts having each 4 core Intel, 24Go RAM, 2x2To disks but 100Mpbs between them.

    Read the article

  • What PowerShell/WSMan clients or queries are consuming more than 1000 requests per 2 seconds?

    - by makerofthings7
    Exchange 2010 remote administration tools are complaining with the following error [txexmb02.ibm.com] Connecting to remote server failed with the following error message : The WS-Management service cannot process the request. The system load quota of 1000 requests per 2 seconds has been exceeded. Send future requests at a slower rate or raise the system quota. The next request from this user will not be approved for at least 558475776 milliseconds. For more information, see the about_Remote_Troubleshooting Help topic. + CategoryInfo : OpenError: (System.Manageme....RemoteRunspace:RemoteRunspace) [], PSRemotingTransportException + FullyQualifiedErrorId : PSSessionOpenFailed VERBOSE: Connecting to TXEXHC02.ibm.com The help document this error referrers to says this is a WS-Man error. We're running SCOM 2007 R2 and am thinking that is increasing the query count, but I need to prove it.

    Read the article

  • Plugin-like architecture in .NET

    - by devoured elysium
    I'm trying to implement a plug-in like application. I know there are already several solution out there but this is just going to be proof of the concept, nothing more. The idea would be to make the application main application almost featureless by default and then let the plugins know about each other, having them have implement all the needed features. A couple of issues arise: I want the plugins at runtime to know about each other through my application. That wouldn't mean that at code-time they couldn't reference other plugin's assemblies so they could use its interfaces, only that plugin-feature initialization should be always through my main app. For example: if I have both plugins X and Y loaded and Y wants to use X's features, it should "register" its interest though my application to use its features. I'd have to have a kind of "dictionary" in my application where I store all the loaded plugins. After registering for interest in my application, plugin Y would get a reference to X so it could use it. Is this a good approach? When coding plugin Y that uses X, I'd need to reference X's assembly, so I can program against its interface. That has the issue of versioning. What if I code my plugin Y against an outdated version of plugin X? Should I always use a "central" place where all assemblies are, having there always the up to date versions of the assemblies? Are there by chance any books out there that specifically deal with these kinds of designs for .NET? Thanks

    Read the article

  • Best server for mailing application [closed]

    - by Cyber Junkie
    My application is similar to a reminder service that reminds users of events that they scheduled. I'm sending emails to users through a PHP script. I'm not sending one email to multiple recipients. Each recipient receives a different message. I plan to use cron jobs every minute and expect the application to send roughly 200 individual emails in 1 hour (for a small user base that may grow). I don't have hosting experience with this type of application. I plan to start on a shared host and move up in the future to vps or dedicated. Most shared hosts that I looked into allow 50-100 emails per hour with delays between mailings. Please kindly inform me what I should look for in web hosts for this kind of application.

    Read the article

< Previous Page | 111 112 113 114 115 116 117 118 119 120 121 122  | Next Page >