Search Results

Search found 5298 results on 212 pages for 'automated deploy'.

Page 125/212 | < Previous Page | 121 122 123 124 125 126 127 128 129 130 131 132  | Next Page >

  • Using Arch Linux computer as a server for Rack Apps

    - by wxl
    What would be the best way to go about using an Arch Linux computer as a Rack (as in Ruby Rack, not an actual rack server) server? Here's what I want to be able to do: Automatically deploy on a git push to the server. (I already have this worked out, on post-receive the server checks out the app to /home/git/app from /home/git/app.git.) Run a Rack server application to serve up this app, one that can be restarted on demand. Run a MongoDB server Be able to access the app by going to my-server.local/app or something similar. (It's really only going to be used on the local network, no port forwarding or outside use) Any ideas would be greatly appreciated. I apologize if this seems too "do it for me".

    Read the article

  • Will a Response.Redirect exception go to global.asax?

    - by mgmedick
    I'm aware that when you call Response.Redirect it fires a ThreadAbortException. A co-worker has demonstrated calling response.redirect and then it goes to the global.asax. For the life of me I cannot get the ThreadAbortException to go to the global.asax, its like it is being suppressed naturally in the system. The reason I'm asking this is we believe the response.redirect is the cause of some automated error emails, but I'm not convinced this is the case especially if I can't even get it to debug into the global.asax. Any Ideas why I can't get the Response.Redirect to fire the global error handler?

    Read the article

  • Web Farm Application deployment best practices

    - by rauts
    Hi All, We are having a web farm which hosts multiple ASP.Net applications. We typically have 4 servers on the farm. The dilemma which i am having is in terms of capacity issue of the farm. Lets say i have currently got 200 apps in total. Should I deploy all 200 apps on all 4 servers (i.e. all the servers in the farm are identical) or should i split the applications between 2 sets of server and create 2 smaller farms so that i can then manage the application based on its criticality and usage etc. Any best practices in this area would be highly appreciated. Thanks Rauts

    Read the article

  • Deploying a Windows 7 image, which way is the fastest?

    - by MatF
    I captured an image of a basic Windows 7 installation with some modifications using imagex. Before the image was captured, I ran sysprep generalizing and selecting to enter OOBE after it's done. Which way would be the fastest to deploy that image again? Using imagex /apply. Or naming the image install.wim and put it in the source folder of a normal installation (on a bootable USB device), running a normal setup afterwards. Currently I have only tried the second approach. However I just found out about the imagex way and wondered if it would be faster. Or are the even more methods that would be better?

    Read the article

  • Can a working Tomcat 6 webapp be turned into a usable .war file?

    - by Bill Cole
    Problem: I have a working webapp on a FreeBSD 8.1 Tomcat 6 test server that I need to move to a production system. The developer who last touched it (and had root on that server) has moved on and isn't helpful. The running app seems to have been deployed from a CVS server that is now unavailable. My thinking is that I would like to find a way to wrap the working webapp into a proper .war so that I can deploy it on a pristine host and (after testing) send the existing system to a very deep bitbucket. But I'm not having luck finding a way to do that. I'm a sysadmin not a developer and don't work much with Tomcat systems so I may be (likely am) overlooking something blindingly simple. I gather that I may be able to just tar up the deployed directory and untar it on the new machine, but I have a nagging feeling that there are pitfalls in that.

    Read the article

  • How to setup a virtual machine in Ubuntu desktop to run Debian Server

    - by stickman
    I want to run a virtual machine in my Ubuntu desktop that runs a Debian server. The purpose of this is to generate Debian packages. I have some C++ applications that were originally developed on my Ubuntu machine, and I need to (re)compile them on a Debian server in order to: build Deb packages for deployment on a Debian server make sure that the applications will definitely work on a debian server The idea is so that I can do 90% of my development on Ubuntu (where I am more comfortable), and deploy a binary package that definitely works on Debian. BTW, I am developing on Karmic Kola (Ubuntu 9.10).

    Read the article

  • Is there a way to control two instantiated systemd services as a single unit?

    - by rascalking
    I've got a couple python web services I'm trying to run on a Fedora 15 box. They're being run by paster, and the only difference in starting them is the config file they read. This seems like a good fit for systemd's instantiated services, but I'd like to be able to control them as a single unit. A systemd target that requires both services seems like the way to approach that. Starting the target does start both services, but stopping the target leaves them running. Here's the service file: [Unit] Description=AUI Instance on Port %i After=syslog.target [Service] WorkingDirectory=/usr/local/share/aui ExecStart=/opt/cogo/bin/paster serve --log-file=/var/log/aui/%i deploy-%i.ini Restart=always RestartSec=2 User=aui Group=aui [Install] WantedBy=multi-user.target And here's the target file: [Unit] Description=AUI [email protected] [email protected] After=syslog.target [Install] WantedBy=multi-user.target Is this kind of grouping even possible with systemd?

    Read the article

  • How to make nginx only respond to one domain?

    - by larryzhao
    I am pretty new to nginx, I host my rails application on nginx+passenger. I want my website to be accessible to only one domain. So I set my nginx conf like the following: server { listen 80; server_name mydomain.com www.mydomain.com; root /var/deploy/myapp/current/public; passenger_enabled on; location ~* \.(js|css|png|jpg|jpeg|gif|ico)$ { expires 1y; add_header Cache-Control public; } } I specify the server_name directive, but still, it answers anything which points to this IP and I could see that in the access.log that it answers to other domain names. Is there anything I am doing wrong?

    Read the article

  • How do you keep up with Nagios/Capistrano configs when using EC2?

    - by imaginative
    I use Amazon EC2 for my mobile app. Depending on load of the application at a given time, I might spawn new instances and then take them down when load is lower to save costs. How does one keep up with Nagios configurations for such a dynamic environment? When one deals with managed hardware, configuration files are predictable. In this case Nagios, Capistrano and a bunch of other configuration files would need to be added. Capistrano needs to know where to deploy a new build to for an app server. Nagios needs to know to remove an existing instance or add a new instance for monitoring. Nagios also needs to know if a node was intentionally taken down or if the host is down due to error. How is this done with the wonderful world of VPS/dynamic instances?

    Read the article

  • Determining End of Line Delimiter

    - by Michael Merchant
    How do you verify what the delimiter for a file is? I'm using eclipse to edit a script file on my windows system that I use to deploy on my unix server. The script specifies file names and I'd like to be certain that eclipse is saving the files with unix end of line delimiters. I've set my preferences to use the Unix Line Delimiter on new files using: Window > Preferences > General > Workspace > "New text file line delimiter" And have changed my script file's line delimiter using. File -> Convert Line Delimiters To Is there a way (preferably with the eclipse IDE) to check that there is no "\r"?

    Read the article

  • Concerns with SoapUI Pro when Updating Web Services

    - by Plenoge
    This may unfortunately be a very open-ended question, but a coworker of mine has shared concerns about using SoapUI Pro as our automated regression testing tool primarily on the basis that seemingly trivial WSDL changes can cause headaches for the QA crew. My coworker's particularly concerned that if we update the WSDL, even with a name change, we'll need to touch and update each existing test. As I understand it SoapUI Pro offers a refactoring tool to handle this for you. What are your experiences with SoapUI Pro and updating your web services? Am I missing something?

    Read the article

  • Deployment from OVA format

    - by Manvendra Bele
    I am deploying a VM using a OVA format. The size of OVA format is 57 GB. Currently free space on my datastore is 388 GB. At the time of selecting Disk Format type if shows me in red that the disk size required is 1 TB therefore you cannot select THICK provisioning. Therefore, i selected THIN provisiong. It THIN provisioing i am showed that Estimated Disk Usage is 112 GB which is less than the free space available. But even after selecting THIN proviosing at the time of deployment it throws an error that it cannot create disk as the size of disk is larger than the maximum specified limit. My block size is of 1 MB. Pasting my exact error here: Failed to deploy OVF package:File [datastore1] IMS Tester 1/IMS Tester1_2.vmdk is larger than maximum size supported by datastore 'datastore1

    Read the article

  • process running on login: can't find in AD or login batch scripts

    - by tombull89
    Hallo, I'm trying to deploy some classroom control software (NetSupport School) to some of the machines on our network but for some reason when you log off and restart the computer any user who logs on ends up re-installing the software while logging on. I spent two hours on the phone to the complanys support and we eventually nailed it down to most likely a setting in Active Directory or in the login.bat (drive mapping and settings) but we can't find anything in those that would say "run this installer at logon". Is there anywhere else on the system that would set something like this? Server 2003/XP. Ta!

    Read the article

  • Printing All Entries in A PHP Table

    - by mgunawan
    I'm trying to insert a php excerpt with SQL (I understand this is outdated, but am trying to grasp the syntax first) into my HTML page, and I've got the following table: ID Name Element1 Element2 0 John John's 1st John's 2nd 1 Bill Bill's 1st Bill's 2nd 2 Steve Steven's 1st Steve's 2nd I'm trying to get the for loop that will essentially print out the following in my html page Name: Name where ID=0 Element1: Element1 where ID=0 Element2: Element2 where ID=0 Name: Name where ID=1 Element1: Element1 where ID=1 Element2: Element2 where ID=1 and so forth. Basically, I am trying to make this process automated so that whenever a new record is added into the table, the HTML page will automatically update with a new "profile". Thank you for your help!

    Read the article

  • mysql error code 13 on windows xampp caused by lower case table names = 0

    - by user127379
    I can import an sql (from test linux server mysql) file if the lower case setting is removed. But then the table names are lower case and the web site doesn't work. Originally it was working (my.ini with the lower case settings), I then exported to a linux server, it was working there. Now importing back to my windows (xampp setup) fails. After wild goose chase looking at disks and permissions, I found that if I remove the lower_case_table_names=0, the import works! But I need the case sensitive command so that I can deploy on the linux server.

    Read the article

  • To be a lazy developer or not to be a lazy developer?

    - by JamesStuddart
    Am I a lazy developer? Is it being lazy to use automated tools, such as code generators and such? Now, I could, if I had to, create all the data layers and entities I needed, but I choose to use CodeSmith to generate my datalayers and entities. I also use Resharper and I would say it fights with MSDeploy as to which gets installed first after Visual Studio. Again if I had to, I could code without it, but prefer not to. Both these tools from my point of view are no brainers as they improve output massively. But is this lazy? I'm sure there are purists out there that would say everything should be wirtten by you so you know what everything is doing, but if you can read through the code and see what is happening is that ok? So am I being lazy or am I just using all the cards in my hand?

    Read the article

  • Windows Task Scheduler

    - by Zulakis
    i am trying to deploy a auto-starting program with Administrator Priviliges on our XP-SP1-machines. For this, i am using the Windows Task Scheduler. Since most of our machines get deployed by using a PXE-imaging-system, the Task fails because the Administrator user entered is for example r126/Administrator. If i only enter Administrator then it automatically changes to machinename/Administrator. Since the machinenames are automatically changes by the imaging-system, the tasks fail run. Any ideas on how to fix that?

    Read the article

  • Deploying Sharepoint Features in a Load Balanced Environment

    - by Adam
    Last night we deployed a new set of Sharepoint features to a load balanced environment. For some reason the new features are on 1 box but are not showing in the sharepoint sites on the others. We have 4 servers and we deployed to them by pulling 1 server out of rotation, stopping the app pool and deploying our new code and the new features. Then we would fire it back up and add it to the rotation. For the remaining servers we would only remove the server from rotation, stop the app pool, and deploy the code, NOT the features, then fire it back up and add it to the rotation. Any thoughts on why the features are not showing up on the other servers? Also, any thoughts on forcing the features to show up? Thanks in advance.

    Read the article

  • how do copyright permission systems for content hosting sites work?

    - by zebraman
    I am wondering about subscription sites that host content, like recorded performances from concerts. I'm sure there is a tangle of copyright permissions that must be granted for these video/audio files to be hosted. For example, if a band plays a cover of another band's song, permission must be obtained from not only the band that performed, but the band that owns the song. Perhaps even from the venue that hosted the performance, to record the video and post the content. I am curious how websites that host content like this work. How might an automated copyright system work to keep track of who has ownership of certain performances and obtain permission from said owners to record and post their content.

    Read the article

  • Amazon EC2 Elastic Load Balancing - strategy for zero downtime server restart

    - by Yoga
    I have 5 web servers (Apache/mod_perl) behind Amazon EC2 Elastic Load Balancing, when I deploy codes to the web servers, I am doing this.. For each machine, shutdown the Apache Update the code Start over the server and proceed to the next server I think when my server is shutdown, ELB will not distribute request to my server, but how about the request still serving? I think a better approach is Stop accepting new request from ELB Sleep for sometimes, shutdown web server only if all requests are responded Update the codes Start the server again But how to perform (1) and (2) from my local sever? Do I need to use AWS API? or other easy way to do it? Thanks.

    Read the article

  • What technologies are appropriate for a human workflow system?

    - by CCw
    I'm researching various workflow architectures and it is overwhelming. The workflow system I am creating will be almost completely human-driven. Very little, if any, asynchronous activity will be taking place. One possibility is to simply use a RDBMS and have a task table, from which stored procedures would be used to enforce synchronous access to each task. This seems very simple, but I'm having a hard time coming up with reasons why I might need to involve a heavier solution. If my system has ~500 concurrent users, and there is very little in the way of automated or asynchronous tasks, should I even consider the various workflow patterns/packages out there like Mule, BPEL/SOA, Spring Work Flow, etc?

    Read the article

  • Update Facebook Page's status using pyfacebook

    - by thornomad
    I am attempting to add functionality to my Django app: when a new post is approved, I want to update the corresponding Facebook Page's status with a message and a link to the post automatically. Basic status update. I have downloaded and installed pyfacebook - and I have read through the tutorial from Facebook. I have also seen this suggestion here on SO: import facebook fb = facebook.Facebook('YOUR_API_KEY', 'YOUR_SECRET_KEY') fb.auth.createToken() fb.login() # THIS IS AS FAR AS I CAN GET fb.auth.getSession() fb.set_status('Checking out StackOverFlow.com') When I get to the login() call, however, pyfacebook tries to open lynx so I can login to Facebook 'via the web' -- this is, obviously, not going to work for me because the system is supposed to be automated ... I've been looking, but can't find out how I can keep this all working with the script and not having to login via a web browser. Any ideas?

    Read the article

  • Routing between two subnets. (Need Solution)

    - by rehanplus
    Need help according to scenario given: Client end PCs: 400 + Network : Server 1 (Linux) : 192.168.2.0/24 (For Application, Internet not working) GW: 192.168.2.1 Clients: 192.168.2.1 - 254 Server 2 (Linux) : 192.168.3.0/24 (For Internet users) GW: 192.168.3.1 Clients: 192.168.3.2 - 254 Server 2 is connected to DSL Broadband. Server 1 and Server 2 both are on same physical network i.e. Same switches. Current issue: i have to deploy a file and print server but this server will be accessed by both (192.168.2.x and 192.168.3.x) one same workgroup. as both subnets are on same switched network. Limitations: Currently there is no hardware routers and firewall. Need to complete this task with Linux / Windows / AD. Tested / Worked so far: Configured one PC with two NIC's With the IPS: NIC 1 : 192.168.3.2 GW: 192.168.3.1 Subnet: 255.255.255.0 NIC 2 : 192.168.2.2 GW: Empty Subnet: 255.255.255.0 Kindly provide any solution what should i do to get sharing enable on both Subnets. Thank you All

    Read the article

  • Crossplatform "jail" for an application

    - by Alexander
    We currently have a variety of systems (Linux, Solarix, *BSD, HP-UX ...) on which we are not allowed to install anything into / (but I have root access. That's strange, I know). But we'd like to run Puppet on all of them. So, the obvious idea is to install Puppet with all prebuilt dependencies into some isolated tree, something like "jail", which will allow to use dependences from some prefix and to access the host system. The big advatanges would be uniform deployment and updates. One solution that came to my mind is to deploy Gentoo Prefix, and install Puppet there with package manager. However, this requires a lot of extra space and some manual patching for each system. Maybe there are some more elegant and simple solutions?

    Read the article

  • Embed Git Commit Log in Rails App?

    - by Andrew
    So, I have a 'development blog' in a rails app I'm working on right now. I'm using Git for version control and deployment (although right now I'm the only person working on it). Now, when I make changes in Git I put a pretty decent log entry about what I've done. I'd love to have the Git commit log automatically posted to the development blog -- or otherwise available for others to read within the deployed site. Is there an automated way to pull the Git Commit Log into a view in a rails app?

    Read the article

< Previous Page | 121 122 123 124 125 126 127 128 129 130 131 132  | Next Page >