Search Results

Search found 19279 results on 772 pages for 'everything'.

Page 306/772 | < Previous Page | 302 303 304 305 306 307 308 309 310 311 312 313  | Next Page >

  • How to adopt scrum agile methodology for a small .Net team

    - by Thabo
    I am working on a small product based company developing .Net applications. There is a small team with 5-6 developers. I am a person responsible for planning everything. But my primary role is Software developer. Now our current project is very unstable because of poor organization. Today my boss called me and told to submit a report about required resources, appropriate methodology, required man power and their salary scales to make the current project success. I know I don’t have enough organization skills and I need to go deep in my programming skills. So I need to focus only in the development. So I can’t manage the project anymore. Now I am searching some other ways to make ongoing development success. My questions are What is the suitable agile methodology to my team? Is Scrum is suitable for above mentioned scenario? If we adopt Scrum, what we have to do next? (I think hiring new one to manage the project is more suitable. So we have to get Scrum master and some other developers.) Are there any resources (books, Blogs and etc) to get some tips and advices to solve this problem? If Scrum is not a suitable methodology for our scenario, what else can be more suitable methodology to adopt? Can anyone give a good solution for my problem?

    Read the article

  • SQL server 2000 reporting bad values to ASP.Net Application

    - by Ben
    I have an instance of SQL server 2000 (8.0.2039) with a rather simple table. We recently had users complain about an application I wrote returning bad values for some of the dates in the databse. When I query the table directly via Server Management Studio, it will return the correct values, however the identical queries from my application report the wrong values, but only for a couple of dates. I have been over the code, and it is solid. If the error was in the code, all of the dates reported should be wrong. I have also run the code on an identical test database, and everything is reported properly. I believe the problem may lie in the sql instance itself, which is why I am posting in Server Fault. My question is, has anyone heard of a database reporting bad (incorrect) date values when queried via web application? It should be noted that this particular server was once manually rebuilt after having a cluster clean run on it.

    Read the article

  • Updating PHP on a Plesk managed Server

    - by mblaettermann
    I just updated PHP and MySQl on my VPS with the current Versions from Atomic Repo. Everything worked out fine so far. From console I get the new PHP 5.3: [root@server phpMyAdmin]# php -v PHP 5.3.16 (cli) (built: Aug 20 2012 11:18:05) Copyright (c) 1997-2012 The PHP Group Zend Engine v2.3.0, Copyright (c) 1998-2012 Zend Technologies with the ionCube PHP Loader v4.0.5, Copyright (c) 2002-2011, by ionCube Ltd. But through Apache I still get the old version (5.1.6). The server is running some old version of crappy Plesk Panel. That gives me the option to choose between Apache Modul, fCGI and CGI-BIN. Any hints, how to update apache, so it will use the new PHP Version? EDIT: I just needed to restart httpd (/etc/init.d/httpd restart)

    Read the article

  • Acrobat 8 on Windows 7 64-bit

    - by user66822
    Here's the situation I'm in. I have Acrobat 8 on Wows 7 (64-bit) and anytime I create PDFs any numbers on the page I'm PDFing come through as garbage. Everything else on the page comes out fine, just the numbers are messed up. They appear as random symbols. I've uninstalled, reinstalled. Applied all updates, not applied updates, nothing seems to help. Mayube I'm a bad searcher, can't put in the right search terms, but I can't find anything that seems to point me in a direction on this. ANy ideas?

    Read the article

  • OS X can't resolve localhost suddenly

    - by Conor
    Last week I fired up a website that I'm currently developing locally only to find out that it wasn't working as it was the night before, (or at all). After an inital stage of panic and 'what did I do' moments... I deduced the problem down to the fact that my OS X now wont resolve localhost properly, so connections to my SQL database were failing. I can still ping localhost in the terminal, but in order to get my websites up and running again, I had to change all the localhost entries to 127.0.0.1 This isn't a huge problem as everything is up and running again, but I would like to try to get to the bottom of it. I have a sneaking suspicion that an apple software update caused this issue, as I don't recall doing anything else that would have had any effect. Other than my hosts file (which looks normal), what else could be causing this? Running OSX 10.6.4

    Read the article

  • What causes Windows Boot to stall?

    - by Nick Berardi
    For about 6 months now I have been having this weird problem where Windows 7 fails to fully boot correctly. What happens is this. Starting Windows shows up on the screen. Then 3 out of 4 times nothing else happens, no Windows Flag animation, just nothing occurs. After 3 or 4 restarts repeating steps 1-2 above, the Windows Flag animation finally shows up and everything works as expected. My question is what is causing this problem in steps 1 and 2? Because I have tried the following with no luck: Error checking and correcting of any disk errors Updating drivers Doing a clean install of Windows 7 My setup is as follows: Windows 7 64-bit Ultimate 8 GB RAM 128 GB Crucial SSD (firmware 0005) Dell Latitude E6410 Intel Wireless and Graphics Other than what I have tried above I am totally out of ideas and looking for some new ones to try.

    Read the article

  • CA For A Large Intranet

    - by Tim Post
    I'm managing what has become a very large intranet (over 100 different hosts / services) and will be stepping down from my role in the near future. I want to make things easy for the next victim person who takes my place. All hosts are secured via SSL. This includes various portals, wikis, data entry systems, HR systems and other sensitive things. We're using self signed certificates which worked o.k. in the past, but are now problematic because: Browsers make it harder for users to understand exactly what is going on when a self signed certificate is encountered, much less accept them. Putting up a new host means 100 phone calls asking what "Add an exception" means What we were doing is just importing the self signed certs when we set up a new workstation. This was fine when we only had a dozen to deal with, but now its just overwhelming. Our I.T. Department has classified this as ya all's problem, all we get from them is support for switch and router configurations. Beyond the user having connectivity, everything else is up to the intranet administrators. We have a mix of Ubuntu and Windows workstations. We'd like to set up our own self signed CA root, which can sign certificates for each host that we deploy on the intranet. Client browsers would of course be told to trust our CA. My question is, would this be dangerous and would we be better off going with intermediate certificates from someone like Verisign? Either way, I still have to import the root for the intermediate CA, so I really don't see what the difference is? Other than charging us money, what would Verisign be doing that we could not, beyond protecting the root CA cert so it can't be used to make forgeries?

    Read the article

  • System time wrong after running ntpdate because DST ignored

    - by Ian Dunn
    When I run ntpdate, my system clock displays the time as an hour behind what it should be. I know that ntpdate does everything in UTC, so I'm guessing there's a timezone setting wrong and it's ignoring Daylight Savings Time, but I can't figure it out. Here's what I've done so far: ln -sf /usr/share/zoneinfo/EST /etc/localtime to set the timezone Set UTC=true in /etc/sysconfig/clock so that DST will be automatically applied date -s hh:mm::ss to set system clock correctly hwclock -systohc --utc to set the hardware clock correctly At this point date and hwclock both display the correct time. But if I then run ntpdate 0.us.pool.ntp.org, the date output is an hour behind what it should be. I've looked at a dozen tutorials and can't figure out what I'm doing wrong. Does anyone have any ideas?

    Read the article

  • Scenes from OpenWorld Day One

    - by Larry Wake
    Sunday's the day that everything comes together, but there's always that last minute scramble. Here are a few peeks at what everyone's doing, and may still be doing far into the night. This is the team putting the final touches on the Hands-On Lab room for  HOL10201, "Reduce Risk with Oracle Solaris Access Control to Restrain Users and Isolate Applications". This should be a great learning experience--plus it's a chance to meet up with some of the top Solaris security people, including Glenn Faden and Darren Moffat. And here's the OTN Garage's own Rick Ramsey, working feverishly to help set up the Oracle Solaris Systems Pavilion. (Moscone South, Booth 733). Several of our featured partners will be demonstrating solutions running on Oracle Solaris systems -- plus, we'll be serving espresso, to help you power through the week. Another panorama shot, courtesy of iOS 6 -- come for the maps, stay for the photos.... Moscone South is also home once again this year to the systems and storage DEMOgrounds. Plenty to learn and see; you might even catch a glimpse of me there on Tuesday afternoon.

    Read the article

  • How to refactor when all your development is on branches?

    - by Mark
    At my company, all of our development (bug fixes and new features) is done on separate branches. When it's complete, we send it off to QA who tests it on that branch, and when they give us the green light, we merge it into our main branch. This could take anywhere between a day and a year. If we try to squeeze any refactoring in on a branch, we don't know how long it will be "out" for, so it can cause many conflicts when it's merged back in. For example, let's say I want to rename a function because the feature I'm working on is making heavy use of this function, and I found that it's name doesn't really fit its purpose (again, this is just an example). So I go around and find every usage of this function, and rename them all to its new name, and everything works perfectly, so I send it off to QA. Meanwhile, new development is happening, and my renamed function doesn't exist on any of the branches that are being forked off main. When my issue gets merged back in, they're all going to break. Is there any way of dealing with this? It's not like management will ever approve a refactor-only issue so it has to be squeezed in with other work. It can't be developed directly on main because all changes have to go through QA and no one wants to be the jerk that broke main so that he could do a little bit of non-essential refactoring.

    Read the article

  • Enter / Return key stops working

    - by andygrunt
    I have a Dell Latitude E5400 laptop running Windows 7. Everything runs fine on it for days then I suddenly notice the enter/return key is no longer working e.g. I could be writing an email, press enter for a new line but it doesn’t ‘register’ – so no new line. I thought it might be some dirt under the key perhaps so I tried running the Windows onscreen keyboard but the Enter key on that doesn’t work either. A reboot always fixes the problem but it’s a pain. Any ideas a) how to fix it and b) out of interest, is there some other key combination that I can use as an alternative to using the Enter key when I need to?

    Read the article

  • Tidying up old apache logs managed by vlogger

    - by Andre Lackmann
    We're using vlogger to manage our apache logs which keeps everything nice and neat but pretty much breaks the ability to use logrotate from as far as I can see. eg. our virtual access.logs each sit within their own dir and are named similar to: /virtual.com/ 20100501-access.log 20100502-access.log 20100503-access.log 20100504-access.log etc.. Has anyone created a cleanup script to go through the /var/log/httpd/ sub dirs and remove old logs? We like using vlogger, but cleaning up olds logs after it is a pain!

    Read the article

  • Internet Explorer defaults to 64-bit version

    - by Tim Long
    My IE8 has suddenly started defaulting to the 64-bit version. I have no idea how or why this has happened, but I suspect it might be linked to the Browser Choice Screen that Microsoft was recently forced to display by EU law. However, many web sites will not display correctly in IE8 x64 (eg. sites that use Adobe Flash or Microsoft Silverlight). I have the 32-bit version of IE pinned to my taskbar and if I launch it manually, everything is fine. But when I click on a URL from another program and IE is not already running, then the 64-bit version gets launched. This really messes with programs like BBC iPlayer which rely heavily on Adbobe Air and Flash. So, how do I get IE8 32-bit version to be the default version again? I've tried using the "default programs" control panel and that doesn;t make any difference (in fact, it doesn't give the choice between x84 and x64 versions, it just lists "internet explorer").

    Read the article

  • Problems connecting Centos on VMware to the network using bridged connection.

    - by Sakin
    Hi, I installed CentOs on VMware running on windows XP. When trying to configure it to connect to the internet in a bridged configuration, I get an error message when trying to bring up the network interface: [root@VMLinux ~]# /et/init.d/network start Bringing up loopback interface: [ OK ] Bringing up interface eth0: Determining IP information for eth0... failed [FAILED] VM is running on a machine that has access to the network, I tried it on two different networks that have DHCP enabled. Everything works fine when using a NAT connection through my host. How can I make the bridge work for me? Thanks.

    Read the article

  • Best practices for setting lm-factor in Squid refresh patterns

    - by Mpentecost
    I am running a Squid (3.1) cache in front of Django. The content of the site does not change very often, so Squid gives our backend much needed breathing room. Currently, this is the refresh pattern that we are using to cache the content: refresh_pattern . 60 100% 60 We basically want to cache everything for at least an hour (and only an hour) before Squid then re-validates the content. My question is on the "100%" parameter, which sets the lm-factor. I'm not sure if setting that to 100% is doing what we want it to. The assumption was that by setting it to 100%, it would ensure that objects stay in the cache for the max cache time. Is this an incorrect assumption? What are the best practices that one should follow when setting up a refresh pattern like this?

    Read the article

  • Application Scope v's Static - Not Quite the same

    - by Duncan Mills
    An interesting question came up today which, innocent as it sounded, needed a second or two to consider. What's the difference between storing say a Map of reference information as a Static as opposed to storing the same map as an application scoped variable in JSF?  From the perspective of the web application itself there seems to be no functional difference, in both cases, the information is confined to the current JVM and potentially visible to your app code (note that Application Scope is not magically propagated across a cluster, you would need a separate instance on each VM). To my mind the primary consideration here is a matter of leakage. A static will be (potentially) visible to everything running within the same VM (OK this depends on which class-loader was used but let's keep this simple), and this includes your model code and indeed other web applications running in the same container. An Application Scoped object, in JSF terms, is much more ring-fenced and is only visible to the Web app itself, not other web apps running on the same server and not directly to the business model layer if that is running in the same VM. So given that I'm a big fan of coding applications to say what I mean, then using Application Scope appeals because it explicitly states how I expect the data to be used and a provides a more explicit statement about visibility and indeed dependency as I'd generally explicitly inject it where it is needed.  Alternative viewpoints / thoughts are, as ever, welcomed...

    Read the article

  • How to build an outbound load balancer with linux?

    - by matnagel
    We have a small house in the countryside and there is no fixed broadband, so we had a mobile flatrate first, and for 2 people with 2 computers it was too slow, so now we have 2 flatrates for 2 client machines. So I pay 2 flatrates and have double bandwith theoretically. There is a local network in the house that connects everything. But when I am alone I wonder how I can use both connections at the same time. I want to build a solution where I can browse the web and page requests are spread between the 2 connections. I imagine there are expensive routers who can split the traffic between 2 lines. But is there a good way to do it with linux? The solution I am looking for will split the requests already for one page (multiple images, css files, javascrfipt files) between the two lines.

    Read the article

  • Greeter login screen cropped login options in 12.04

    - by ammianus
    I have a pretty newly installed Ubuntu 12.04, using Unity. My external monitor is 1920x1080 max resolution. In the Unity desktop itself everything looks great. I have an NVidia graphics card. When I start my computer and get to the Unity greeter login screen the display is oddly formatted and the resolution seems off. It looks like a zoomed view on the larger 1920x1080 screen. As such it crops the login options off to the left hand side of the screen. So I can only just see the edge of the password box for the user I want to log in with. I can log in with one account by default by blindly typing the password, but I am unable to switch to other accounts. Is there anything I can do to fix the log in screen display so that I can see the normal login options? Note: I first noticed it when I changed my desktop background and the next time I logged in I saw the issue.

    Read the article

  • Will disabling NAT on my wireless router cause a performance hit on my DSL router?

    - by user117313
    I have a Thomson TG508v2 router/modem and a brand new AirPort Extreme connected to it. At first, I set the AirPort Extreme to "share a public IP", which enables DHCP/NAT. Everything works great but I was having the Double NAT error. So I set the AirPort Extreme as bridge mode, which disables DHCP/NAT and let my crappy modem handle my internal network IPs. Will this cause any performance hit in my network? I'm worried because I'll outsource this job from AirPort Extreme (premium hardware) to the crappy modem. Before you suggest, I tried setting my modem as a bridge and configuring PPPoE on the AirPort Extreme, however it wouldn't connect to the internet, so I just dismissed leaving the modem as bridge.

    Read the article

  • Internet Explorer 8/9 javascript disable, need to enable back

    - by Shiro
    Dear all, This is not a programming issue, just want to ask any one having Internet Explorer Javascript disabled issue. I try go to Internet Option enable everything, including javascript restart it. But all the javascript still disabled. I go to jQuery API website, test the javascript, at the Demo part nothing show. Even I write a alert("hello"); in a html file, nothing come out. I am not sure what happen to my Internet Explorer. Initially I am using IE8, I try to update to IE9Beta, IE9RC, the problem start from IE8-IE9RC still haven't resolved. I am not sure what application crash with the IE. Please advice! It is Urgent. Thanks!

    Read the article

  • Choose an output audio device different from the default on WMP 11

    - by GetFree
    I like to play my music through a Hi-Fi audio equipment and everything else (like windows sounds, web videos and such) through my default PC speakers. On WIndows XP I had WMP 9 and I could do that with no problems since I can choose what audio device (which sound card) to use, and that selection is for WMP only, which can be different from Windows' default audio device. But now that I have Windows Vista and WMP 11 I cannot longer choose an audio device just for WMP, or at least I can't find a way to do it (the control in the options dialog is no longer there). Was this useful feature really removed from WMP 11? or there is some other way to do it?

    Read the article

  • Bluetooth not working on Acer 3820tg (Ubuntu 12.04)

    - by Iohannes
    After upgrading to 12.04 (it was done by a complete reinstallation), bluetooth doesn't work anymore. Before (I think it was Ubuntu 11.04), I just followed instructions on this page and copied /lib/firmware/ath3k-2.fw to /lib/firmware/ath3k-1.fw and everything was fine. But in 12.04, there isn't any ath3k-2.fw in the /lib directory, so this method is not possible. Also, pressing Fn+F3 doesn't work. I would appreciate any help. Some further information (I hope it's enough): dmesg [ 23.063710] Bluetooth: Core ver 2.16 [ 23.063740] NET: Registered protocol family 31 [ 23.063743] Bluetooth: HCI device and connection manager initialized [ 23.063746] Bluetooth: HCI socket layer initialized [ 23.063747] Bluetooth: L2CAP socket layer initialized [ 23.063972] Bluetooth: SCO socket layer initialized [ 23.066930] Bluetooth: BNEP (Ethernet Emulation) ver 1.3 [ 23.066934] Bluetooth: BNEP filters: protocol multicast [ 23.068422] Bluetooth: RFCOMM TTY layer initialized [ 23.068428] Bluetooth: RFCOMM socket layer initialized [ 23.068430] Bluetooth: RFCOMM ver 1.11 lsusb Bus 001 Device 001: ID 1d6b:0002 Linux Foundation 2.0 root hub Bus 002 Device 001: ID 1d6b:0002 Linux Foundation 2.0 root hub Bus 001 Device 002: ID 8087:0020 Intel Corp. Integrated Rate Matching Hub Bus 002 Device 002: ID 8087:0020 Intel Corp. Integrated Rate Matching Hub Bus 002 Device 003: ID 046d:c05f Logitech, Inc. Bus 002 Device 004: ID 046d:c31c Logitech, Inc. Keyboard K120 for Business Bus 002 Device 005: ID 0402:9665 ALi Corp. Gateway Webcam hcitool dev Devices:

    Read the article

  • Using default parameters for 404 error (PHP with mvc design)?

    - by user1175327
    I have a custom made Framework (written in PHP). It all works very good, but i have some doubts about a certain thing. Right now when a user call this url for example: http://host.com/user/edit/12 Which would resolve to: user = userController edit = editAction() in userController 12 = treated as a param But suppose the controller 'userController' doesn't exist. Then i could throw a 404. But on the other hand, the url could also be used as params for the indexController (which is the default controller). So in that case: controller = indexController user = could be an action in indexController, otherwise treated as a param edit = treated as a param 12 = treated as a param That is actually how it works right now in my framework. So basically, i never throw a 404. I could ofcourse say that only params can be given if the controller name is explicitly named in the URL. So if i want the above url: http://host.com/user/edit/12 To be invoked by the indexController, in the indexAction. Then i specifically have to tell what controller and action it uses in the URL. So the URL should become: http://host.com/index/index/user/edit/12 index = indexController index (2nd one) = the action method user = treated as a param edit = treated as a param 12 = treated as a param That way, when a controller doesn't exist, i don't reroute everything as a param to the index controller and simply throw a 404 error. Now my question is, which one is more preffered? Should i allow both options to be configurable in a config file? Or should i always use one of them. Simply because that's the only and best way to do it?

    Read the article

  • 504 Gateway Time-out after php fatal error

    - by tiagojsag
    I'm using nginx and php-fpm to develop a symfony2 based website, under ubuntu 12.10 (yes, I know I'm using a beta OS). Everything was working out fine until, due to an error on my code, I called an unexisting function, and got the following: Fatal error: Call to a member function (....) This isn't a problem (it's a bug in my code, easily fixable), but after this, no other page loads. My browser just keeps trying to load the page from the webserver, until nginx timeouts (after +- 30s, which should be some default timeout) and returns: 504 Gateway Time-out Restarting php-fpm solves the issue. Nginx logs show a timeout message, and nothing appears on php-fpm logs, even if I set them to debug level. I tried switching from fpm to fastcgi, and the same thing happens. I've looked around, but all similar error are related to big requests/file handling, which isn't the case. All the pages on my website load in a few seconds, even under development conditions (no caching, etc).

    Read the article

  • Writing Acceptance test cases

    - by HH_
    We are integrating a testing process in our SCRUM process. My new role is to write acceptance tests of our web applications in order to automate them later. I have read a lot about how tests cases should be written, but none gave me practical advices to write test cases for complex web applications, and instead they threw conflicting principles that I found hard to apply: Test cases should be short: Take the example of a CMS. Short test cases are easy to maintain and to identify the inputs and outputs. But what if I want to test a long series of operations (eg. adding a document, sending a notification to another user, the other user replies, the document changes state, the user gets a notice). It rather seems to me that test cases should represent complete scenarios. But I can see how this will produce overtly complex test documents. Tests should identify inputs and outputs:: What if I have a long form with many interacting fields, with different behaviors. Do I write one test for everything, or one for each? Test cases should be independent: But how can I apply that if testing the upload operation requires that the connect operation is successful? And how does it apply to writing test cases? Should I write a test for each operation, but each test declares its dependencies, or should I rewrite the whole scenario for each test? Test cases should be lightly-documented: This principles is specific to Agile projects. So do you have any advice on how to implement this principle? Although I thought that writing acceptance test cases was going to be simple, I found myself overwhelmed by every decision I had to make (FYI: I am a developer and not a professional tester). So my main question is: What steps or advices do you have in order to write maintainable acceptance test cases for complex applications. Thank you.

    Read the article

< Previous Page | 302 303 304 305 306 307 308 309 310 311 312 313  | Next Page >