Search Results

Search found 22653 results on 907 pages for 'robert may'.

Page 483/907 | < Previous Page | 479 480 481 482 483 484 485 486 487 488 489 490  | Next Page >

  • A key principle of Scrum...

    - by AndyScott
    "A key principle of Scrum is its recognition that during a project the customers can change their minds about what they want and need (often called requirements churn), and that unpredicted challenges cannot be easily addressed in a traditional predictive or planned manner. As such, Scrum adopts an empirical approach—accepting that the problem cannot be fully understood or defined, focusing instead on maximizing the team’s ability to deliver quickly and respond to emerging requirements." I have been working in a SCRUM environment, with 4-6 week cycles, for about 6 months now and have been very pleased with the impact that it has had on my life (regular work hours, seeing my family, etc).  But was looking up the criteria for a 'Certified Scrum Master' and came across the SCRUM definition on Wikipedia, and started reading the actual definition.  My first thought was "hey, this development methodology actually allows you to deal with what happens in the real world (i.e. customers changing requirements); but is this "selling out" on solid requirements? I understand that this works in the environment that I am currently working in, where there are deep pockets paying the bills, and also making the descisions on what requirements to change / impliment; but is this a recepie for success in smaller or simply more budget concious environments?  Having the ability to be completely flexible when the client wants the product changed.   The more I think about it, the more I feel that SCRUM development may be better suited for an environment where a team is taking over a project from another team (bringing some outside development in-house or something of that ilk), as opposed to ground up development. What do you think?

    Read the article

  • Convert old videos to have smaller sizes

    - by Tim
    I have some videos from a few years ago,with various formats, such as avi, mpg, wmv, rm, rmvb, .... Their sizes are huge(more than 500 MB, and sometimes 1GB). Given there may likely be some advance in data compression, I would like to know which file formats and compression methods are recommended these days, by the standard that without losing obvious data, while achieving big size reduction. How can I perform the file format conversion and data compression in Ubuntu 12.04? Command line and batch ways would be the most convenient, although GUI ways are also appreciated. Thanks and regards!

    Read the article

  • Limiting my heavy thinking to my job [closed]

    - by Robin Castlin
    This might be a weird problem which is only to a half relevant to actual programming, but hopefully there are people here that knows what I'm talking about. Basicly I'm proud of how I can deal with coding problems and fix them in short notice and many other aspects like building new systems and such. I'm fast on finding solutions and I often think about the impact my changes does to existing systems and so on, therefor preventing problem from arising at all and such. I am simply happy with how my mind operates when it comes to programming and I wouldn't want to change it at all. The problem, however is when I'm not programming. I find myself rather limited in social situations. I can't determine if it is through programming, but I sometimes think way to much about the consequences when it comes to being social. I know from own experience that most times you earn by not thinking about consequences, but it's hard for me not to. Often my friends tells me "I think too much" and even though I agree, I can't seem to change this behavior. My brain wants to think, and it likes to overthink simple stuff. Does anyone recognize the bad habit of not leaving advanced thinking at work, and in what way do you deal with it? If this isn't a suitable place to ask this question, I apologize and hope you may point me to the right site.

    Read the article

  • Multi Threading - How to split the tasks

    - by Motig
    if I have a game engine with the basic 'game engine' components, what is the best way to 'split' the tasks with a multi-threaded approach? Assuming I have the standard components of: Rendering Physics Scripts Networking And a quad-core, I see two ways of multi-threading: Option A ('Vertical'): Using this approach I can allow one core for each component of the engine; e.g. one core for the Rendering task, one for the Physics, etc. Advantages: I do not need to worry about thread-safety within each component I can take advantage of special optimizations provided for single-threaded access (e.g. DirectX offers a flag that can be set to tell it that you will only use single-threading) Option B ('Horizontal'): Using this approach, each task may be split up into 1 <= n <= numCores threads, and executed simultaneously, one after the other. Advantages: Allows for work-sharing, i.e. each thread can take over work still remaining as the others are still processing I can take advantage of libraries that are designed for multi-threading (i.e. ... DirectX) I think, in retrospect, I would pick Option B, but I wanted to hear you guys' thoughts on the matter.

    Read the article

  • Help looking before I leap! I need expert guidance...

    - by Ellen Reddick
    27" iMac running win7 under bootcamp (slick! ). I have Access 2003 program with files linked through ODBC used by 4 installations (all with Access 2003 installed). I want to buy Access 2010 and try it under virtual PC (under Bootcamp). Will it work (since I have to install the ODBC drivers)? If I decide after this trial that I like what it does, can I then install it under the Windows 7 bootcamp partition (with or without uninstalling the virtual PC) without using up the 2nd allowed installation? Also, I see that MS allows an Office Pro 2010 trial download good for 60 days. Would this work in Windows 7 Virtual PC and would it be a better way to go, followed by a legitimate purchase of Access 2010 for the Windows 7? This is not an Access programming question--I realize there may be some tweaks necessary in the program to run it under 2010 and I can handle that part.

    Read the article

  • Why does deleting from the command line take significantly less time than from a GUI?

    - by Jordan Plahn
    So this is probably the dumbest question you'll read today, but it's something I just wondered about as I was deleting a dozen or so images from my computer. With a quick rm -rf command on the directory's contents, all the images were gone in a snap. When I drag the same dozen or so images to a trash can/recycle ban, it takes sometimes 10 seconds or more. Now I'm sure some of it comes from the overhead of the GUI and such, and some of it may be the fact that the file still "exists" in some form if it's put into the recycle bin, but is there anything else that accounts for such a huge time disparity? Are "rm" and "delete" just such fundamentally different commands so I'm trying to compare apples and oranges? Enlighten me, please!

    Read the article

  • Upgraded from 11.4 to 11.10, There was an error, now the system won't initiialize

    - by Eric
    This morning the system gave me a message that my Version (11.4) was no longer supported, and I took the 'upgrade' option (- 11.10). While installing the various components I encountered a message to the effect that the there was an error and the system may have become unusable. Among the messges: E:Sup-process /usr/bin/dpkg received a segmentation fault...returned an error code (1). I was given an option to do several things, one of which seemed to mean that it would attempt to roll back to the previous version (the default), which I took. After the process ran it said the upgrade process had finished, but there were errors. I attempted to initialize a console so I could enter ubuntu-bug update-manager /var/log/dist-upgrade, per the instructions I received when I received the error message, but the console failed during initialization. I restarted the machine, and the screen has stopped with the following contents: * Starting bluetooth * Stopping save kernel messages * Starting CUPS printing spooler/server * PulseAudio configured per-user sessions saned disabled: edit /etc/default/saned $starting up Cisco VPN daemon *Starting anac(h)ronistic cron *Stopping anac(h)ronistic cron Each of these steps followed by [ OK ] What are my options? Any help appreciated!

    Read the article

  • suggest AMD machine for software developement [closed]

    - by superb1
    I need a AMD machine (Processor, Motherboard, Ram, HDD ) for software developement .. I will mainly use microsoft visual studio 2010. Operating system would be windows 7 . Suggest decent machine but i dont want to overspend....i guess the rating system of ecommerce websites may help you to suggest !? I want you to suggest from this list : http://www.flipkart.com/computers/components/processor-20246?ref=bae2643a-2cc8-4710-942f-79e7fbc1b416 plz note basic criteria for vs 2010 is 1.6GHz or faster processor 1 GB (32 Bit) Ram 5400 RPM hard disk DirectX 9 capable video card running at 1024 x 768 or higher-resolution

    Read the article

  • Real-time Image Resize, Cropping and Caching Server Product

    - by Elijah
    I'm investigating what products are out there that will allow you to request images through a HTTP API in arbitrary image sizes. The server would behind a CDN but would still need to be able to handle a fair bit of traffic and be possibly load-balanced. I've been tasked with writing such a service, but I wanted to do some due diligence to see what commercial or open source solutions are out there. Google has not been particularly helpful. It may be because I have been searching for the wrong term. Third-party sites and services are out of the question because of corporate policies.

    Read the article

  • After deleting a local machines offline file cache, the same user's "my documents" no longer redirects to the network location.

    - by stead1984
    One of my apprentices was tasked with clearing out unused local profiles and clearing the offline file cache. After he cleared the offline file cache and rebooted the machine, he would log in as himself and no longer have his "my documents" redirected to the set network location. More over this seemed to then affect ANY other networked machine he logged into, except his own laptop. All our standard workstations run Windows XP Service Pack 3, the apprentice's laptop runs Windows 7 Professional. I can understand how clearing the offline file cache after deleting old local profiles could cause this issue but draw a complete blank as to why it would affect all networked machines. It's a strange one so this question may be a little hard to understand so any questions or further understanding required please ask.

    Read the article

  • Do I expect too much work from an employer? [closed]

    - by Ant
    I recently switched jobs because I was not challenged enough, the work would come in waves, and I HATED the people I worked with. I am a recent college grad, May 2009, and based off the 3 internships I had, and 2 full time jobs I obtained, I am finding that employers can not keep me satisfied with the amount of work. At my new job, I like the people I work with, I am challenged, but I still do not get enough work. I hate down time. I always want to have something to work on AT LEAST 6 out of the 8 hours. I was surprised that my new employer actually hired me because the majority of the technologies they implement, I had minimal exposure to. I never programmed in the technologies they use outside of one class in college. My greatest strength is that I am an extremely fast learner. I can pick up new technologies with relative ease. They gave me a project to work on by myself and I think they assumed it would take me longer to complete. Now that I finished that app, they are struggling to find something for me to do. I am not sure if it is bad timing being close to the holidays, my manager dealing with personal issues at home, how quickly I finished the first project, or that I expect too much out of an employer? If so, what are good things to do on all this downtime?! EDIT: Thanks for all the feedback! EDIT 2: I am going to "unaccept" the answer in an effort to keep the question open. As a few people have mentioned, this is a great discussion on how to grow as a new worker in the programming field. EDIT 3: I am attempting to revive this question so the moderators will see the support to re-open it.

    Read the article

  • Server 2008 SP1 VSS Writers Not Responding

    - by Jason
    I've got a Windows Server 08 box on SP1 that is having some problems. We've experienced backup problems and I've traced it down to VSS Writers not responding. From the command line, if I type vssadmin list providers, I get Provider name: 'Microsoft Software Shadow Copy provider 1.0' Provider type: System Provider Id: {b5946137-7b9f-4925-af80-51abd60b20d5} Version: 1.0.0.7 If I type vssadmin list writers, I get this vssadmin 1.1 - Volume Shadow Copy Service administrative command-line tool (C) Copyright 2001-2005 Microsoft Corp. Waiting for responses. These may be delayed if a shadow copy is being prepared. I could wait this out for hours and it won't move. I looked up how Server 2008 handles VSS writers, and you can't reregister them like you could in Server 2003 http://social.technet.microsoft.com/Forums/en/windowsserver2008r2general/thread/062cc52c-899b-45f3-8d0c-798b92363f41 Does anyone know how to fix something like this or where to turn next?

    Read the article

  • nginx + reverse proxy question

    - by Joe Pilon
    Hello, I am using nginx right now for our production sites with the reverse proxy to apache that's on the same server and it works fantastic. I'm wondering if I can do this: Install nginx on box #1 in say Canada and have it reverse proxy http requests to box #2 in a datacenter in the USA. I know there may be some latency or delays in loading the page etc but that would probably be not noticable to the end user especially if both servers have 100mb ports. Box #2 only does the apache requests, all images are served from box #1 via nginx. Now, would the end visitor be able to tell in any which way that there are 2 boxes being used? Box #2 has sensitive data which we can't have stolen in the event of hacking etc, so this method helps keep things a bit more secure. Anyone know if this is possible or have done something similar?

    Read the article

  • Get Zipped Logs from a Remote Server

    - by Jonathan
    I am tasked with trying to find a way to download zipped logs from a remote server. There are quite a bit of these logs and they are constantly created. I do have limited ssh access to the remote server and can scp or rsync the files. However, due to the sheer size of these logs file, I do not want to rsync all of them. The logs could get to terabytes and for rsync to compare them may take some time. I only want to get any new file that was created/last updated an hour ago. I also am worried that I will rsync logs that are in the process of being created, so I was thinking to only rsync files that were last modified 3-5 minutes ago. Would anyone be so kind as to help me with such a process? Thank you in advance.

    Read the article

  • Ubuntu 12.04 - Custom Quick Launcher Menu

    - by gotqn
    I have been using Ubuntu for few weeks now and I am have been trying to create custom menu which to add in the launcher. Please, note that my version is 12.04 and I have not got any experience with Linux. I have checked a lot articles and questions likes this one, but have not been able to find what I am looking for. I want to create a custom quick launcher menu and to add it the the standard launcher, but I want to add links to other programs or folders in it. For example: A browsers menu - when it is open (with right mouse click on it), my browsers will be listed: chrome safari opera firefox I even may add their icons there. Music menu - some shortcuts to my favorite music folders.

    Read the article

  • How do you synchronise huge sparse files (VM disk images) between machines?

    - by chrisdew
    Is there a command, such as rsync, which can synchronise huge, sparse, files from one linux server to another? It is very important that the destination file remains sparse. It may be longer (but not bigger) than the drive which contains it. Only changed blocks should be sent across the wire. I have tried rsync, but got no joy. groups.google.com/group/mailing.unix.rsync/browse_thread/thread/94f39271980513d3 If I write a programme to do this, am I just reinventing the wheel? http://www.finalcog.com/synchronise-block-devices Thanks, Chris.

    Read the article

  • TCP/IP communication between Hyper-V host and guests

    - by Tedd Hansen
    This may be a simple one. :) I have a simple Hyper-V setup with a few guest os running. The host has 1 physical network adapter with a static IP assigned to it. The guests have network adapters assigned to "Internet" (Hyper-V network) which is bound to the physical host network adapter (Hyper-V "External" connection type). I am not able to communicate (ping or anything else) between guests and host. I've checked firewall and it seems fine (ports open from anywhere still don't work). I'm trying to communicate with the hosts IP assigned to the same physical interface that the guests are sharing. Guests can communicate between them just fine. I can't seem to find any relevant setting (I might just be missing it). So my questions: How do I fix it so host and guests can communicate?

    Read the article

  • How to force Windows to call the partition it installs on C:?

    - by maaartinus
    I'm going to install Windows XP from a CD and want to make sure it calls the partition it lands on C:. The target partition is the first one on a SATA disk, which is not the first one in my computer. There's no IDE disk there. I don't think I can swap the disks, as I'm using a fake RAID and really don't want to get problems with it. The target partition is on a normal disk. I know it may be unimportant, but I don't want to run in any problems I can avoid. I've seen a question slightly related to it, but I'm not going to install from a USB.

    Read the article

  • How can I play 24Bit 96000Hz vinyl rips?

    - by Peter
    I am getting music through the Nvidia HDMI but it is all down-sampled to 44000Hz. I have spent at least 3 hours searching for an answer yesterday. I use Pulse Volume and I even changed the default settings in the Pulse folder to 96000 but it does not work. Though 5.1 sound works perfectly. They all work well in Windows 7, but I prefer to use Ubuntu. All the config files for all the programs are in the etc folder, then look in the pulse folder, and there are 3 three files, 2 of these I changed the permissions and I changed 16 and 44000 to 24 and 96000 as I had read in another forum, but it worked for them and not for me. Though 5.1 works as I have special mp4 file that can check each channel. I did alot of restarts and checking after these changes as nothing had happened, my amp tells me the freq of the signal, which does work which I have tested on Windows 7. I have also gone to systems to download the latest driver for mu Nvidia card. But even if it does not work, I guess I can always play Cd's. But I think my player have become more unstable since I changed the configs, so I may uninstall and install Pulse Volume Control again.

    Read the article

  • More productive alone than in a team?

    - by Furry
    If I work alone, I used to be superproductive, if I want to be. Running prototypes within a day, something that you can deploy and use within a few days. Not perfect, but good enough. I also had this experience a few times when working directly with someone else. Everybody could do the whole thing, but it was more fun not to do it alone and also quicker. The right two people can take an admittedly not too large project onto new levels. Now at work we have a seven person team and I do not feel nearly as productive. Not even nearly. Certain stuff needs to be checked against something else, which then needs to also take care of some new requirement, which just came in three days ago. All sorts of stuff, mostly important, but often just a technical debt from long ago or misconception or different vocabulary for the same thing or sometimes just a not too technically thought out great idea from someone who wants to have their say, and so on. Digging down the rabbit hole, I think to myself, I could do larger portions of this work faster alone (and somewhat better, too), but it's not my responsibility (someone else gets paid for that), so by design I should not care. But I do, because certain things go hand in hand (as you may experience it, when you done sideprojects on your own). I know this is something Fred Brooks has written about, but still, what's your strategy for staying as productive as you know you could be in the cubicle? Or did you quit for some related reason; and if so where did you go?

    Read the article

  • What are solutions and tradeoffs to maintain search result consistency in a web application

    - by iammichael
    Consider a web application with a custom search function that must display the results in a paged manner (twenty per page with up to hundreds of thousands of total results) and the ability to drill down to individual results that maintain next/previous links to navigate through the results. Re-executing the search on each page request to get the appropriate results for that page of data can be too expensive (up to 15s per search). Also, since the underlying data can change frequently (e.g. addition of new results), re-executing could cause the next/previous functionality to result in inconsistent behavior (e.g. the same results reappearing on a later page after having been viewed on an earlier page). What options exist to ensure the search results can be viewed across multiple pages in a consistent manner, and what tradeoffs does each option have in terms of network, CPU, memory, and storage requirements? EDIT: I thought caching the query search results was an obvious necessity. The question is really asking about where to cache the result set and what tradeoffs might exist to each. For example, storing the ids of the entities in the result set on the client, or storing the IDs of the entities themselves in the users session on the web server, or in a temporary table in the database. I'm not looking specifically for a single solution as different scenarios may result in different approaches (and such a question would be more suited for stackoverflow.com rather than here), but more of a design comparison between the possible approaches.

    Read the article

  • Intermittent freezing on a Macbook Pro 7,1

    - by LR0
    I recently installed Ubuntu 12.10 on my MacBook Pro 7, 1. I used to have 11.10 on it. I've encountered a problem were I can boot up and the login screen works flawlessly. When I attempt to log into my account I set up during install it would only show the default wallpaper and the mouse was frozen. In other words it was logging on but froze up. Is what it seems like. When I rebooted I attempted to log into the guest account. I was able to log in and it seemed everything worked so I tried to update everything figuring it may be the problem. After updating and restarting as prompted I tried to login to my account. It still froze up but this time showing "glitched up" (you can see the dock and status bar but they don't look like they should) and still freezes. This is all happening on the guest account to. I was unable to find a similar article on the Internet therefore I am asking; Any thoughts Ubuntu?

    Read the article

  • how to report a malicious site (http://newss.gr) to google, microsoft and mozilla so that they will prompt

    - by Jayapal Chandran
    Hi, I completed a project an year ago. Now a few modification were needed. While i try to test there was an index.html with a malicious script which had an iframe to this site's jar file. and kaspersky anti virus blocked it. So i browsed the ftp to find the file and i deleted it. and also disabled directory listing. May be the ftp details of the site owner would have been hacked. I want to report this site to google, msn and mozilla and other antivirus programs. How to do that. any idea? I hope kaspersky would have updated it in their database but still i want to explicitly inform it about this. here is the popup kaspersky showed.

    Read the article

  • MacBook Air with Bootcamp - How to partition?

    - by Andrew
    I want to buy a MacBook Air for my wife with a 128GB SSD. She has to use Windows 7 but I would like to keep OS X for myself to use somtimes. Using Bootcamp, is it feasible to install the following? Mac partition: 36GB with Mac OS X and Microsoft Office 2011 Word, Excel & Powerpoint and Skype. (minimal use) Windows partition: 92GB with Windows 7 professional and Microsoft Office 2010 Word, Excel & Powerpoint, and Skype (daily use) Media to be kept on SD card or external USB3 drive. (Note: Using Parrallels may save space, but my wife won't go for the user experience)

    Read the article

  • Intermittent HTTP 401 errors

    - by forthrin
    I am using an Intranet solution which requires basic HTTP login. However, there is an intermittent error which requires me to log in again, and then the server says "Forbidden" whether I give the correct login information or not. To add insult to injury, Safari (and Chrome) seems to show the login dialog for every included resource in the HTML, and it's impossible to cancel this modal dialog sequence, so the whole browser is blocked until I've pressed Esc some 30 odd times. After an hour, I may gain access again, without having really done anything. My questions: What could cause temporal 401 errors? Why do the browsers show the login dialog 30 times per page load (assumedly for every included resource in the HTML from the same domain)?

    Read the article

< Previous Page | 479 480 481 482 483 484 485 486 487 488 489 490  | Next Page >