Search Results

Search found 19871 results on 795 pages for 'commit log'.

Page 76/795 | < Previous Page | 72 73 74 75 76 77 78 79 80 81 82 83  | Next Page >

  • How should I log time spent on multiple tasks?

    - by xenoterracide
    In Joel's blog on evidence based scheduling he suggests making estimates based on the smallest unit of work and logging extra work back to the original task. The problem I'm now experiencing is that I'll have create object A with subtask method A which creates object B and test all of the above. I create tasks for each of these that seems to be resulting in ok-ish estimates (need practice), but when I go to log work I find that I worked on 4 tasks at once because I tweak method A and find a bug in the test and refactor object B all while coding it. How should I go about logging this work? should I say I spent, for example, 2 hours on each of the 4 tasks I worked on in the 8 hour day?

    Read the article

  • Things I've noticed with DVCS

    - by Wes McClure
    Things I encourage: Frequent local commits This way you don't have to be bothered by changes others are making to the central repository while working on a handful of related tasks.  It's a good idea to try to work on one task at a time and commit all changes at partitioned stopping points.  A local commit doesn't have to build, just FYI, so a stopping point doesn't mean a build point nor a point that you can push centrally.  There should be several of these in any given day.  2 hours is a good indicator that you might not be leveraging the power of frequent local commits.  Once you have verified a set of changes works, save them away, otherwise run the risk of introducing bugs into it when working on the next task.  The notion of a task By task I mean a related set of changes that can be completed in a few hours or less.  In the same token don’t make your tasks so small that critically related changes aren’t grouped together.  Use your intuition and the rest of these principles and I think you will find what is comfortable for you. Partial commits Sometimes one task explodes or unknowingly encompasses other tasks, at this point, try to get to a stopping point on part of the work you are doing and commit it so you can get that out of the way to focus on the remainder.  This will often entail committing part of the work and continuing on the rest. Outstanding changes as a guide If you don't commit often it might mean you are not leveraging your version control history to help guide your work.  It's a great way to see what has changed and might be causing problems.  The longer you wait, the more that has changed and the harder it is to test/debug what your changes are doing! This is a reason why I am so picky about my VCS tools on the client side and why I talk a lot about the quality of a diff tool and the ability to integrate that with a simple view of everything that has changed.  This is why I love using TortoiseHg and SmartGit: they show changed files, a diff (or two way diff with SmartGit) of the current selected file and a commit message all in one window that I keep maximized on one monitor at all times. Throw away / stash commits There is extreme value in being able to throw away a commit (or stash it) that is getting out of hand.  If you do not commit often you will have to isolate the work you want to commit from the work you want to throw away, which is wasted productivity and highly prone to errors.  I find myself doing this about once a week, especially when doing exploratory re-factoring.  It's much easier if I can just revert all outstanding changes. Sync with the central repository daily The rest of us depend on your changes.  Don't let them sit on your computer longer than they have to.  Waiting increases the chances of merge conflict which just decreases productivity.  It also prohibits us from doing deploys when people say they are done but have not merged centrally.  This should be done daily!  Find a way to partition the work you are doing so that you can sync at least once daily. Things I discourage: Lots of partial commits right at the end of a series of changes If you notice lots of partial commits at the end of a set of changes, it's likely because you weren't frequently committing, nor were you watching for the size of the task expanding beyond a single commit.  Chances are this cost you productivity if you use your outstanding changes as a guide, since you would have an ever growing list of changes. Committing single files Committing single files means you waited too long and no longer understand all the changes involved.  It may mean there were overlapping changes in single files that cannot be isolated.  In either case, go back to the suggestions above to avoid this.  Committing frequently does not mean committing frequently right at the end of a day's work. It should be spaced out over the course of several tasks, not all at the end in a 5 minute window.

    Read the article

  • How To Log Into The Desktop, Add a Start Menu, and Disable Hot Corners in Windows 8

    - by Chris Hoffman
    If you don’t have a touchscreen computer and spend all your time on the desktop, Windows 8’s new interface can seem intrusive. Microsoft won’t allow you to disable the new interface, but Classic Shell provides the options Microsoft didn’t. In addition to providing a Start button, Classic Shell can take you straight to the desktop when you log in and disable the hot corners that activate the charms and metro app switcher. There are other programs that do this, but Classic Shell is free and open-source. Many of the alternatives, such as Start8 and RetroUI, are commercial apps that cost money. We’ve covered Classic Shell in the past, but it’s come a long way since then. How Hackers Can Disguise Malicious Programs With Fake File Extensions Can Dust Actually Damage My Computer? What To Do If You Get a Virus on Your Computer

    Read the article

  • Changed username. Now I cannot log on or view my previous files

    - by Lauren
    I want to change my username and followed the instructions from How do I change my username? by creating a temp user with admin privileges. While logged in as temp, I did : sudo usermod -l newname oldname sudo usermod -d /home/newname -m newname Now I cannot log in under newname and /home only lists newname and temp Reading through other sites now, it seems I should have used usermod -d /home/newname -m oldname Based on this, I think I may have deleted the contents of my previous home folder?? I'm sure I'm not the first person to do some stupid while changing username, but any help is greatly appreciated. Thank you!

    Read the article

  • Why am I getting messages from cloudfront in my error log?

    - by JK01
    I frequently have messages like this in my websites error log: "Script error.". URL: https://e3m4drct5m1ays.cloudfront.net/items/loaders /loader_21.js?pid=21&systemid=13504281c5a501837196c23300f84e66&aoi=1327214632& zoneid=16620&cid=HK&rid=Hong%20Kong%20(general)&ccid=Kowloon&dma=0. Line number: 0 Error name: Stack: Now I don't actually know what cloudfront is or what it does. And I do not refer to this script in my site. So why would I be getting js error logged as if it was a script being run on my own site? This is using elmah logging.

    Read the article

  • Is there a pattern or logical structure I can follow for Event Log Numbers?

    - by makerofthings7
    What are some ideas or structure I can use when assigning EventID to events that will be saved to the Event Log? Some options I've considered Sequential (0... int.Max) Multiple of 10, where the "0" is replaced with how noisy the debugLevel is set. xxx0 may represent exceptions, critical information, start, stop etc. ...? What numbering approach gives you the most insight when a user describes the event in an email or phone? What is the most useful to support staff?

    Read the article

  • Run a script with user interaction on log out / shutdown?

    - by lumbric
    I'd like to run a script on shut down, which interrupts the logout process and and pops up a window with zenity. My target is to get autofsck working with lightdm. It seems to work with 12.04 after installing the old *.deb file, if one runs the check script manually. In order to use it, it should run automatically on log out and ask the user if she/he wants to check the disc on shutdown. There is the option session-cleanup-script in the file /etc/lightdm/lightdm.conf which seems to work, if a bash file with full path is used (I can't place the command directly there). But if I press shutdown, there is no time for a user choice. Is there any other option to solve this problem?

    Read the article

  • How to log frame times in an existing OpenGL game? [on hold]

    - by J Collins
    I have been using FRAPS for some time to benchmark instantaneous frame rates in an OpenGL game for which I am creating maps. Until recently it had been quite reliable. Now however, the bench marking shortcut has been unresponsive and I can't explain why. Ideally I could have a logging system automatically start logging whenever the game had focus, but can't find a good tool to do so. So option a) find out how to make FRAPS reliable again or b) find a new tool. Could one of you kind folks help me? Edit: Concise questions Is there a widely recognised tool to log frame drawing times and rates for compiled applications? If the answer is universally the FRAPs tool, are there any clear cases in which logging will not or should not be expected to work?

    Read the article

  • After installing ubuntu, all my partitions are gone. Boot-repair log

    - by user211079
    I have an HP Pavilion Sleekbook that came with windows 8 pre-installed. I had trouble dual booting after installing ubuntu, so I disabled safeboot on bios and proceeded to try boot-repair, nothing happened. No dual boot yet. So I tried to reinstall ubuntu, but without the manual partitioning. So I chose to erase ubuntu 13 and reinstall it. Instead it deleted all my HP recovery partitions and windows as well. Here is the log of boot-repair. http://paste.ubuntu.com/6354919/ Gparted and fdisk only show one partition: /dev/sda1 I am wondering if you could suggest any way of recovering my windows partition and have a working windows 8 again? I need some information there with urgency. If you could help me I will be welcome. I am desperated. Thanks

    Read the article

  • Why am I missing the Ubuntu icon on the log-in screen?

    - by dtdlurch
    I am running Ubuntu 12.10. On the log-in screen, there is no Ubuntu icon for me to click on to change desktop environments. I know where it should be, but it is simply not there. Why is this? And how can I get it back??? I just installed Cinnamon and I would very much like to switch to this environment. I apologize if this is has been addressed elsewhere, but I honestly did try to research before asking. I found several similar questions, but nothing that has answers that helped my situation. And an hour of Googling did not produce anything helpful at all. Thank you for your help.

    Read the article

  • What else is needed to get iptables to log into this file I created?

    - by anthony01
    I want to create the logging of iptables DROP's and intrusion attemps. First, I put --log-prefix "iptables: " at the end of every iptables rules in my iptables rules file. But this doesn't work, as it says there is a syntax error. So where should I put that command? (I would want to have it included in the saved rules file) Secondly, I created a file iptables.conf within /etc/rsyslog.d/, and I put the following inside of it: :msg, startswith, "iptables: " -/var/log/iptables.log & ~ I assume that at this stage, I'm supposed to restart the rsyslog daemon. What else is needed to do what I'm attempting? Thanks a lot

    Read the article

  • How to analyze CBS.log file by SFC command to detect corrupt files in Windows 7?

    - by energydream2007
    Hey I have some problems with my Win7 so I've ran sfc /scannow and I got this message: "Windows Resource Protection found corrupt files and was unable to fix some of them..." I also told to look at the folder of the log file to find the cbs.log. After that I've ran this command to pull out the actual problems/corrupt files: findstr /c:"[SR]" %windir%\Logs\CBS\CBS.log >"%userprofile%\Desktop\sfcdetails.txt" This command have created an sfcdetails text file (download). Can someone help me to analyze this file? I haven't found a detailed article about it so far.

    Read the article

  • Why am I not getting text into /var/log/messages?

    - by dafydd
    Good Morning, Today, I went to look at a backup RHEL5 server that doesn't usually need to get looked at very often. I found that /var/log/messages existed, but had a size of zero. Similarly, messages.1, messages.2, messages.3, and messages.4 are also all empty. /var/log/messages has its correct permissions. (755) Reboot - No help. rm -f /var/log/messages* and service syslog restart - No help. What am I missing?

    Read the article

  • How to use Binary Log file for Auditing and Replicating in MySQL?

    - by Pranav
    How to use Binary Log file for Auditing in MySQL? I want to track the change in a DB using Binary Log so that I can replicate these changes to other DB please do not give me hyperlinks for MySQL website. please direct me to find the solution EDIT I have looked for auditing options and created a script using Triggers for that, but due toi the Joomla DB structure it did'nt worked for me, hence I have to move on to Binary Log file concept now i am stucked in initiating the concept as I am not getting the concept of making the server master/slave, so can any body guide me how to actually initiate it via PHP?

    Read the article

  • Windows Malicious Software Removal Tool log says it can't do all required actions. Should I be conce

    - by Tom
    Here's what the log file c:/Windows/debug/mrt.log of my Windows 7 install says: WARNING: Security policy doesn't allow for all actions MSRT may require. ->Scan ERROR: resource process://pid:6080 (code 0x00000005 (5)) ->Scan ERROR: resource process://pid:5300 (code 0x00000057 (87)) ->Scan ERROR: resource process://pid:3512 (code 0x00000057 (87)) I use the default setup. I didn't change anything. This is the first time I checked the log file and this warning is in there from the start. Can I do something about it? Or I shouldn't be concerned, because it can do everything what's necessary anyway? Do you have this warning in your logfile?

    Read the article

  • How to use Binary Log file for Auditing and Replicating in MySQL?

    - by Pranav
    How to use Binary Log file for Auditing in MySQL? I want to track the change in a DB using Binary Log so that I can replicate these changes to other DB please do not give me hyperlinks for MySQL website. please direct me to find the solution I have looked for auditing options and created a script using Triggers for that, but due toi the Joomla DB structure it did'nt worked for me, hence I have to move on to Binary Log file concept now i am stucked in initiating the concept as I am not getting the concept of making the server master/slave, so can any body guide me how to actually initiate it via PHP?

    Read the article

  • Is there anyway I can secure my connection when I try to log in to my router remotely?

    - by HardwareMuch
    I'm trying to configure my desktop to be accessed remotely. Here's what I've done so far: enabled wake on lan / remote wake up on all interfaces. I'm using logmein as my remote desktop application. setup DDNS so that I can log in to my router. When I try to remotely log in to my router it says that it is not encrypted there anyone can see my log in information. What can I do to make this a more secure setup? Any other suggestions or different methods will be greatly appreciated.

    Read the article

  • Is there anyway I can secure my connection when I try to log in to my router from remotely?

    - by HardwareMuch
    I'm trying to configure my desktop to be accessed remotely. Here's what I've done so far: enabled wake on lan / remote wake up on all interfaces. I'm using logmein as my remote desktop application. setup DDNS so that I can log in to my router. When I try to remotely log in to my router it says that it is not encrypted there anyone can see my log in information. What can I do to make this a more secure setup? Any other suggestions or different methods will be greatly appreciated.

    Read the article

  • Apache crashing at random intervals. Can not find a reason in log files

    - by Nick Downton
    We are having an issue with a VPS running plesk 9.5 on ubuntu 8.04 At seemingly random intervals Apache will disappear and needs to be started manually. I have checked the apache error log, /var/log/messages, individual virtual host apache error files and cannot find anything that coincides with the time of the failure. dmesg is empty which is a bit odd. We have also had the psa service go down for no apparent reason but apache stay up. I'm at a loss to diagnose this really because all the log files I can find do not point to any issues. Are there any others I can look at? Memory usage sits at about 55% (out of 400mb) and it isn't a particularly high trafficed server. Any pointers as to where else I can find out what is going on would be very much appreciated. Nick

    Read the article

  • Automatically Log off Google when logging in using Google OpenID?

    - by Ross Charette
    I use Google as my OpenID provider. Once you log into a website with Google's OpenID I am then logged into Google as well. I do not desire this. Can I somehow automatically log off my Google account, or prevent Google from logging in every time I use my Google OpenID? I prefer not to have my personal google account always logged in. It's not a big deal going to gmail and clicking log off, but if there is a simpler way that would be good. Note to admins: this is not just for stack overflow, please don't close the question.

    Read the article

  • Iptables: how do I LOG what's not being ACCEPTED and limit what gets logged?

    - by Kris
    How do I log what's not being accepted by the following rule: iptables -A OUTPUT -p icmp --icmp-type 3 -m -limit --limit 10/minute -j ACCEPT And how do I limit what's being logged because I don't want to log 1000s of pings? My first thought was: iptables -A OUTPUT -p icmp --icmp-type 3 -m -limit --limit 50/day -j LOG iptables -A OUTPUT -p icmp --icmp-type 3 -m -limit --limit 10/minute -j ACCEPT But that doesn't seem right to me. I think this limits the logging to 50/day but not necessarily what is not being accepted, or am I wrong?

    Read the article

  • Is there a log showing why a Windows server did not restart SQL Server after a reboot?

    - by MerlinMags
    Our server was rebooted after a Windows Update scheduled for 1am, but after the restart SQL Server did not start up, so our websites were unable to display. Usually this process happens with no manual intervention. Is there a log somewhere which might indicate the reason why the Windows startup process did not call SQL Server to get going again? I've looked in the Event Viewer (Application Log) and SQL's own file E:\MSSQL\MSSQL10.MSSQLSERVER\MSSQL\Log\ERRORLOG* but these only contain records of successful startup operations....nothing mentions a failed attempt to start a service or anything like that.

    Read the article

  • Windows CE Remote Kernel Tracker - gathering data in one (more) file during a log period of time

    - by Nic
    I'm using the "Windows CE Kernel Tracker" tool to gather data from my embedded device. This is working fine for short period of time. It seems that the tool is getting data in memory and not on disk. I'm wondering if there is a way to take the data from the device and log it in one or more file on my development computer. This could be useful for long time test period : for instance, one night or one entire day. Any ideas? p.s. I don't want to log on to the device, I want to log on my development PC.

    Read the article

  • MSSQL, ASP.NET, IIS. SQL Server perfmon log question

    - by Datapimp23
    Hi, I'm testing a web application that runs on a hypervisor. The database server and the webserver are seperate vm's that run on the same hypervisor. We did some tests and the functions perform ok. I want you guys to look at a screenshot of a permon log of the sql 2005 server on the busiest moment. The webserver perfmon log looks fine and it's obvious that we have enough resources to present the page in a timely fashion. http://d.imagehost.org/view/0919/heavyload http://d.imagehost.org/0253/heavyloadz.jpg Zoomed out The striped blue line maxing out is the Processor que length (scale 100,0) The green line at around value 30 is Available MBytes (scale 0,01) The rest of the counters are visible on the screenshot. The sql server machine has no CPU limitations on the hypervisor resources and has 5 vcpu's and 5 GB RAM. Can someone help me to interpret this log. Thanks

    Read the article

  • Grep /var/log for hacker/script kiddy activity and e-mail?

    - by Jason
    CentOS 6 Apache Server version: Apache/2.2.15 (Unix) Thinking about how to automatically, once a day, grep all the logs in /var/log/httpd for hacker, phishing, etc activity and e-mail it to myself so I can evaluate what I might need to do. But what are the patterns I can look for? IE, we dont run Wordpress and we see a lot of attempts to access Wordpress related content, obviously for an exploit. Same with PHPMyAdmin. I could do something like repeatedly, matching common patterns we see. # grep -r -i wp-content /var/log/httpd/ # grep -r -i php-my-admin /var/log/httpd/ How do I e-mail myself this the results of each grep command or better yet all Grep results in a single e-mail?

    Read the article

< Previous Page | 72 73 74 75 76 77 78 79 80 81 82 83  | Next Page >