Search Results

Search found 15906 results on 637 pages for 'scott and the dev team'.

Page 219/637 | < Previous Page | 215 216 217 218 219 220 221 222 223 224 225 226  | Next Page >

  • How are Reads Distributed in a Workload

    - by Bill Graziano
    People have uploaded nearly one millions rows of trace data to TraceTune.  That’s enough data to start to look at the results in aggregate.  The first thing I want to look at is logical reads.  This is the easiest metric to identify and fix. When you upload a trace, I rank each statement based on the total number of logical reads.  I also calculate each statement’s percentage of the total logical reads.  I do the same thing for CPU, duration and logical writes.  When you view a statement you can see all the details like this: This single statement consumed 61.4% of the total logical reads on the system while we were tracing it.  I also wanted to see the distribution of reads across statements.  That graph looks like this: On average, the highest ranked statement consumed just under 50% of the reads on the system.  When I tune a system, I’m usually starting in one of two modes: this “piece” is slow or the whole system is slow.  If a given piece (screen, report, query, etc.) is slow you can usually find the specific statements behind it and tune it.  You can make that individual piece faster but you may not affect the whole system. When you’re trying to speed up an entire server you need to identity those queries that are using the most disk resources in aggregate.  Fixing those will make them faster and it will leave more disk throughput for the rest of the queries. Here are some of the things I’ve learned querying this data: The highest ranked query averages just under 50% of the total reads on the system. The top 3 ranked queries average 73% of the total reads on the system. The top 10 ranked queries average 91% of the total reads on the system. Remember these are averages across all the traces that have been uploaded.  And I’m guessing that people mainly upload traces where there are performance problems so your mileage may vary. I also learned that slow queries aren’t the problem.  Before I wrote ClearTrace I used to identify queries by filtering on high logical reads using Profiler.  That picked out individual queries but those rarely ran often enough to put a large load on the system. If you look at the execution count by rank you’d see that the highest ranked queries also have the highest execution counts.  The graph would look very similar to the one above but flatter.  These queries don’t look that bad individually but run so often that they hog the disk capacity. The take away from all this is that you really should be tuning the top 10 queries if you want to make your system faster.  Tuning individually slow queries will help those specific queries but won’t have much impact on the system as a whole.

    Read the article

  • Microsoft MVP for year 2011

    - by imran_ku07
        This is great news for me that I become Microsoft Most Valuable Professional (Microsoft MVP) and Most Valuable Blogger. It feels so great for me. I am very thankful to all friends, community member and team members. Special thanks to ASP.NET MVC Team members Rick Anderson, Levi Broderick, Brad Wilson and Marcin Dobosz. I learn very much from these guys. They are indeed wizards and very much deserve their positions   

    Read the article

  • ClearTrace Performance on 170GB of Trace Files

    - by Bill Graziano
    I’ve always worked to make ClearTrace perform well.  That’s probably because I spend so much time watching it work.  I’m often going through two or three gigabytes of trace files but I rarely get the chance to run it on a really large set of files. One of my clients wanted to run a full trace for a week and then analyze the results.  At the end of that week we had 847 200MB trace files for a total of nearly 170GB. I regularly use 200MB trace files when I monitor production systems.  I usually get around 300,000 statements in a file that size if it’s mostly stored procedures.  So those 847 trace files contained roughly 250 million statements.  (That’s 730 bytes per statement if you’re keeping track.  Newer trace files have some compression in them but I’m not exactly sure what they’re doing.)  On a system running 1,000 statements per second I get a new file every five minutes or so. It took 27 hours to process these files on an older development box.  That works out to 1.77MB/second.  That means ClearTrace processed about 2,654 statements per second. You can query the data while you’re loading it but I’ve found it works better to use a second instance of ClearTrace to do this.  I’m not sure why yet but I think there’s still some dependency between the two processes. ClearTrace is almost always CPU bound.  It’s really just a huge, ugly collection of regular expressions.  It only writes a summary to its database at the end of each trace file so that usually isn’t a bottleneck.  At the end of this process, the executable was using roughly 435MB of RAM.  Certainly more than when it started but I think that’s acceptable. The database where all this is stored started out at 100MB.  After processing 170GB of trace files the database had grown to 203MB.  The space savings are due to the “datawarehouse-ish” design and only storing a summary of each trace file. You can download ClearTrace for SQL Server 2008 or test out the beta version for SQL Server 2012.  Happy Tuning!

    Read the article

  • Fixed Bid vs. T&amp;M &ndash; Take 2

    - by AjarnMark
    One of my most popular blog entries of all time is my Contracting Tips: Fixed Bid vs. T&M post from January, 2004.  This post consistently shows up in my referrers list, usually coming from a search engine.  Recently, Brent Ozar (@BrentO) wrote a great argument for why he always bills by the hour (a.k.a. Time & Materials or T&M) which itself was a response to Mark Richman’s (@mrichman) post on why he never bills by the hour (fixed bid).  Each article has good arguments, and I encourage you to read them both and choose the best approach for you. As for me, my experience parallels Brent’s and I historically have leaned toward the Time & Materials model.

    Read the article

  • ubuntu 12.10 graphic does not works correctly

    - by Amin
    I have installed Ubuntu 12.10 but my graphic does not works correctly. I used this command to installation: sudo apt-get install fglrx fglrx-amdcccle fglrx-dev But after the restarting my system, the side panel and upper one does not appear any more! I guessed it maybe that I used some incompatible packages so I removed the graphic card driver by this command: sudo apt-get purge fglrx fglrx-amdcccle fglrx-dev and tried it from graphical way from system setting > Software Sources > Aditional drivers and choosed the second option then applied change. But the resault waas such as before way!! I belive it is because of the Ubuntu does not know my graphic card. I am using VAIO VPCEA2TGX (N/A) and my graphic cart version is Mobility Radeon HD 5400 Series if it is matter. So now what is the exact problem and what I have to solve this?

    Read the article

  • WebCenter in Action: ResCare

    - by Kellsey Ruppel
    Register Now for this webcast. ResCare Solves Content Lifecycle Challenges with Oracle WebCenter Complex documents must be created, assembled, reviewed, and tracked. To avoid fragmented, chaotic information processes, organizations must adopt an integrated set of strategies, standards, best practices, and technologies for managing information. Attend this webcast to learn how Oracle WebCenter has allowed ResCare to: Solve content lifecycle challenges Reduce compliance and business risks Increase adoption of intranet as primary business communication tool Register now for this webcast.  REGISTER NOW Register now for this exclusive event. Tuesday, October 30, 2012 10:00 a.m. PT / 1:00 p.m. ET Presented by: Joe Lichtefeld, VP of Application Services & PMO, ResCare Wayne Boerger, Product Manager, TEAM Informatics Doug Thompson, EVP Global Development, TEAM Informatics Presented by : Copyright © 2012, Oracle and/or its affiliates. All rights reserved. Contact Us | Legal Notices and Terms of Use | Privacy Statement

    Read the article

  • SQL Source Control Contest

    - by Ajarn Mark Caldwell
    If you’re a regular reader of this blog, you know that I have written several posts about how important I think it is to protect your source code, to version it, and in particular, all the aspects I like about Red Gate’s SQL Source Control product.  But for a moment, let’s take a break from my writing and I want to hear your stories.  What nightmare situation are you in, or can you imagine, where source control for your database would save the world.  Or maybe your life is not so dramatic, but you do see a challenge that, if you just had a good tool like SQL Source Control, it would go much smoother.  What’s your pain?  You have read my writings, now tell me your story, and be in the running for a free copy of SQL Source Control from Red Gate. Yes, that’s right.  Although I am just a fan of Red Gate, they have authorized me to give out a handful of licenses to blog readers who are willing to share their story by posting a comment to this blog entry.  Simply add your comment below (be sure to include a valid email address in the box that asks for that) to be entered.  The contest starts immediately and over the next few days, the best stories will win.

    Read the article

  • How to modify grub entry for supporting KGDB kernel image?

    - by Nishant
    I am trying to update target m/c grub.cfg file for KGDB setup but while booting the m/c it got hung completely and not asking/waiting for remote gdb connection. Following is the entry which I added:- menuentry 'Ubuntu, with Linux 2.6.32-24-kgdb' --class ubuntu --class gnu-linux --class gnu --class os { recordfail insmod ext2 set root='(hd0,1)' search --no-floppy --fs-uuid --set 12878c3b-c553-4b4b-986a-6e32daea3ad1 linux /vmlinuz-2.6.32-kgdb root=/dev/mapper/ubuntu-root ro kgdbwait [email protected]/,@192.168.140.158/ quiet initrd /initrd.img-2.6.32-24-server } I have also compiled and copied /boot/vmlinuz-2.6.15.5-kgdb & /boot/System.map-2.6.15.5-kgdb to target m/c from devlopement m/c. STD entry before adding KGDB in grub.cfg was:- menuentry 'Ubuntu, with Linux 2.6.32-24-server' --class ubuntu --class gnu-linux --class gnu --class os { recordfail insmod ext2 set root='(hd0,1)' search --no-floppy --fs-uuid --set 12878c3b-c553-4b4b-986a-6e32daea3ad1 linux /vmlinuz-2.6.32-24-server root=/dev/mapper/ubuntu-root ro quiet initrd /initrd.img-2.6.32-24-server } Please suggest how to get rid of this problem.

    Read the article

  • Where does a "Technical Programmer" fit in, and what does the title mean? [closed]

    - by Mike E
    Was: "What is a 'Technical Programmer'"? I've noticed in job posting boards a few postings, all from European companies in the games industry, for a "Technical Programmer". The job description was similar, having to do with tools development, 3d graphics programming, etc. It seems to be somewhere between a Technical Artist who's more technical than artist or who can code, and a Technical Director but perhaps without the seniority/experience. Information elsewhere on the position is sparse. The title seems redundant and I haven't seen any American companies post jobs by that name, exactly. One example is this job posting on gamedev.net which isn't exactly thorough. In case the link dies: Subject: Technical Programmer Frictional Games, the creators of Amnesia: The Dark Descent and the Penumbra series, are looking for a talented programmer to join the company! You will be working for a small team with a big focus on finding new and innovating solutions. We want you who are not afraid to explore uncharted territory and constantly learn new things. Self-discipline and independence are also important traits as all work will be done from home. Some the things you will work with include: 3D math, rendering, shaders and everything else related. Console development (most likely Xbox 360). Hardware implementations (support for motion controls, etc). All coding is in C++, so great skills in that is imperative. Revised Summarised Question: So, where does a programmer of this nature fit in to software development team? If I had these on my team, what tasks am I expecting them to complete? Can I ask one to build a new level editor, or optimize the rendering engine? It doesn't seem to be a "tools programmer" which focuses on producing artist tools, often in high-level languages like C#, Python, or Java. Nor does it seem to be working directly on the engine, nor a graphics programmer, as such. Yet, a strong C++ requirement, which was mirrored in other postings besides this one I quoted. Edited To Add As far as it being a low-level programmer, I had considered that but lacking from the posting was a requirement of Assembly. Instead, they tend to require familiarity with higher-level hardware APIs such as DirectX, or DirectInput. I wasn't fully clear in my original post. I think, however, that Mathew Foscarini has it right in his answer, so barring someone who definitely works with or as a "Technical Programmer" stepping in to provide a clearer explanation, I'll go with that. A generalist, which also fits the description of a more-technical-than-artist TA.

    Read the article

  • How can I install an old version of libc on 12.04 and is it safe to do so?

    - by mathematician1975
    I am building an application on 12.04 and I need to run it on an embedded device. The device has libc-2.8.90.so on it and my dev machine has libc-2.15.so on it. I would like to install libc-2.8.90 onto my dev machine and attempt to link it to my application. I have searched at the Ubuntu software centre for libc-2.8.90 but I cannot find anything resembling it. Is there a way to install this on my machine from command line?? Also will my system be safe having 2 installed versions of libc at the same time? Can it lead to any instability?

    Read the article

  • Useful git commit messages for merged branches

    - by eykanal
    As a follow-up to this question: If I'm working on a team by myself, I can maintain useful commit messages when merging branches by squashing all the commits to a single diff and then merging that diff. That way I can easily see what changes were introduced in the branch, and I have a single summary describing the feature/change/whatever that was accomplished in that branch when browsing the master branch. My question now is, how can I accomplish this when working with a team? In that situation, the branches will be pushed to a remote repository, meaning that I can't squash all the commits in the branch down to a single commit. If the branch is public, can I still have a single useful merge commit in the master branch? (By "useful" I mean that the commit in the master line tells me (1) a useful summary of what was done in the branch and (2) diffs of the same.)

    Read the article

  • Continuous Deployment to Azure powered by Git

    Today Scott Guthrie announced several updated capabilities for Azure Web Sites. Announcing: Great Improvements to Windows Azure Web Sites I recommend you checkout the full post there are some really cool improvements. My favorite is the ability to enable Continuous Deployment from your CodePlex project into Azure. David Ebbo has a great video walk-through: (Please visit the site to view this video)

    Read the article

  • Help make the next Summit even better

    - by Bill Graziano
    After the Summit we send out a survey to capture feedback.  We ask a consistent set of questions so we get good year over year results.  I’ve watched blog posts and email threads with ideas for a better Summit.  I got to sit with Denny and crew again on Saturday night and talk about what worked and what didn’t.  We’d like to capture those ideas in a way that you can vote on what’s important to you.  Please take a second and visit http://feedback.sqlpass.org/.  You can make suggestions, vote on the ideas already posted and add your own comments.  Help PASS make next year’s Summit “The Best Summit Ever!”

    Read the article

  • How to get innovation in product development?

    - by simonsabin
    I was at Redgate last week and say plastered all over the canteen wall flip chart sheets with ideas scrawled over them. I was going to take a photo but thought it might not be allowed. I asked what it was all about and they told me that ALL the development team are going to stop working on what they are doing and do whatever they want. Whatever they want? yes So thats the whole team working on something? No, it can be individual or in groups. Having worked in product teams before I and the others...(read more)

    Read the article

  • DotNetNuke 5.4.1 Released

    I am happy to announce the release of DotNetNuke 5.4.1 which corrects the major issues which slipped through the QA process for 5.4. While we try to do a good job in testing our releases, our recent efforts for 5.3 and 5.4 have fallen short of the mark. We are currently working with a small team of commercial module developers and the core team to put a better public beta testing process in place that will help augment our own internal testing. Ultimately, community testing is the only testing that...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Installing linux on OCZ RevoDrive3 x2

    - by user2101712
    First of all, here is the configuration of my computer: Motherboard: Asus H87Plus RAM: Corsair Vengeance 32GB Processor: Intel i7 4770 Drive: OCZ RevoDrive 3 x2 (240 GB) (OCZ Revodrive3 is a PCIe module) I am trying to install the latest version of Ubuntu Desktop (13.10). The problem is that in the UEFI (bios) the drive shows up as a 240 GB drive, but in the Ubuntu installer it shows up as two 120 GB drives. If I install Ubuntu in any of these two drives, it never boots. The screen flickers a few times and comes back to the UEFI menu. I have tried reading up and have come across information that the drive has a "fakeraid", and the solution is to use dmraid. However, when I give the following commands in the terminal (from live CD): # modprobe dm_mod # dmraid -ay it says: no raid disks. And the following command: # ls -la /dev/mapper/ just shows /dev/mapper/control How can I install Ubuntu on my computer? what is the correct method?

    Read the article

  • Sharing ideas without risk of leaking

    - by eversor
    As freelancers, we meet somewhere and chat about a new idea for a project, brainstorm etc. Up to this point, we have taken notes of the ideas, but we would like to be able to share more efficiently the ideas with each other. However, I fear that if I use some online product (such as Google Docs) these ideas could be seen by people outside the team (employees of the company of the online product, other users...). I am not sure if I am being a little Paranoid parrot... One solution that we have considered is to install a Subversion with just one ideas.txt. But that would require a server in one of our houses, which is a little unconfortable. So how do you share your ideas for a new project with your team without risking the ideas to be stolen?

    Read the article

  • WinTV-HVR-1900 no picture in VLC

    - by nLinked
    I've connected a Hauppauge WinTV-HVR-1900 to my Ubuntu 10.10 PC and used the instructions here to install the firmware and test that it is detected using dmesg. Detection is successful. The WinTV is connected to my TV using component connections. I run this command to enable the component interface: v4l2-ctl -i 1. Then I use cat /dev/video0 test.mpg, and that file is created and plays perfectly with video and sound. However, when I try to play /dev/video0 in VLC Player (and other players), I just get a black screen. I also try the v4l2-ctl -i 1 command before I try it in VLC, but I still get a black screen, and when I use v4l2-ctl -I, it shows it has connection 0 again (Television), so it reverts back to connection 0 instead of connection 1 (component). So I'm a little stuck with getting VLC to get a picture, even though the CAT command works fine. Anything I can try?

    Read the article

  • Dealing with technical debt

    - by Desolate Planet
    This is a question that I often ask myself when working with developers. I've worked at four companies so far, and I've noticed a lack of attention to keeping code clean and dealing with technical debt that hinders future progress in a software app. For example, the first company I worked for had written a database from scratch rather than take something like MySQL and that created hell for the team when refacoring or extending the app. I've always tried to be honest and clear with my manager when he discusses projections, but management doesn't seem interested in fixing what's already there and it's horrible to see the impact it has on team morale and in their attitude towards others. What are your thoughts on the best way to tackle this problem? What I've seen is people packing up and leaving and the company becomes a revolving door with developers coming and and out and making the code worse. How do you communicate this to management to get them interested in sorting out technical debt?

    Read the article

  • Why do some open-source projects NOT have a public issue tracker?

    - by linquize
    Why do some open source projects not allow to report issues in a public issue tracker? Those projects require the issues to be reported via email, and the issues may be forwarded to people in a mailing list. Users may repeatedly report the same issue if there is no public issue tracker, as they have no easy way to know what has been reported before. The project team members need to spend extra time answering those repeated issues. Some projects do have a public issue tracker but the issues are still reported through email and they are posted by the project team only after filtering. It does not allow anyone to report directly in issue tracker (example: SVN). Such arrangement is not transparent nor open, which I think violates the philosophy of open source. And it is outdated.

    Read the article

  • What software development model has worked best for software teams with heavy dependancy on hardware teams?

    - by MasterDIB
    So, let me explain more. There are a number of competing best practices for software development. I can find that many teams have benefited from Agile practices in some cases. In some other cases, using the Unified Process has been championed by large companies like IBM. The common themes that I find seemed to work well for teams that mainly develop software. I am interested to know what has worked best for people who have worked in shops where there is a team on the other side that produce the hardware that your software is running on. For example, one team puts together a crate with several custom hardware on it; while you need to develop the software that would run on those crates. I can't find a development model (agile, spiral ...) that works best in this case. Any wisdom is this area will be well appreciated.

    Read the article

  • Corticon provides Business Rules Engines for Silverlight, WCF and .NET developers

    Now Corticon Business Rules Engines and Business Rules Management Systems users can enjoy support for the Windows 7 operating system, and for Silverlight and Windows Communication Foundation developers. The new Corticon 4.3 provides numerous performance, usability, and integration enhancements and provides the industry-first cloud deployment option for a business rules engine. span.fullpost {display:none;}

    Read the article

  • What to do with my "unmounted drive"?

    - by Taylor Guistwite
    I just recently followed the tutorials on http://www.ubuntu.com/download/ubuntu/download for installing the ubuntu server onto my 1TB Seagate External. I was planning on using this to install it on my macbook and in these instructions it states to preform this line of code Run diskutil unmountDisk /dev/diskN (replace N with the disk number from the last command; in the previous example, N would be 2) Now my HD prompts "The disk you inserted was not readable by this computer". Would I just run diskutil mountDisk /dev/diskN in order to be able to access all my files again? here is a screenshot to the instructions i followed http://i17.photobucket.com/albums/b97/hello_screamo/Screenshot2011-11-11at113914AM.png

    Read the article

  • Floppy Drive not recognised

    - by Art
    I am a newbie with Ubuntu during the past two weeks, and I believe, that I have made good progress so far. Current setup tasks are completed, like: "Wireless Network, Printer, Thunderbird + Firefox, e-mail", also my local weather station - all setup. Updated files and rearranged the desktop to my liking. The problem I have now is my floppy drive which is not recognized, but the activity light is shown on constantly. I have downloaded the "MakeFloppyDriveAvailableToEveryone" and i am confident that I can follow the instructions to change the floppy drive. However, I do not have an existing fd0 file in /dev/???, but only /dev/fd/ files 0 to 46. Please advise.

    Read the article

  • How can we unify business goals and technical goals?

    - by BAM
    Some background I work at a small startup: 4 devs, 1 designer, and 2 non-technical co-founders, one who provides funding, and the other who handles day-to-day management and sales. Our company produces mobile apps for target industries, and we've gotten a lot of lucky breaks lately. The outlook is good, and we're confident we can make this thing work. One reason is our product development team. Everyone on the team is passionate, driven, and has a great sense of what makes an awesome product. As a result, we've built some beautiful applications that we're all proud of. The other reason is the co-founders. Both have a brilliant business sense (one actually founded a multi-million dollar company already), and they have close ties in many of the industries we're trying to penetrate. Consequently, they've brought in some great business and continue to keep jobs in the pipeline. The problem The problem we can't seem to shake is how to bring these two awesome advantages together. On the business side, there is a huge pressure to deliver as fast as possible as much as possible, whereas on the development side there is pressure to take your time, come up with the right solution, and pay attention to all the details. Lately these two sides have been butting heads a lot. Developers are demanding quality while managers are demanding quantity. How can we handle this? Both sides are correct. We can't survive as a company if we build terrible applications, but we also can't survive if we don't sell enough. So how should we go about making compromises? Things we've done with little or no success: Work more (well, it did result in better quality and faster delivery, but the dev team has never been more stressed out before) Charge more (as a startup, we don't yet have the credibility to justify higher prices, so no one is willing to pay) Extend deadlines (if we charge the same, but take longer, we'll end up losing money) Things we've done with some success: Sacrifice pay to cut costs (everyone, from devs to management, is paid less than they could be making elsewhere. In return, however, we all have creative input and more flexibility and freedom, a typical startup trade off) Standardize project management (we recently started adhering to agile/scrum principles so we can base deadlines on actual velocity, not just arbitrary guesses) Hire more people (we used to have 2 developers and no designers, which really limited our bandwidth. However, as a startup we can only afford to hire a few extra people.) Is there anything we're missing or doing wrong? How is this handled at successful companies? Thanks in advance for any feedback :)

    Read the article

< Previous Page | 215 216 217 218 219 220 221 222 223 224 225 226  | Next Page >