Search Results

Search found 9044 results on 362 pages for 'bad sector'.

Page 13/362 | < Previous Page | 9 10 11 12 13 14 15 16 17 18 19 20  | Next Page >

  • Rapid Application Development, good, bad or ugly?

    - by chrisw
    I have been working for such a shop for the past three years and I know deep down it cannot be like this everywhere. When I think of Rapid Application Development I immediately think programming without fore-thought. For example, when my company decides to come out with a new product, they don't do any type of relationship mapping, no ER diagrams, no round table discussions on expandability. No, the senior developer that ends up working on the product puts together a screen shot walk-through of the application to show to the client. Once the client signs off on the project work is underway by the senior developer. Now you have a senior developer (I use that term "senior" loosely) coding the application in under a week with no unit testing. Well I guess the good to this is it keeps programmers employed due to the enormous amount of unforeseen "features" in the newly created application. Have any of you dealt with a company like this? If you did how did you preserve your sanity?

    Read the article

  • Can too much abstraction be bad?

    - by m3th0dman
    As programmers I feel that our goal is to provide good abstractions on the given domain model and business logic. But where should this abstraction stop? How to make the trade-off between abstraction and all it's benefits (flexibility, ease of changing etc.) and ease of understanding the code and all it's benefits. I believe I tend to write code overly abstracted and I don't know how good is it; I often tend to write it like it is some kind of a micro-framework, which consists of two parts: Micro-Modules which are hooked up in the micro-framework: these modules are easy to be understood, developed and maintained as single units. This code basically represents the code that actually does the functional stuff, described in requirements. Connecting code; now here I believe stands the problem. This code tends to be complicated because it is sometimes very abstracted and is hard to be understood at the beginning; this arises due to the fact that it is only pure abstraction, the base in reality and business logic being performed in the code presented 1; from this reason this code is not expected to be changed once tested. Is this a good approach at programming? That it, having changing code very fragmented in many modules and very easy to be understood and non-changing code very complex from the abstraction POV? Should all the code be uniformly complex (that is code 1 more complex and interlinked and code 2 more simple) so that anybody looking through it can understand it in a reasonable amount of time but change is expensive or the solution presented above is good, where "changing code" is very easy to be understood, debugged, changed and "linking code" is kind of difficult. Note: this is not about code readability! Both code at 1 and 2 is readable, but code at 2 comes with more complex abstractions while code 1 comes with simple abstractions.

    Read the article

  • The old "do as I say, not as I do" problem

    - by AaronBertrand
    Microsoft is often considered a leader, an innovator, a trend-setter. The same could be said for Apple, Google, and a host of other tech companies. And each of those has its set of critics as well, who think that the company is the opposite - or worse. Some people think it is a good idea to model their own code, architecture or applications after things that these companies have done, but this is not always the best approach. Humans work at these companies too, and everyone is prone to mistakes,...(read more)

    Read the article

  • Bad texture on model with different GPU

    - by Pacha
    I have some kind of distortion on the texture of my 3D model. It works perfectly well on an AMD GPU, but when testing on a integrated Intel HD graphics card it has a weird issue. I don't have a problem with the rest of my entities as they are not scaled. The models with the problems are scaled, as my engine supports different sizes for the platforms. I am using Ogre3D as rendering engine, and GLSL as shader language. Vertex shader: #version 120 varying vec2 UV; void main() { UV = gl_MultiTexCoord0; gl_Position = gl_ModelViewProjectionMatrix * gl_Vertex; } Fragment shader: #version 120 varying vec2 UV; uniform sampler2D diffuseMap; void main(void) { gl_FragColor = texture(diffuseMap, UV); } Screenshot (the error is on the right and left side, the top and bottom part are rendered perfectly well):

    Read the article

  • Test-only Members: Good or Bad?

    In the article, Dino focuses on two particular situations: dealing with dependencies and testing private members. He'll be discussing this in the context of ASP.NET MVC and MSTest, but without any significant loss of generality.

    Read the article

  • Are null references really a bad thing?

    - by Tim Goodman
    I've heard it said that the inclusion of null references in programming languages is the "billion dollar mistake". But why? Sure, they can cause NullReferenceExceptions, but so what? Any element of the language can be a source of errors if used improperly. And what's the alternative? I suppose instead of saying this: Customer c = Customer.GetByLastName("Goodman"); // returns null if not found if (c != null) { Console.WriteLine(c.FirstName + " " + c.LastName + " is awesome!"); } else { Console.WriteLine("There was no customer named Goodman. How lame!"); } You could say this: if (Customer.ExistsWithLastName("Goodman")) { Customer c = Customer.GetByLastName("Goodman") // throws error if not found Console.WriteLine(c.FirstName + " " + c.LastName + " is awesome!"); } else { Console.WriteLine("There was no customer named Goodman. How lame!"); } But how is that better? Either way, if you forget to check that the customer exists, you get an exception. I suppose that a CustomerNotFoundException is a bit easier to debug than a NullReferenceException by virtue of being more descriptive. Is that all there is to it?

    Read the article

  • Poor, Bad SEO Techniques

    In today's competitive web market place good search engine and Google rankings are so important when it comes to the fine lines of success or not. There are many SEO companies out there that can offer advice and/or optimise your site for you but when are we going too far?

    Read the article

  • self referencing tables, good or bad?

    - by NimChimpsky
    Representing geographical locations within an application, the design of the underlying data model suggests two clear options (or maybe more?). One table with a self referencing parent_id column uk - london (london parent id = UK id) or two tables, with a one to many relationship using a foreign key. My preference is for one self-refercing table as it easily allows to extend into as many sub regions as required. IN general do people veer away from self referencing tables, or are they A-OK ?

    Read the article

  • Mouse pointer hides at bad times after updating to 13.10

    - by Richard
    A bug, I think, though I can't even seem to find any information when searching for it. The mouse pointer hides itself, but it is still possible to use it. Whenever I start watching a video in VLC, or start playing KSP through Steam my mouse pointer disappears. It is not returned upon exiting the application. I have not tried with other games in Steam. The mouse pointer do not disappear when watching videos online through YouTube, though it properly hides itself after a few seconds. The bug came about after I updated to 13.10, it was not present with the same setup in 13.04. I have tried removing ~/.compiz and restarting the kernel module psmouse. Going to tty1 and back to the desktop reveals the pointer, though this fix is bothering me. How do I fix this problem?

    Read the article

  • EAV - is it really bad in all scenarios?

    - by Giedrius
    I'm thinking to use EAV for some of the stuff in one of the projects, but all questions about it in stackoverflow end up to answers calling EAV an anti pattern. But I'm wondering, if is it that wrong in all cases? Let's say shop product entity, it has common features, like name, description, image, price, etc., that take part in logic many places and has (semi)unique features, like watch and beach ball would be described by completely different aspects. So I think EAV would fit for storing those (semi)unique features? All this is assuming, that for showing product list, it is enough info in product table (that means no EAV is involved) and just when showing one product/comparing up to 5 products/etc. data saved using EAV is used. I've seen such approach in Magento commerce and it is quite popular, so may be there are cases, when EAV is reasonable?

    Read the article

  • Ubuntu does not boot after bad shutdown

    - by Molly Gibson
    I recently swapped from Windows7 to Ubuntu 11.10 due to problems with my Windows. Ubuntu worked great for a few days with no problem, This morning I ran out of battery and the laptop shut down. I rebooted it and now it just displays an error message ( black screen with coding commands) it doesn't even go onto the purple screen anymore. Any help on this would be much appreciated.. Ps: would installing Ubuntu again make any difference?

    Read the article

  • Remove a bad/erroneous WebPart from a SharePoint page

    - by KunaalKapoor
    If you've added a poorly written webpart to your 'default.aspx' page, the consequence of this action will be that you won't be able to load the page anymore... Don't be sad, there is still a way to remove the webpart from the page :) (Yes, even removing the webpart from the webpart gallery would not solve this issue).Steps to fix this:1. Append the following query string to your URL: ?Contents=1.Once you've added Contents=1 as a query string to the webpart page's URL it will display the Webpart Maintenance Page. Example: http://mysharepointserver/default.aspx?contents=12. On that page you can now see the webparts added to the page, delete the problematic webpart.Now try reloading the default.aspx page... Tadaaa!!! you can view your page again :)3. Leave a thank you note @ comments section :)

    Read the article

  • Blocked Sites at work (that aren't even bad)

    - by Mercfh
    So here recently, i've been using google to look up information for basically random programming things (i was just hired on a month or so ago). So here recently I was actually looking up some information about RAW_SOCKETS (but thats beside the point) Anyways some of the tutorials sites/explaining how to use them and explaining the protocol sites are actually blocked. (and our manager sent out an email saying that if u run into a site just to email her just in case). Now obviously...w/e sys admins probably see these 'blocked' sites in their reports. But should I be worried? I mean....I literally am not trying to be devious Im just trying to learn stuff. I guess programming websites are sometimes labeled as "hacking". sometimes blogs get labeled like that, but alot of the time blogs have USEFUL information. This apparently happens alot of my other co-workers and they don't even bother emailing our manager.....but should I be worried? Or has this happened to you guys before?

    Read the article

  • How to deal with bad developers who hold back the project

    - by ILovePaperTowels
    We're at the end of a project, but we continue to run into issues because of a single piece of the project. This piece is handled by a specific developer. Finally, we grabbed latest and started reviewing it. It's just horrendous code! Trying to step through it was difficult and it's a relatively simple workflow. The point of this question is, how to deal with this situation. The developer has a hard time accepting criticism (constructive or otherwise) and feels he is more knowledgeable than others on the team who are, well, highly decorated, experienced and accomplished developers. It's difficult to even get into a topic about development because it turns into "I know what I'm talking about and you're just wrong!" type of conversation. A request has already been put in to replace this developer but it is a hard sell since devs are in short supply where we are and this is a corporation with a LOT of political bs. Management has been notified a few times but nothing is happening.

    Read the article

  • Tracking 502 bad gateway error

    - by dasickle
    I moved my Wordpress site to WP Engine and now I constantly get 502 errors. I spoke with support and they said that its because I have a lot of DB queries. I ran some tests and my frontpage only has 95 queries and page size is about 500kb. Most inner pages are around 60 queries. All queries are very short. Some people tell me its common with WP Engine because they run nginx. Why do I keep getting these errors and is there a way to track how many of them happen on daily basis? P.S. WP Engine log is empty so cant see the 502's there.

    Read the article

  • Why is verbosity bad for a programming language?

    - by frowing
    I have seen many people around complaining about verbosity in programming languages. I find that, within some bounds, the more verbose a programming language is, the better it is to understand. I think that verbosity also reinforces writing clearer APIs for that particular language. The only disadvantage I can think of is that it makes you type more, but I mean, most people use IDEs that do all that work for you. so, What are possible downsides to a verbose programming language?

    Read the article

  • Caching: the Good, the Bad and the Hype

    One of the more important aspects of the scalability of an ASP.NET site is caching. To do this effectively, one must understand the relative permanence and importance of the data that is presented to the user, and work out which of the four major aspects of caching should be used. There is always a compromise, but in most cases it is an easy compromise to make considering its effects in a heavily-loaded production system

    Read the article

  • Is Joomla CMS Good Or Bad For SEO?

    Many articles have been written for and against the use of Joomla and it's effectiveness regarding Search Engine Optimization. It is one of the most popular and widely used Content Management Systems out there. Joomla can be Search Engine friendly more so than other content management systems provided it is handled by SEO experts with experience in optimizing Joomla pages. The fact is some of the best ranking sites on Google are those that run on popular CMS like Joomla and blogs.

    Read the article

  • Improve Bad testing

    - by SetiSeeker
    We have a large team of developers and testers. The ratio is one tester for every one developer. We have full bug tracking and reporting systems in place. We have test plans in place. Every change to the product, the testing team is involved in the design of the feature and are included in the development process as much as possible. We build in small iterative blocks, using scrum methodology and every scrum they are included in, including the grooming sessions etc. But every release of the product, they miss even the most simple and obvious defects. How can we improve this?

    Read the article

  • The 5 stages reviewing bad TSQL

    - by Mike Femenella
    I'm working with an app team that is light on TSQL expertise this week and couldn't help but draw a parallel to the 5 stages of grieving. Denial: There’s nothing wrong with the code SQL Server has a bug in it. There is a network problem. Anger: You’re doing what in your code?! Why on earth are you doing that? That’s crazy. Bargaining: Fine you can keep your cursor but let’s speed things up a bit. Depression: Ugh, this is so horrible I’m never going to be able to fix all of it. Acceptance: Ok, we’re screwed and we know we’re screwed. This is going to hurt…

    Read the article

< Previous Page | 9 10 11 12 13 14 15 16 17 18 19 20  | Next Page >