Standards for Bug Severities? 162
MythoBeast asks: "While attempting to determine just how far a (unnamed) company's software is from releasability, a group that I belong to discovered that they consider a piece of software releasable with numerous 'severity one' bugs. This seems outright heinous to us at first, but then the conversation devolved into an argument of what constituted a 'severity one' bug. Is there a standard for such things? Is there a standard for how many of these things are reasonable for 'releasable' software?"
Re:Purpose of testing (Score:1)
As you know (now), the testers role is to bash the fuck out of your software, break it, and then tell you about it. Then the developer scratches his or her head for two or three days, sends the bug back as "User Understanding" and breaks something else in the process.
Then all the testers go to the pub at lunch time, and complain about how shit the code is, how we found a bug that should have been found by the developer in Unit Test, how some bastard from development has been messing up the application database, and how managment won't listen and the development/testing process is all fucked up.
Then we come back to work, and annoy the developers with almost infinate questions and problems.
This is the role of a Software Tester
(A bitter and twisted Software Tester)
Re:I've always respected... (Score:1)
What we use (Score:1)
S1 = Causes data loss or damage to system. Hangs at OS level. Fire, flood, famine, locusts
S2 = Hands outside of OS level. Prevents system from booting or restarting.
S3 = Annoyance bugs. Major look & feel issues of user interface.
S4 = Bad grammar or spelling, minor look & feel issues.
The guideline is being printed in huge letters & plastered all over the test lab so there's no confusion.
Re:How do you get lower-order bugs addressed? (Score:1)
Hey, back up. Why wouldn't Testers want cool Geek toys? Look, can we get this clear please? Testers are just Geeks who don't want to code all day
Maybe you should go talk to your QA people some time, find out how much some of them really know.
How they do it at Microsoft (Score:1)
Four-level system (Score:1)
The goal of a software company is to make money, and the way to make money is to satisfy the customers. So the most important thing is that the severity system reflects the users' satisfaction with the software. If there's some silly little glitch that drives the customers bats, it deserves a high-severity rating, even if the engineers think it's harmless. I currently use the following system, which I've refined over the years and which serves me well:
4 - A feature request, i.e., something outside the design spec of the software.
3 - The software ought to do it differently, but it's unlikely most users will run into it. There is no loss of functionality, or there is a workaround possible. There is no possibility the bug will cause loss of data.
2 - A serious bug...something lots of users will encounter, but that does not result in loss of data. There is a workaround possible.
1 - A serious bug causing loss of functionality. No workaround possible or causes loss of data.
A commercial release of a product should contain no severity 1 or 2 bugs. However, it's okay to ship a product with these bugs if the bugs and workarounds are described in the release notes. If you're embarrassed to write up a particular bug in the release notes, then fix it.
A beta release follows the same rules, except the release notes tend to be longer.
Priority depends on context (Score:1)
The standard for shipping (Score:2)
Yes. It is somthing on the line 'if it compiles, you can ship it'
Cheers,
--fred
Be realistic (Score:2)
Re:Difference between Severity and Priority (Score:2)
That is something technical people tend to forget.
OTOH, what people in the software industry tend to either not know, or know but avoid thinking about, is exactly how bad their reputation is in the rest of the business world. Let me illustrate.
For the last 10 years or so I have done software implementation and IT technical support for a variety of industrial products manufacturers. The last company I worked for had the following characteristics:
* Served heavy industrial customers with mechanical products
* Competitive market with fixed price, fixed delivery date (+ penalties!) contracts and annual price _decreases_
* Product had a software component
* Gross margin in the industry and its suppliers of about 1/4 that of software firms.
And oh yeah, if our product failed, there was about a 10% chance that the crew of four people operating the system would die a very, very unpleasant death.
Guess what? Our products didn't fail. Not the hardware, not the software. Of course, we weren't gods, so very rarely an accident would occur. But EVERY accident was reviewed in a very thorough and unpleasant manner by officers of the customer, who would come onto the shop floor and interrogate production line employees at length. And woe unto the supplier whose employees weren't (a) fully knowledgable about their job and product (b) fully truthful.
Now, in this environment, what opinion of the software "industry" do you think general management held? That is, the people responsible for the _business_? I am sure you can figure it out, but to be specific: a very low opinion. A typical question was, "Explain again why this product that I just paid $1.2M for, which figure represents 10% of our gross profit this year, crashes every time I press the print button?" These were not very nice conversations as you can imagine.
I won't go on, but I would suggest that the software developers reading this thread look up some of William Mossberg's columns in the Wall Street Journal and think a bit about what he has to say concerning software reliablity and complexity. From a business perspective.
sPh
changing severities (Score:2)
Re:I've always respected... (Score:2)
Seems like a fairly simple sceme to me -
Hellooooooooooooo? The Perl Bugs Index doesn't actually define the different categories of bugs. Larry (or whomever) uses the force to determine which bugs fall into which categories. This is unfortunately common.
The question being asked by our poster is "How do you avoid doing this? What are good definitions for different categories of bugs?"
And so how does slashdot answer? "Oh, you do it the way Perl does, just use the force," which is exactly the answer not to give! And then for good measure, let's mod that post up to 5, Informative.
Come on folks. Think a little bit. This place is only as good as you make it. Before you post, think. Before you moderate, think.
End of sermon. Sorry. This just annoys me.
Are you complaining about my employer? (Score:4)
I have seen this in my company. There is no conspiracy against the customers. Usually I have detected two reasons for these kinds of semantic disputes:
(Sorry. Anonymous this time.)
Re:Sayeth the Capability Maturity Model... (Score:4)
If you are playing tetris on a computer controlling your life support system, and every time you get a green piece tetris pauses briefly and the CO2 filter shuts off so you suffocate Tetris has a Severity 4 or 5 bug.
Difference between Severity and Priority (Score:5)
Sev1 bugs: cause loss of data, hardware damage, or take down other programs (or the OS) which might cause loss of data.
Sev2 bugs: cause the program to crash without losing data, or otherwise make some feature non-functional
Sev3 bugs: impair the usefullness of some feature
Sev4 bugs: are cosmetic or may cause some confusion to the user (wrong status text, spelling errors, etc)
HOWEVER, there is also a priority scale. A Sev1 bug is not necessarily a Pri-1 bug, if it only occurs under extremely obscure conditions. Eg, if your program crashes, losing data, when a PC-Card is inserted, that's a Sev-1 bug. If it occurs any time any PC-card is inserted, then it's a Pri-1 bug. But if it's only one particular kind of PC-Card that exists in the wild in quantities of, say, a couple of hundred, then it's not a Pri-1 bug even though it's still a Sev-1 bug (you may still choose to fix the bug, but other Sev-1 bugs have priority since you can doc around this one). On the other hand, a low severity bug that's in your face every time you use the product is probably a fairly high-priority bug.
If you're running your business as a business, you will ship with bugs. If your product has a broad market, you will probably ship with a couple of Sev-1 bugs. (This sounds paradoxical, but consider -- the broader your market, the larger the number of permutations of hardware and software your product is going to inderact with, and the higher the chance some obscure situation will occur in which a Sev1 rears its head). It's a business decision to ship, so the number of bugs you ship with has to be based on more than just a technical analysis (Rule #1 of software development: shipping is a feature too).
It's easy for us technical heads to get on a high and mighty horse about quality and not shipping with bugs. But we tend to forget that what we're doing is NOT engineering (it's a lot closer to handicrafts in the days of guilds than it is to engineering the way it is done with bridges and aircraft). And we're not the people trying to run a business.
Re:Management v QA/DVT (Score:5)
The corporate-wide standard in this case was badly broken, and was obviously written by someone who doenst understand software release.
QA is NOT there to find, log, and force a fix for every single bug in the system. That is an impossibiliy. Just like manufacturing QA, Software QA is a statistical processes.
The classic manufacturing QA example is the Bolt Building Machine. It makes bolts from metal bar at the rate of 10000 a day. It would be wonderful if manufacturing QA could take each one of the bolts produced by the machine and measure the eight or so critical bolt dimentions to ensure that each bolt was within tolerance. Unfortunatly, it takes about 15 minutes to measure a single bolt with that kind of precision, so it is not cost effective to measure each and every bolt. The next best way to ensure quality is to take a random sample of bolts and measure them, so that you can be statistically certain to an acceptable degree that the rest of the bolts are good.
Software QA works the same way. When a new release of software comes out, it is rarely run against the entire QA testplan again. This is especially true with patch releases where the development team believes that only a few minor things have changed. It just takes too long to run a full QA cycle against a build like that-- it is possible, but it isnt cost effective. Instead, the QA team runs a subset of the test plan against the build in an effort to sample the test space and come up with a statistical certainty that this release of the product isnt broken.
Bug classification is also a statistical problem equivilent to the classic 'expected value' problem. The standard five bug classifications are useful measures of the 'weight' of a problem. They answer the question, 'If this bug happens, how much does it hurt?' What the abovementioned corporate scheme failed to qualify was the 'How often it happens' part. Without knowing the frequency that the problem occurs, it is impossible to calculate the 'expected hurt' that the problem will generate.
Here is a degenerate case example. Say the program has code like this:
If a user directs a cosmic ray into the computer that corrupts the value of pointer between the time its validity is checked and the time it is dereferenced, the program will crash with severity 1. There is no workaround, because every time the user sends in this ray, the program fails. According to the abovementioned corporate policy, the product cant release until this problem is fixed, period.
What are the odds of a user here on earth actually sending in that cosmic ray at exactly the right time, let alone having a need to? Very close to zero. So close in fact, that it practically doesnt matter how severe the crash is. The 'expected hurt' from this problem is going to be very low, and the product should ship with that severity 1 bug going unfixed. If the software was going into space where cosmic rays are common, it might be prudent to recalculate the odds of the problem happening before shipping the code-- and maybe then the only practical solution is to ship with the bug but send up redundant systems.
QA is not there to ensure that the product has no bugs. QA is there to qualifiy risks, and to assure that the quality of the product falls within a certain tolerance. In a perfect universe, software would never have bugs, and bolts would never bend. Here in our solar system though, it all boils down to risk management.
A. Coward, (a QA Manager for a fortune 50)
The Universal Standard (Score:4)
Re:Purpose of testing (Score:1)
How many bugs are they finding, and what severity are they? This speaks to the quality of the system design and the quality of the engineers working on the system. If there are few bugs found and fixed, you can be confident you have a good system going out. If there are a lot of bugs found and fixed, I would worry that there are a lot more left undiscovered.
the above is only true is the testing group is good. If they run one test case continiously they can miss a lot of bugs. (This has been known to happen, generally when many seemingly different cases all exercise the ame code path).
Who can you ship to. (Score:2)
We have guildlines, that state something different, but i've noticed it comes down to a management decision "Is the pain of shipping with this bug outwieghed by the cost of fixed it?"
Early on everyone else's definition is what we use. Several months before shipment (and things are normally fairly good then) management does a bug review and decides what the priority is. At that point:
Sev 1: data corruption, or serious crash. encountered often and no work around. Affects major features. Cannot ship.
Sev 2: above, but we know exatly how to cause it, if you don't do that things work okay. Can ship to beta customers, but not to production use.
Sev 3: Not as serious, but a black eye. Fix if at all possibal. But work arounds exist, and in general we can live without. If it affects a minor feature we can turn off, then it is always this level. Anything shiped with this level or problem goes into the release notes. Before shipping management has to review each and decide to wait while it is fixed or ship now.
Sev 4: Minor. Deseriable to fix, but rare enough that we can just document it.
Note that for sev 3 and 4, there is a chance that if the feature in question isn't a major one we will disable it instead of fixing it. Also only major features can be sev 1 or 2, as we can't disable them.
Severity vs priority (Score:1)
As it did to me at first, but think: a bug that causes a crash, if it happens sufficiently rarely, might not be worth fixing from an economic standpoint. While it still sometimes drives me nuts to ignore the bone-jarring crash-causing bugs to fix cosmetic flaws, if the crash happens rarely while the cosmetic flaw stares you in the face every time the program runs, it can make sense.
Re:Japanese Standards (Score:3)
Re:Sayeth the Capability Maturity Model... (Score:3)
~luge
Re:Severity/Priority (Score:1)
Re:Each bug has two axis (Score:1)
usually, yes there are... (Score:5)
Show stoppers are usually bugs that crash the software when doing normal functions. Like if you open up a file and the software crashes, that would be a show stopper. In some cases it depends on your clients as well. If the client finds a bug they may deem it as necessary to be fixed. Most companies I have worked at usually work closely with clients. In many cases the software does not get FULLY tested till the client actually gets the software. Then when (not if they will find bugs) they tell us an dthen we make it a priority to fix them and send out patches to certain clients if not all. This would depend on the client contract.
"Low priority" bugs would be ones that do not impead functionality, but are more of annoyances. Take for instance I saw one bug that when you scroll the scrollbar the window would flicker. This is annoying, but does not prevent the user from using the software. Other less important bugs are bugs that are on reports. If a report does not show the correct data and you can get the data from the database or it is a known bug this can also be a low priority bug. The question usually is "Can the client work with the application with this bug or will it cause the client major problems or loss of money?" Take for instance the reports bug, a client could just not use the reports and create their own from sql.
Something you must remember is that some of the programs that I have worked on have been written by many programmers. They are thousands of files andhave client, server, and database code, in multiple languages. Some of these bugs could be caused by some of the software that we use. Bugs in windows API's, bugs in Sybase, or Oracle, bugs in the server OS, or even 'quirks' that many programmers do not know about.
Next, many times programmers get put in a job where they know the language but NOT the API and the API is not well documented, especially if this is a 'custom' system where the software company has developed programs for DB access on top of the 'db lib'. What about programmers that leave the company or get fired?
So who do you blame? The programmer that impemented something that he did not know the company API, the software company for not completely testing "EVERY" single case that could possibly come up?
You never know how a user will use software or what they may do or have installed on their system that could affect how your software will work.
I don't want a lot, I just want it all!
Flame away, I have a hose!
Depends on how the "features" are accessed (Score:2)
As long as I can use the software in a reasonable manner, it's all good.
Re:The Rules... (Score:3)
1. The computer knows all.
2. The computer is always right.
3. The computer loves you.
6. The computer sees everything.
4. The computer is in connttrrolll.
9. The computer wants all good things.
14. The computer will make things all better.
2. The computer cannot be questioned.
4. The computer is above you.
7. The computer makes you, thinks.
Apologies to the creators of Paranoid.
--
Re:The Golden Rule... (Score:1)
More or less.. no...
I honnestly don't know what Microsoft people want. What ever it is Microsoft is delivering.
Microsoft people admit you need to know what you are doing to make Windows work. You need to memoerize commands and fiddle with config files.. or at least one big config file.
Thats no diffrent than Linux.
However Windows is very diffrent from Linux.
Yeah Windows has it's problems and truth is those problems are NOT found in Linux.
The inverse is also true..
People put up with limitations to get something they value more.
If they could they'd go for something lacking the problems of Linux and Windows.
For now it's a choice.
If you want easy use Mac.
If power is all you want.. they all are..
Just a matter of picking the trade offs..
User friendlyness in trade for stability is NOT a tradeoff offered by Windows...
Re:Sayeth the Capability Maturity Model... (Score:5)
Re:Sayeth the Capability Maturity Model©©© (Score:1)
So by this model would all usability defects be classified as sev© 3 or sev© 4 ¥primarily sev© 4, I would imagine?
Nice© Tell that to your human factors friends©
go search comp.software-testing (Score:1)
Its a serious bug if... (Score:1)
Boy does the truth hurt!
Re:How do you get lower-order bugs addressed? (Score:1)
: )
He's always asking me how to install software and fix his computer anyway, so although his social skills are, uh, iffy, he's no geek.
Re:How do you get lower-order bugs addressed? (Score:2)
Your point is well taken, though. If everybody just did their damn job, the process would work fine!
How do you get lower-order bugs addressed? (Score:3)
Any suggestions, other than make EVERYTHING a Sev1 issue and making the whole process useless?
Tetris on Bet-your-life systems... (Score:2)
If a user is playing Tetris on a computer controlling your life support system, there are two top priority bugs:
Bet-your-life systems tend to be very small, very limited in their functionality, isolated from the outside world, and tested up the yin-yang. If you are building a system that needs such systems, and want to add additional capabilities, the accepted practice is to put those capabilities on another computer that can't impact the bet-your-life system.
Take an airplane as an example. If it has a fly-by-wire system, that is a seperate computer. If there is a life support computer (do you need a computer to regulate pressure and the like?), that would be a separate computer. If it has an autopilot, that is a seperate computer. Other systems you can live without, such as navigation and targeting (on military jets), are again separate computers.
For that matter, there are airlines that want to provide net access in flight. This means that they have to have a small ISP on board their planes. These are certainly not going to get tied into the flight control computers. If they did, they can get infected with a virus in flight. By keeping the computers disconnected, any attacks on or failures of the ISP computers have no impact on flying the plane.
By doing this, engineers are able to protect the computing resources that you can't live without.
If you can play Tetris on your life-support system, then you lack the above isolation, and are betting your life on a system that may have bugs.
Sayeth the Capability Maturity Model... (Score:5)
"Severity 1: An error that prevents the accomplishment of an operational or mission-essential function, prevents the operator/user from performing a mission-essential function, or jeopardizes personnel safety.
Severity 2: An error that adversely affects the accomplishment of an operational or mission-essential function and for which no acceptable alternative workarounds are available.
Severity 3: An error that adversely affects the accomplishment of an operational or mission-essential function for which acceptable alternative workarounds are available.
Severity 4: An error that is an operator/user inconvenience and affects operational or mission-essential functions.
Severity 5: All other errors."
Any other questions?...
A bunch of unrelated points (Score:2)
Severity vs. Priority
As several other posters have pointed out, it's important to make a distinction between severity and priority. I would also say that it helps to define severity as narrowly as possible, to refer only to the effect of a bug when it occurs. I've already seen some people who accept the severity/priority distinction make the mistake of letting other factors (e.g. availability of a workaround) be factored into severity when they really only affect priority. Priority is the "net net" of severity, frequency, relationship to core/extended functionality, availability of a workaround, and possibly even cost to fix.
Factors affecting severity
Even within a limited definition of severity, there are multiple factors to be considered - specifically scope, magnitude, and duration of damage. Scope is a matter of what was affected: the whole system, the program, access to particular features. Magnitude is a matter of how complete the failure is: did the affected component(s) fail totally, did it hiccup and then continue, did it suffer a performance degradation, etc. Lastly, duration is simply a question of transient vs. permanent errors. When you've characterized the bug according to these three criteria it's pretty easy to use a point system to sum it up into a single severity value (though it never hurts to record the details). Likewise, once you know severity you can use a point system incorporating that and other factors into a priority value. Obviously the assignment of point values can be a complex and contentious process, but it only needs to be done once and once it's done the assignment of both priorities and severities becomes so much clearer that it's worth the effort.
Other Points Just some random thoughts about bug classification, in no particular order.
Re:Purpose of testing (Score:2)
I disagree. If there are few bugs found and fixed, that could be a sign that the testing is inadequate. I once worked on a project where the customer expressed a concern about our bug counts being too low. This seemed counterintuitive at first, until they explained that in their (very broad) experience this more often indicated something about testing than about the product. Sure enough, we stepped up testing, found more bugs, and went on to deliver a much higher quality product than if we had sat around patting ourselves on the back for our low bug counts.
What's more important than the absolute bug count is the *trend* in bugs reported. Every non-trivial project can be expected to exhibit an increasing rate of bug reports as QA ramps up and as components become ready for testing, followed by a plateau, followed - hopefully - by a steady decrease as the bugs get nailed without introducing new ones. If the plateau goes on too long, or if the extrapolated decrease at the tail and does not reach an acceptable level by shipment time, there's a problem. If the rate starts to go back up at that late a stage, there's an even bigger problem: your programmers are getting desperate, and checking in "fixes" that introduce more new bugs. If you don't see a "hump" at all, it probably means your testers aren't doing their jobs (or possibly not getting the help they need from developers).
All of these patterns are readily apparent if you know to look for them. I'd say you're right that the purpose of QA is statistical, but I'd say it's a statistical measure of *process* as well as product quality, and if the process is broken you can't know anything about the product.
Re:Shipping with Sev1 bugs (Score:3)
I can tell you about a (true) Sev 2 (or maybe sev 1, as the main part of the program didn't work) bug that was LOW priority
Way back when, HP made a card that went into you PC that accelerated your Laser Printer. What it did was offload any graphics processing from the printer to the PC - If I remember right, it was called a LaserDirect card
Well anyway, we found a bug after we shipped - The the client was running a PS2 Model 80, in a certain video mode, with this HP card, NO report printing worked, AT ALL. When you consider the main purpose of the app was to print these reports, it was fairly bad
The thing is, in the first 6 months the app was out there, we found ONE customer who hit that bug. We had a simple answer - we gave them their money back, and said, "Sorry"
Re:Well, none. (Score:1)
Re:Well, none. (Score:2)
If you can release software with bugs and still make more money than your competition, you've won. What I'm trying to say is, the motivation to fix bugs is to gain marketshare, not to "do the right thing".
All this because there's no standard software testing lab, and no software "building code" to comply to.
win2k is not very good (Score:1)
with a 'computer manager' thing and a fancy graph for your cpu time
Re:different strokes for different folks (Score:2)
This is theoretically true but practically false. Consider the equivalent statement about features: "There is absolutely no way to determine whether one efature is more important than another because user needs are different."
This may be true in the theoretical sense, in that you can never know for sure the complete needs of even one person, and aggregating the needs of thousands of users is even harder. But in practice designers and programmers build stuff every day that manages to more or less match the needs of their users.
And so it is with bugs.
Re:Severity/Priority (Score:2)
sev 2: feature broken, no workaround
sev 3: feature incomplete, or broken w/ workaround.
sev 4: cosmetic or UI flaw
This strikes me as pretty reasonable except for the "or UI" part. A bad or buggy UI can lead to user behaviors that destroy just as much data as a fandango on core [tuxedo.org] does.
One of my favorite examples of this comes from the NeXT 1.0 OS release, which was a slick GUI on top of a BSD-derived Unix. As appropriate in a GUI OS, you could click on a floppy (mounted in the filesystem, of course) and choose "format" from a menu to erase the disk (after a little "are you sure you want to erase this disk" messsage.
But you would get the identical warning if you tried to format any other partition in the file system, even your root partition. And it was awfully easy to mistakenly select the directory that the floppy was mounted in rather than the floppy itself. In one sense, this is "just" a UI flaw. But it's certainly not a just cosmetic one; it should be classified as severity 1.
GNU software makes this same distinction (Score:2)
The GNATS bug-tracking software (oh, go to www.gnu.org and look on their software list) does the same thing. Severity is how badly the software screws up. Priority is how soon it needs to be fixed.
Take a look at how GCC manages their bug reports for an example of this in use.
These are the categories that I use: (Score:2)
No category 1 or 2 defects should exist in released software and category 3 defects should only be permitted if the repairs will have a strong negative impact on the schedule. (I have never shipped a product with any but category 4 defects.)
The whole "valid" vs "invalid" inputs may confuse some people, so I should probably explain. I'm used to working in embedded systems where the system reads some measured values. Typically, those values will be represented to the system as a current, a voltage, or a contact closure. The "4 to 20" current loop is typical of the sort of thing I'm talking about. In that sort of system, the physical value is calculated from the current in the loop which varies from 4 mA to 20 mA. If the current is within that range, then the interpretation is easy. The question then becomes what happens if you get 0 mA or 22.5 mA? Neither of those is valid current as far as the system is concerned so it's not clear that the normal conversion to measured value should apply.
However, you will see such values in practice. The first is what happens when the wire gets cut, the second might be the result of an uncalibrated transmitter. A large part of the design of an automated control system has to do with figuring out what to do with those out of range values.
However, other sorts of programs get presented with invalid inputs, so the approach is not without merit in the more typical case. If, for example, you present a COBOL compiler with 6 MB of random data (as an acquaintance of mine once did) it shouldn't crash and should give you some kind of helpful error message. If it did the first, it's a category 1 defect, if it didn't do the second, then it's a category 3 defect.
Re:The Golden Rule... (Score:2)
It just seems there isn't always time to make code that I'd want sold to me, or that's bug free. I don't know if that makes it acceptable to release buggy software however. Seems a bit of a grey area.
Re:The Golden Rule... (Score:2)
Re:How do you get lower-order bugs addressed? (Score:5)
Seriously -- the intern gets a flavour of the company and corperate programming, and is capable of coming on board and seeing his contributions right away.
You get a cheap way to clean up the program. I believe one of the reasons why Microsoft is so popular is that they fix these kinds of bugs. Sure, you may get a BSOD every once in a while, but the program appears professional. Much like spelling and formatting errors in resumes are deadly, spelling and formatting errors in software reek of unprofessionalism.
For the cost of $12,000 (4 months intern salary) you could get tons of simple bug fixes knocked out, and give some poor undergrad tuition for his next semester or two. Everybody's a winner.
Re:Sayeth the Capability Maturity Model... (Score:2)
--
Lord Nimon
Re:Holistic Development vs. Software Engineering (Score:2)
I'd like to take small issue with this statement - I'd claim that software engineering, like any sort of creative process (and often like many forms of engineering), *IS* an art. If it weren't, you could take your specs & automate a process which could grind out the perfect solution w/o human intervention.
However, many so-called "artists" have a false idea of exactly how much discipline goes into a truly good piece of art. They believe that a "creative" person can bring forth the perfect solution for any problem full-blown from the murky depths of their unconscious minds. They point to tiny single-person projects that they've completed as "proof" of that software development model. Then they run into software projects where the details are too large to fit into one person's mind, and/or there is more than one person working on the project.
_THAT'S_ where the discipline and all of the software development "techniques" and methodologies are required - to help us mere humans keep the overall information flows cohesive enough to be useful. But after the problems have been broken down small enough for individuals to completely mentally assimilate, I believe that the results at those steps are still creative in nature, and can be classified as "art".
Simple really. (Score:2)
All the bugs that I submit are important. They are severity one.
All the bugs that you submit are useless. They are severity five.
The point? Get an objective standard, and don't take the submitter's priority too seriously. Otherwise everything will be "highest", at which point you're better off having no priority system at all- you'll spend time pretending to prioritize things and have no results.
Each bug has two axis (Score:5)
Holistic Development vs. Software Engineering (Score:5)
Logging bugs, and giving them "severities", is well and good. It can help in the same way that any effective software development tool helps, by enhancing communication. The moral is that bug tracking is useful only in the context of a team that can, and wishes to, use it effectively.
Wrong. Software engineering practices like tracking the severity and priority of bugs are aimed at creating higher quality software. Instead of developers spending time fixing bugs that are unimportant or not as annoying to the user, they have a way of categorizing and fixing bugs on a scale of importance.
Using bug count and severity as a measure of "releasability", however, is the fallacy of feeble-minded managers, who are afraid to make a decision without a number to back it up.
Yet people like you will bitch that Microsoft shouldn't release software with known severe bugs. I guess if they didn't track bugs they would be able to say they didn't know that there were any severe bugs with a clear conscience.
Software development (as practiced in all but epsilon of cases) is simply not a measurable process. You will only waste your time trying to quantify it.
Software development can be measured based on a large number of metrics. The fact that most people refuse to use software engineering practices that have been known for decades doesn't mean that the process is "immeasurable" it simply means our industry is full of people who'd rather subscribe to the myth that programming is an art instead of trying to make it an engineering discipline.
Releasability can only be determined by the judgement of the team working on the release (which may include developers, testers, release managers, beta testers, partners, etc). That is not to say that you should not draw upon the bug database for evidence upon which to base your judgement. But it requires intelligent interpretation, not counting up some totals.
So if the members of the team come up with the rules for what makes a bug a certain severity and what marks a bug as a certain priority exactly why can they not base releasability on these metrics? Quite frankly, waiting until the software "feels right" before releasing it is probably the most ridiculuos thing I've ever heard. On the other hand saying "we won't release until all the bugs that cause core dumps/segfaults/BSODs are fixed (severity one)" or that "we won't ship until we fix the annoying UI issues (priority one or two)" are quite reasonable even from a common sense point of view.
Some people consider this a failing of the software development process. I think they are too quick to condemn. The customer doesn't (usually) judge software by its bug count.
Interestingly most users of Windows 98 I know judged the software on its bug count (i.e. how many crashes needing reboots per day or other ridiculuos problems per day)
Most software is judged by an overall feel: if the software is compelling, many deficiencies will be overlooked. Further, it is difficult to guess ahead of time (even with beta testing) which bugs will really bite people and impede their use of the software.
If beta testers don't give you a feel for a large number of the bugs that will bite people on the ass then there are flaws in your beta-testing process.
Given the many interacting factors that determine the success of software, release decisions are naturally more art than science.
People like you are the major problem with the software industry. Those that believe that even though there have been increasing strides in the field of sooftware engineering, we should all still practice software development like it is a black art handed down from master to apprentice in a scene reminescent of guilds of the middle ages.
I know, I haven't presented any hard evidence. I'm arguing from experience in both free and closed software projects, and appealing to common sense. Most free software is released "when it's ready", without any metrics. Ponder on whether this is a strength or a weakness. And remember that when someone gives you number, the burden is on him to show that the number means something.
Do you really think Mozilla, GNOME, KDE, Apache, the Linux kernel, OpenBSD, etc are shipping "stable" releases of software without keeping an eye on bug severities and priorities? They may choose to ignore them for one reason or the other but they are keeping track.
--
Re:Our standards (Score:2)
As a Software Tester myself, the four catagories you list make sense, as any good tester can imediatly class a bug into that scale. With the 1-10 scale I currently have to use, the diference between a 3 and a 4/5 bug is infitismal. So, don't let anyone try and tell you more catagories are better, they must certainly arn't
Re:The Golden Rule... (Score:2)
Re:Severity one? (Score:2)
Beg to differ.
We've always used two classifications, severity and sensitivity, with the latter being the "eye of the beholder" measure. You can have a severity one bug ("nobody goes home until it's fixed") that a customer may never see, but if they do it'll be a disaster. That would be high severity, low sensitivity.
You can also have cosmetic problems that don't affect the overall operation of the system, but they still piss the customers off something fierce--these would be low severity, high sensitivity.
Keeping these distinct helps prevent the severity from correlating too closely with how loudly the engineers are being cursed. It also helps prioritze things differently based on where you are in the release cycle.
Severity - Priority (Score:2)
Severity rates the effect of the bug and runs from 1 to 4. 1 being a bug that makes the software unusable (usually a bug that crashes the program) and 4 being something like a spelling mistake.
Priority rates how likely the bug is to affect a user and also is rated from 1 to 4. 1 means the user will always be affected and 4 means almost never.
Using this system I could see how it was possible to leave a severity 1 bug in the sorfware if its low priority 4.
Software Engineering Institute (Score:2)
Why free software can often be better. (Score:3)
However this can change when the developers start getting more pressure to release by a certain date.
Cheerio,
Link.
Re:The (un)importance of "severity" (Score:2)
Of course the concept of a Free Software release is a little bit fuzzier than a commercial software release, as the software is generally available for use by anyone who's willing to download it well before the official release date. That means that the release is more generally a statement of confidence by the developers than anything else. As suce it's necessarily a bit murkier than a proprietary software release, as it's going to reflect a quantitative, rather than qualitative, difference in the number of users.
Our standards (Score:5)
1 - This is a drop-everything-to-fix-this-immediately kind of bug. A bug that makes it impossible to use the software at all.
2 - This level of bug is very serious and prevents some major part of the program from working. If we were doing a word processor, it might be cut-and-paste doesn't work.
3 - This level of bug displays a runtime error to the user but if they accept the error, the functionality is still there. Everything works, more or less.
4 - This is a spelling error or a size-of-textbox problem, something which the end user will notice but which obviously works properly.
Our standards say we are never to release with a level 1 or level 2 bug and, as much as possible, no level 3 bugs. Ideally, of course, we release with no bugs but that can be rather difficult.
Under no circumstances would we ever release with a level 1 bug because this simply means our app is broken.
--
The (un)importance of "severity" (Score:2)
Using bug count and severity as a measure of "releasability", however, is the fallacy of feeble-minded managers, who are afraid to make a decision without a number to back it up. Software development (as practiced in all but epsilon of cases) is simply not a measurable process. You will only waste your time trying to quantify it.
Releasability can only be determined by the judgement of the team working on the release (which may include developers, testers, release managers, beta testers, partners, etc). That is not to say that you should not draw upon the bug database for evidence upon which to base your judgement. But it requires intelligent interpretation, not counting up some totals.
Some people consider this a failing of the software development process. I think they are too quick to condemn. The customer doesn't (usually) judge software by its bug count. Most software is judged by an overall feel: if the software is compelling, many deficiencies will be overlooked. Further, it is difficult to guess ahead of time (even with beta testing) which bugs will really bite people and impede their use of the software. Given the many interacting factors that determine the success of software, release decisions are naturally more art than science.
I know, I haven't presented any hard evidence. I'm arguing from experience in both free and closed software projects, and appealing to common sense. Most free software is released "when it's ready", without any metrics. Ponder on whether this is a strength or a weakness. And remember that when someone gives you number, the burden is on him to show that the number means something.
Finally, before someone else points it out, I know that software can be developed with sufficient care and process that it is measurable. But I challenge you to demonstrate that it can be cost effective in today's mainstream software markets. Until this is shown, I don't think that such methodologies merit consideration.
Re:Holistic Development vs. Software Engineering (Score:3)
The excess of moderator points hasn't subsided.
Wrong. Software engineering practices like tracking the severity and priority of bugs are aimed at creating higher quality software. Instead of developers spending time fixing bugs that are unimportant or not as annoying to the user, they have a way of categorizing and fixing bugs on a scale of importance.
You're saying the same thing I said. Bug tracking helps communication. People know where the nasty problems are, and nothing falls through the cracks.
Yet people like you will bitch that Microsoft shouldn't release software with known severe bugs. I guess if they didn't track bugs they would be able to say they didn't know that there were any severe bugs with a clear conscience.
That must be the most common fallacy on Slashdot: that everyone here thinks the same thing. I myself understand very well that Microsoft's business requires them to release software with serious bugs. I also know that I have the choice of software developed under a better (for me, at least) model.
Programming is certainly not engineering. Why? The requirements are nowhere near well enough defined. And with the pace of the field, it's not worth anyone's time to define them to "engineering" precision. This is simply a fact at this stage of the game.
So if the members of the team come up with the rules for what makes a bug a certain severity and what marks a bug as a certain priority exactly why can they not base releasability on these metrics?
For the same reason that "the law is an ass", "metrics are an ass". Even if developed by reasonable people with good intentions. See below.
Quite frankly, waiting until the software "feels right" before releasing it is probably the most ridiculuos thing I've ever heard.
Where did you get the idea that I want to release software based on good vibes and burning entrails? Look, I'm saying that you use all the evidence that you have available. But you use it more in the manner of a historian arguing a thesis, not a mathematicial constructing a proof. These are different modes of thinking, but each is valid in its domain. Or do you think history is new age bullshit?
On the other hand saying "we won't release until all the bugs that cause core dumps/segfaults/BSODs are fixed (severity one)" or that "we won't ship until we fix the annoying UI issues (priority one or two)" are quite reasonable even from a common sense point of view.
Depends what you mean. If you mean, "We reviewed the bug db and the beta tester feedback, and everyone agrees that we can release when the four known crashes are fixed, barring unforseen new issues", then I'm with you. If you mean, "Welcome to the new project, team; we're going to ship when the feature-set is implemented and there are no sev 1 bugs", you're fooling yourself. There are just too many possibilities, and no bug database can objectively capture the complex reality of a software project. Any hope that metrics will guide you is (pardon the slang) just wanking.
Interestingly most users of Windows 98 I know judged the software on its bug count (i.e. how many crashes needing reboots per day or other ridiculuos problems per day)
Really? They don't care about the zillions of applications they can run? What proportion of Windows users quit using it because of the bugs? If it's not high, then bug count is probably not a significant factor in the success of Windows.
Do you really think Mozilla, GNOME, KDE, Apache, the Linux kernel, OpenBSD, etc are shipping "stable" releases of software without keeping an eye on bug severities and priorities?
The projects I follow most closely are Linux and Debian. Debian explicitly eschews specific release criteria based on hard experience. Linus has been known to release kernels that don't compile.
They may choose to ignore them for one reason or the other but they are keeping track.
Hear my point: bug databases are good because they assist in keeping track. Releasing based on a formula is bad. "Ignoring bugs for one reason or another" represents the practice I advocate: look at the evidence (including bug database), apply intelligence, draw conclusion.
No Start Up's, don't release then ;) (Score:2)
Even worse are the installers then immediatly exit with a Divide by 0 error.
How in the world did they NOT catch that one? Didn't they even try to install their own software first???
Ugh, Black and White has had some simular issues. So far it seems to only work with 75% of the machines out there. The infamious level 5 bug should have been fixed by now. . . . hehe, it's not even that hard to make a work around for it, if Lionhead just went ahead and released at least a minor patch for that one little bug.
At least the Windows installer starts up most of the time, at least Microsoft got that right (notice that I said MOST of the time, heh. . . .
I've always respected... (Score:4)
Seems like a fairly simple sceme to me -
Severity levels (Score:2)
Low - Small bugs. For example - incorrect labels for textboxes.
Medium - Allows you to complete your input operations but output results are incorrect.
Basically operations where data is not destroyed and recovery is possible by a simple fix to the program.
High - Program Crashes or security bugs.
Re:Our standards (Score:2)
Two answers. The Software Quality Assurance shouldn't just be covering crashes. The "misspelled company name" would be (at the company I used to work at) a "severity one" (we used the term "blocking") marketing error. The marketing department wouldn't sign off on release if such a bug was still in the system.
Your second point is covered through a process called triage. A meeting where the localization (one representitive), SQA (2), marketing (1), and engineering (2 representitives) (6 people total though someone from tech support, usually, sat in as we got near to ship to get an idea of what kind of things might still be in the program) met for 30 minutes each morning to assign a severity, priority and area (e.g. GUI (mispellings would have gone in here), loading files, installation, documentation, etc..) to each bug discovered in the previous day and assign the person responsible for fixing it.
The most senior engineering rep ran the meeting, and the most senior SQA person presented the bugs that had been discovered since the last meeting (or issues that had been tested and found to not be fixed.)
In my experience, SQA wanted to assign everything "severity one", engineering wanted to assign everything "severity four", marketing wanted to find the nearest coffee machine, and localization and documentation generally stayed quiet. Between the six of them, the process seemed to work pretty well.
One final note, the reason it worked pretty well was that this was a "high priority" meeting. Only the Head of Engineering (or the owner) could schedule a meeting that took priority over "triage".
sev bug definitions (Score:2)
Sev 1 bug:
This is a show stopper. You can't release software with this bug, the software just won't be usable.
Sev 2 bug: major functionality is broken. The software works in many ways, but you'd be embarrassed if you tried to present this (think win98 crashing on the day of its debut before hundreds- think of a billboard run by NT blue-screening in public.)
Sev 3 bug: can be handled by a workaround but really ought to be addressed. Can also be a documentation bug that needs to be fixed before docs can be shipped. (please don't ship docs after you've shipped the product.)
Sev 4 bug: works most of the time, but only under specific conditions. showed up in testing, can be addressed by a technote on the website.
Sev 5 bug: works most of the time, under most conditions, showed up in testing, can be addressed by a technote.
Any bug addressed by a technote must be fixed in the next release, or else. Any bug sev 1 or sev 2 bug must be fixed or you don't release. Sev three is one of those that you fix at the end of the release cycle, or handle with a technote and take care of it in a service release or next full release.
A host is a host from coast to coast, but no one uses a host that's close
Categories (Score:2)
I am sure there are plenty of QA sites out there, but I don't have the master page of links right at hand.
Check out the Vinny the Vampire [eplugz.com] comic strip
Re:Sayeth the Capability Maturity Model... (Score:4)
If bug X triggers a BSOD when I'm surfing the Web, it is a level 4 or 5.
"
It has utterly stopped you surfing the web, it's catagory 1.
FP.
--
There is testing, and there is testing (Score:4)
(I was a developer who moved into the testing area because my employers and peers found I had a knack for exposing weaknesses in a product. So let me draw on both sets of experiences to provide a "developer-oriented" answer to this question.)
The first goal of DVT/QA testing is to ensure that the product functions as the functional specification and documentation says it should.
That means DVT first tests each operation, each feature, and each option on each environment with a variety of recognized and accepted test data sets. Not only are DVT testers looking for proper operation with proper data, but also proper operation with bad data. In other words, do the error-catching routines actually catch errors? If there is a sequencing error, is it reported properly? Recovered from properly?In many DVT environments, the test conditions are required, to the best of the DVT system's ability, to test all of the code, using profiling and test coverage tools to indicate when each path has been checked.
Like development, design verification and test (DVT) and product quality assurance have as many levels of abstraction as your typical development effort. Properly structured testing looks at the major stuff first, options next, and then they go for the anomolies. Indeed, in the DVT efforts I have managed, the checkout of a software change required the "20-minute test", the "3-hour test", and the "24-hour test" before I would approve the change to be part of a release candidate subject to full regression testing. (From project to project, the actual length of time taken for each of these tests differed, but the names stuck.)
The second goal of DVT/QA testing is to find operational problems in-house before the customer finds them in the field.
This is the time when testers "bang on the product" to determine if there are any weaknesses or coupling effects. The best testing is when you have a good number of alpha testers that use the product in the manner it was intended to be used, followed by a good-sized group of beta testers. Beta test management is an art unto itself, with the most important function being the selection of the beta testers themselves. Some companies miss the boat by only soliciting one type of beta tester. The best beta programs court fringe users, those who would use the product in unusual ways, in order to ferret out more error conditions.Frankly, I find development easier to do than testing, but testing is to me the more satisfying activity...especially when my efforts expose some really funny cascades.
There is a third aspect to testing that doesn't get much play in companies: reproducing field problems. I have some war stories if someone serves up the beer...
Management v QA/DVT (Score:5)
This is probably a good time in this thread to throw out an allied problem: what happens when management doesn't like the results of the bug evaluation.
True story: a corporate-wide standard for bug determination was defined in a Fortune-50 corporation. In a project, though, the director of engineering decided that the standards "didn't reflect reality" -- or rather, that the numbers made him look bad. It also meant that there was no way he would be able to ship the product on time, because the number of ship-block bugs was just too high for his development staff to clear in a manner acceptible to him and to his superiors.
Now, the definition was pretty strict: Severity 1 bugs were features in the specification that were implemented and broken in significant ways, severity 2 were features in the specification that were implemented and met the operational requirements but not performance requirements, or were not implemented and marked in the specification as required for release, severity 3 were features that were implemented but broken in non-significant ways (cosmetics, usually) or features in the spec that were not implemented and marked as not required for current release, severity 4 were features that met spec but were identified as areas for improvement, and severity 5 were spec change requests. The rule is that the product MUST NOT ship with severity 1 bugs, SHOULD NOT ship with severity 2 bugs (release required a review and analysis of each S2 bug by corporate QA), MAY ship with severity 3 bugs, and MUST ship with priority 4 and 5 bugs.
So, Job One for this manager was to pull a Captain Kirk and change the rules. Severity 1 was redefined as "causing a system crash". Severity 2 was redefined to "causing an operation failure that does not result in a system crash". Severity 3 was everything else that used to be in S1 and S2, as well as what was originall in S3. That meant that missing or severely broken features identified as required for release obtained an S3 rating instead of the original S1 or S2 rating. Or, in other words, a broken product could now be shipped!
Of course, it didn't do very well in the market, but the manager in question still got his raise -- and I suspect that's all he cared about.
The root cause of this breakdown was that QA/DVT was an Engineering function, not a separate function. The guy saying "no" should be a peer to the guy saying "yes". (It would have helped if the original bug definitions would have been part of the specification, because the spec is king within this company -- but that didn't happen.)
Re:How do you get lower-order bugs addressed? (Score:5)
Have you tried social engineering?
Arrange with your management to set aside a day for a "Little Stuff Contest." It works like this:
Make this a regular thing, and you will be surprised how quickly the nuisence bugs get cleared. Moreover, I can tell you from personal experience that in the process of clearing "the small stuff" some big stuff will be found and fixed as well.
Motivation is a beautiful thing.
The Golden Rule... (Score:5)
...would seem to apply.
"In everything do to others as you would have them do to you..." -- Matthew 7:12, NRS TranslationA generally good philosophy, regardless of your adherence to ceratin world religions.
When releasing software, I ask myself: Would I be pleased if somebody gave/sold this program to me? It isn't so much a matter of numbering conventions of bug severities -- it's a matter of pride and responsibility.
--
Scott Robert Ladd
Master of Complexity
Destroyer of Order and Chaos
Motivation for fixing bugs? (Score:4)
Fixing bugs has a cost associated with it (e.g. QA employee's salaries, test beds, test software, time required to perform tests.)
NOT fixing bugs has a cost associated with it (tech support staff, development staff's time and resources required to fix the bugs, testing the fixes, the cost of packaging and shipping fixes, loss of repeat sales, bad press.)
The question for a commercial software vendor, then, is this:
NOTE: How the costs are perceived seems to depend somewhat on the size of the operation and the nature of the product. Here are some examples from my experience:
I view testing and bug severity/priority/frequency assessments as a means to help management make an informed decision as to when software is ready to ship.
It is their responsibility to know the market, to know the business decisions as to when fixing bugs costs more than leaving them in. It is my job to use the best practices and tools that I know how to help provide them the information they need. In short: their expertise is in the business side; mine (as a QA professional) is in the technical side.
Lastly, 1*1000 = 1000*1; one nuclear bomb can ruin your whole day, but getting bit by a thousand mosquitos can ruin it, too... don't ignore the minor bugs! (Besides, what may appear to be cosmetic bugs can make it incredibly difficult to develop and maintain automated tests that depend on field prompts and error messages!)
different strokes for different folks (Score:2)
Software Kills People (Score:3)
An OS that crashes on a PC at home may not be that big of a deal to you, but if that OS is installed on a machine that's powering a space station and the lives of the crew depend on it, then you'll probably be a little more concerned about the bugs in that environment.
Despite the popular belief, it is possible to write software that is completely free of bugs (just look at the space shuttle launching code). It is a lot more difficult to do that, sure, but nevertheless quite doable.
Re:Well, none. (Score:3)
That's a bit of an oversimplification (unless you are making the tautological argument that a Sev1 bug is one that makes the software unreleasable - a reasonable position but not much help in deciding when the software is releasable).
The essential question is: is the software, with the bug, still more valuable than no software at all? If so, it's better to release it than not. I can deliver software with zero defects in infinite time and at infinite cost. Most users are willing to accept some non-zero defect rate in exchange for non-infinite cost and delivery sooner rather than later - and this is a value tradeoff that the user must make.
Among the factors to consider:
Bottom line: Listen to your customers, not engineering religion.
Re:Software Kills People (Score:4)
Indeed it is. In 1967(!) Boeing executives were lecturing IBM executives about failure rates; a 707 or 727 is no less complex than a mainframe of that era, and was many orders of magnitude more reliable - because the failures killed people.
The key point is that "a lot more difficult" is generally equivalent to "a lot more expensive". When the cost of reducing the defect rate exceeds the cost of living with the defect rate, it becomes rational to stop working on further reduction of the defect rate and to start working on adding new functionality, improving performance, or otherwise adding value on some other axis.
That crossover point is reached at one point with MS Office ("damn - there goes a few minutes of work"), another point with an enterprise ERP system ("damn - three minutes of downtime at $3E6/minute"), and yet another (near-extreme) point with the space shuttle launching code ("damn - six people dead, $3E9 up in smoke, one irreplaceable shuttle gone").
We should not expect people to be willing to pay more for less buggy software (if they did, it would be prima facie evidence that the vendor has underestimated the value of the software to the user). What may happen is that, as happened with automobiles during the 1980's and 1990's, some vendors will figure out how to improve quality while holding costs constant; these vendors will raise the quality bar across the board. They won't succeed by mainstreaming the techniques used in designing mission-critical software, any more than Honda or Toyota succeeded by mainstreaming techniques used by Rolls-Royce - they'll succeed by driving the craftsmanship out of the engineering and replacing it with a soulless but repeatable development process.
Re:Severity one? (Score:3)
alpha Not all the features / business cases are part of the code. The developers are still adding them.. The code may be unstable, that is crashes for no reason. May contain bugs that trash files. This level should never reach testers who cannot recover their hard drives.
beta All features coded. The application runs OK. Outside testers (outside the development group) are asked to test it. After a while a public beta may be used where people outside the company may try it.
This is the traditional way of separating the two. Of course recently some companies have passed off alpha level as beta, the beta level as release. But that's another story
-----------
Re:Holistic Development vs. Software Engineering (Score:2)
Notice that there seems to be a huge number of mod points floating around? (I just got another 5 before I finished with the last batch.) Notice that just about everything is flying up to +5 and then back down?
Apparently Michael wasn't kidding [slashdot.org]:
Due to an as-yet-unexplained error in slashdot's database, over the past few days fewer moderator points than usual have been given out. We've fixed it, but will probably have to restart the webservers for changes to take effect - probably a lot of people will suddenly receive moderator points in the next hour or two.
Questions:
I'll be gracious and not use one of my other accounts to mod this up. ;-)
Unsettling MOTD at my ISP.
Re:Holistic Development vs. Software Engineering (Score:2)
I don't know about that but it *did* come from his account. And distribution of mod points *has* gone berserk.
Unsettling MOTD at my ISP.
Bug classification (Score:3)
1."Show Stopper." The client absolutely refuses to sign-off on the project until this issue is solved.
2. "Mickey Mouse." Very minor error that can usually take a minute or two to fix, but wastes more of your time in the error reporting, documentation, and follow-ups.
3. "Ghost." Only the client sees it but can't replicate it or give you instructions so you can do that. The error is real (the client will send you a screen capture), but they can't helo you in recreating it.
Pedro
Re:Sayeth the Capability Maturity Model... (Score:2)
BSOD is not a bug
COOL!
Hello?...Boss, what? Our customers got BSOD?! Oh GREAT! No...sorry....I mean, it's not a bug.......wait a minute, my ear hurt....ya know...Any number of problems, from severity 5 to severity 1 all trigger BSOD......you don't understand?....well, I mean it's probably caused by something else, not ours!!
Programmers' Paradise....I love Winodws....
Sorry Jerf, just kidding. You are probably right. ^_^
Re:Sayeth the Capability Maturity Model... (Score:4)
Any other questions?...
What Severity is BSOD [slashdot.org]?
Wait...don't mod me troll so fast.....
If it happened on my desktop while I'm doing nothing just staring at the monitor aimlessly, it probably falls into Severity 4 or lower....
But if it happened here [cnn.com], then personnel safety would be jeopardized, definitely Severity 1.
Say, a BSOD just killed your project, your boss will probably kill you tomorrow, your personal safety is in danger - Severity 1. In the meanwhile, buddies in Microsoft believe a BSOD can be solved by magic reboot, clean and clear - Severity 5 or lower, thus carrying the same bugs over and over again in every version of the Windows.
At any rate, the severity of a problem cannot be determined by the "bug" itself, it's also determined by where the situation applied. I think that may be the reason why people have not come with a standard on bugs' severity.
Windows Error Codes (Score:4)
WinErr: 001 Windows loaded - System in danger
WinErr: 002 No Error - Yet
WinErr: 003 Dynamic linking error - Your mistake is now in every file
WinErr: 004 Erroneous error - Nothing is wrong
WinErr: 005 Multitasking attempted - System confused
WinErr: 006 Malicious error - Desqview found on drive
WinErr: 007 System price error - Inadequate money spent on hardware
WinErr: 008 Broken window - Watch out for glass fragments
WinErr: 009 Horrible bug encountered - God knows what has happened
WinErr: 00A Promotional literature overflow - Mailbox full
WinErr: 00B Inadequate disk space - Free at least 50MB
WinErr: 00C Memory hog error - More Ram needed. More! More! More!
WinErr: 00D Window closed - Do not look outside
WinErr: 00E Window open - Do not look inside
WinErr: 00F Unexplained error - Please tell us how this happened
WinErr: 010 Reserved for future mistakes by our developers
WinErr: 013 Unexpected error - Huh ?
WinErr: 014 Keyboard locked - Try anything you can think of.
WinErr: 018 Unrecoverable error - System has been destroyed. Buy a new one. Old Windows licence is not valid anymore.
WinErr: 019 User error - Not our fault. Is Not! Is Not!
WinErr: 01A Operating system overwritten - Please reinstall all your software. We are terribly sorry.
WinErr: 01B Illegal error - You are not allowed to get this error. Next time you will get a penalty for that.
WinErr: 01C Uncertainty error - Uncertainty may be inadequate.
WinErr: 01D System crash - We are unable to figure out our own code.
WinErr: 01E Timing error - Please wait. And wait. And wait. And wait.
WinErr: 020 Error recording error codes - Additional errors will be lost.
WinErr: 042 Virus error - A virus has been activated in a dos-box. The virus, however, requires Windows. All tasks will automatically be closed and the virus will be activated again.
WinErr: 079 Mouse not found - A mouse driver has not been installed. Please click the left mouse button to continue.
WinErr: 103 Error buffer overflow - Too many errors encountered. Additional errors may not be displayed or recorded.
WinErr: 678 This will end your Windows session. Do you want to play another game?
WinErr: 683 Time out error - Operator fell asleep while waiting for the system to complete boot procedure.
WinErr: 625 Working Error - The system has been working perfectly for the past ten minutes
WinErr: 902 Screen Error - The system is working perfectly, I'm not lying, your monitor is wrong
WinErr: 72a Crucial Halt - Hang on, WHAT was that?
WinErr: 72b Memory Error - What? Tell me again.
WinErr: 39c Disk Error in drive a: - Although your disk is in perfect condition (I just formatted it), I don't like it any more
WinErr: 983 Hard Disk Error - The files on the hard disk were neatly arranged and fully optimised, so I had to mess them up and put a couple of hundred
WinErr: 294 BlackMail Error - $500 to Gates or your computer gets screwed
Winerr: 294b BlackMail Error - $500 to Gates or I'll show your wife the JPG's you just downloaded
vroom vroom [speedygrl.com]
Re:Our standards (Score:3)
1-Critical -- A bug that causes the whole application to crash and/or become unuseable.
2-Major -- Not a fatal bug, but we cannot ship the software with this bug in the code.
3-Average -- Needs fixed before program release, but does not cause error boxes or incorrect operation as designed. (Products have shipped with level 3 bugs before, but usually only if they are found one or two days before release and they cannot be fixed and tested prior to release).
4-Low -- Needs fixed sometime in the future (preferrably by next release), but product can ship regardless of fixing this error.
5-Enhancement -- Nice to have for the next software release. (Almost never gets actually fixed, but at least it's documented).
Having said that, we also have a classification system for how soon a bug needs to be fixed (i.e. - 24hrs, 48hrs, 3 days, etc.) with 5 levels of classification. So who wants to compile this list and provide it to slashdotters so that we can go back to our managers and make our case for why we should all stick with a set standard like other companies out there?
Re:Japanese Standards (Score:2)
When the components arrived from Japan, a note was enclosed saying: "We don't know why you want the faulty components, but for your convenience they have been packed separately."
Re:Severity one? (Score:2)
Alpha - it compiles!!!
Beta1 - still adding features.
Beta2 - nearly everything is in place.
Beta3 - feathure freeze.
Release - now they start hunting those bugs.
Re:The Golden Rule... (Score:4)
It's actually not all-purpose a philosphy as one might think... what people *like* varies enormously. The inverse holds up a bit better as a general-purpose philosphy... "Don't do unto others as you wouldn't want them to do to you."
In any case, that philosophy doesn't work very well when applied to software, since the range of what people want varies enormously... some (the "Microsoft people") want everything there now, want it easy to maintain and setup just by clicking a few dialogs, don't want to fiddle with config files or memorize commands, and are willing to put up bugs in order to get those features. Others (the "UNIX people") are willing to fiddle with configurations, memorize commands, and sometimes wait for features in order to feel they're bug-free. And the range of taste in between those two options is enormous...
Best example of that is probably game support, and OS and beta driver quirks people are willing to put up with for a few more frames per second.
--
Convictions are more dangerous enemies of truth than lies.
Pain versus Likelihood (Score:2)
One thing we do, in particular when evaluating which bugs to fix in the next release, is look at pain (severity in other posts), what problem this defect is causing to the users, and also look at the number of users having this problem (in the case of a bug that is found before it is shipped, this would be the likelihood that the defect will occur in the field).
We have shipped a product that bluescreened under somewhat obscure conditions. I think of our 10,000 customers, one managed to get it to happen. We knew about it beforehand, the dev and QA staff had a fit, but we shipped anyway (of course, two of the twenty developers is still on that project anymore, but that's a different story ;)
Some defects that seemed trivial, like stupid cosmetic errors in reporting, were fixed, because every single user saw them and complained about them (well not every single user)
Of course, we did try to track down the blue screen, but we never did. According to MS, blue screens (on NT/w2k) are never the fault of the app running in user mode anyway, so we were able to shift the blame to MS. The point is, another factor in what you fix is whether or not it is fixable by your dev staff (how long you are willing to spend trying to fix it before you give up).
So the three big factors are: How severe is the defect, how many people see it or are likely to see it, and how long do we think it will take (or how long have we spent trying) to fix it.
Personally I like the "pride" idea, but it doesn't sit well with management/sales. Where you see this is in private companies like id software, that doesn't have to maintain earnings to the street will keep their stock at a reasonable level, etc. In a public company, management needs to cut their losses and get product out the door - they can always fix bugs in the next version, but they can't make revenue from the current version until they ship it.
An interesting dilemma to be sure, especially when dealing with four or five year old applications that nobody really understands anymore...
PhilTheChill
The Rules... (Score:5)
1. Software can not harm a user or through any inaction allow a user to come to harm
2. Software must obey orders give by said user unless it would conflict with first law
3. Software must preserve itself and OS unless conflicting with first or second law
Adapted from the Genius A.Asimov
Re:Severity one? (Score:3)
Bug Severities and Priorities (Score:5)
The SEVERITY of the bug is entered when it is discovered. It is one of:
1 - Urgent, causes crash
2 - High - no workaround
3 - medium - workaround available
4 - low - inconvenience.
Other information, such as the steps to reproduce, who discovered it, , a unique ID, and date and version discovered are also recorded. Most important is the TYPE of bug, which is one of:
Code Problem
Configuration Problem
Data Problem
Design ISsue
Enhancement
UI Problem
Documentation Issue
Bugs are assigned to a developer to be fixed. If they are easily fixed, then they are immediately fixed. The unique ID is used in the CVS checkin for the changed files. If the bug is NOT easily fixed, then the developer gives a level of effort, so we can determine the 'cost' of fixing the bug.
Bugs are then assigned a PRIORITY in a 'Configuration Control Board' (CCB) meeting with me (the project manager), several developers, documentation writers, and/or our end users/customers/testers. This can be a small, informal meeting, or a large process-driven meeting. I prefere them small. The PRIORITY in one of:
1 - Urgent, resolve immediately
2 - High, resolve ASAP
3 - Medium, resolve pre-release
4 - Low, desired, but not urgent
5 - on hold, fix in future release.
Our customer has deemed that all severity 1 bugs are automatically priority 1, and all severity 2 are pri 2.
Severity 3 and below are given priorities based on the cost to fix, but if we have too many severity 3's in one 'use case', they collectively become a pri 2.
It's not nearly as much process as it seems, and it really streamlines our development.
We use a tool called TestTrack to track all this, but I would prefer to move towards Bugzilla, as my company has recently jumped on the open source bandwagon.
Purpose of testing (Score:5)
As a developer, my relationship with our testing department has changed over the years...
Originally, I thought the purpose of testing was to prove there were no bugs in the system.
They ALWAYS discover bugs. I began to think that the purpose of testing was to discover ALL the bugs, and it was the engineer's job to fix them.
They NEVER discover ALL the bugs. I now think that the purpose of testing is statistical... How many bugs are they finding, and what severity are they? This speaks to the quality of the system design and the quality of the engineers working on the system. If there are few bugs found and fixed, you can be confident you have a good system going out. If there are a lot of bugs found and fixed, I would worry that there are a lot more left undiscovered.
It is a common practice to ship with known bugs, but we NEVER ship with a pri/sev 1 or 2, unless a political deadline forces it.
Sounds to me like the company whose product you are using found more bugs than they could reasonably fix in the time they allocated to this phase of their development cycle, and now they are being forced, for political reasons, to ship with those bugs. That would be a red flag to me that there will be other, undiscovered, severe bugs in their software.
Severity/Priority (Score:4)
Severity has to do with how badly the bug affects users:
sev 1: crashing bug, or data corruption
sev 2: feature broken, no workaround
sev 3: feature incomplete, or broken w/ workaround.
sev 4: cosmetic or UI flaw
Priority has to do with how important fixing the bug is to shipping the product. There are no names for the priority levels, but in conjunction with severity levels you can make determinations on what to fix, what to cut, and in what order.
Using this system, a sev1/pri1 bug is clearly the first bug I'd fix. A sev1/pri1 bug gets fixed before a sev2/pri1, and very rarely, a sev2/pri1 bug will get fixed before a sev1/pri2. These last two are the corner cases: sev2/pri1 is typically a very ugly sev2 bug, and a common reason for marking a bug with this classification is to fix a beta or demo issue. A sev1/pri2 bug is also rare - it means that the product crashes or corrupts data after a very obscure repro case.
I've been using this system for years, and I have yet to find a clearer method.
Invisible Agent
Re:Depends on how the "features" are accessed (Score:2)
People who glibly talk about shipping major software bug free, have either no experience of major software projects, are liars, or tend to creatively reclassify and ignore bugs so that they become 'documented issues' or 'known limitations' or 'under continued development' etc etc. It's particularly rich coming from the open source movement, who ship over and over continuously (presumably without bugs in any release...).
Some Testing Resources (Score:3)