2008/11/30

Finance, FMAA & ANAO and Good Management: Never any excuse for repeating known errors

In light of the Sir Peter Gershon's Review of the Australian Government’s use of Information and Communication Technology, here's an email I sent to Lindsay Tanner (Finance Minister) prior to the 24-Nov-07 election of the Rudd ALP government. Edited lightly, formatting only.

Date: Sun, 11 Nov 2007 15:02:40 +1100
From: steve jenkin 
To:  lindsay.tanner.mp@aph.gov.au
Subject: Finance, FMAA & ANAO - Good Management: Never any excuse for repeating known errors

Here is something very powerful, but simple to implement & run, to amplify your proposed review of government operations and can be used to gain a real advantage over the conservative parties. On 8-Nov I wrote a version via the ALP website.


Headline:
The Libs talk about being Good Managers, but they have been asleep at the wheel for the last 10+ years.

It's not "efficient, effective or ethical" to allow public money to be wasted by repeating known mistakes.

Nothing new needs to be enacted - only the political will to demand Good Governance from bureaucrats and the 'ticker' to follow through.


2008/11/29

Gershon Report - Review of Australian FedGovt ICT

The Gershon Review is good solid stuff that doesn't rock the boat, doesn't challenge current methods & thinking, nor show deep understanding of the field.

It has a major omission - it addresses ICT inputs only.
ICT is useful only in what it enables others to do or improve - measuring & improving ICT outputs is completely missing from 'Gershon'.

It doesn't examine the fundamentals of ICT work:
  • What is that we do?
    How is Computing/IT special or different to anything else?

  • Why do we do it?
    Who benefits from our outputs and How?
Here are my partial answers to these questions:
  1. Computing is a "Cognitive Amplifier" allowing tasks to be done {Cheaper, Better, Quicker, More/Bigger}.

  2. IT is done for a Business Benefit.
    Like Marketing, defining how outputs & outcomes are measured and assessed - both in the macro and micro - is one of the most important initial tasks.

Gershon doesn't address outstanding issues of the IT Profession:
  • improving individual, organisational and general professional competence and performance.
  • Reducing preventable failures, incompetence/ignorance and under-performance.
  • Deliberate, directed & focussed effort is required to institute and maintain real Improvement of the Profession. (vs 'profession-al improvement' of practitioners)
After ~60 years of Commercial Computing:
  • Are there any new ways to stuff things up?
  • Is it "efficient, effective, ethical" to allow known Errors, Mistakes, Failures to recur without consequences? [see FMAA s44]
It isn't like the Government isn't aware of the processes and instruments needed to avoid repeating Known Errors, nor the benefits of doing so.

Aviation is controlled by ATSB (Australian Transport Safety Bureau, previously Bureau of Air Safety Investigation [BASI]) and CASA (Civil Aviation Safety Authority). The USA's FAI publishes hard data on all aspects of Aviation - and mostly they improve on every measure every year. This isn't just due to the march of technology - the figures for 'General Aviation' (as opposed to Regular Passenger Transport) plateaued decades ago... This is solid evidence that Aviation as a Profession takes itself seriously - and that commercial operators in one of the most competitive and cut-throat industries understand the commercial imperative of reducing Known Errors.

Aviation shows that profession wide attention to Learning and Improvement isn't just about Soft benefits, but translates into solid business fundamentals. You make more money if you don't repeat Know Errors/Mistakes.

ATSB investigates incidents and looks for Root Causes.
CASA takes these reports and turns them into enforceable guidelines - with direct penalties for individuals, groups and organisations. CASA is also responsible for the continual testing and certification of all licensed persons - pilots, Aircraft Engineers, ...

There are 4 specific areas Gershon could've included to cause real change in the IT Profession - to start the inculturation of Learning & Improvement and the flow-on business gains.
Federal Government accounts for 20% of total Australian IT expenditure. It is the single largest user and purchaser of IT - and uniquely positioned to redefine and change the entire IT profession in Australia.
  • Lessons Learned - Root Cause Analysis of Failures/Problems
    Dept. Finance 'Gateway Review Process' on Projects.
    Needs equivalent of CASA - inspection and enforcement of standards plus penalties/sanctions - Not just reviews and suggested guidelines.
    Not just ICT staff, not just FedGovt but their suppliers/vendors/contractors as well.
    Without real & timely (personal and organisational) consequences, nothing changes.

  • Standish 'Chaos Report' equivalent - real stats on IT Projects.
    Without solid numbers, nothing can change.

  • Operational Reviews.
    How well does an IT organisation do its work?
    Critical Self-assessment isn't possible - exactly the reason work needs to be cross-checked for errors/mistakes/omissions/defects.
    C.f. Military Operational Readiness Reviews - done by specialist, impartial experts.

  • Individual Capability Assessment - equivalent of on-going Pilot etc recertification.

  • Research: Quantifying & standardising metrics and models for "Effectiveness".
    DCITA/DBCDE on macro-economic results.


The ACS describes Gerhon's recommendations as "all aimed at addressing the efficiency of ICT":
  • governance,
  • capability,
  • ICT spending,
  • skills,
  • data centres
  • sustainable ICT
Note the issue of Reducing Faults/Failures/Errors/Mistakes doesn't make the list.
Nor does the idea of institutionalising the building/improving the Profession of IT and increasing the Capability/Performance of IT Professionals.

By the DCITA/DBCDE own reports, ICT contributes 75% of productivity improvements: ICT is still the single greatest point of leverage for organisations reducing costs and improving output.

Does getting IT right in Federal Government matter?
Absolutely.

Gershon delivers 'more of the same' and could conceivably achieve its targets of 5% & 10% cost improvement

2008/07/15

Bad Science or Science Done Badly?

Is 'Science', as practiced by Academic Researchers, executed poorly?

More specifically:
Is the practice of Research as undertaken by Academics, as effective as it could be?

This posits that an aspect of "Professional Research" is intentionally increasing your capability and effectiveness.

Computing/Information Technology is a Cognitive Amplifier - exactly suited to central parts of "Professional Research" - e.g. learning, recalling and searching published papers and books.

If an individual researcher can increase their "knowledge uptake" just 7% in a year, after a decade they know twice as much, given uptake builds on existing knowledge.

What is Research about if not Knowledge: Gathering, Analysis, Representation, Taxonomy/Ontology, Management and Communication?
This field began in 1995 and is broadly known as "Knowledge Management".

2008/05/28

I.T. Strategic Planning Failures

Sue Bushell asked on "LinkedIn": What are the most common failures in strategic IT planning and how are these best avoided? What best practices in strategic planning are most effective?

My answer:

1. There are no I.T. projects - only Business Projects.
Hence changing the premise of your question:
What are the most common business process failures around I.T. solutions?
[A: Make the business run the project and take the rap if it fails.]

2. I.T. is an Industry, not a Profession.
Proof: Professions Learn: repeating Known and avoidable Errors/Mistakes isn't consequence free, as it is within I.T.

3. The complete lack of History in I.T. - both on macro and micro scales.
  • Show me any large organisation that can even list all its current projects, which is a necessary starting point for:

  • Formal "Lessons Learned" from projects and operations - known problems are avoided, known effective practices are used.

  • Jerry Weinberg wrote definitive works on Software Quality Management and 35 years ago proved that focusing on Quality results in better code, written far faster & cheaper. And it is much more reliably and consistently produced!

  • Jim Johnson of Standish Group, nearly 15 years ago started definitive research on what proportion of IT Business Projects fail and the causes of failure. This work is fundamental to advancing the Profession - but nobody else studies this field so his results can't be verified or refuted. Nor have organisations or practitioners, by-and-large, acted on this knowledge. People do argue that his results are suspect because other single-shot reports don't agree. But nothing happens to resolve this fundamental issue!

  • Software ReUse is notable in how little it is practiced. Can it be possible that nearly ever problem is completely new? Not in my experience.

4. The fundamental reason IT is used: It's a "cognitive amplifier".
Computing amplifies the effort and output of people, providing results 'Cheaper, Better, Faster'.

On the micro scale, no organisation I've heard of measures this. It's quantitative and should be calculable by any half-reasonable Management Accountant.

On the macro scale, the 'Profession' doesn't have or publish benchmarks on results (i.e. from across many organisations).

5. The 'Profession' doesn't even have a taxonomy of jobs and tasks, let alone any consistent method for evaluating and reporting the competence of, and skill level of, practitioners.
  • In a construction project you wouldn't specify "10 vehicles needed", you say "6 5-tonne trucks, 2 utes, a 20-tonne tip-truck and a bobcat".

  • If the profession can't distinguish between the speciality, competence and skill levels of its practitioners, how can the business folk?

  • If project plans don't identify the necessary the precise skills needed - implying some way to assess and rate the 'degree of difficulty' of individual tasks/components - then the right 'resources' can't be applied.

6. The almost complete disconnect between research results and practice. Enough said.

7. [Added]. The general capability of the Profession in general and young I.T. practitioners has declined greatly.
Proof: The increasing number of failed projects attempting to replace 'Legacy Systems'.

E.g. The failed A$200M Federal Government ADCNET project. I worked on the original IBM mainframe system, then found myself 15 years later sitting in the same awful basement not 50 feet away, coding it's replacement. The IBM system took 30-35 man-years (in structured assembler), just the second phase of the ADCNET system had a team of 70 for 1-2 years - and was abandoned. The best description of it is the Federal Court Judgment:
GEC Marconi Systems Pty Limited v BHP Information Technology Pty Limited
Federal Court of Australia
12 February 2003 and 14 July 2003
[2003] FCA 50; [2003] FCA 688

8. [Added] Creating Software is a performance discipline.
You have to both know the theory and be able to create good software.
Who are the Great Heros of Open Source? The guys that demonstrate they can code well.

Like Music, Surgery and Architecture, software requires head and hands to do it well.


9. [Added] Design is Everything.
This is what the Bell Labs Computing Research guys understood and what Microsoft doesn't. They invented the most cloned Operating System in the world - Unix, and then went onto build Plan 9, it's replacement 20 years later - with around 20 man-years. It was created portable and scalable, running on 6 different platforms from day 1. Of course it was incredibly small and blindingly fast. Time has shown it was robust and secure as well.

Not an accident that 15 years later Microsoft spent around 25,000 man-years on 'Longhorn', and then threw it all away! (The infamous 'Longhorn Reset' on 23-Sept-2005 by Jim Allchin)
Then spent the same again to create 'Vista' afresh from the 'Windows Server 2003' codebase.

How could Microsoft not understand what was well known 15 years prior, especially as Microsoft ported Unix to Intel in 1985?


There's more, but that will do for now.


"I.T. Governance" may be part of the Solution, but standards like AS8015 are primarily aimed at allocating blame or pushing all responsibility for failure onto I.T. and abnegating from I.T. any successes.

The 'root cause' of all I.T. failures is trivial to identify, but probably exceedingly hard to fix. These days, almost no projects should fail due to technology limitations - only practitioner and management failures.

The 'root cause' is: Business Management.

Yes, there are many problems with I.T. practitioners, but think about it...

Around 1950, Commercial Computing was born.
Some projects worked, in fact succeeded brilliantly: Man went to the moon on the back of that work just 2 decades later.

And then we have the majority or 'ordinary' projects that fail to deliver, are abandoned or under-deliver...

The first time 'management' commissioned a bunch of 'Bright Young Things' to build The Very Best Computer System Ever, they would naturally believe the nerds and their self-confidence.

After that effort failed, what would the rational approach be to the next project?

Not the usual, "do whatever you want and we'll see", but "you didn't do so well last time, how about we try smaller pieces or doing it differently?"

And when lining up for the third go-round, you'd think competent business managers (the ones writing the cheques) would put the brakes on and say "you haven't shown you can deliver results, we have to manage you closely for your own sakes."

"Fool me once, shame on you. Fool me twice, shame on me."

And who's the cause on the third, fifth, hundredth or thousandth repetition?
The people who keep paying for the same 'ol, same 'ol.




2008/03/09

Videos on Flash Memory Cards - II

My friend Mark expanded on my idea of "HD DV being irrelevant" - like phone SIM's, video stores can sell/rent videos on flash cards (like SD) sealed in a credit-card carrier.

The issues are more commercial than technical. 8Gb USB flash memory might hit the A$50 price point this year - and A$30 next year. There is a 'base price' for flash memory - around $10-$15.

This inverts the current cost structure of expensive reader/writer and cheap media. Which is perfect for rental/leasing of media - a refundable 'media deposit' works. An added bonus for content owners is a significant "price barrier" for consumers wanting to make a copy. If a 'stack' of 100 SD cards costs $1500 (vs $100 for DVDs), very few people will throw these around 'like candy'.

Mark's comments:

Y'know, the more I think of it, the more the SD-embedded-in-a-credit-card has a lot of appeal when the availability and price point for 8Gb SDs is right. It makes it easy to print a picture, title and credits/notices etc on the 'credit card' - something big enough to be readable and a convenient display format and, as you say, nicely wallet-sized. Snap off the SD and you've agreed to the conditions etc, plus the media is now obviously 'used'.

It's a useful format for other distributions too - games, software, etc (Comes to mind that SAS media still comes on literally dozens of CDs in a cardboard box the size of a couple of shoe boxes).

My complete collection of "Buffy" would come in something the size of a can of SPAM or smaller, rather than something the size of a couple of house bricks for the DVD version, or something still the size of a regular paperback for the Blu-Ray version. For collectors of such things, the difference between having many bookshelves taken up by the complete set of Vs a small box of credit card (or smaller) sized objects is significant. The ability to legally re-burn or replace and re-burn the media when it fails is critical though.
SJ: Because of the per-copy encoding to a 'key', stealing expensive collections isn't useful, unless the key is also taken. So those 'keys' have to be something you don't leave in the Video player.

You've covered the DRM aspects and better alternatives to DRM - which also means that I can burn and sign the media I might produce and distribute myself without needing to involve the likes of Sony or Verisign - although that is possible also - which protects the little producer. Include content in Chrissy and Birthday cards - you've seen those Birthday cards with a CD of songs from your birth year - why not a sample of the movies from that year, plus newsreels etc. Good for things like audio books - whole collections. And if the content on an SD gets destroyed, as long as the media is OK, it would be possible to re-burn it. Most current DVD players now also have SD readers as standard.

Surely someone has thought of it already! Part of the attraction of DVD over storing your library on a 2TB USB disk from Dick Smith is the problem of backups. DVD is perceived, incorrectly, as permanent storage. Though I notice some external USB drives now have built-in RAID 1 or RAID 5, but Joe public doesn't see the need (how come I bought a 2TB drive and I only get 1TB?).

Yeah, I think the proposition that SD or similar will become the ubiquitous preferred standard portable, point-of-sale, recording and backup storage media for photos, movies and music, has some credence. There is something to be said for - "you pick it up in your hand; you buy it; it's yours" - over - "downloading and buying some limited 'right to use' ".

2008/03/07

Service Desk and Politician e-mail

Over the last year I've penned 6+ e-mails to various Labor Party politicians - including one of my local representatives who've I dealt with for ~10 years.

And not one reply. Zero, Zip, Nada...

Rang the Good Person's electoral office today - and got various run-around responses. "Oh, I've been on holiday", "Oh, can they call you" and "they are booked solid for a month".

Yeah, right.

I first contacted my rep. last December saying "this can wait until after the School Holidays". January came and went, no reply... A follow-up email yielded nothing... A note to the support staff was replied to: "I've moved. XXX is responsible".

What I originally wanted to talk about was 3 emails I'd sent various members without even getting acknowledged. Which is strange, because in the media I've seen reports that Political Parties are now tracking every contact from a voter. Putting together, apparently, impressive profiles - and all completely legit under the Privacy Laws.

For a new Government this seems a pretty poor response, doubly so for one that prides itself on 'listening'.

The solution that I wanted to put forward to my Rep:

Use HelpDesk Software to manage constituent contacts.
Not just piecemeal, but an integrated system for all participating elected members.

Not all that hard.
It scales. It goes across the whole Party. It covers both 'aph.gov.au' contacts and via other email addresses. It copes with email, phone, fax, mail and personal contacts - and the worst of all "voice prompt systems".

The software is well known, there are many vendors and trained consultants and the marketplace is competitive. As consumers and office workers, most of us are used to the concepts and who these systems all work.

It creates a definite process - with self-imposed rules & priorities that are checked and enforced.

AND it ensures that little people like me don't just fall between the cracks.
Or if some 'critical person' falls down - work queues can get given to those who can best deal with them.

Imagine getting a tracking number back from your local Pollie, and being able to automatically check where it is up to - and just when you should expect an answer. Wow! Just like they worked for us and were trying to use the technology responsibly...

It would do a service for our erstwhile representatives - you know, the ones we pay to work for us:
  • They could become more efficient - by delegating work, not needing to deal with "whatever happened to" requests, and identifying common themes and selecting the most efficient way to respond.
  • They could make a very exact case for additional clerical support from the Parliament - or even have a pool of paid staff doing the grunt work.
So I'm not holding my breath while waiting for anything different to happen.

The Internet Changes Everything - but Politicans and their ways.

2008/03/06

Who cares about HD DV?

Talking to a friend at lunch today, the topic of "Blu-Ray" vs "HD DV" formats came up...

I think "Blu-Ray" may take the market, but it won't be much of a market.
There are just too many competitors for moving around video files:
  • DVD format disks - still good for 8Gb (dual layer). Drives & media are cheap.
  • flash memory - 2008 sees A$50 for 8Gb on USB (less on SD card)
  • A$300 for 750-,1000Gb USB hard-drives. Under $1/DVD.
  • Internet download. With ADSL 2+ giving 5-10Mbps for many.
My thesis to my friend was "Video stores may well go for SD cards". Pay a refundable deposit for the flash card, and a fee (rental or ownership) for the content. Video stores can pre-burn large numbers of movies - and if you want a 'special' - they can make one for you in 20 minutes.

His response: "they could package them like SIMs - in a snap-off credit card-sized holder". Which is better than any idea I've had on packaging.
And it fulfills the most important criteria:
fits comfortably in a pocket (now a wallet)


Practical problems:
  • How to stop people copying the flash and resealing it?
  • Some sort of effective copy-protection system would be good.
  • Flagging 'ownership' or usage conditions of a movie. Not so much DRM, but 'this is property of XXX'
These problems can be nicely solved by users having their own "Key Card" with a digital identity and an encryption key.

The flash needs a 'fuse' that is broken when the card is freed. Preferably an on-chip use counter that can only be factory reset.

To issue a movie to a customer, the encoding key of the video (if present) would be combined with the users key - and the resulting unique key written on the card. Players need both the card and user key to decode and play the movie.

That same process also tags the card with the current owner.
You lose it, it can come home to you.

Because the content can be locked to a particular ID, the raw content can be stored on disk without the movie studios giving away their birth right.

Summary:
I think 120mm disks are going to follow the floppy disk into the technology graveyard.
They will have certain uses - like posting something on cheap, robust media.

With the convergence of PC displays and Home Theater, the whole "Hi-Def TV" problem is morphing. Blu-Ray - can't wait to not buy one.

2008/02/08

The Open Source Business Model

This post by Dana Blankenhorn on ZDnet is the best answer I've seen to the question "Why Open Source?".

He says 'plumbing', I'd say '(Unix|Open Source) is the Universal Glue'.
And the on-going Open Source Business Model is "support" for those that need/want 'certainty'.

Which if you are the CIO (read: 'my arse is on the line') for somewhere with a high dependence on I.T., is only Good Governance (or "common sense"). You can't make key staff stay, nor mandate they never get sick or burn-out and "go sit on a beach" - and after '9/11', all Business Continuity plans have to account for covering people as well as systems and networks.

That's it - Business I.T. is all about the Data (or "all about XXX, stupid" to be Clintonesque).
Open Source tools are usually about manipulating data or providing services - like Apache, e-mail, DNS, firewalls and IDS, ...

Open Source is here to stay: use it, don't deny or fight it.

This Business Model, 'support for essential tools', is robust and on-going.
Whatever systems you use in the Data Center, you'll always have the need to provide many services and interface disparate systems and data formats.

The model also applies to embedded 'Appliances' and dedicated devices, like firewalls - or commercial web-hosting services. They are based in whole or part on Open Source.

You'll note this model has very limited application to the client-side - the 'Desktop' or End-User compute platform.

"Free Software" from GNU et al is about an ideological stance and subsumes all other goals to this.

"Open Source" is pragmatic and about getting on with the job. It makes sense for large vendors, like IBM and HP, to support it. Customers can feel confident and secure - because the source and tool-chain are freely available from multiple sites, they cannot be held to ransom or 'orphaned' by unpredictable events or capricious decisions.

"Open Source" starts from the premise that "IT is done for a Business Benefit" - that you build software, systems and services for the use of others, not your own amusement and benefit.

Business supporting software has to meet Professional standards/criteria - good design, clear documentation, reliability, robustness and very few errors/defects - with the unstated driver of Continuous Improvement.

Never new features for their own sake or to create 'forced upgrades', always making the code more stable, usable and useful.
Commercial considerations, by definition, are always subsidiary to technical. If the user community doesn't like changes - they aren't forced to upgrade and in an extreme case, can 'fork' the code, internally or publicly: just do it how they want.

2008/01/19

Human Response to Cognitive Workload

Context: This piece started as a question to a researcher in Psychology.

There's a topic I've been trying to find research results for some time.
I call it "(Human) Cognitive Response to Workload".

There is a bunch of qualitative data for "Physiological response to Workload" available - e.g. US Navy for "stokers" working in different levels of heat.

I found Prof. Lisanne Bainbridge in the UK. She's retired now. Her field is 'mental load' and couldn't point me at research in the area or help me properly phrase my question.
She pointed me at Penny Sanderston, now Prof. at University of Queensland.

What I'm interested in is any information to apply to Software Developers and other Knowledge workers:
  • In the short, medium & longer term (day, week, year) how do you maximise cognitive output?
  • What roles do sleep, recreation & holidays play in 'recharging' cognitive abilities?
  • For different levels (degrees of difficulty) of cognitive task (or skilled manual task) what are the optimum work rates and duty cycles? (ratio of work/rest)
Related areas are Rates of Error & 'tiredness' effect on maximum cognitive task.
[James T. Reason has very good work on "Human Error" and "Organisational Error". His work is used extensively in Aviation and Nuclear safety. He originated "the swiss-cheese" model of accidents.]

2008/01/01

Solving 'Spam'

It never ceases to amaze me, the Politician attitude to Porn and 'Spam' & it's friend, malware.

Porn is "bad, bad, bad" and Pollies show very high interest - including policy & legislation.

Lots of angst & trashing around about eradicating something that 2,000+ years of writing/publishing shows can't be controlled/legislated away. The physical publishing world & (cable) TV show that the only effective is means of control is to allow-but-license.

Same as tobacco. Never going to eradicate it, only control it.

'Restricted Content' access can only be controlled iff:
  • every page is 'classified' at source (meta-tags),
  • an unforgeable Internet 'proof-of-age' card/system is created,
  • there are criminal penalties for subverting the system, forging identities or misclassifying pages,
  • there are no legal jurisdictions outside 'the system' [e.g. on the high-seas],
  • all browsers enforce 'the rules',
  • and browsers can't be built/written to ignore 'the rules'.
i.e. It is impossible to eliminate 'restricted content', and possibly provably so...


2007/12/29

IBM, Outsourcing and the IT Profession

This is a reaction to Robert X. Cringely's "Pulpit" of 28-Dec-2007:

Leaner and Meaner Still: IBM's U.S. operations continue to shrivel.

There are 3 parts to my comments:
  1. Will IBM Survice?
  2. Outsourcing
  3. IT as a Profession

They are interlinked. Lou Gertsner set IBM on the road on "Services" and away from Mainframes. It looked promising.
IT Services look very appealing on the Balance Sheet - nearly no investment (no tangible assets) and what seem to be good profits from turnover. The ROA and ROI (Return on Assets and Return on Investment) look great - until you take some other factors into account.

  • Barriers to Entry for competitors are low.
    EDS under Ross Perrot came from nowhere to define and dominate the field - so can the next giant in the field.
    If your business model is "hire cattle and drive them till they drop" - you have no market differentiation.
    Same cattle, same drivers, same pay - same 'ol, same 'ol... The cattle aren't loyal, motivated or engaged.
    Writing new contracts is a matter of perception, influence and contacts.
    There is so much feeling against IT Outsourcers in business at the moment, the first company to come along and tell a better story will take the field.
    The change won't be overnight, but fast enough that the incumbents won't notice until too late.

  • Whilst only tangible assets appear on the Balance Sheet, IT Services are driven by your Human Capital and some Intellectual Capital embodied in your processes, branding and IP, such as trademarks and patents.
    What value is let in the offices when everybody has gone home? Very, very little.
    What is the business risk of a large, sudden exodus of your staff? A competitor may deliberately poach enough to put you in trouble.
    It's a failing of the Board not understand this and institute appropriate metrics, accounting and management rewards.

  • Profit based on Operations turnover are very fragile/volatile.
    Income and Expenses are very large numbers with a small difference. Expenses are mainly employees - which you may not be able to shed as quickly as service contracts expire.
    Tendering for new contracts implies you have, or can quickly get, the resources to fulfill the contract. That's an extreme business risk.
    The key figures-of-merit are Income/Employee and Profit/Employee.
    We don't see those reported or obviously managed.

  • IT Services work is Knowledge Work - it is mostly invisible and intangible.
    Driving IT staff like unskilled labourers with threats/punishment to lift performance is anti-productive.
    Unhappy staff withdraw and pushback. At best they aren't engaged or motivated. They 'do the minimum' - a grudging compliance.
    They stop caring about their work, the customer and their employer. And if you are lucky, it stops there.
    Hiring bright, capable people doing intangible work and treating them badly is not just a recipe for disaster, it is foolishness writ large.

  • IT is a cognitive amplifier and this can be leveraged both within the business and internally in IT.
    The only sustainable strategy to deliver improved profits is through investment.
    • Automating tasks.
      Applying our own technology to our jobs to make tasks, not jobs, redundant.
      Investing in tools and hardware to increase the both Quality of work
    • Building Human Capital.
      Investing in the people at the work-face to build their capability and performance.
      The SEI's Barry Boehem created COCOMO - a quantitative model for estimating Software costs.
      Experienced, competent practitioners not only produce better work, fewer defects, faster - they are cheaper.
    • Actively reducing Errors.
      Consciously reducing waste, rework and wrong work.
      Quality is not about 'doing the minimum', it's a mindset where Errors are allowed, but their repetition is anathema.
      High Quality performances are only achieved with deliberate, focussed intention. Not blaming and denial.
      Quality Systems only goal is to make it difficult for good people to make mistakes.
      Deming said it all with "Plan-Do-Check-Act", or in new-speak: "Preparation - Execution - Review & Evaluation - Improvement"
    • Learning is central to improving Quality, Performance, Security & Safety and Usability.
      Learning systems, processes and procedures takes an investment of time, tools and technology.
      Failing to build teams and their capability will decrease expenses in the short-run and will increase them in the long-run.

  • Resiling from the classic adversarial stance of IT Outsourcing.
    IT is a business enabler. It is now central to normal business operations. It is still where 80% of efficiency improvements arise.
    Every act that hurts the client will turn-around and hurt the provider, but more.
    The client is earning the income that pays for the IT.
    More income, more IT, more outsourcing revenue and profits. A simple equation that seems lost on Outsourcing managers.
    Outsourcing contracts need to align the internal management rewards with improving business outcomes for the Client.
    What's anathema to current management - reducing Client costs - must be aggressively pursued to create a long-term Outsourcing business.

Quality is not 'gold-plating' - it is central to improving productivity, reducing waste and fulfilling customer expectations. These are the drivers for growth, profitability and sustainability - not penny-pinching and cost-cutting.

IT Services companies cannot, and will not, pursue Excellence & Quality if they are not driven to it.
It is only their Clients who can hold them accountable and force a change.

Concurrently, IT has to evolve from an Industry to a Profession so that managers can realistically evaluate the performances of different practitioners. It's not hard to win new business and make good profits if your employees are 10 times more productive than your competitions.


Will IBM Survive?


Answering the poll question: Will IBM survive?

Lou Gertsner turned IBM around, starting 1993.
It took an outsider to do it - and the board knew that.

His legacy, after leaving in 2002, should've been a company with a solid future. Five years on, it appears not so - that can only be "Corporate Culture".

IBM is far too important to be let fail and broken up in a firesale.

But we have a perfect model for the future of Cringely's "lumbering giant": Unisys.

In 1986, Numbers 2&3 in the market (Burroughs & Sperry Univac) combined and produced a dud. It's still alive, but failing. Because enough people use their mainframes (2200's and A-series), they can't be allowed to die. Slowly withering on the vine seems to be fine.

Fujitsu is the perfect vacuum-cleaner to buy the hardware business in the final break-up.


Outsourcing


IBM GSA and the other 'Tier 1' outsourcers operate from the same playbook - a version of 'bait and switch'. Also known as "The Value Prevention Society".

I've worked with and for all the major outsourcers in Australia. They all bid low to win contracts and adopt a dual strategy of "controlling costs" and price gouging for "variations".

'Controlling costs' is reducing staff, replacing competent staff with 'cheap and cheerful' newbies, not performing maintenance and avoiding capital investment.

What's wrong with a 5-10 year-old system? Nothing if you don't have to suffer the performance and other problems!

They routinely ignore contract provisions - like scheduled roll-outs of new desktops, upgrades and system performance targets.

The problems are at least three-fold:
- inequality of parties (Outsourcer vs Client)
- internal 'manager' performance has no upside, only downside
- no impartial umpire and effective 'stick' to enforce system performance targets

Inequality
Every company that signs an IT Outsourcing agreement signs just one. The outsourcers has done this many, many times before.
Clients also don't factor in the increased staff and reporting costs - each side needs additional staff for 'contract management'.

The Client thinks it has stitched up an iron-clad contract and they forecast a bountiful harvest... Which doesn't happen.

Service degrades, minor works become hugely expensive, major works take forever and often don't get implemented.
The business people give-up and adapt around it.

In Australia, all the major EDS contracts let around 10 years ago are now being re-tendered - with EDS getting very little of the new work.
Are they the worst? Hard to say...

Aligning internal rewards with Client Needs
Outsourcer 'managers' can only be assessed on monetary performance. With fixed price contracts, base income is fixed.

If a manager reduces costs 5% one year, this becomes the expectation for every following year - it is not seen as a 'one-off'. Without significant staff training and capital expenditure, this quickly becomes impossible without sacrificing service quality. Commercial systems are quite reliable these days. For existing stable systems, 'Do nothing' is good for at least 3 years - then you are in deep trouble.

The only ways to increase profits are to reduce expenses or increase non-base income.

Every service request is deemed a 'change' and subject to the full, heavyweight, project evaluation methodology. No project, not even buying a simple standalone appliance, takes under 4 man-weeks ($20-50,000). For the client, this stifles change/innovation (or forces it underground) and these additional costs overshadow most systems costs.

Capital expenditures are worse. Payback has to be within 12-18 months - and it has to beat 'do nothing'.
Since the 2003 slowdown in Moores' Law for CPU speed, the problem has compounded.

Take a 5 year-old file server that is now close to saturated most of the day. It is not yet 'end of life' and maintenance costs still low.
Because file open/close, read/write performance is not specified and the system is "available" during work hours, the Client cannot complain.
The Operating System (O/S) may be old and need constant attention, updates and reboots - but they are part of the normal admin workload, so not an 'additional' cost. Salaried staff as 'professionals' must work any unpaid overtime that is demanded.

Any proposal to replace the server or upgrade it has to pass a simple, and reasonable, test:
How much extra revenue will we make? How long will the payback period be?

'Do nothing' is the benchmark - for zero capital expenditure and a few extra unpaid admin hours, a service is provided that brings in the service full revenue - and will continue to do so. That's a very tough argument to beat.

Only when the client funds the replacement, hardware maintenance costs are high enough, an O/S upgrade is required for security or compatibility or qualified admin staff move on will the system be upgraded. And then it will begin the same inevitable slide into entropy and uselessness.

Finding solutions that benefit the customer and reduce operating expenses are career suicide for outsourcing staff in a culture focussed on increasing billables.

For example: a major Australian bank replaced all the local file servers with small Network Appliance NAS's. These are the most expensive product per Gb available. The outsourcer had charged ~$2,500/month to 'administer' these systems. The bank paid for the change in under a year, increased availability and performance and solving many other issues to boot.

If the client gives all its IT staff to the outsourcer, who is going to seek out, design and implement new cost saving technology/systems?
Not the outsourcer - it's not in the contract and not in its (short term) interests.
The client has no IT staff - so it cannot and doesn't happen.

Audits and an Impartial Umpire

Who reports to the Client on the performance of their systems?
Who has the training/qualifications to check and asses the metrics and reports?
Who maintains & audits the basis of payments - the asset register?

Only the Outsourcer.

What are the downsides to the Outsourcer of a major failure in Prime Time?
A small number of 'service credits'.
Meanwhile, the Client suffers real costs and potentially large losses.

The Client wears all the business and financial risk with only minor penalties to the Outsourcer.
We are yet to see a corporate collapse due to an outsourcers IT failures - but it is only a matter of time.

There is a clear conflict of interest, or an real Agency Theory problem.
The outsourcer is Judge, Jury and Executioner...
There is no way to hold them to account or dispute their figures.


The Profession of IT

Contributors Michael Ellis, BJ, Kevin James, Richard Steven Hack,... started a thread about the 'value'/competency of individual IT practitioners.

The huge (100+:1) variability in individual competence and the inability to measure it is one of the worst problems in our industry.

IT is not a 'Profession'. It, like 'Management', fail a very simple test:

What are the personal and organisational consequences of repeating, or allowing to be repeated, a known error, fault or failure??
[Do your mistakes have clear 'consequences' professionally?']

Mostly it is "fire/blame the innocent, promote the guilty". The exact inverse of what you'd want.
People may trump technology and process, but Politics trumps everything...

And our Professional Bodies don't help.

The only real research into the causes of Project Failure are by consultancies - who are driven by the ability to sell their products, not what will benefit the Profession.

The ACM, IEEE, IFIP and friends have abrogated their responsibilities. We on the firing line, get to suffer their inaction.

Managers have to go with what they can quantify and inspect. Good managers will see through the B/S - but mostly too little, too late. Mostly, office politics, influence and self-promotion rule.

The adversarial nature of Outsourcing and the seemingly universal decline in code and service Quality stems from this failure of IT as a Profession.



Steve Jenkin 29-December-2007

2007/10/20

MinWin: 25Mb WIndows. Hypervisor expected?

Could this be the start of a real change at MSFT? [i.e. doing software 'properly' (small, fast, secure)]

First question:
  • What if they pick up GPL or similar code & illegally include it.
  • How would that be detected??

Timeframe for 'commercial' MinWin is 2010.
The real news here is MSFT's focus on virtualisation...
With their purchase of "Virtual PC", they have the tools to build their next-gen O/S products around VM's.
Licensing???

Another question:
  • If MS kernels ship with a hypervisor, how do we dual-boot or run our own VM like XEN?
  • Would they be stupid enough to "embrace & extend" the VMI API/paravirt_ops?

The actual talk was on virtualisation and its impacts.. [143Mb]
<http://www.acm.uiuc.edu/conference/2007/video/UIUC-ACM-RP07-Traut.wmv>


<http://blogs.zdnet.com/microsoft/?p=842>

Traut spent most of his time describing Microsoft’s thinking around virtualization, and how virtualization can be used to ease backwards compatibility and other problems Windows users incur.


Microsoft has created a stripped-down version of the Windows core, called MinWin, that will be at the heart of future Windows products, starting with Windows 7, the Windows client release due in 2010.


MinWin is 25 MB on disk; Vista is 4 GB, Traut said. (The slimmed-down Windows Server 2008 core is still 1.5 GB in size.)

but no graphics subsystem


The MinWin core is 100 files total, while all of Windows is 5,000 files in size.

Runs in 40Mb memory. Ascii only

MinWin will be at the heart of future versions of Windows Media Center, Windows Server, embedded Windows products and more.

First good MSFT decision I've heard in ages

Traut said he is running a team of 200 Windows engineers working on the core kernel and Windows virtual technologies.

C.f. 10,000 total on Longhorn/Vista. Say 3,000 coders

(he) said that Microsoft is operating under the premise that “at some point, we’ll have to replace it (the kernel),” given that it “doesn’t have an unlimited life span.

That's important news

2007/10/05

Open Source - Barriers to Entry

Open Source - Barriers to Entry


I think I have a short, coherent description of the underlying cause of
the barriers to adoption to Open Source:

"Some Thinking/Expertise Required"
(as in "Some Assembly Required" or "Batteries not included")

It stems from:
Is IT well-managed?

Which leads to:
Is "Mangement" generally practiced well??

To both of these, my answer is a strong "NO" - it's all about failure of
management.

The Usual Management Method


I've seen very consistent behaviours, attitudes and approaches across
every organisation I've worked in [a very large cross-section]. I don't
know where they arise or how - but there are best described as
'unschooled' or 'hard knocks'.
Certainly not 'insightful', educated nor informed... That appears to be
anathema.

I've met precious few managers that I'd call competent, let alone good.
And very few who'd bothered to train in their work.
One (a scientist in charge of 100 ppl and a $30M budget) bragged "I've
never done *any* management training".
His PhD in biology qualified him for everything...
[The subtitle of "Other Peoples Money" is: 'Arrogance, Ignorance and
Self-Delusion'. Wide-spread and

Perhaps this one point, consistent management training, is the reason
IBM dominated the computing industry for 3 decades...
[And their avarice/denial brought them undone]

Professional Management & Management Profession


'Management' doesn't qualify as a 'Profession' under my definition:
  • an identified & testable set of practices, skills, competencies
  • (behaviour?) [think pilot or surgeon]
  • means to provide barriers to entry and disqualification/discipline
  • Improvement/Learning mechanisms:
  • by invention/discovery
  • by incremental improvement
  • analysis of failure & root-cause analysis + corrective actions (think 'bridge falling down' or plane crash)

IT Management and general Management


Without Professional & competent business managers, there can be good
management of IT.
Without good IT management, good practices and competent practitioners
are rare and can't be maintained...

Summary:
IT is populated mostly by a bunch of rowdy, undisciplined 'cowboys'
that are set in their ways and do what they please.
IT management is about politics, influencing and pleasing, not any
rational, objective measures.

That explains the Fads & Fashions of Management, and the almost
universal CIO mantra "nobody got fired for buying <fad/fashion>".
And of course:
  • Risk Avoidance & Blame Shifting [consultants & outsourcers]
  • CYA

Implications for Open Source business


How to use this premise?
  • Wait for the current fashion to collapse [or have cracks]. All fads & fashions change.
  • Find the few competent business & IT managers out there and sell to them...
  • Sell them camouflaged/disguised systems - like embedded devices or appliances (e.g. network, storage, security)

2007/06/07

Why IPTV won't work in Australia - ISP Volume Charging.

It's all about cost...

A phone line costs $25-$35 just to have. Add ADSL plans on top.

ADSL plan costs vary with speed and pre-paid download volume.
Reasonable TV needs around 4Mbps these days - the highest speed and cost plans.


2007/06/03

Turnarounds

My previous post on 'Digging Out', a methodology built from experience in a number of turnarounds, can't stand alone without some justification:
What have I done to be credible?

Here's a sample taken from a version of my CV:
Completing large business critical projects on-time and on-spec. In complex political environments achieving all client outcomes without formal authority.
  • ABN online registration system - Business Entry Point, DEWRSB.
  • Y2K conversion - Goodman Fielders (Milling and Baking)
  • Y2K remediation, CDM
  • DFAT (ADCNET Release 3, ADCNET build/release system)
  • TNT
  • Unisys/Customs (EDI & Finest/Nomad)
  • CSIRO - all Australian daily weather database
  • Diskrom ($350k project income in a year)
ABN registrations:
The ATO paid for and ran the project. The software contractor was combative and unhelpful. The environment complex - around 6 different organisations for this one project, and another 10 projects. To get anything done required a huge amount of effort, negotiation and discussion.

The software contractor hadn't built any load monitoring and response time facilities into the system nor made any provision for capacity planning and performance analysis.

On my own recognisance, I designed and built some simple but sufficient tools, did the performance analysis - and accurately predicted the final peak load - 20 times the design load, and after diagnosing a catastrophic failure mode, designed a simple solution.

This 100% availability over 3 months was not accidental and directly contributed to 600,000 registrations of 3.3M being done on-line (around 15 times the estimate) and the site not attracting bad press like other aspects of the operation. Definitely a high-profile site.

The software contractor had to be dragged kicking and screaming all the way through the process. But I got the outcome the client needed - the site kept running with good response time on the busiest day. Some years later I analysed the performance data from the time and uncovered a few more nascent problems that we'd skirted around.

Goodman Fielders (Milling and Baking)
This was a Y2K project - they needed to migrate a legacy (Universe/Pick) application to a Y2K compliant platform and simultaneously upgrade the software and remerge all the 6 variants.

The application ran their business - ordering, tracking, accounting - the whole shooting match.
And for accounting systems, the Y2K deadline was the start of the financial year - 1-July-1999.

The work got done, I contributed two major items: deferring non-accounting conversion and moving the new single system to a properly managed facility.


DFAT (ADCNET Release 3, UNCLgate, ADCNET build/release system)
This was bizzare, I ended up sitting 20' from the desk I used when I worked on the system being replaced ('the IBM'), when it was being commissioned.

ADCNET failed and went to trial with the developer losing on all counts. It's worth reading the decision on 'austlii.edu.au' [Federal Court]. That account is certainly more detailed than the ANAO report. It was obvious in 1995 that the project could never deliver, let alone by the deadline. So I did my tasks and went my way.

To be called back again to degug and test an email gateway between the IBM and ADCNET (R2) for Unclassified messages. This was the first time I realised that being better than the incumbent staff in their own disciplines was 'a career limiting move'. Showing experienced, supposedly expert, programmers how to read basic Unix 'man' pages and act on them was a real lesson. A major problem that caused queued messages to be discarded was found and fixed by my testing - along with a bunch of the usual monitoring, administration and performance issues being solved.

I was called back again to help with the Y2K converstion of ADCNET (departmental staff were doing it). The system was over a million lines of code and the release/development environment bespoke. And required maintenance work on the dependencies/make side of the software had never been done. A few months part-time work saw all that tamed.

TNT
Went for a year as an admin. Did what I could, but they were past redemption... Bought out by the Dutch Post Office (KPN) soon after I'd arrived.
Presented a paper at a SAGE-AU conference detailing my experience with their 'technical guru' - who'd considered himself "World's Greatest Sys Admin". Google will find the paper for you.
It was so woeful, it defies description.

Unisys/Customs (EDI & Finest/Nomad)
In early 1995 was called in to replace a SysAdmin for 8 weeks on the Customs "EDI Gateway" project. The project and site were a disaster - so much so that Unisys listed it as a "World Wide Alert" - the step before they lost the customer and hit the law courts.

In two months the team stabilised the systems, going from multiple kernel 'panics' [a very bad thing] per week, 8-10 hour delays in busy hour and lost messages - to 100% uptime over 6 systems, 1-2 second turnarounds and reasonable documentation, change processes and monitoring/diagnosis tools. The Unisys managers were very appreciative of my efforts and contributions. This same sort of chaos that was evident in the 2005 Customs Cargo clearance System debacle. [The 'COMPILE' system ran on Unisys 2200 and was being replaced over a 10-year period. It was the back-end for the EDI systems I worked on.]

So much so, that I was called back for another few months to stabilise another system running ADABAS/Natural legacy applications that provided the Financial Management & Information Systems and Payroll/Personnel system. Another high-profile, critical system.

CSIRO - all Australian daily weather database
The research scientists on the project I worked for created some tools to analyse weather data - and had found a commercial partner to sell them. The partner was not entirely happy due to extended delays and many unkept promises. I'd been told that to buy the entire dataset - a Very Good Thing for the commercial partner - was not affordable, around $20,000 for the 100 datasets from the Bureau of Meteorology. When I contacted the BoM, they not only provided the index in digital form for free, but the whole daily datasets would cost around $1,750. I scammed access to another machine with the right tape drive, wrote scripts and did magic - and stayed up overnight reading the 70-odd tapes. In pre-Linux days, there was no easy way to compress the data and move it around.

The whole dataset as supplied was 10Gb raw - and I only had a 1Gb drive on my server [$3,000 for the drive!].

It took 6 weeks to fully process the data into their file format. And of course I had to invent a rational file scheme and later wrote a program to specifically scan and select datasets from the collection.

The Commercial Partner got to release the product at the ABARE 'Outlook' conference with a blaze of CSIRO publicity. Don't know what the sales were - but they were many times better.
The research scientist got a major promotion directly as a result, and I was forced to leave for having made it all possible.

Diskrom ($350k project income in a year)
In under a year I learnt SGML, hypertext and enough about handling large text databases [1991 - before the Web had arrived], took over and completed 3-4 stalled and failing projects, migrated datasets and ssytems, designed tools and file structures/naming conventions and completed the first merge of the Income Tax Assessment Act with the Butterworths commentary, speedup processing of a critical step by 2,000 times - all of which directly contributed $350,000 in revenue [apportioned to my effort] - or around 12 times my salary.

So it's natural that everybody else in the office was given a pay rise, I was told that I was technically brilliant but not worthy of a rise and one of the 'political players' was promoted to manage the group. With a number of other key technical 'resources' I left to pursue other avenues.

Diskrom was shut down just a few years later when a new chief of the AGPS (Aus. Gov. Printing Service) reviewed the contract and decided they were being ripped off. They'd provided all the infrastructure and services, with the commercial partner paying for staff and computers - and despite lucrative contracts and overseas work, never seen any return.

Basis of Quality - What I learnt in my first years at work

On the 17th of January 1972, with 70+ others, I started as a cadet at CSR. This early work experience set the stage for how I approached the rest of my life at work. I'm come to the conclusion that these early formative experiences are crucial to a persons' performance and attitude for their entire working life. Breaking these early patterns can be impossible for some.

The cadets of my intake had the privilege to have unrivaled experiential life lessons in quality, safety, team building & working, skills development and personal responsibility.

The lessons I took with me into my career:
  • Quality workmanship and precision techniques
  • Formal quality techniques and analyses
  • Following precise analytical processes
  • Managing high work loads and meeting tight deadlines
  • Responsibility for high-value processes/outcomes
  • Satisfying stringent customer requirements
  • Respect for, coping with and managing work in dangerous
    environments - and experience in handling/responding to hazardous
    incidents.
  • And doing time sheets and "filling in the paper work" as a natural part of things.

My first 2 years of full time work as a cadet chemical engineer at CSR Ltd saw me personally responsible every day for $10M’s of product – performing sugar analysis for the whole of the Australian sugar crop. At the time the price of sugar was at an all time high. Each of us played our part in process - no one person did it all, but one person could botch the work of a whole team, or even a whole days' work.

At these NATA certified laboratories, we were trained in Chemical Analysis - but also safety and quality methods - with lives and ‘real money’ at stake.

Routinely large quantities of flammable and explosive alcohols, and highly toxic and corrosive acids were used and safely disposed of.

Deadlines were tight and fixed – each day samples for 50-100,000 tonnes of raw sugar and full analysis had to be delivered with a very high degree of accuracy and certainty that same day.

Speed, precision and absolute dependability were instilled in us alongside a clear knowledge of the value and consequences of our work - and mistakes.

We were tutored in analytical techniques, trained in reading and following exactly processes, statistical analysis, fire and safety (OH&S) skills, certified first-aid and our duties and responsibilities to our clients - the sugar producers.

It was expected that "people make mistakes" - the first rule of any precise analysis is the error range (+/- ??). The system was designed consistently to produce accurate, repeatable results with a very low margins of error. Calibrated samples were fed through the process alongside multiple samples and any 'repeats' that had failed the checking process. The performance of individual analysts and groups was tracked and monitored. People were assigned to tasks suited to their particular talents - based on objective results, not presumption or basis.

We all acquired robust team working skills. Along with how to do the boring things like time sheets with exactitude.

The next year I spent working as an Analyst in the Pyrmont Sugar Refinery.

Lots of routine and tight deadlines - and the same attention to detail, importance of results and a necessity to understand 'the big picture'.

There'd been an analyst, an ex-secretary, who'd dropped her (mercury) thermometer into the 'product stream'. She hadn't realised it was a problem, and a few hundred tons of sugar had to be thrown away and the process shutdown for cleaning. A tad expensive - many more times her wage and the cost of training and supervision to prevent.

My routine work led to uncovering a systematic problem with a very simple and preventable cause. We had a power station on-site - a real live coal burning, electricity and high-pressure steam (50 atmosphere?) producing plant. It used enough coal that when a fire in the furnace hopper caused a malfunction, the whole of the Sydney CBD was covered in a thick pall of black smoke - which made the TV news.

The plant feed-water was supposed to be closely maintained near a neutral pH - not acidic. A test station was on the factory floor, with a reagent bottle next to it. The reagent ran out, so I made a new batch - only to find that the measurements were now grossly wrong. The bottle was stored without a lid and slowly evaporated and concentrated. Over a couple of years, the pH measurement had slowly drifted and the water pushed way out of spec.

Serendipitously a major explosion or equipment failure was avoided. Replacement of the power-station, and shutting down the whole of the Pyrmont facility dependent on it for a couple of years, would've seriously impacted the whole company.

Digging Out - Turning around challenged Technical Projects/Environments

Something I wrote in 2002:

‘Digging Out’ - 7 Steps to regaining control

This is a process to regain administrative control of a set of systems. It can be practised alone or by groups and does not require explicit management approval, although that will help.

‘Entropy’ is the constant enemy of good systems administration – if it has blown out of control, steps must be taken to address it and regain control. The nature of systems administration is that there is always more than can be done, so deciding what not to do, where to stop, becomes critical in managing work loads. The approach is to ‘work smarter, not harder’. Administrators must have sufficient research, thinking & analysis time to achieve this – about 20% ‘free time’ is a good target.

This process is based on good troubleshooting technique, the project management method (plan, schedule, control) and the quality cycle (measure, analyse, act, review).

The big difference from normal deadline based project management is the task focus, not time. Tasks will take whatever time can be spared from the usual run of crises and ‘urgent’ requests until the entropy is under (enough) control.

Recognition

Do you have a problem? Are you unable to complete your administration tasks to your satisfaction within a reasonable work week? Most importantly, do you feel increasing pressure to perform, ‘stressed’?

Gather

The Quality Cycle first step is ‘Measure’. First you have to consciously capture all the things that 1) you would like to do to make your life easier and 2) take up good chunks of your time.

The important thing is to recognise and capture real data. As the foundation, this step requires consistent, focussed attention and discipline.

The method of data capture is unimportant. Whatever works for the individual and fits naturally in their work cycle – it must NOT take significant extra time or effort.

Analyse

Group, Rewrite, Prioritise.

Create a ‘hard’ list of specific tasks that can be implemented as mini projects that can be self managed. Individual tasks must be achievable in reasonable time – such as 1-2 days effort. Remember you are already overloaded and less than fully productive from accumulated over stress.

Order the list by 1) business impact and 2) Daily Work-time gained.

The initial priority is to gain some ‘freeboard’ – time to plan, organise and anticipate, not just react.

Prioritisation can be done alone if there is not explicit management interest.

It will surprise you what management are prepared to let slide – this can save you considerable time and angst.

Act


Having chosen your first target, create time to achieve it. This requires discipline and focus. Every day you will have to purposefully make time to progress your goal. This means for a short period spending more time working or postponing

Do not choose large single projects initially, break them into small sub projects.

When you start, schedule both regular reviews and a ‘drop-dead’ review meeting – a time by which if you haven’t made appreciable progress on your task to review

Review

How did it go? Did you achieve what you wanted? Importantly, have you uncovered additional tasks? Are some tasks you’ve identified not necessary.

If your managers are involved, regular meetings to summarise and report on progress and obstacles will keep both you and them focussed and motivated.

‘Lightweight’, low time-impact processes are the watchword here. You are trying to regain ‘freeboard’, you do NOT need additional millstones dragging you further into the quagmire.

Iterate

Choose what to do next. If you’ve identified extra or unnecessary work items, re-analyse.

When do you stop this emergency management mode? When you’ve gained enough freeboard to work effectively.

A short time after the systems are back in control and you are working (close to) normal hours, you should consider scheduling a break. You’ve been overworking for some time and have lost motivation and effectiveness. A break should help you freshen up, gain some perspective and generate ideas for what to do next.

Maintain

What are you and your managers going to do to keep on top of things? How did you slide into the ‘tar pit’ in the first place? What measures or indicators are available to warn if this repeats.

How will you prevent continuous overload from recurring?

2007/06/02

Commercial Software is Good - because you Have Someone To Sue.

A friend came back from a ITIL Practitioners course with an interesting story:
  • The course was mostly about the complexities of handling Commercial Licenses, no two of which are the same.
  • The course provider made the not unexpected statement about Open Source:
    "Don't use it because you have nobody to sue".
    And they went onto ask "Why use OSS?"
His response was: "Because it's best of breed, especially rock-solid utilities & tools".
And they continued to not listen...

This note is NOT about that particular mindset [Risk Avoidance, not Risk Management].

I'd like to give him, and other technical people like him, a "slam dunk" one-liner response to each of these questions:
  • Why OSS - because it's best of breed!
    Why use a bug-ridden, poor functioning piece of commercial software when the best there is is rock-solid, secure & Free and Open?
    Not only do you remove the need to sue *anybody*, you get the best tool for the job and know it will never be orphaned, withdrawn or torpedoed.

    Or you may be held to ransom with enormous support costs - the Computer Associates model of buying 'mature' software and raising the support costs to turn a profit until the customer base bails out.
  • Using rock-solid OSS apps. means you are unlikely to need to sue anybody. It "just works", not "works just".
    And if you have concerns over "prudent commercial Risk Management",just hire an OSS support organisation who's got both "Professional Indemnity" and OSRM insurance.

And I tried to quickly find *two* significant lists for him:
  • widely used open-source software [Apache, Samba, Perl-PHP-Python-Ruby, gcc/make/cvs/subvers, Eclipse, ...]

    The caveat on this list, is that I need estimates of the market share or extent of use of the software. Viz: For apache, the netcraft survey:

  • OSS support organisations. [remember Linuxare ?]
If you have pointers to that data, I've love to hear from you.

Who can we sue? Or - the Myth of Riskless I.T. Management

This started as a conversation on an Open Source list - on how to respond to people that assert:

"We can't use Open Source because there's Nobody to Sue".

2007/06/01

Why "IT Service Delivery and Management" should be an Academic Discipline

IT Service Delivery is where "the rubber hits the road". Without sufficiently capable and reliable IT infrastructure every other piece of the IT equation - Architecture and Design, Business Analysis, Project Management, Software Engineering, Software Maintenance, Information Management, Data Modelling. ... - becomes irrelevant. All other effort is wasted is the services don't run for the users.



All the potential benefits of IT, the 'cognitive amplifier' effects and leveraging people's skills and experience, rely on the delivery of IT Services.



Where is the academic discipline that:

  • Defines IT Service Delivery (and it's OAM components - Operations, Administration and Maintenance)
  • Provides a framework to compare and audit the performance of IT Service Delivery in an organisation to benchmarks relative for the industry.

  • Defines absolute and relative performance of individuals, teams and IT organisations.
  • Defines and explores performance, effectiveness, utilisation and 'service benefit conversion' metrics?



If the goal of IT/IS is to "deliver a business benefit" - but the benefits aren't trivially measurable, the deep knowledge/experience of the discipline of Marketing can be brought to bear. The first step in every project is to define 'the desired benefit', how it will be measured and reported, and the break-even or cancel point.



The academic discipline that informs practitioners, management and the profession on how to actually realise the benefits of IT/IS systems in practice.





ITIL and ISO 20,000



"ITIL" (IT Infrastructure Library) was first formulated in 1989 by the UK OGC (Office of Govt. Computing) to provide a common language and framework or conceptual model for IT Operations (now 'service management') - necessary for the process of tendering and outsourcing.



In 1999 it was re-released ('version 2') and new books written.

Mid-2007 sees the release of 'version 3' - another rethink and rewrite.



In 2000 ITIL spawned a British Standard, BS15000, revised and updated in 2002. In Dec 2005 BS15000 was adopted internationally as IEC/ISO20,000. It sits alongside "Information Security Management" ISO17799 [formerly BS7799:2002] and "Information Security" ISO27001. BS25999 address "Business Continuity".



Forrester Research in early 2007 reports (in "CIOs: Reduce Cost By Scoring Applications" by Phil Murphy) that 'IT Service Delivery' (Forrester calls it "lights on" operations and maintenance) is accounting for a rising percentage of IT budgets. Referenced in "Maintenance Pointers".



Reasons for a discipline of IT Service Delivery and Management



  • The Forrester survey of October 2006 reports IT Service Delivery consumes 80% of more of IT budgets - up from 60-65% ten years ago.
  • 100% of the User Utilisation of IT Software, Systems and Services is mediated by Service Delivery. It's where "the rubber hits the road".

  • IT is useful in business because it's a cognitive amplifier - it amplifies the amount of useful work that people can perform/process. IT provides "cheaper, better, faster, more, consistent/correct".
  • Business and Government are now dependent on their IT. We've crossed the event horizon where [in the 'developed' world] it's possible to resile from IT systems.
  • IT is arguably still the greatest single point of leverage [staff effectiveness amplifier] available to organisations.
  • Service Delivery is the anchor point for "What Value does IT Deliver?"



Where the 'IT Services' discipline belongs

There are two requirements for a faculty teaching 'IT Services' or 'IT Service Management'

  • Business and Management focus, and
  • Ready access to large, complex "IT Services" installations
Traditional computing and IT faculties are focussed on the internal technical aspects of computing. 'IT Services and Management' is about delivering and realising Business Benefits - the managerial focus. The necessary disciplines and knowledge/expertise already exist in Business/Commerce/Management Schools - and are somewhat foreign to traditional CS/ISE/IT Schools.



Canberra, accounting for 20% of the IT expenditure in Australia, is well placed to initiate 'IT Service Delivery and Management' in this country.