Skip to main content

The Obscure One

Heraclitus wrote these words 2500 years ago: "Ever-newer waters flow on those who step into the same rivers." or paraphrased in more colloquial English: You never stand in the same river twice.


Known as the "The obscure one" to some of his contemporaries, he was known to make statements that were considered paradoxical and sometimes unhelpfully contradictory. I don't know about you  - but sometimes when discussing testing feedback - I feel like I am channeling the ghost of Heraclitus.

His comments regarding walking through rivers are an apt description of our work with software and its versioning. Do we ever play with the same app twice? On a trivial level, we do. When we widen our view we can see that the waters have moved on.

For example, 

  • The time has changed. It may even have gone back to a previous date and time. 
  • The code is probably located in a different memory location. 
  • The app and operating system are probably facing different types of automated attack from various pieces of malware on the same device and across the network
  • Your typing / swiping speed has changed
  • Libraries used by the app may have been patched
  • The OS has probably been patched
  • Your training/test datasets' random split is slightly different
  • Other programs or apps have probably been patched.

Etc.

Add to this the changes in your test tools and test data, then an application some might naively see as the 'same version', can behave quite differently. 

So when someone says a test is flaky or the app is flaky, think about the river of change that's running through your app. Smile to yourself, and know that people have been explaining what you are about to explain for at least 2500 years. 

Comments

Popular posts from this blog

The gamification of Software Testing

A while back, I sat in on a planning meeting. Many planning meetings slide awkwardly into a sort of ad-hoc technical analysis discussion, and this was no exception. With a little prompting, the team started to draw up what they wanted to build on a whiteboard.

The picture spoke its thousand words, and I could feel that the team now understood what needed to be done. The right questions were being asked, and initial development guesstimates were approaching common sense levels.

The discussion came around to testing, skipping over how they might test the feature, the team focused immediately on how long testing would take.

When probed as to how the testing would be performed? How we might find out what the team did wrong? Confused faces stared back at me. During our ensuing chat, I realised that they had been using BDD scenarios [only] as a metric of what testing needs to be done and when they are ready to ship. (Now I knew why I was hired to help)



There is nothing wrong with checking t…

Software development is in the Doldrums

"Don't get off the boat."

"Seriously, never get off the boat," The instructor said, leaning forward and looking at each of us in turn.

"But surely if it's sinking..." We reply, somewhat confused and slightly incredulous. We've seen Titanic, we think to ourselves, we know how this sea survival stuff works...

"OK" He concedes, If things get really bad, "Get on the life raft if you can step-up from the boat to the life raft".

"But, But... the yacht is like 37ft long, Do we want to wait until that whole boat is lower than the life-raft? When less than 1ft of the yacht is above the surface? Meanwhile all the time the life raft is just there... floating happily alongside."

"Pretty much, yes," he said nodding.


That was about 15 years ago. Not much has changed since. The reasons are manifold. Firstly, the yacht is a decent shelter. The thin plastic of a legal minimum life-raft isn't going to protect you fro…

A h̶i̶t̶c̶h̶h̶i̶k̶e̶r̶'s̶ software tester's guide to randomised testing - Part 1

Mostly Harmless, I've talked and written about randomisation as a technique in software testing several times over the last few years. It's great to see people's eyes light up when they grok the concept and its potential. 
The idea that they can create random test data on the fly and pour this into the app step back and see what happens is exciting to people looking to find new blockers on their apps path to reliability.
But it's not long before a cloud appears in their sunny demeanour and they start to conceive of the possible pitfalls. Here are a few tips on how to avert the common apparent blockers. (Part 1) Problem: I've created loads of random numbers as input data, but how will I know the answer the software returns, is correct? - Do I have to re-implement the whole app logic in my test code?
Do you remember going to the fun-fair as a kid? Or maybe you recall taking your kids now as an adult? If so then you no doubt are familiar with the height restriction -…