through ok which is handy as Bobby (my other test colleague) doesn't want to deploy the latest code to our QA environment just yet as he is in the middle of a meaty test. So in the meantime I can check the story Mahmesh finished yesterday afternoon on my local environment. Had a brief look at his unit tests and then double-checked the selenium test implementation. I had written the draft of the test upfront so I was happy with the flow of the test but I needed to make sure Mahmesh was using the correct data and values and he hadn't made any significant changes to the test. Of course if he had I'm pretty confident he would have made me aware beforehand. Unfortunately he neglected to call me over to do a quick walk through of the completed test and story prior to him checking in yesterday but I forgive him as he did fix my build.
Was actually a very productive day in the end. Besides us three testers there was only a couple of devs in. So no story huddles, planning sessions, contention on environments or any of the other distractions we normally have in the week. We also had the CI build pretty much to ourselves. It was green all day so we were able to check in a few selenium script changes. Bobby broke the build at one point but we managed to figure out the problem fairly quickly (we needed to reset some test data at the end of an automated test). I do enjoy the 'buzz' of an agile project but sometimes its nice to just put our heads down and test without being disturbed.
Wednesday–Day 41–Iteration 9
Lots of preparation sessions going on today for the next iteration. This is where the business analysts and the product owner (along with the project manager, iteration manager and other interested parties) discuss what stories they would like to see the devs work on next. If the story is ready for play then it will be considered for the upcoming iteration planning session. We try to send at least one tester to these sessions so we can get a bit of a heads up as to what may be coming next and then work on completing our test scenarios and draft automation tests against each stories acceptance criteria. On the Friday during the iteration planning session the devs decide which of the selected stories they can complete against their allocated budgets (pair days). We attend that one too, which can be long and torturous as it involves a lot of 'discussion' on the estimate for each story and the brief implementation. We also give a short summary on how we expect the story to be tested. I guess it's good to have this debate with your peers up front but I’m always glad when the Iteration manager bangs the table three times before everyone puts their 'fingers in the wind' to estimate how many days it will take to complete the story.
As we came out of one prep session Lindsay started having kittens about the index cards on the wall. We have hundreds of cards with all our stories, task and defects split across our various workflows. She's convinced that someone is moving them around. 'These cards should be in the dev stream not the QA stream,' she said.'They haven't finished developing these stories yet.' She wasn't too impressed when I mentioned the possibility that the cards may have fallen off the wall and the cleaner may have stuck them back up