15
DEALING WITH LARGE DATA SET AND COMPLEXITY IN YOUR TESTING Jae-Jin Lee

Dealing with large data set and complexity in your testing

  • Upload
    renate

  • View
    44

  • Download
    0

Embed Size (px)

DESCRIPTION

Dealing with large data set and complexity in your testing. Jae-Jin Lee . Search results (Facts). Google/Bing Possible number of inputs is close to infinity There are huge amount of data source Algorithms (placement) are very complex Expedia - PowerPoint PPT Presentation

Citation preview

Page 1: Dealing with large data set and complexity in your testing

DEALING WITH LARGE DATA SET AND COMPLEXITY IN YOUR TESTINGJae-Jin Lee

Page 2: Dealing with large data set and complexity in your testing

Search results (Facts) Google/Bing

Possible number of inputs is close to infinity There are huge amount of data source Algorithms (placement) are very complex

ExpediaPossible inputs are not as huge as Google, but the

same input can return different results based on dates, traveler info and other factors.

There are huge amount of inventoriesAlgorithms are very complex and direct impact to the business

Page 3: Dealing with large data set and complexity in your testing

Search results (Facts) Google/Bing

Possible number of inputs is close to infinity There are huge amount of data source Algorithms (placement) are very complex

Expedia Possible inputs are not as huge as Google, but the

same input can return different results based on dates, traveler info and other factors.

There are huge amount of data source(world, inventories)

Algorithms are very complex and direct impact to the business

Page 4: Dealing with large data set and complexity in your testing

Testing Challenges Test input selection

Data is not organized in a way to be tested Equivalent partitioning is hard Randomness? Coverage?

Verifying mechanism How do we get expected result? RE-implement the algorithm? 474,000,000 results for "Seattle“ search

Page 5: Dealing with large data set and complexity in your testing

Good news Algorithms are complex but defined We have a full access to data source Historical data/statistics are available Not all the results are equally important

Page 6: Dealing with large data set and complexity in your testing

Risk analysis / assessment

Page 7: Dealing with large data set and complexity in your testing

Risk analysis / assessment Question the project (Is it feasible to do it?)Practical risk analysis

Understand the risk on business perspectiveUnderstand the likelihood of faults from

development perspectiveValidations to be doneTest cases

Come up with list and reviewed by entire project team

Page 8: Dealing with large data set and complexity in your testing

Risk analysis / assessment Question the project (Is it feasible to do it?) Practical risk analysis

Understand the risk from business perspective Understand the likelihood of faults from

development perspective Test cases Validations to be done

Come up with list and reviewed by entire project team

Page 9: Dealing with large data set and complexity in your testing

Risk analysis / assessment Question the project (Is it feasible to do it?) Practical risk analysis

Understand the risk from business perspective Understand the likelihood of faults from

development perspective Test cases Validations to be done

Summarized it and reviewed by entire project team

Page 10: Dealing with large data set and complexity in your testing

Understand data source Data source is trusted source of test

case validation / verification mechanism Modifying data source should be piece of

cake Insert, delete, update rows or execute

sprocs Setup and tear down

No assumption on data source

Page 11: Dealing with large data set and complexity in your testing

Test input selection Historic data and statistics Priority from risk analysis Creativity and product knowledge to

break Radom valid inputs from bucketing (do

as much as you can and log the useful details)

Hard-coded data

Page 12: Dealing with large data set and complexity in your testing

Decompose the algorithm Exercise each logic separately by

controlling data source and dependencies

Working with dev for testability or hooks (architecture, logs, and etc.)

If possible, implement algorithms for happy path in your test automation

Page 13: Dealing with large data set and complexity in your testing

Heuristic approach helps Is there a place where good enough

result acceptable? Seatttle (three ‘t’s) Is that in the list? Is that in the first 10 results?

Page 14: Dealing with large data set and complexity in your testing

Hybrid approach (manual + automation)

Integration environment Combine human’s intuition/product

knowledge and machine’s powerful diligence

Execute manually and validate using test validation code (turning on logs)

Requires decoupled class design in your automation

UI(JavaScript) broke the functionality

Page 15: Dealing with large data set and complexity in your testing

Your thought?