Difference between revisions of "Performance testing, automating"
From CitconWiki
Jump to navigationJump to search (Created page with "Foo.") |
|||
| Line 1: | Line 1: | ||
| − | + | TPCH | |
| + | * Improving an analytics platform | ||
| + | * unnamed database vendor, data, scale, queries against data | ||
| + | |||
| + | Write performance stories | ||
| + | * We have a speculation that makes a story | ||
| + | * Some teams are unable to estimate a story for performance | ||
| + | * PO should give acceptance criteria performance | ||
| + | * Should each story allow for x time to find performance issues | ||
| + | |||
| + | Make it work | ||
| + | * Make it work well | ||
| + | * Make it work fast | ||
| + | |||
| + | JMeter | ||
| + | * How does a process scale with # of users, throughput | ||
| + | * Finding limits of the system, 20 second responses, for example, is that too slow? or is that part of subjective metrics? | ||
| + | * Definition of | ||
| + | |||
| + | Solution: | ||
| + | * Set baselines for metrics | ||
| + | * At each release, we want to find out if we slowed down or sped up. | ||
| + | |||
| + | When does do we evaluate performance? | ||
Revision as of 13:01, 24 August 2013
TPCH
- Improving an analytics platform
- unnamed database vendor, data, scale, queries against data
Write performance stories
- We have a speculation that makes a story
- Some teams are unable to estimate a story for performance
- PO should give acceptance criteria performance
- Should each story allow for x time to find performance issues
Make it work
- Make it work well
- Make it work fast
JMeter
- How does a process scale with # of users, throughput
- Finding limits of the system, 20 second responses, for example, is that too slow? or is that part of subjective metrics?
- Definition of
Solution:
- Set baselines for metrics
- At each release, we want to find out if we slowed down or sped up.
When does do we evaluate performance?