THIS SECTION IN PROGRESS AS OF 11/29/2006 - Tom Brennan
Writing the report is easy if you have been collecting the information during each stage of testing. If is important for the customer to always get a systematic assessment looking for classes of issues in unknown developer code.
Questions I always ask is, is this project a baseline report or the 12th report of a single application/system in the current year?
Either way the report needs to be consistent in the delivery to customers, management etc. The reason this is important is that the 1st rule of security is to use multiple vendors for testing security right, but unless you specify you threat model type (see 5.1 How to value the real risk ) each report will have its own interpretation of risk and threats. If it does happen to be the 5 report on the same system, I typically submit a example report, and then ask the customer to compare apples to apples. If he liked certain areas of the other firms report, I want to review how it will look on our report ahead of time. Its all about meeting expectations. On the technical level, I am a fan of using the VISA/Mastercard PCI standard for host reporting.
The following sections are typically standard
I. Executive Summary - It this report is for many systems then there should be a summary paragraph that outlines the total devices tested, the dates of the testing and a summary of the results. I think we all know that management likes visual graphics to illustrate problems so why not provide some to do that for all of the systems tested.
II. Systems Summary – If more than 1 application or system is tested, we recommend that the systems summary defined the systems and the systems boundaries and a description of the systems purpose. Each “System” may contain many sub-systems etc., it is important to provide both Executive Summary about systems as well as within the systems summary. Appendix A, below would be ideal to illustrate per system the status of a Application Assessment and if you were to add a few items, you could include the host/network level as well.
III System Detail – This is a details on each system test and details of the individual hosts or applications again Appendix A would serve this purpose as a summary on a individual system level and then detailed information and recommended corrective actions to the found issues.
Within the system detail I like to utilize the following sections:
Observations Findings Recommendations
Keeping in mind as an assessment person that you can find the biggest issues in the world, but it is the customer/employer etc... That has to accept the facts and elect to fix them. So the report is exactly that, as a subject matter expert you are providing the facts good or bad. Think in context of testifying in a court of law as to what you observed, what you found and what you recommended this document is your professional deliverable.
IIII Toolbox This is the area that is used to describe to the customer the commerical and open-source tools that were used in conducting the assessment. When custom scripts/code are utilized during the assessment, it should be disclosed in this section. The customer paid for the assessment and should be provided with a list of the tools used to provide the service in addition to the digital knowledge of the tester.
|Category||Ref Number||Name||Finding||Affected Item||Comment/Solution||Risk Value|
|Information Gathering||Application Discovery|
|Spidering and googling|
|Analisys of error code|
|DB Listener Testing|
|File extensions handling|
|Old, backup and unreferenced files|
|Business logic testing|
|Authentication Testing||Default or guessable account|
|Bypassing authentication schema|
|Directory traversal/file include|
|Vulnerable remember password and pwd reset|
|Logout and Browser Cache Management Testing|
|Session Management Testing||Session Management Schema|
|Session Token Manipulation|
|Exposed Session Variables|
|Data Validation Testing||Cross site scripting|
|HTTP Methods and XST|
|Stored procedure injection|
|Denial of Service Testing||Locking Customer Accounts|
|User Specified Object Allocation|
|User Input as a Loop Counter|
|Writing User Provided Data to Disk|
|Failure to Release Resources|
|Storing too Much Data in Session|
|Web Services Testing||XML Structural Testing|
|XML content-level Testing|
|HTTP GET parameters/REST Testing|
|Naughty SOAP attachments|
|AJAX Testing||AJAX Vulnerabilities|
OWASP Testing Guide v2
Here is the OWASP Testing Guide v2 Table of Contents