merge from zack

This commit is contained in:
soulgalore 2014-02-11 18:56:14 +01:00
commit 73e1808f46
5 changed files with 22 additions and 13 deletions

View File

@ -1,5 +1,4 @@
CHANGELOG sitespeed.io
version 2.6-wip
------------------------
* Upgraded to YSlow 3.1.8 (with configurable CDN), removed PhantomJS har files (will be included using BrowserTime) and with basic auth support
@ -10,12 +9,18 @@ version 2.6-wip
* Bug fix: If an error happens when crawling, log that to the error.log #378
* Bug fix: User defined measurements get Velocity code as description #366
* Bug fix: Show one decimal for time metrics on pages #363
* Bug fix: Connect via any SSL protocol #379 thanks @tollmanz
version 2.5.6
------------------------
* New BrowserTime version 0.6 that fix crash while trying to run resource timing measurements in Firefox,
see the list of changes here: https://github.com/tobli/browsertime/releases/tag/browsertime-0.6
version 2.5.5
------------------------
* Bug fix: New version of the crawler, the proxy support was broken in the last release
* Added proxy support when collecting Navigation Timing metrics #351
* Added support for local configuration wher you can override default configuration (thanks @AD7six)
* Added support for local configuration where you can override default configuration (thanks @AD7six)
version 2.5.4
------------------------

View File

@ -3,16 +3,18 @@ Sitespeed.io continues to evolve thanks to people who contributes, so please ple
## Making changes
If you want help out, that's great! Check the issue list and see if there's something there you want to do and drop me a note so I know and I can help you get into the project.
One **important** thing: The YSlow file `dependencies/yslow-X.Y.Z-sitespeed.js` is generated from the project https://github.com/sitespeedio/yslow meaning all pull requests for that file
needs to be sent to that project and then we will include the finalized javascript file in this project.
## Add a defect
First, check the [defect/bug list](https://github.com/sitespeedio/sitespeed.io/issues?labels=bug&page=1&state=open) to make sure that it hasn't been filed yet.
If you find a defect, please file a bug report. Include the following:
- Explain the bug/defect and what you where doing.
- OS & versions
- Always add the URL of the page you where analyzing (if it is secret, drop me an email peter**at**soulgalore.com and send me the address).
- Add a screenshot and clearly point out where the defect is.
- Add a log of the output and include the content of the data/error.log file in a [gist](https://gist.github.com/) and attach it to the issue.
- OS & versions
- Always add the URL of the page you where analyzing (if it is secret, drop me an email peter**at**soulgalore.com and send me the address).
- Add a screenshot and clearly point out where the defect is (if applicable)
- Include the content of the data/error.log file in a [gist](https://gist.github.com/) and attach it to the issue.
If you have the skills & the time, it is perfect if you send a pull request with a fix, that helps me a lot!
@ -22,4 +24,4 @@ If you have an idea or something that you need sitespeed.io to handle, add an is
Thanks for your time & support!
Peter
Peter & Tobias

View File

@ -7,4 +7,7 @@ Many many many thanks to:
* Peter Van Gils (@kab12)
* Juga Paazmaya (@paazmaya)
* Ajay Jain
* Andy Dawson (@AD7six)
* Andy Dawson (@AD7six)
* Chris Southam (@ChrisSoutham)
* JoongSeob Vito Kim (@dorajistyle)
* Zack Tollman (@tollmanz)

View File

@ -8,6 +8,7 @@ You can use it on Mac, Linux & Windows (using Git Bash). And collect timing metr
Run on Mac OS X
=============
```bash
$ brew tap tobli/browsertime
$ brew install sitespeedio/sitespeedio/sitespeed.io
$ sitespeed.io -h
```

View File

@ -781,8 +781,7 @@ for url in "${URLS[@]}"
do
local imagefilename=$(get_filename $url $runs)
echo "Creating screenshot for $url $REPORT_IMAGE_PAGES_DIR/$imagefilename.png "
phantomjs --ignore-ssl-errors=yes $PROXY_PHANTOMJS $DEPENDENCIES_DIR/screenshot.js "$url" "$REPORT_IMAGE_PAGES_DIR/$imagefilename.png" $width $height "$USER_AGENT" true $BASIC_AUTH_USER_PASSWORD > /dev/null 2>&1
phantomjs --ssl-protocol=any --ignore-ssl-errors=yes $PROXY_PHANTOMJS $DEPENDENCIES_DIR/screenshot.js "$url" "$REPORT_IMAGE_PAGES_DIR/$imagefilename.png" $width $height "$USER_AGENT" true $BASIC_AUTH_USER_PASSWORD > /dev/null 2>&1
if [ "$PNGCRUSH_EXIST" = "true" ]
then
pngcrush -q $REPORT_IMAGE_PAGES_DIR/$imagefilename.png $REPORT_IMAGE_PAGES_DIR/$imagefilename-c.png
@ -858,8 +857,7 @@ function analyze() {
echo "Analyzing $url"
## Removing HAR functionality from phantomjs, will be included in browsertime -n "$REPORT_DATA_HAR_DIR/$pagefilename.har"
phantomjs --ignore-ssl-errors=yes $PROXY_PHANTOMJS $YSLOW_FILE -d -r $RULESET $BASIC_AUTH_PHANTOMJS -f xml $CDN --ua "$USER_AGENT_YSLOW" $VIEWPORT_YSLOW "$url" >"$REPORT_DATA_PAGES_DIR/$pagefilename.xml" 2>> $REPORT_DATA_DIR/phantomjs.error.log || echo "PhantomJS could not handle $url , check the error log: $REPORT_DATA_DIR/phantomjs.error.log"
phantomjs --ssl-protocol=any --ignore-ssl-errors=yes $PROXY_PHANTOMJS $YSLOW_FILE -d -r $RULESET $BASIC_AUTH_PHANTOMJS -f xml $CDN --ua "$USER_AGENT_YSLOW" $VIEWPORT_YSLOW "$url" >"$REPORT_DATA_PAGES_DIR/$pagefilename.xml" 2>> $REPORT_DATA_DIR/phantomjs.error.log || echo "PhantomJS could not handle $url , check the error log: $REPORT_DATA_DIR/phantomjs.error.log"
local s=$(du -k "$REPORT_DATA_PAGES_DIR/$pagefilename.xml" | cut -f1)
# Check that the size is bigger than 0
if [ $s -lt 10 ]
@ -868,7 +866,7 @@ function analyze() {
## do the same thing again but setting console to log the error to output
log_error "Could not analyze $url unrecoverable error when parsing the page"
log_error "Input parameters: $INPUT"
phantomjs --ignore-ssl-errors=yes $PROXY_PHANTOMJS $YSLOW_FILE -d -r $RULESET -f xml "$USER_AGENT_YSLOW" $VIEWPORT_YSLOW "$url" -c 2 2>&1 >> $REPORT_DATA_DIR/$ERROR_LOG
phantomjs --ssl-protocol=any --ignore-ssl-errors=yes $PROXY_PHANTOMJS $YSLOW_FILE -d -r $RULESET -f xml "$USER_AGENT_YSLOW" $VIEWPORT_YSLOW "$url" -c 2 2>&1 >> $REPORT_DATA_DIR/$ERROR_LOG
## write the error url to the list
echo "sitespeed.io got an unrecoverable error when parsing the page,$url" >> $REPORT_DATA_DIR/errorurls.txt