|
|
|
The read timeout has already been set to 1 hour, and if it still result in a read timeout, please examine your artifact publish step to make sure not too many files are being published all together. Also it can take a lot of time if you specify a pattern searching in a large directory.
We use QB in a corporate environment, and our builds produce relatively large images which need to be transferred back to our company. This might change after
Therefore the best option would be exposing the value as a configuration parameter. OK. Will make it configurable in next patch release.
Checked this issue again. Although we can make this configurable, but setting it to a very high value can cause artifact sending threads not being able to quit timely in case of network connection problems.
Also the socket read timeout does not mean that the file transfer has to be finished within that period of time. It means the maximum time when no data is being sent. So the reason for the timeout should not be caused by the large artifact, instead, it should be caused by the factor that QB is spent a lot of time searching files with the specified patterns in specified directory, before sending any data to server. For instance, if workspace contains many many files, and if you specify pattern as "**/*.zip" will cause QB spending a lot of time searching for zip file recursively in workspace, however, if you can make sure that your zip files sits inside several small directories, you can limit the search scope by specifying pattern as for instance: dir1/*.zip, dir2/*.zip If it is caused due to search time for patterns, I would expect similar behaving in setup without EC2 as well. But without EC2 we don't see this pattern of socket time out.
I agree with Maikel that it is not about IO operations. Because indeed IO operation on Amazon ephemeral storage are slower, but not that slow. In this case we are talking about publishing a directory containing built images and various build artifacts. In total 2194 files (587 MB).
When I executed _tar_ utlility to create and archive. With compression it took 1 m 20 sek. Without - 5 secs. And the publish step breaks exactly after 1 hour, after transferring 100-300 MB. Can you please send me full log of server and the agent running the artifact publish step (exists in logs directory of server and agent installation) when this error happens? Please also send me the build log of that failed build.
If possible, please upgrade to 5.1.7, and enable debug logging (conf/log4j.properties) on server and the problematic agent and then collect relevant logs when the timeout occurs.
Hi Robin, we need to schedule the upgrade, so it is not so easy to get to 5.1.7. I will provide you with logs from 5.1.6, unless you think the extended logging feature is essential in this case.
BTW we have raised several EC2 related requests: do you think there is a change to have them addressed in 5.1.8 ? This would make for me easier getting all the approvals for QB downtime related to upgrade. Regards, Lukasz We need to investigate these features and priority them among other features. For now, we can not give an estimation of the date getting them implemented.
|
https://gist.github.com/anonymous/8282692#file-gistfile1-txt
Could you please instruct how to configure the timeout value?