Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
20 changes: 17 additions & 3 deletions docker/client.py
Original file line number Diff line number Diff line change
Expand Up @@ -209,9 +209,23 @@ def _get_raw_response_socket(self, response):

def _stream_helper(self, response):
"""Generator for data coming from a chunked-encoded HTTP response."""
for line in response.iter_lines(chunk_size=32):
if line:
yield line
socket_fp = self._get_raw_response_socket(response)
socket_fp.setblocking(1)
socket = socket_fp.makefile()
while True:
# Because Docker introduced newlines at the end of chunks in v0.9,
# and only on some API endpoints, we have to cater for both cases.
size_line = socket.readline()
if size_line == '\r\n':
size_line = socket.readline()

size = int(size_line, 16)
if size <= 0:
break
data = socket.readline()
if not data:
break
yield data

def _multiplexed_buffer_helper(self, response):
"""A generator of multiplexed data blocks read from a buffered
Expand Down
6 changes: 3 additions & 3 deletions tests/integration_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,7 @@

import time
import base64
import json
import io
import os
import signal
Expand Down Expand Up @@ -666,10 +667,8 @@ def runTest(self):
self.assertIn('Images', info)
img_count = info['Images']
stream = self.client.pull('joffrey/test001', stream=True)
res = u''
for chunk in stream:
res += chunk
self.assertEqual(type(res), six.text_type)
json.loads(chunk) # ensure chunk is a single, valid JSON blob
self.assertEqual(img_count + 3, self.client.info()['Images'])
img_info = self.client.inspect_image('joffrey/test001')
self.assertIn('id', img_info)
Expand Down Expand Up @@ -762,6 +761,7 @@ def runTest(self):
stream = self.client.build(fileobj=script, stream=True)
logs = ''
for chunk in stream:
json.loads(chunk) # ensure chunk is a single, valid JSON blob
logs += chunk
self.assertNotEqual(logs, '')

Expand Down