在 Tornado 中获取 python 子进程作业的输出

75 阅读2分钟

在 Tornado 中,您希望从正在运行的 python 子进程中获取输出,就像 Travis CI 中一样。您将在管理页面中启动作业,服务器将收到请求并启动子进程。此子进程将执行一些数据挖掘并将一些日志馈送到字符串缓冲区。您将通过带有 settimeout 或 websocket 的 Ajax 获取这些日志,并将这些日志输出到页面上。即使用户关闭页面并稍后返回,也将有日志记录,并且通常会更新。

2. 解决方案

使用这种方法来实现: stefaanlippens.net/python-asyn…

博文中展示了一种在异步读取进程的输出时防止死锁的方法,该方法是通过异步读取 stdout 和 stderr 来实现的。您可以将 main 中的 producer 命令替换为您喜欢的任何命令,并将 print 语句替换为 Tornado 中处理输的代码。该解决方案包括一个 AsynchronousFileReader 类,用于在单独的线程中异步读取标准输出和错误管道,并将每行放在队列中。主线程可以通过监视队列中出现的行来监视子进程。

代码示例:

import sys
import subprocess
import random
import time
import threading
import Queue

class AsynchronousFileReader(threading.Thread):
    '''
    Helper class to implement asynchronous reading of a file
    in a separate thread. Pushes read lines on a queue to
    be consumed in another thread.
    '''

    def __init__(self, fd, queue):
        assert isinstance(queue, Queue.Queue)
        assert callable(fd.readline)
        threading.Thread.__init__(self)
        self._fd = fd
        self._queue = queue

    def run(self):
        '''The body of the tread: read lines and put them on the queue.'''
        for line in iter(self._fd.readline, ''):
            self._queue.put(line)

    def eof(self):
        '''Check whether there is no more content to expect.'''
        return not self.is_alive() and self._queue.empty()

def consume(command):
    '''
    Example of how to consume standard output and standard error of
    a subprocess asynchronously without risk on deadlocking.
    '''

    # Launch the command as subprocess.
    process = subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE)

    # Launch the asynchronous readers of the process' stdout and stderr.
    stdout_queue = Queue.Queue()
    stdout_reader = AsynchronousFileReader(process.stdout, stdout_queue)
    stdout_reader.start()
    stderr_queue = Queue.Queue()
    stderr_reader = AsynchronousFileReader(process.stderr, stderr_queue)
    stderr_reader.start()

    # Check the queues if we received some output (until there is nothing more to get).
    while not stdout_reader.eof() or not stderr_reader.eof():
        # Show what we received from standard output.
        while not stdout_queue.empty():
            line = stdout_queue.get()
            print('Received line on standard output: ' + repr(line))

        # Show what we received from standard error.
        while not stderr_queue.empty():
            line = stderr_queue.get()
            print('Received line on standard error: ' + repr(line))

        # Sleep a bit before asking the readers again.
        time.sleep(.1)

    # Let's be tidy and join the threads we've started.
    stdout_reader.join()
    stderr_reader.join()

    # Close subprocess' file descriptors.
    process.stdout.close()
    process.stderr.close()

def produce(items=10):
    '''
    Dummy function to randomly render a couple of lines
    on standard output and standard error.
    '''
    for i in range(items):
        output = random.choice([sys.stdout, sys.stderr])
        output.write('Line %d on %s\n' % (i, output))
        output.flush()
        time.sleep(random.uniform(.1, 1))

if __name__ == '__main__':
    # The main flow:
    # if there is an command line argument 'produce', act as a producer
    # otherwise be a consumer (which launches a producer as subprocess).
    if len(sys.argv) == 2 and sys.argv[1] == 'produce':
        produce(10)
    else:
        consume(['python', sys.argv[0], 'produce'])