runtime.py 18 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479
  1. import os
  2. import tempfile
  3. import threading
  4. import time
  5. import uuid
  6. from zipfile import ZipFile
  7. import docker
  8. import requests
  9. import tenacity
  10. from openhands.core.config import AppConfig
  11. from openhands.core.logger import openhands_logger as logger
  12. from openhands.events import EventStream
  13. from openhands.events.action import (
  14. ActionConfirmationStatus,
  15. BrowseInteractiveAction,
  16. BrowseURLAction,
  17. CmdRunAction,
  18. FileReadAction,
  19. FileWriteAction,
  20. IPythonRunCellAction,
  21. )
  22. from openhands.events.action.action import Action
  23. from openhands.events.observation import (
  24. ErrorObservation,
  25. NullObservation,
  26. Observation,
  27. UserRejectObservation,
  28. )
  29. from openhands.events.serialization import event_to_dict, observation_from_dict
  30. from openhands.events.serialization.action import ACTION_TYPE_TO_CLASS
  31. from openhands.runtime.builder import DockerRuntimeBuilder
  32. from openhands.runtime.plugins import PluginRequirement
  33. from openhands.runtime.runtime import Runtime
  34. from openhands.runtime.utils import find_available_tcp_port
  35. from openhands.runtime.utils.runtime_build import build_runtime_image
  36. class LogBuffer:
  37. """
  38. Synchronous buffer for Docker container logs.
  39. This class provides a thread-safe way to collect, store, and retrieve logs
  40. from a Docker container. It uses a list to store log lines and provides methods
  41. for appending, retrieving, and clearing logs.
  42. """
  43. def __init__(self, container: docker.models.containers.Container):
  44. self.client_ready = False
  45. self.init_msg = 'Runtime client initialized.'
  46. self.buffer: list[str] = []
  47. self.lock = threading.Lock()
  48. self.log_generator = container.logs(stream=True, follow=True)
  49. self.log_stream_thread = threading.Thread(target=self.stream_logs)
  50. self.log_stream_thread.daemon = True
  51. self.log_stream_thread.start()
  52. self._stop_event = threading.Event()
  53. def append(self, log_line: str):
  54. with self.lock:
  55. self.buffer.append(log_line)
  56. def get_and_clear(self) -> list[str]:
  57. with self.lock:
  58. logs = list(self.buffer)
  59. self.buffer.clear()
  60. return logs
  61. def stream_logs(self):
  62. """
  63. Stream logs from the Docker container in a separate thread.
  64. This method runs in its own thread to handle the blocking
  65. operation of reading log lines from the Docker SDK's synchronous generator.
  66. """
  67. try:
  68. for log_line in self.log_generator:
  69. if self._stop_event.is_set():
  70. break
  71. if log_line:
  72. decoded_line = log_line.decode('utf-8').rstrip()
  73. self.append(decoded_line)
  74. if self.init_msg in decoded_line:
  75. self.client_ready = True
  76. except Exception as e:
  77. logger.error(f'Error streaming docker logs: {e}')
  78. def __del__(self):
  79. if self.log_stream_thread.is_alive():
  80. logger.warn(
  81. "LogBuffer was not properly closed. Use 'log_buffer.close()' for clean shutdown."
  82. )
  83. self.close(timeout=5)
  84. def close(self, timeout: float = 10.0):
  85. self._stop_event.set()
  86. self.log_stream_thread.join(timeout)
  87. class EventStreamRuntime(Runtime):
  88. """This runtime will subscribe the event stream.
  89. When receive an event, it will send the event to runtime-client which run inside the docker environment.
  90. """
  91. container_name_prefix = 'openhands-sandbox-'
  92. def __init__(
  93. self,
  94. config: AppConfig,
  95. event_stream: EventStream,
  96. sid: str = 'default',
  97. plugins: list[PluginRequirement] | None = None,
  98. env_vars: dict[str, str] | None = None,
  99. ):
  100. self.config = config
  101. self._port = find_available_tcp_port()
  102. self.api_url = f'http://{self.config.sandbox.api_hostname}:{self._port}'
  103. self.session = requests.Session()
  104. self.instance_id = (
  105. sid + '_' + str(uuid.uuid4()) if sid is not None else str(uuid.uuid4())
  106. )
  107. self.docker_client: docker.DockerClient = self._init_docker_client()
  108. self.base_container_image = self.config.sandbox.base_container_image
  109. self.runtime_container_image = self.config.sandbox.runtime_container_image
  110. self.container_name = self.container_name_prefix + self.instance_id
  111. self.container = None
  112. self.action_semaphore = threading.Semaphore(1) # Ensure one action at a time
  113. self.runtime_builder = DockerRuntimeBuilder(self.docker_client)
  114. logger.debug(f'EventStreamRuntime `{sid}`')
  115. # Buffer for container logs
  116. self.log_buffer: LogBuffer | None = None
  117. if self.config.sandbox.runtime_extra_deps:
  118. logger.info(
  119. f'Installing extra user-provided dependencies in the runtime image: {self.config.sandbox.runtime_extra_deps}'
  120. )
  121. if self.runtime_container_image is None:
  122. if self.base_container_image is None:
  123. raise ValueError(
  124. 'Neither runtime container image nor base container image is set'
  125. )
  126. self.runtime_container_image = build_runtime_image(
  127. self.base_container_image,
  128. self.runtime_builder,
  129. extra_deps=self.config.sandbox.runtime_extra_deps,
  130. )
  131. self.container = self._init_container(
  132. self.sandbox_workspace_dir,
  133. mount_dir=self.config.workspace_mount_path,
  134. plugins=plugins,
  135. )
  136. # will initialize both the event stream and the env vars
  137. super().__init__(config, event_stream, sid, plugins, env_vars)
  138. logger.info(
  139. f'Container initialized with plugins: {[plugin.name for plugin in self.plugins]}'
  140. )
  141. logger.info(f'Container initialized with env vars: {env_vars}')
  142. @staticmethod
  143. def _init_docker_client() -> docker.DockerClient:
  144. try:
  145. return docker.from_env()
  146. except Exception as ex:
  147. logger.error(
  148. 'Launch docker client failed. Please make sure you have installed docker and started docker desktop/daemon.'
  149. )
  150. raise ex
  151. @tenacity.retry(
  152. stop=tenacity.stop_after_attempt(5),
  153. wait=tenacity.wait_exponential(multiplier=1, min=4, max=60),
  154. )
  155. def _init_container(
  156. self,
  157. sandbox_workspace_dir: str,
  158. mount_dir: str | None = None,
  159. plugins: list[PluginRequirement] | None = None,
  160. ):
  161. try:
  162. logger.info(
  163. f'Starting container with image: {self.runtime_container_image} and name: {self.container_name}'
  164. )
  165. plugin_arg = ''
  166. if plugins is not None and len(plugins) > 0:
  167. plugin_arg = (
  168. f'--plugins {" ".join([plugin.name for plugin in plugins])} '
  169. )
  170. network_mode: str | None = None
  171. port_mapping: dict[str, int] | None = None
  172. if self.config.sandbox.use_host_network:
  173. network_mode = 'host'
  174. logger.warn(
  175. 'Using host network mode. If you are using MacOS, please make sure you have the latest version of Docker Desktop and enabled host network feature: https://docs.docker.com/network/drivers/host/#docker-desktop'
  176. )
  177. else:
  178. port_mapping = {f'{self._port}/tcp': self._port}
  179. if mount_dir is not None:
  180. volumes = {mount_dir: {'bind': sandbox_workspace_dir, 'mode': 'rw'}}
  181. logger.info(f'Mount dir: {sandbox_workspace_dir}')
  182. else:
  183. logger.warn(
  184. 'Mount dir is not set, will not mount the workspace directory to the container.'
  185. )
  186. volumes = None
  187. if self.config.sandbox.browsergym_eval_env is not None:
  188. browsergym_arg = (
  189. f'--browsergym-eval-env {self.config.sandbox.browsergym_eval_env}'
  190. )
  191. else:
  192. browsergym_arg = ''
  193. container = self.docker_client.containers.run(
  194. self.runtime_container_image,
  195. command=(
  196. f'/openhands/miniforge3/bin/mamba run --no-capture-output -n base '
  197. 'PYTHONUNBUFFERED=1 poetry run '
  198. f'python -u -m openhands.runtime.client.client {self._port} '
  199. f'--working-dir {sandbox_workspace_dir} '
  200. f'{plugin_arg}'
  201. f'--username {"openhands" if self.config.run_as_openhands else "root"} '
  202. f'--user-id {self.config.sandbox.user_id} '
  203. f'{browsergym_arg}'
  204. ),
  205. network_mode=network_mode,
  206. ports=port_mapping,
  207. working_dir='/openhands/code/',
  208. name=self.container_name,
  209. detach=True,
  210. environment={'DEBUG': 'true'} if self.config.debug else None,
  211. volumes=volumes,
  212. )
  213. self.log_buffer = LogBuffer(container)
  214. logger.info(f'Container started. Server url: {self.api_url}')
  215. return container
  216. except Exception as e:
  217. logger.error('Failed to start container')
  218. logger.exception(e)
  219. self.close(close_client=False)
  220. raise e
  221. @tenacity.retry(
  222. stop=tenacity.stop_after_attempt(10),
  223. wait=tenacity.wait_exponential(multiplier=2, min=10, max=60),
  224. reraise=(ConnectionRefusedError,),
  225. )
  226. def _wait_until_alive(self):
  227. logger.debug('Getting container logs...')
  228. # Print and clear the log buffer
  229. assert (
  230. self.log_buffer is not None
  231. ), 'Log buffer is expected to be initialized when container is started'
  232. # Always process logs, regardless of client_ready status
  233. logs = self.log_buffer.get_and_clear()
  234. if logs:
  235. formatted_logs = '\n'.join([f' |{log}' for log in logs])
  236. logger.info(
  237. '\n'
  238. + '-' * 35
  239. + 'Container logs:'
  240. + '-' * 35
  241. + f'\n{formatted_logs}'
  242. + '\n'
  243. + '-' * 80
  244. )
  245. if not self.log_buffer.client_ready:
  246. attempts = 0
  247. while not self.log_buffer.client_ready and attempts < 5:
  248. attempts += 1
  249. time.sleep(1)
  250. logs = self.log_buffer.get_and_clear()
  251. if logs:
  252. formatted_logs = '\n'.join([f' |{log}' for log in logs])
  253. logger.info(
  254. '\n'
  255. + '-' * 35
  256. + 'Container logs:'
  257. + '-' * 35
  258. + f'\n{formatted_logs}'
  259. + '\n'
  260. + '-' * 80
  261. )
  262. response = self.session.get(f'{self.api_url}/alive')
  263. if response.status_code == 200:
  264. return
  265. else:
  266. msg = f'Action execution API is not alive. Response: {response}'
  267. logger.error(msg)
  268. raise RuntimeError(msg)
  269. @property
  270. def sandbox_workspace_dir(self):
  271. return self.config.workspace_mount_path_in_sandbox
  272. def close(self, close_client: bool = True):
  273. if self.log_buffer:
  274. self.log_buffer.close()
  275. if self.session:
  276. self.session.close()
  277. containers = self.docker_client.containers.list(all=True)
  278. for container in containers:
  279. try:
  280. if container.name.startswith(self.container_name_prefix):
  281. logs = container.logs(tail=1000).decode('utf-8')
  282. logger.debug(
  283. f'==== Container logs ====\n{logs}\n==== End of container logs ===='
  284. )
  285. container.remove(force=True)
  286. except docker.errors.NotFound:
  287. pass
  288. if close_client:
  289. self.docker_client.close()
  290. def run_action(self, action: Action) -> Observation:
  291. # set timeout to default if not set
  292. if action.timeout is None:
  293. action.timeout = self.config.sandbox.timeout
  294. with self.action_semaphore:
  295. if not action.runnable:
  296. return NullObservation('')
  297. if (
  298. hasattr(action, 'is_confirmed')
  299. and action.is_confirmed
  300. == ActionConfirmationStatus.AWAITING_CONFIRMATION
  301. ):
  302. return NullObservation('')
  303. action_type = action.action # type: ignore[attr-defined]
  304. if action_type not in ACTION_TYPE_TO_CLASS:
  305. return ErrorObservation(f'Action {action_type} does not exist.')
  306. if not hasattr(self, action_type):
  307. return ErrorObservation(
  308. f'Action {action_type} is not supported in the current runtime.'
  309. )
  310. if (
  311. hasattr(action, 'is_confirmed')
  312. and action.is_confirmed == ActionConfirmationStatus.REJECTED
  313. ):
  314. return UserRejectObservation(
  315. 'Action has been rejected by the user! Waiting for further user input.'
  316. )
  317. logger.info('Awaiting session')
  318. self._wait_until_alive()
  319. assert action.timeout is not None
  320. try:
  321. response = self.session.post(
  322. f'{self.api_url}/execute_action',
  323. json={'action': event_to_dict(action)},
  324. timeout=action.timeout,
  325. )
  326. if response.status_code == 200:
  327. output = response.json()
  328. obs = observation_from_dict(output)
  329. obs._cause = action.id # type: ignore[attr-defined]
  330. return obs
  331. else:
  332. error_message = response.text
  333. logger.error(f'Error from server: {error_message}')
  334. obs = ErrorObservation(f'Command execution failed: {error_message}')
  335. except requests.Timeout:
  336. logger.error('No response received within the timeout period.')
  337. obs = ErrorObservation('Command execution timed out')
  338. except Exception as e:
  339. logger.error(f'Error during command execution: {e}')
  340. obs = ErrorObservation(f'Command execution failed: {str(e)}')
  341. return obs
  342. def run(self, action: CmdRunAction) -> Observation:
  343. return self.run_action(action)
  344. def run_ipython(self, action: IPythonRunCellAction) -> Observation:
  345. return self.run_action(action)
  346. def read(self, action: FileReadAction) -> Observation:
  347. return self.run_action(action)
  348. def write(self, action: FileWriteAction) -> Observation:
  349. return self.run_action(action)
  350. def browse(self, action: BrowseURLAction) -> Observation:
  351. return self.run_action(action)
  352. def browse_interactive(self, action: BrowseInteractiveAction) -> Observation:
  353. return self.run_action(action)
  354. # ====================================================================
  355. # Implement these methods (for file operations) in the subclass
  356. # ====================================================================
  357. def copy_to(
  358. self, host_src: str, sandbox_dest: str, recursive: bool = False
  359. ) -> None:
  360. if not os.path.exists(host_src):
  361. raise FileNotFoundError(f'Source file {host_src} does not exist')
  362. self._wait_until_alive()
  363. try:
  364. if recursive:
  365. # For recursive copy, create a zip file
  366. with tempfile.NamedTemporaryFile(
  367. suffix='.zip', delete=False
  368. ) as temp_zip:
  369. temp_zip_path = temp_zip.name
  370. with ZipFile(temp_zip_path, 'w') as zipf:
  371. for root, _, files in os.walk(host_src):
  372. for file in files:
  373. file_path = os.path.join(root, file)
  374. arcname = os.path.relpath(
  375. file_path, os.path.dirname(host_src)
  376. )
  377. zipf.write(file_path, arcname)
  378. upload_data = {'file': open(temp_zip_path, 'rb')}
  379. else:
  380. # For single file copy
  381. upload_data = {'file': open(host_src, 'rb')}
  382. params = {'destination': sandbox_dest, 'recursive': str(recursive).lower()}
  383. response = self.session.post(
  384. f'{self.api_url}/upload_file', files=upload_data, params=params
  385. )
  386. if response.status_code == 200:
  387. return
  388. else:
  389. error_message = response.text
  390. raise Exception(f'Copy operation failed: {error_message}')
  391. except requests.Timeout:
  392. raise TimeoutError('Copy operation timed out')
  393. except Exception as e:
  394. raise RuntimeError(f'Copy operation failed: {str(e)}')
  395. finally:
  396. if recursive:
  397. os.unlink(temp_zip_path)
  398. logger.info(f'Copy completed: host:{host_src} -> runtime:{sandbox_dest}')
  399. def list_files(self, path: str | None = None) -> list[str]:
  400. """List files in the sandbox.
  401. If path is None, list files in the sandbox's initial working directory (e.g., /workspace).
  402. """
  403. self._wait_until_alive()
  404. try:
  405. data = {}
  406. if path is not None:
  407. data['path'] = path
  408. response = self.session.post(f'{self.api_url}/list_files', json=data)
  409. if response.status_code == 200:
  410. response_json = response.json()
  411. assert isinstance(response_json, list)
  412. return response_json
  413. else:
  414. error_message = response.text
  415. raise Exception(f'List files operation failed: {error_message}')
  416. except requests.Timeout:
  417. raise TimeoutError('List files operation timed out')
  418. except Exception as e:
  419. raise RuntimeError(f'List files operation failed: {str(e)}')