pymllm.engine.launch ==================== .. py:module:: pymllm.engine.launch Attributes ---------- .. autoapisummary:: pymllm.engine.launch.HAS_BANNER_LIBS pymllm.engine.launch.logger Classes ------- .. autoapisummary:: pymllm.engine.launch.Engine Module Contents --------------- .. py:data:: HAS_BANNER_LIBS :value: True .. py:data:: logger .. py:class:: Engine .. py:property:: is_healthy :type: bool True if engine and all subprocesses are alive. .. py:method:: launch() .. py:method:: generate(prompt = None, sampling_params = None, input_ids = None, image_data = None, audio_data = None, video_data = None, return_logprob = None, logprob_start_len = None, top_logprobs_num = None, lora_path = None, session_params = None, stream = False, rid = None, **kwargs) Synchronous, non-streaming generation entry point. Accepts a single prompt (``str``) or a batch (``List[str]``). Returns a single result dict for single inputs and a list of result dicts for batch inputs, preserving the input order. .. py:method:: generate_async(prompt = None, sampling_params = None, input_ids = None, image_data = None, audio_data = None, video_data = None, return_logprob = None, logprob_start_len = None, top_logprobs_num = None, lora_path = None, session_params = None, stream = False, rid = None, **kwargs) :async: Asynchronous generation entry point. For a **single** request and ``stream=False`` yields one final result dict; with ``stream=True`` yields incremental chunks. For a **batch** request the iterator yields the final result for each sub-request as it completes (order not guaranteed); streaming mode yields incremental chunks from all sub-requests interleaved. .. py:method:: shutdown() Terminate all subprocesses.