from typing import Any, Dict, Optional, Union from huggingface_hub.hf_api import InferenceProviderMapping from huggingface_hub.inference._common import RequestParameters, _as_dict from ._common import BaseConversationalTask, BaseTextGenerationTask, filter_none _PROVIDER = "featherless-ai" _BASE_URL = "https://api.featherless.ai" class FeatherlessTextGenerationTask(BaseTextGenerationTask): def __init__(self): super().__init__(provider=_PROVIDER, base_url=_BASE_URL) def _prepare_payload_as_dict( self, inputs: Any, parameters: Dict, provider_mapping_info: InferenceProviderMapping ) -> Optional[Dict]: params = filter_none(parameters.copy()) params["max_tokens"] = params.pop("max_new_tokens", None) return {"prompt": inputs, **params, "model": provider_mapping_info.provider_id} def get_response(self, response: Union[bytes, Dict], request_params: Optional[RequestParameters] = None) -> Any: output = _as_dict(response)["choices"][0] return { "generated_text": output["text"], "details": { "finish_reason": output.get("finish_reason"), "seed": output.get("seed"), }, } class FeatherlessConversationalTask(BaseConversationalTask): def __init__(self): super().__init__(provider=_PROVIDER, base_url=_BASE_URL)