@@ -119,8 +119,6 @@ def __new__(
119119 spark_provider : Optional ["BaseSparkSessionProvider" ] = None ,
120120 code_type : str = "script" ,
121121 ) -> Client :
122- if "function" in code_type :
123- return cls ._new_function_client ()
124122
125123 if cls ._instance is None :
126124 cls ._instance = super ().__new__ (cls )
@@ -175,16 +173,6 @@ def __new__(
175173 raise ValueError ("Cannot set reader or writer after client is initialized" )
176174 return cls ._instance
177175
178- @classmethod
179- def _new_function_client (cls ) -> Client :
180- cls ._instance = super ().__new__ (cls )
181- cls ._instance ._proxy = (
182- config .proxy_config .to_object () # type: ignore
183- if config .proxy_config is not None
184- else None
185- )
186- return cls ._instance
187-
188176 def read_dlo (self , name : str ) -> PySparkDataFrame :
189177 """Read a DLO from Data Cloud.
190178
@@ -195,7 +183,7 @@ def read_dlo(self, name: str) -> PySparkDataFrame:
195183 A PySpark DataFrame containing the DLO data.
196184 """
197185 self ._record_dlo_access (name )
198- return self ._reader .read_dlo (name )
186+ return self ._reader .read_dlo (name ) # type: ignore[no-any-return]
199187
200188 def read_dmo (self , name : str ) -> PySparkDataFrame :
201189 """Read a DMO from Data Cloud.
@@ -207,7 +195,7 @@ def read_dmo(self, name: str) -> PySparkDataFrame:
207195 A PySpark DataFrame containing the DMO data.
208196 """
209197 self ._record_dmo_access (name )
210- return self ._reader .read_dmo (name )
198+ return self ._reader .read_dmo (name ) # type: ignore[no-any-return]
211199
212200 def write_to_dlo (
213201 self , name : str , dataframe : PySparkDataFrame , write_mode : WriteMode , ** kwargs
@@ -220,7 +208,7 @@ def write_to_dlo(
220208 write_mode: The write mode to use for writing to the DLO.
221209 """
222210 self ._validate_data_layer_history_does_not_contain (DataCloudObjectType .DMO )
223- return self ._writer .write_to_dlo (name , dataframe , write_mode , ** kwargs )
211+ return self ._writer .write_to_dlo (name , dataframe , write_mode , ** kwargs ) # type: ignore[no-any-return]
224212
225213 def write_to_dmo (
226214 self , name : str , dataframe : PySparkDataFrame , write_mode : WriteMode , ** kwargs
@@ -233,17 +221,17 @@ def write_to_dmo(
233221 write_mode: The write mode to use for writing to the DMO.
234222 """
235223 self ._validate_data_layer_history_does_not_contain (DataCloudObjectType .DLO )
236- return self ._writer .write_to_dmo (name , dataframe , write_mode , ** kwargs )
224+ return self ._writer .write_to_dmo (name , dataframe , write_mode , ** kwargs ) # type: ignore[no-any-return]
237225
238226 def call_llm_gateway (self , LLM_MODEL_ID : str , prompt : str , maxTokens : int ) -> str :
239227 if self ._proxy is None :
240228 raise ValueError ("No proxy configured; set proxy or proxy_config" )
241- return self ._proxy .call_llm_gateway (LLM_MODEL_ID , prompt , maxTokens )
229+ return self ._proxy .call_llm_gateway (LLM_MODEL_ID , prompt , maxTokens ) # type: ignore[no-any-return]
242230
243231 def find_file_path (self , file_name : str ) -> Path :
244232 """Return a file path"""
245233
246- return self ._file .find_file_path (file_name )
234+ return self ._file .find_file_path (file_name ) # type: ignore[no-any-return]
247235
248236 def _validate_data_layer_history_does_not_contain (
249237 self , data_cloud_object_type : DataCloudObjectType
0 commit comments