Skip to content

Commit b8349e9

Browse files
Fixing lint errors
1 parent 731f9ec commit b8349e9

4 files changed

Lines changed: 21 additions & 14 deletions

File tree

src/datacustomcode/client.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -196,7 +196,7 @@ def read_dlo(self, name: str, row_limit: int = 1000) -> PySparkDataFrame:
196196
A PySpark DataFrame containing the DLO data.
197197
"""
198198
self._record_dlo_access(name)
199-
return self._reader.read_dlo(name, row_limit=row_limit)
199+
return self._reader.read_dlo(name, row_limit=row_limit) # type: ignore[no-any-return]
200200

201201
def read_dmo(self, name: str, row_limit: int = 1000) -> PySparkDataFrame:
202202
"""Read a DMO from Data Cloud.
@@ -209,7 +209,7 @@ def read_dmo(self, name: str, row_limit: int = 1000) -> PySparkDataFrame:
209209
A PySpark DataFrame containing the DMO data.
210210
"""
211211
self._record_dmo_access(name)
212-
return self._reader.read_dmo(name, row_limit=row_limit)
212+
return self._reader.read_dmo(name, row_limit=row_limit) # type: ignore[no-any-return]
213213

214214
def write_to_dlo(
215215
self, name: str, dataframe: PySparkDataFrame, write_mode: WriteMode, **kwargs
@@ -222,7 +222,7 @@ def write_to_dlo(
222222
write_mode: The write mode to use for writing to the DLO.
223223
"""
224224
self._validate_data_layer_history_does_not_contain(DataCloudObjectType.DMO)
225-
return self._writer.write_to_dlo(name, dataframe, write_mode, **kwargs)
225+
return self._writer.write_to_dlo(name, dataframe, write_mode, **kwargs) # type: ignore[no-any-return]
226226

227227
def write_to_dmo(
228228
self, name: str, dataframe: PySparkDataFrame, write_mode: WriteMode, **kwargs
@@ -235,17 +235,17 @@ def write_to_dmo(
235235
write_mode: The write mode to use for writing to the DMO.
236236
"""
237237
self._validate_data_layer_history_does_not_contain(DataCloudObjectType.DLO)
238-
return self._writer.write_to_dmo(name, dataframe, write_mode, **kwargs)
238+
return self._writer.write_to_dmo(name, dataframe, write_mode, **kwargs) # type: ignore[no-any-return]
239239

240240
def call_llm_gateway(self, LLM_MODEL_ID: str, prompt: str, maxTokens: int) -> str:
241241
if self._proxy is None:
242242
raise ValueError("No proxy configured; set proxy or proxy_config")
243-
return self._proxy.call_llm_gateway(LLM_MODEL_ID, prompt, maxTokens)
243+
return self._proxy.call_llm_gateway(LLM_MODEL_ID, prompt, maxTokens) # type: ignore[no-any-return]
244244

245245
def find_file_path(self, file_name: str) -> Path:
246246
"""Return a file path"""
247247

248-
return self._file.find_file_path(file_name)
248+
return self._file.find_file_path(file_name) # type: ignore[no-any-return]
249249

250250
def _validate_data_layer_history_does_not_contain(
251251
self, data_cloud_object_type: DataCloudObjectType

src/datacustomcode/config.py

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -36,10 +36,7 @@
3636
# This lets all readers and writers to be findable via config
3737
from datacustomcode.io import * # noqa: F403
3838
from datacustomcode.io.base import BaseDataAccessLayer
39-
from datacustomcode.io.reader.base import BaseDataCloudReader # noqa: TCH001
40-
from datacustomcode.io.writer.base import BaseDataCloudWriter # noqa: TCH001
4139
from datacustomcode.proxy.base import BaseProxyAccessLayer
42-
from datacustomcode.proxy.client.base import BaseProxyClient # noqa: TCH001
4340
from datacustomcode.spark.base import BaseSparkSessionProvider
4441

4542
DEFAULT_CONFIG_NAME = "config.yaml"
@@ -48,6 +45,10 @@
4845
if TYPE_CHECKING:
4946
from pyspark.sql import SparkSession
5047

48+
from datacustomcode.io.reader.base import BaseDataCloudReader
49+
from datacustomcode.io.writer.base import BaseDataCloudWriter
50+
from datacustomcode.proxy.client.base import BaseProxyClient
51+
5152

5253
class ForceableConfig(BaseModel):
5354
force: bool = Field(

src/datacustomcode/io/reader/query_api.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -179,7 +179,7 @@ def read_dlo(
179179
self, "_sf_cli_reader", None
180180
)
181181
if sf_cli_reader is not None:
182-
return sf_cli_reader.read_dlo(name, schema, row_limit)
182+
return sf_cli_reader.read_dlo(name, schema, row_limit) # type: ignore[no-any-return]
183183

184184
query = SQL_QUERY_TEMPLATE.format(name, row_limit)
185185

@@ -214,7 +214,7 @@ def read_dmo(
214214
self, "_sf_cli_reader", None
215215
)
216216
if sf_cli_reader is not None:
217-
return sf_cli_reader.read_dmo(name, schema, row_limit)
217+
return sf_cli_reader.read_dmo(name, schema, row_limit) # type: ignore[no-any-return]
218218

219219
query = SQL_QUERY_TEMPLATE.format(name, row_limit)
220220

src/datacustomcode/llm_gateway/base.py

Lines changed: 9 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -15,13 +15,19 @@
1515
from __future__ import annotations
1616

1717
from abc import abstractmethod
18+
from typing import TYPE_CHECKING
1819

19-
from datacustomcode.llm_gateway.types.generate_text_request import GenerateTextRequest
20-
from datacustomcode.llm_gateway.types.generate_text_response import GenerateTextResponse
20+
if TYPE_CHECKING:
21+
from datacustomcode.llm_gateway.types.generate_text_request import (
22+
GenerateTextRequest,
23+
)
24+
from datacustomcode.llm_gateway.types.generate_text_response import (
25+
GenerateTextResponse,
26+
)
2127

2228

2329
class LLMGateway:
24-
def __init__(self):
30+
def __init__(self) -> None:
2531
pass
2632

2733
@abstractmethod

0 commit comments

Comments
 (0)