@@ -25,11 +25,11 @@ class MockDataCloudReader(BaseDataCloudReader):
2525
2626 CONFIG_NAME = "MockDataCloudReader"
2727
28- def read_dlo (self , name : str ) -> DataFrame :
28+ def read_dlo (self , name : str , schema = None , row_limit : int = 1000 ) -> DataFrame :
2929 df = MagicMock (spec = DataFrame )
3030 return df
3131
32- def read_dmo (self , name : str ) -> DataFrame :
32+ def read_dmo (self , name : str , schema = None , row_limit : int = 1000 ) -> DataFrame :
3333 df = MagicMock (spec = DataFrame )
3434 return df
3535
@@ -153,7 +153,7 @@ def test_read_dlo(self, reset_client, mock_spark, mock_proxy):
153153 client = Client (reader = reader , writer = writer , proxy = mock_proxy )
154154 result = client .read_dlo ("test_dlo" )
155155
156- reader .read_dlo .assert_called_once_with ("test_dlo" )
156+ reader .read_dlo .assert_called_once_with ("test_dlo" , row_limit = 1000 )
157157 assert result is mock_df
158158 assert "test_dlo" in client ._data_layer_history [DataCloudObjectType .DLO ]
159159
@@ -166,7 +166,7 @@ def test_read_dmo(self, reset_client, mock_spark, mock_proxy):
166166 client = Client (reader = reader , writer = writer , proxy = mock_proxy )
167167 result = client .read_dmo ("test_dmo" )
168168
169- reader .read_dmo .assert_called_once_with ("test_dmo" )
169+ reader .read_dmo .assert_called_once_with ("test_dmo" , row_limit = 1000 )
170170 assert result is mock_df
171171 assert "test_dmo" in client ._data_layer_history [DataCloudObjectType .DMO ]
172172
@@ -238,7 +238,7 @@ def test_read_pattern_flow(self, reset_client, mock_spark, mock_proxy):
238238 df = client .read_dlo ("source_dlo" )
239239 client .write_to_dlo ("target_dlo" , df , WriteMode .APPEND )
240240
241- reader .read_dlo .assert_called_once_with ("source_dlo" )
241+ reader .read_dlo .assert_called_once_with ("source_dlo" , row_limit = 1000 )
242242 writer .write_to_dlo .assert_called_once_with (
243243 "target_dlo" , mock_df , WriteMode .APPEND
244244 )
@@ -253,13 +253,41 @@ def test_read_pattern_flow(self, reset_client, mock_spark, mock_proxy):
253253 df = client .read_dmo ("source_dmo" )
254254 client .write_to_dmo ("target_dmo" , df , WriteMode .MERGE )
255255
256- reader .read_dmo .assert_called_once_with ("source_dmo" )
256+ reader .read_dmo .assert_called_once_with ("source_dmo" , row_limit = 1000 )
257257 writer .write_to_dmo .assert_called_once_with (
258258 "target_dmo" , mock_df , WriteMode .MERGE
259259 )
260260
261261 assert "source_dmo" in client ._data_layer_history [DataCloudObjectType .DMO ]
262262
263+ def test_read_dlo_with_row_limit (self , reset_client , mock_spark , mock_proxy ):
264+ """Test that row_limit parameter is passed through to reader."""
265+ reader = MagicMock (spec = BaseDataCloudReader )
266+ writer = MagicMock (spec = BaseDataCloudWriter )
267+ mock_df = MagicMock (spec = DataFrame )
268+ reader .read_dlo .return_value = mock_df
269+
270+ client = Client (reader = reader , writer = writer , proxy = mock_proxy )
271+ result = client .read_dlo ("test_dlo" , row_limit = 500 )
272+
273+ reader .read_dlo .assert_called_once_with ("test_dlo" , row_limit = 500 )
274+ assert result is mock_df
275+ assert "test_dlo" in client ._data_layer_history [DataCloudObjectType .DLO ]
276+
277+ def test_read_dmo_with_row_limit (self , reset_client , mock_spark , mock_proxy ):
278+ """Test that row_limit parameter is passed through to reader."""
279+ reader = MagicMock (spec = BaseDataCloudReader )
280+ writer = MagicMock (spec = BaseDataCloudWriter )
281+ mock_df = MagicMock (spec = DataFrame )
282+ reader .read_dmo .return_value = mock_df
283+
284+ client = Client (reader = reader , writer = writer , proxy = mock_proxy )
285+ result = client .read_dmo ("test_dmo" , row_limit = 100 )
286+
287+ reader .read_dmo .assert_called_once_with ("test_dmo" , row_limit = 100 )
288+ assert result is mock_df
289+ assert "test_dmo" in client ._data_layer_history [DataCloudObjectType .DMO ]
290+
263291
264292# Add tests for DefaultSparkSessionProvider
265293class TestDefaultSparkSessionProvider :
0 commit comments