diff --git a/python/pyspark/sql/tests/test_dataframe.py b/python/pyspark/sql/tests/test_dataframe.py index e686fa9e929fd..bd2f1cb75b7af 100644 --- a/python/pyspark/sql/tests/test_dataframe.py +++ b/python/pyspark/sql/tests/test_dataframe.py @@ -1119,10 +1119,10 @@ def test_to_pandas(self): pdf = self._to_pandas() types = pdf.dtypes self.assertEqual(types[0], np.int32) - self.assertEqual(types[1], np.object) - self.assertEqual(types[2], np.bool) + self.assertEqual(types[1], object) + self.assertEqual(types[2], bool) self.assertEqual(types[3], np.float32) - self.assertEqual(types[4], np.object) # datetime.date + self.assertEqual(types[4], object) # datetime.date self.assertEqual(types[5], "datetime64[ns]") self.assertEqual(types[6], "datetime64[ns]") self.assertEqual(types[7], "timedelta64[ns]") @@ -1181,7 +1181,7 @@ def test_to_pandas_avoid_astype(self): df = self.spark.createDataFrame(data, schema) types = df.toPandas().dtypes self.assertEqual(types[0], np.float64) # doesn't convert to np.int32 due to NaN value. - self.assertEqual(types[1], np.object) + self.assertEqual(types[1], object) self.assertEqual(types[2], np.float64) @unittest.skipIf(not have_pandas, pandas_requirement_message) # type: ignore @@ -1242,8 +1242,8 @@ def test_to_pandas_from_null_dataframe(self): self.assertEqual(types[3], np.float64) self.assertEqual(types[4], np.float32) self.assertEqual(types[5], np.float64) - self.assertEqual(types[6], np.object) - self.assertEqual(types[7], np.object) + self.assertEqual(types[6], object) + self.assertEqual(types[7], object) self.assertTrue(np.can_cast(np.datetime64, types[8])) self.assertTrue(np.can_cast(np.datetime64, types[9])) self.assertTrue(np.can_cast(np.timedelta64, types[10]))