File tree Expand file tree Collapse file tree 2 files changed +1
-17
lines changed Expand file tree Collapse file tree 2 files changed +1
-17
lines changed Original file line number Diff line number Diff line change @@ -1732,22 +1732,6 @@ def __getitem__(
1732
1732
)
1733
1733
)
1734
1734
else :
1735
- # TODO: revisit classic Spark's Dataset.col
1736
- # if (sparkSession.sessionState.conf.supportQuotedRegexColumnName) {
1737
- # colRegex(colName)
1738
- # } else {
1739
- # ConnectColumn(addDataFrameIdToCol(resolve(colName)))
1740
- # }
1741
-
1742
- # validate the column name
1743
- if not hasattr (self ._session , "is_mock_session" ):
1744
- from pyspark .sql .connect .types import verify_col_name
1745
-
1746
- # Try best to verify the column name with cached schema
1747
- # If fails, fall back to the server side validation
1748
- if not verify_col_name (item , self ._schema ):
1749
- self .select (item ).isLocal ()
1750
-
1751
1735
return self ._col (item )
1752
1736
elif isinstance (item , Column ):
1753
1737
return self .filter (item )
Original file line number Diff line number Diff line change @@ -133,8 +133,8 @@ def test_access_column(self):
133
133
self .assertTrue (isinstance (df ["key" ], Column ))
134
134
self .assertTrue (isinstance (df [0 ], Column ))
135
135
self .assertRaises (IndexError , lambda : df [2 ])
136
- self .assertRaises (AnalysisException , lambda : df ["bad_key" ])
137
136
self .assertRaises (TypeError , lambda : df [{}])
137
+ self .assertRaises (AnalysisException , lambda : df .select (df ["bad_key" ]).schema )
138
138
139
139
def test_column_name_with_non_ascii (self ):
140
140
columnName = "数量"
You can’t perform that action at this time.
0 commit comments