Skip to content

Commit dd2982b

Browse files
committed
test
1 parent f8ab74e commit dd2982b

File tree

2 files changed

+21
-24
lines changed

2 files changed

+21
-24
lines changed

python/pyspark/sql/connect/dataframe.py

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -1703,10 +1703,10 @@ def __getattr__(self, name: str) -> "Column":
17031703
errorClass="JVM_ATTRIBUTE_NOT_SUPPORTED", messageParameters={"attr_name": name}
17041704
)
17051705

1706-
if name not in self.columns:
1707-
raise PySparkAttributeError(
1708-
errorClass="ATTRIBUTE_NOT_SUPPORTED", messageParameters={"attr_name": name}
1709-
)
1706+
# if name not in self.columns:
1707+
# raise PySparkAttributeError(
1708+
# errorClass="ATTRIBUTE_NOT_SUPPORTED", messageParameters={"attr_name": name}
1709+
# )
17101710

17111711
return self._col(name)
17121712

@@ -1739,14 +1739,14 @@ def __getitem__(
17391739
# ConnectColumn(addDataFrameIdToCol(resolve(colName)))
17401740
# }
17411741

1742-
# validate the column name
1743-
if not hasattr(self._session, "is_mock_session"):
1744-
from pyspark.sql.connect.types import verify_col_name
1745-
1746-
# Try best to verify the column name with cached schema
1747-
# If fails, fall back to the server side validation
1748-
if not verify_col_name(item, self._schema):
1749-
self.select(item).isLocal()
1742+
# # validate the column name
1743+
# if not hasattr(self._session, "is_mock_session"):
1744+
# from pyspark.sql.connect.types import verify_col_name
1745+
#
1746+
# # Try best to verify the column name with cached schema
1747+
# # If fails, fall back to the server side validation
1748+
# if not verify_col_name(item, self._schema):
1749+
# self.select(item).isLocal()
17501750

17511751
return self._col(item)
17521752
elif isinstance(item, Column):

sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/ColumnResolutionHelper.scala

Lines changed: 9 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -622,18 +622,15 @@ trait ColumnResolutionHelper extends Logging with DataTypeErrorsBase {
622622
p: LogicalPlan,
623623
currentDepth: Int): (Option[(NamedExpression, Int)], Boolean) = {
624624
val (resolved, matched) = if (p.getTagValue(LogicalPlan.PLAN_ID_TAG).contains(id)) {
625-
val resolved = try {
626-
if (!isMetadataAccess) {
627-
p.resolve(u.nameParts, conf.resolver)
628-
} else if (u.nameParts.size == 1) {
629-
p.getMetadataAttributeByNameOpt(u.nameParts.head)
630-
} else {
631-
None
632-
}
633-
} catch {
634-
case e: AnalysisException =>
635-
logDebug(s"Fail to resolve $u with $p due to $e")
636-
None
625+
val resolved = if (!isMetadataAccess) {
626+
p.resolve(u.nameParts, conf.resolver)
627+
} else if (u.nameParts.size == 1) {
628+
p.getMetadataAttributeByNameOpt(u.nameParts.head)
629+
} else {
630+
None
631+
}
632+
if (resolved.isEmpty) {
633+
throw QueryCompilationErrors.cannotResolveDataFrameColumn(u)
637634
}
638635
(resolved.map(r => (r, currentDepth)), true)
639636
} else {

0 commit comments

Comments
 (0)