Skip to content

Commit

Permalink
Fix the logging for pushdown filters (#1201)
Browse files Browse the repository at this point in the history
  • Loading branch information
isha97 committed Mar 21, 2024
1 parent eb21717 commit f20529f
Show file tree
Hide file tree
Showing 2 changed files with 10 additions and 4 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -90,10 +90,16 @@ public ReadSessionResponse create(
TableId table, ImmutableList<String> selectedFields, Optional<String> filter) {
Instant sessionPrepStartTime = Instant.now();
TableInfo tableDetails = bigQueryClient.getTable(table);

TableInfo actualTable = getActualTable(tableDetails, selectedFields, filter);

BigQueryReadClient bigQueryReadClient = bigQueryReadClientFactory.getBigQueryReadClient();
log.info(
"|creation a read session for table {}, parameters: "
+ "|selectedFields=[{}],"
+ "|filter=[{}]",
actualTable.getFriendlyName(),
String.join(",", selectedFields),
filter.orElse("None"));

String tablePath = toTablePath(actualTable.getTableId());
CreateReadSessionRequest request =
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -111,14 +111,14 @@ public RDD<Row> buildScan(String[] requiredColumns) {

@Override
public RDD<Row> buildScan(String[] requiredColumns, Filter[] filters) {
compiledFilter = getCompiledFilter(filters);
log.info(
"|Querying table {}, parameters sent from Spark:"
+ "|requiredColumns=[{}],"
+ "|filters=[{}]",
+ "|filter=[{}]",
getTableNameForLogging(),
String.join(",", requiredColumns),
Arrays.stream(filters).map(f -> f.toString()).collect(Collectors.joining(",")));
compiledFilter = getCompiledFilter(filters);
compiledFilter);
ReadSessionCreator readSessionCreator =
new ReadSessionCreator(
options.toReadSessionCreatorConfig(), bigQueryClient, bigQueryReadClientFactory);
Expand Down

0 comments on commit f20529f

Please sign in to comment.