|
||||
0001 /* 0002 * Licensed to the Apache Software Foundation (ASF) under one or more 0003 * contributor license agreements. See the NOTICE file distributed with 0004 * this work for additional information regarding copyright ownership. 0005 * The ASF licenses this file to You under the Apache License, Version 2.0 0006 * (the "License"); you may not use this file except in compliance with 0007 * the License. You may obtain a copy of the License at 0008 * 0009 * http://www.apache.org/licenses/LICENSE-2.0 0010 * 0011 * Unless required by applicable law or agreed to in writing, software 0012 * distributed under the License is distributed on an "AS IS" BASIS, 0013 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 0014 * See the License for the specific language governing permissions and 0015 * limitations under the License. 0016 */ 0017 0018 package org.apache.spark.sql.connector.catalog; 0019 0020 import org.apache.spark.annotation.Evolving; 0021 import org.apache.spark.sql.connector.read.Scan; 0022 import org.apache.spark.sql.connector.read.ScanBuilder; 0023 import org.apache.spark.sql.util.CaseInsensitiveStringMap; 0024 0025 /** 0026 * A mix-in interface of {@link Table}, to indicate that it's readable. This adds 0027 * {@link #newScanBuilder(CaseInsensitiveStringMap)} that is used to create a scan for batch, 0028 * micro-batch, or continuous processing. 0029 * 0030 * @since 3.0.0 0031 */ 0032 @Evolving 0033 public interface SupportsRead extends Table { 0034 0035 /** 0036 * Returns a {@link ScanBuilder} which can be used to build a {@link Scan}. Spark will call this 0037 * method to configure each data source scan. 0038 * 0039 * @param options The options for reading, which is an immutable case-insensitive 0040 * string-to-string map. 0041 */ 0042 ScanBuilder newScanBuilder(CaseInsensitiveStringMap options); 0043 }
[ Source navigation ] | [ Diff markup ] | [ Identifier search ] | [ general search ] |
This page was automatically generated by the 2.1.0 LXR engine. The LXR team |