@@ -23,7 +23,7 @@ use clap_complete::{generate, Shell};
23
23
#[ derive( Parser ) ]
24
24
#[ clap( version) ]
25
25
struct Cli {
26
- /// Only print errors to standard error stream. Supresses warnings and all other log levels
26
+ /// Only print errors to standard error stream. Suppresses warnings and all other log levels
27
27
/// independent of the verbose mode.
28
28
#[ arg( short = 'q' , long) ]
29
29
quiet : bool ,
@@ -133,10 +133,10 @@ pub struct QueryOpt {
133
133
row_groups_per_file : u32 ,
134
134
/// Then the size of the currently written parquet files goes beyond this threshold the current
135
135
/// row group will be finished and then the file will be closed. So the file will be somewhat
136
- /// larger than the threshold. All furthrer row groups will be written into new files to which
136
+ /// larger than the threshold. All further row groups will be written into new files to which
137
137
/// the threshold size limit is applied as well. If this option is not set, no size threshold is
138
138
/// applied. If the threshold is applied the first file name will have the suffix `_01`, the
139
- /// second the suffix `_2` and so on. Therfore the first resulting file will be called e.g.
139
+ /// second the suffix `_2` and so on. Therefore, the first resulting file will be called e.g.
140
140
/// `out_1.par`, if `out.par` has been specified as the output argument.
141
141
/// Also note that this option will not act as an upper bound. It will act as a lower bound for
142
142
/// all but the last file, all others however will not be larger than this threshold by more
@@ -152,16 +152,16 @@ pub struct QueryOpt {
152
152
///
153
153
/// This is useful in situations there ODBC would require us to allocate a ridiculous amount of
154
154
/// memory for a single element of a row. Usually this is the case because the Database schema
155
- /// has been ill defined (like choosing `TEXT` for a user name , although a users name is
155
+ /// has been ill- defined (like choosing `TEXT` for a username , although a users name is
156
156
/// unlikely to be several GB long). Another situation is that the ODBC driver is not good at
157
- /// reporting the maximum length and therfore reports a really large value. The third option is
157
+ /// reporting the maximum length and therefore reports a really large value. The third option is
158
158
/// of course that your values are actually large. In this case you just need a ton of memory.
159
159
/// You can use the batch size limit though to retrieve less at once. For binary columns this is
160
- /// a maximum element length in bytes. For text columns it depends wether UTF-8 or UTF-16
161
- /// encoding is used. See documentation of the `encondig ` option. In case of UTF-8 this is the
160
+ /// a maximum element length in bytes. For text columns it depends on whether UTF-8 or UTF-16
161
+ /// encoding is used. See documentation of the `encoding ` option. In case of UTF-8 this is the
162
162
/// maximum length in bytes for an element. In case of UTF-16 the binary length is multiplied by
163
163
/// two. This allows domain experts to configure limits (roughly) in the domain of how many
164
- /// letters do I expect in this column, rather than to care about wether the command is executed
164
+ /// letters do I expect in this column, rather than to care about whether the command is executed
165
165
/// on Linux or Windows. The encoding of the column on the Database does not matter for this
166
166
/// setting or determining buffer sizes.
167
167
#[ arg( long) ]
@@ -195,7 +195,7 @@ pub struct QueryOpt {
195
195
/// been introduced in an effort to increase the compatibility of the output with Apache Spark.
196
196
#[ clap( long) ]
197
197
prefer_varbinary : bool ,
198
- /// Specify the fallback encoding of the parquet output column. You can parse mutliple values
198
+ /// Specify the fallback encoding of the parquet output column. You can parse multiple values
199
199
/// in format `COLUMN:ENCODING`. `ENCODING` must be one of: `plain`, `delta-binary-packed`,
200
200
/// `delta-byte-array`, `delta-length-byte-array` or `rle`.
201
201
#[ arg(
@@ -204,18 +204,18 @@ pub struct QueryOpt {
204
204
action = ArgAction :: Append
205
205
) ]
206
206
parquet_column_encoding : Vec < ( String , Encoding ) > ,
207
- /// Tells the odbc2parquet, that the ODBC driver does not support binding 64 Bit integers (aka
207
+ /// Tells the odbc2parquet, that the ODBC driver does not support binding 64- Bit integers (aka
208
208
/// S_C_BIGINT in ODBC speak). This will cause the odbc2parquet to query large integers as text
209
- /// instead and convert them to 64 Bit integers itself. Setting this flag will not affect the
210
- /// output, but may incurr a performance penality . In case you are using an Oracle Database it
211
- /// can make queries work which did not before, because Oracle does not support 64 Bit integers.
209
+ /// instead and convert them to 64- Bit integers itself. Setting this flag will not affect the
210
+ /// output, but may incur a performance penalty . In case you are using an Oracle Database it
211
+ /// can make queries work which did not before, because Oracle does not support 64- Bit integers.
212
212
#[ clap( long) ]
213
213
driver_does_not_support_64bit_integers : bool ,
214
214
/// The IBM DB2 Linux ODBC drivers have been reported to return memory garbage instead of
215
215
/// indicators for the string length. Setting this flag will cause `odbc2parquet` to rely on
216
216
/// terminating zeroes, instead of indicators. This prevents `odbc2parquet` from disambiguating
217
217
/// between empty strings and `NULL``. As a side effect of this workaround empty might be mapped
218
- /// to NULL. Currently this workaround is only active if UTF-8 is used. This should be the case
218
+ /// to NULL. Currently, this workaround is only active if UTF-8 is used. This should be the case
219
219
/// on non-window platforms by default, or if the `System` encoding is active.
220
220
#[ clap( long) ]
221
221
avoid_decimal : bool ,
@@ -236,7 +236,7 @@ pub struct QueryOpt {
236
236
output : IoArg ,
237
237
/// Query executed against the ODBC data source. Question marks (`?`) can be used as
238
238
/// placeholders for positional parameters. E.g. "SELECT Name FROM Employees WHERE salary > ?;".
239
- /// Instead of passing a query verbatum , you may pass a plain dash (`-`), to indicate that the
239
+ /// Instead of passing a query verbatim , you may pass a plain dash (`-`), to indicate that the
240
240
/// query should be read from standard input. In this case the entire input until EOF will be
241
241
/// considered the query.
242
242
query : String ,
@@ -251,7 +251,7 @@ pub struct InsertOpt {
251
251
connect_opts : ConnectOpts ,
252
252
/// Encoding used for transferring character data to the database.
253
253
///
254
- /// `Utf16`: Use 16Bit characters to send text text to the database, which implies the using
254
+ /// `Utf16`: Use 16Bit characters to send text to the database, which implies the using
255
255
/// UTF-16 encoding. This should work well independent of the system configuration, but requires
256
256
/// additional work since text is always stored as UTF-8 in parquet.
257
257
///
@@ -409,7 +409,7 @@ fn open_connection<'e>(
409
409
DriverCompleteOption :: NoPrompt
410
410
} ;
411
411
412
- // We are not interessted in the completed connection string, beyond creating a connection, so
412
+ // We are not interested in the completed connection string, beyond creating a connection, so
413
413
// we pass an empty buffer.
414
414
let mut completed_connection_string = OutputStringBuffer :: empty ( ) ;
415
415
0 commit comments