The file was built for DuckDB version 'v1.1.3', but we can only load extensions built for DuckDB version '19864453f7'.
by using nightly repo for extensions by default.cached_connection()
helper will configure a
temporary storage location by default. It also now supports all options
supported by duckdb::duckdb()
for connection creation.as_dataset()
utility copies a local in-memory
data.frame into the connection.**
. Also, trying to read from an existing
local file won’t try and append recursive search even when
given the default recursive=TRUE option.open_dataset()
uses random table name by
default, avoid naming collisions.write_dataset()
no longer adds **
into paths when writing some partitions.open_dataset()
gains the ability to read spatial
vector data formats (objects read by sf
) using
format="sf"
default geometry column in to_sf()
is now termed
geom
, to match the default used in duckdb
’s
st_read()
function.
open_dataset()
now tries to guess the data format
instead of defaulting to parquet when no format is explicitly
provided.
a new function, spatial_join()
, allows a variety of
spatial joins.
a new function, st_read_meta()
, exposes the spatial
metadata of remote spatial objects.
new helper function, as_view()
, creates a temporary
view of a query.
write_dataset()
now understands lazy queries, not just
lazy tables.duckdbfs now has spatial data query support! Users can leverage
spatial data operations like st_distance()
and
st_area()
and request return values as sf
objects. Supports network-based access too. See README.md
Added write_dataset()
which can write to
(potentially partitioned) parquet to local directories or remote (S3)
buckets.
The S3 interface supports arrow
-compatible URI
notation:
s3://userid:secret_token@bucket-name?endpoint_override=data.ecoforecast.org
*
(match any file) or
**
(recursive search) and just supply a path. Recursive
search is then assumed automatically. Note: unlike arrow
,
this still supports the use of globs (*
) elsewhere in the
path, e.g. s3://bucket/*/path
duckdb_s3_config
gains argument
anonymous
allowing users to ignore existing AWS keys that
may be set in environmental variables or AWS configuration files. This
can also be passed as the username position in URI notation, e.g.
s3://anonymous@bucket_name
.
open_dataset
drops use of endpoint
as
an argument. Instead, alternative S3 endpoints can be set either by
using the URI query notation or calling duckdb_s3_config()
first. Additionally, any arguments to duckdb_s3_config()
,
including s3_endpoint
, can now be passed to
open_dataset
through the ...
. Note these
settings will override any set by the URI notation.