ODBC.jl provides access to any database that supports an ODBC driver; Postgres has excellent support for ODBC. For windows, Iâd downloads the msi folder version for an easy installer, for other unix platforms, just grab the tarball from the src directory. You can setup an ODBC dsn file like the one used for testing, which can be registered w/ your system ODBC via the command odbcinst -i -s -h -f ./test/pgtest.odbc.ini.
As noted elsewhere, it takes a bit more setup, but once youâve been thru it once or twice, I find it much easier to setup any other database.
ERROR: LoadError: LoadError: LoadError: UndefVarError: @isdefined not defined
Stacktrace:
[1] include_from_node1(::String) at ./loading.jl:576
[2] include(::String) at ./sysimg.jl:14
[3] include_from_node1(::String) at ./loading.jl:576
[4] include(::String) at ./sysimg.jl:14
[5] include_from_node1(::String) at ./loading.jl:576
[6] eval(::Module, ::Any) at ./boot.jl:235
[7] _require(::Symbol) at ./loading.jl:490
[8] require(::Symbol) at ./loading.jl:405
There seem to be an issue with your active environment Project.toml. Use a new environment for that project or alternatively, clean up that Project.toml to figure out the bas entries.
res = execute(wrds, âselect column_name
from information_schema.columns
where table_schema=âcrspâ
and table_name=âmsfâ
order by column_nameâ)
crsp_msf_var = DataFrame(columntable(res))
i need to specify the number of rows to be downloaded. sometimes the data could be too large, and i just need to see a part of it. for example, something like this:
res = execute(wrds, âselect column_name
from information_schema.columns
where table_schema=âcrspâ
and table_name=âmsfâ
order by column_name
limit 10â)
thank you so much! this may be unrelated to this post, just want to know if DataFrame is optimized for good performance or if there is any other fast data structure for tabular data in Julia.
DataFrames.jl is one of the oldest packages in the data ecosystem and is currently the flagship when it comes to in-memory tabular data. It is quite efficient and has tons of features. However, for efficiency the optimal course would be to run operations in the performant database when possible (e.g., a relational database such as Postgres with relevant indices). For example, run any select/filter/joins/sort operations in the database when possible and work with the results with DataFrames. Same advice applies to most tabular ecosystem no matter how efficient those are (e.g., same applies to Râs data.table).