JuliaDB.jl icon indicating copy to clipboard operation
JuliaDB.jl copied to clipboard

Parallel analytical database in pure Julia

Results 104 JuliaDB.jl issues
Sort by recently updated
recently updated
newest added
trafficstars

Like what's in pandas, http://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.DataFrame.diff.html, or more generally http://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.DataFrame.shift.html

This seems to be more useful generally! I think the cost incurred in the extra allocation is negligible. We should only keep the table indexed by only those columns for...

enhancement
api

These is a method like that for `ndsparse` tables: ```julia merge(left::DNDSparse{I1,D1}, right::DNDSparse{I2,D2}; agg=IndexedTables.right) ``` but not for indexed tables. As a side-note, there seem to be a bunch methods that...

I want to create the `max(v1)` and `sum(v2)` grouped by `id`. This is my code ```julia groupreduce(max, mytable, :id, select = :v1) groupreduce(+, mytable, :id, select = :v2) ``` but...

docs

Cannot add a row to a loaded table: ```julia using JuliaDB t = table((x = [1], y = [2])) save(t, "tmp.jdb") t = load("tmp.jdb") push!(rows(t), (x = 3, y =...

It's pretty common that an empty string (e.g. `x,,z`) would signify a missing value. I tried that with no success. But then I discovered that specifying a specific string doesn't...

Even though the Doc reads that `Union{Missing, T}` is the default type, the `table` function converts that Union to DataValue. ```julia julia> df = DataFrames.DataFrame(mis = Union{Missing, Int}[missing]) 1×1 DataFrames.DataFrame...

Hey, being new to JuliaDB I though using its save function would be straight-forward and following the rules of FileIO, thus the argument order would be `save(::IO, obj)` which let...

Using a `remotecall` for distributed table operations appear to keep some reference around and temporary objects are not getting GCed correctly. @nkottary has already looked into this a little and...

I tried several times to see if the total memory usage is smaller than the worker number * chunk size. Basically, it has NO limit. I tried to use MemPool.max_memsize...