Thanks, let me try it.

I hope, the suggestion is this

sql>insert into t2 ( id, name )  select id, name from t1;
2 affected rows (121.998ms)
sql>drop t1;

Regards,
Manish


On Sun, Mar 26, 2017 at 3:11 PM, Stefan Manegold <Stefan.Manegold@cwi.nl> wrote:
Being a database system, MonetDB keeps persistent data on disk.
Being a memory-optimized databse system, MonetDB tries to use as much
memory as available to acheive high query processing performance.
However, it should not "demand" more than available.

A binary copy of your table in MonetDB will be smaller (and faster)
than serializing to text (CSV).

Stefan

----- On Mar 26, 2017, at 9:05 AM, Manish gupta gahoimnshg@gmail.com wrote:

> Hi Ying,
> Thanks for response. Would not this method need 2 copies of same data in DB at
> the same time ( until original table is finally deleted fully )? Would it
> demand more memory, as my dataset is quite large.
>
> Regards,
> Manish
>
> On Sun, Mar 26, 2017 at 12:02 PM, Ying Zhang < Y.Zhang@cwi.nl > wrote:
>
>
>
>> On 26 Mar 2017, at 08:04, Manish gupta < gahoimnshg@gmail.com > wrote:
>>
>> OK,
>> I got there is no vacuum cleaning algo implemented yet.
>> https://www.monetdb.org/Documentation/Manuals/SQLreference/Transactions
>>
>> Would drop the records after ascii dump and recreate table.
>
> You can also temporarily store the data of the <table to be cleaned> in a
> <backup table>.
> Then do a “delete from <table to be cleaned>”, which simply wipe out all data,
> incl. garbage from <table to be cleaned>.
> Finally, put all data from <backup table> back to <table to be cleaned>.
>
> This is mostly probably (much) faster than an ASCII dump and restore.
>
>>
>> Regards,
>> Manish
>>
>> On Sat, Mar 25, 2017 at 5:25 PM, Manish gupta < gahoimnshg@gmail.com > wrote:
>> The problem is that I am having ~70M rows in a table, but its bat storage size
>> is showing 170M rows. Although the count * still returns 70M, but memory is
>> increasing very heavy each day. As described in previous mail, each day, ~1M
>> rows are deleted from table and ~1.1M added, but it actually increases the size
>> by 1.1M.
>> Please let me know if something wrong in this way of updation of monetDB tables.
>>
>> Regards,
>> manish
>>
>> On Sat, Mar 25, 2017 at 2:05 PM, Manish gupta < gahoimnshg@gmail.com > wrote:
>> Dear All,
>> Is there a way to truncate the table size when deleting some of the records? In
>> my application, I delete and insert ~1 million records daily. But effectively,
>> it should result in few hundreds additional records. But, I see the size of the
>> table increases by a millions ( that can be seen in storage(), as well as hard
>> disk size ). While the "select count(*) from table" reflects correct number. Is
>> there a way, I can stop increase in table size in my scheme?
>>
>>
>> Regards,
>> Manish
>>
>>
>> _______________________________________________
>> users-list mailing list
>> users-list@monetdb.org
>> https://www.monetdb.org/mailman/listinfo/users-list
>
> _______________________________________________
> users-list mailing list
> users-list@monetdb.org
> https://www.monetdb.org/mailman/listinfo/users-list
>
>
> _______________________________________________
> users-list mailing list
> users-list@monetdb.org
> https://www.monetdb.org/mailman/listinfo/users-list

--
| Stefan.Manegold@CWI.nl | DB Architectures   (DA) |
| www.CWI.nl/~manegold/  | Science Park 123 (L321) |
| +31 (0)20 592-4212     | 1098 XG Amsterdam  (NL) |
_______________________________________________
users-list mailing list
users-list@monetdb.org
https://www.monetdb.org/mailman/listinfo/users-list