Here\'s my problem. I want to ingest lots and lots of data .... right now millions and later billions of rows.
I recently started dealing with large sets of data stored in XML files.I\'ve always wondered how Facebook and other networking sites store all the information associated with individual profiles (name
I\'ve been given the task to extract all google analytics data for our website and analyse it and put in the database and then producing some reports on that data.
I would like to create huge data sets (25 ints a row, 30 rows per second, multiply that by 60). On the other hand, I want to query it for rows that matc开发者_运维问答h a certain condition (e.g. rows
I\'m looking into chunking my data source for optimial data import into solr and was wondering if it was possible to use a master url that chunked data into sections.
Faster way to insert rows into a table from a select statement and insert into? Insert into PartyMain Select [PermID]
We are looking for some recommendations around systematically re-indexing in Solr an ever growing corpus of documents (tens of millions now, hundreds of millions in than a year) without taking the cur
I might be going about this all wrong, but here\'s my problem and proposed solution: You have a 50+ gigabyte file with hundreds开发者_StackOverflow社区 of millions of independent records that need to
This was a pretty good question that was posed to me recently. Suppose we have a hypothetical (insert your favori开发者_开发技巧te data storage tool here) database that consists of the names, ages and
Is it possible to read a f开发者_如何学Cile line-by-line with Amazon S3?I\'m looking to let people upload large files somewhere, then have some code (probably running on Amazon) read their file line-b