There are a number of ways to get data into BigQuery but one of the most convenient as far as Apps Script is concerned is to write it straight there from data generated during some Apps Script project. There are limits on Drive file size that's it's hard to get past - so you sometimes can't just dump it out and load it using the bigquery console tools, and also BigQuery takes something called delimited JSON as input - so you do have to do a little tweaking. Apps Script does have a BigQuery advanced service, but it also has a 10mb limit on data that can be loaded. However it's easy to load data in multiple stages. Here's how. Creating a datasetYou have to enable billing to use BigQuery, so use a developer console project that has it enabled. Just select the project in the big query console at https://bigquery.cloud.google.com Add a dataset and pick an Id. In this example, I've created a dataset called jeopardydata. It's going to hold more or less all the jeopardy questions ever asked . I came across it here on Reddit. Creating a tableWithin datasets. you can have tables,and you'll be tempted to create one here. However, I'll create a table when I load the data so we're done in the bigquery console for now. Creating a table will need a table reference object - it looks like this. By default, a load operation will create the table if it doesn't already exist. var tableReference = { projectId: "lateral-command-416", datasetId: "jeopardydata", tableId: "questions" }; Enabling BigQueryYou'll need to enable the advanced service for BigQuery, and also enable it in the target project. The SchemaWhen creating a table, you need to also make a schema to describe the data to expect. The data types don't exactly match the JavaScript types, but for my data, which is flat and simple, I only need to enable 3 types, driven by this object. // These are the data types I'll support for now. var bqTypes = { number:"FLOAT", string:"STRING", bool:"BOOLEAN" }; The DataMy source data is a 53mb Drive file, and is a JSON array of objects that looks like this. [{ "category": "HISTORY", "air_date": "2004-12-31", "question": "'For the last 8 years of his life, Galileo was under house arrest for espousing this man's theory'", "value": "$200", "answer": "Copernicus", "round": "Jeopardy!", "show_number": "4680" }, ... ] BigQuery needs JSON delimited data, so I'll use Apps Script to transform and load the data directly. I'm confident that the objects in each array element are consistent, so I'll just use the first row to create the fields for the schema var model = data[0]; var fields = Object.keys(model).reduce(function(p,c) { var t = typeof model[c]; if (!bqTypes[t]) { throw 'unsupported type ' + t; } p.push ( {name:c, type: bqTypes[t]} ); return p; },[]); The JobsSince my data is larger than 10mb, I have to load it in chunks, whilst at the same time transforming it from an array of objects, into new line delimited JSON. Here's how - I've actually limited my chunk size to 8mb in this example. // there is a max size that a urlfetch post size can be var MAX_POST = 1024 * 1024 * 8; // now we can make delimted json, but chunk it up var chunk = '', jobs = []; data.forEach (function (d,i,a) { // convert to string with "\n" var chunklet = JSON.stringify(d) + "\n"; if (chunklet.length > MAX_POST) { throw 'chunklet size ' + chunklet.length + ' is greater than max permitted size ' + MAX_POST; } // time to flush? if (chunklet.length + chunk.length > MAX_POST) { jobs.push ( BigQuery.Jobs.insert(job, tableReference.projectId, Utilities.newBlob(chunk)) ); chunk = ""; // after the first , we move to append job.configuration.load.writeDisposition = "WRITE_APPEND"; } // add to the pile chunk += chunklet; }); // finish off if (chunk.length) { jobs.push ( BigQuery.Jobs.insert(job, tableReference.projectId, Utilities.newBlob(chunk)) ); } ProgressOnce kicked off, you can find progress of your jobs in the BigQuery console. It looks like this when complete. Each load looks like this And here's a preview of the data in BigQuery The codeThe entire code is below. The file shown is public so you can use it if you like. You'll need to change the project id to one of your own. It takes quite a while so you may need to modify it to do it a section at a time. function makeBigQueryFileFromDrive() { // get the file questions var file = DriveApp.getFileById("0B92ExLh4POiZbXB0X2RDWXZqLTA"); // get the data var data = JSON.parse (file.getBlob().getDataAsString()); // this will create this table if needed var tableReference = { projectId: "lateral-command-416", datasetId: "jeopardydata", tableId: "questions" }; // load it to bigquery var jobs = loadToBigQuery (tableReference, data); Logger.log(jobs); // keep an eye on it Logger.log('status of job can be found here: https://bigquery.cloud.google.com/jobs/' + tableReference.projectId); } function loadToBigQuery (tableReference,data) { // These are the data types I'll support for now. var bqTypes = { number:"FLOAT", string:"STRING", bool:"BOOLEAN" }; // figure out the schema from the JSON data- assuming that the data is all consistent types var model = data[0]; var fields = Object.keys(model).reduce(function(p,c) { var t = typeof model[c]; if (!bqTypes[t]) { throw 'unsupported type ' + t; } p.push ( {name:c, type: bqTypes[t]} ); return p; },[]); // the load job var job = { configuration: { load: { destinationTable: tableReference, sourceFormat: "NEWLINE_DELIMITED_JSON", writeDisposition: "WRITE_TRUNCATE", schema:{ fields:fields } }, } }; // there is a max size that a urlfetch post size can be var MAX_POST = 1024 * 1024 * 8; // now we can make delimted json, but chunk it up var chunk = '', jobs = []; data.forEach (function (d,i,a) { // convert to string with "\n" var chunklet = JSON.stringify(d) + "\n"; if (chunklet.length > MAX_POST) { throw 'chunklet size ' + chunklet.length + ' is greater than max permitted size ' + MAX_POST; } // time to flush? if (chunklet.length + chunk.length > MAX_POST) { jobs.push ( BigQuery.Jobs.insert(job, tableReference.projectId, Utilities.newBlob(chunk)) ); chunk = ""; // after the first , we move to append job.configuration.load.writeDisposition = "WRITE_APPEND"; } // add to the pile chunk += chunklet; }); // finish off if (chunk.length) { jobs.push ( BigQuery.Jobs.insert(job, tableReference.projectId, Utilities.newBlob(chunk)) ); } return jobs; } For more like this, see Google Apps Scripts snippets. Why not join our forum, follow the blog or follow me on twitter to ensure you get updates when they are available. You want to learn Google Apps Script?Learning Apps Script, (and transitioning from VBA) are covered comprehensively in my my book, Going Gas - from VBA to Apps script, available All formats are available now from O'Reilly,Amazon and all good bookshops. You can also read a preview on O'Reilly. If you prefer Video style learning I also have two courses available. also published by O'Reilly. |
Services > Desktop Liberation - the definitive resource for Google Apps Script and Microsoft Office automation > Google Apps Scripts snippets >