diff --git a/demo/js/app.js b/demo/js/app.js index e421d894..95817c8e 100755 --- a/demo/js/app.js +++ b/demo/js/app.js @@ -45,13 +45,12 @@ function demoDataset() { , {id: 5, x: 6, y: 12, z: 18} ] }; - // this is all rather artificial here but would make more sense with more complex backend - var backend = new recline.Model.BackendMemory({ - metadata: metadata, - data: indata - }); - recline.Model.setBackend(backend); - var dataset = backend.getDataset(datasetId); + var dataset = new recline.Model.Dataset(metadata); + dataset.backendConfig = { + type: 'memory' + // deep copy so we do not touch original data ... + , data: $.extend(true, {}, indata) + }; return dataset; } @@ -63,11 +62,11 @@ function setupLoadFromWebstore(callback) { e.preventDefault(); var $form = $(e.target); var source = $form.find('input[name="source"]').val(); - var backend = new recline.Model.BackendWebstore({ + var dataset = new recline.Model.Dataset(); + dataset.backendConfig = { + type: 'webstore', url: source - }); - recline.Model.setBackend(backend); - var dataset = backend.getDataset(); + }; callback(dataset); }); } diff --git a/src/backend.js b/src/backend.js index 85d6091e..7fae84b6 100644 --- a/src/backend.js +++ b/src/backend.js @@ -9,71 +9,57 @@ this.recline = this.recline || {}; this.recline.Model = this.recline.Model || {}; (function($, my) { - // ## Convenience function - my.setBackend = function(backend) { - Backbone.sync = backend.sync; - }; + my.backends = {}; + + Backbone.sync = function(method, model, options) { + return my.backends[model.backendConfig.type].sync(method, model, options); + } // ## BackendMemory - uses in-memory data + // + // To use you should: + // + // A. provide metadata as model data to the Dataset + // + // B. Set backendConfig on your dataset with attributes: + // + // - type: 'memory' + // - data: hash with 2 keys: + // + // * headers: list of header names/labels + // * rows: list of hashes, each hash being one row. A row *must* have an id attribute which is unique. + // + // Example of data: + // + //
+ // {
+ // headers: ['x', 'y', 'z']
+ // , rows: [
+ // {id: 0, x: 1, y: 2, z: 3}
+ // , {id: 1, x: 2, y: 4, z: 6}
+ // ]
+ // };
+ //
my.BackendMemory = Backbone.Model.extend({
- // Initialize a Backend with a local in-memory dataset.
- //
- // NB: We can handle one and only one dataset at a time.
- //
- // :param dataset: the data for a dataset on which operations will be
- // performed. Its form should be a hash with metadata and data
- // attributes.
- //
- // - metadata: hash of key/value attributes of any kind (but usually with title attribute)
- // - data: hash with 2 keys:
- // - headers: list of header names/labels
- // - rows: list of hashes, each hash being one row. A row *must* have an id attribute which is unique.
- //
- // Example of data:
- //
- // {
- // headers: ['x', 'y', 'z']
- // , rows: [
- // {id: 0, x: 1, y: 2, z: 3}
- // , {id: 1, x: 2, y: 4, z: 6}
- // ]
- // };
- initialize: function(dataset) {
- // deep copy
- this._datasetAsData = $.extend(true, {}, dataset);
- _.bindAll(this, 'sync');
- },
- getDataset: function() {
- var dataset = new my.Dataset({
- id: this._datasetAsData.metadata.id
- });
- // this is a bit weird but problem is in sync this is set to parent model object so need to give dataset a reference to backend explicitly
- dataset.backend = this;
- return dataset;
- },
sync: function(method, model, options) {
var self = this;
if (method === "read") {
var dfd = $.Deferred();
- // this switching on object type is rather horrible
- // think may make more sense to do work in individual objects rather than in central Backbone.sync
if (model.__type__ == 'Dataset') {
var dataset = model;
- var rawDataset = this._datasetAsData;
- dataset.set(rawDataset.metadata);
dataset.set({
- headers: rawDataset.data.headers
+ headers: dataset.backendConfig.data.headers
});
- dataset.docCount = rawDataset.data.rows.length;
+ dataset.docCount = dataset.backendConfig.data.rows.length;
dfd.resolve(dataset);
}
return dfd.promise();
} else if (method === 'update') {
var dfd = $.Deferred();
if (model.__type__ == 'Document') {
- _.each(this._datasetAsData.data.rows, function(row, idx) {
+ _.each(model.backendConfig.data.rows, function(row, idx) {
if(row.id === model.id) {
- self._datasetAsData.data.rows[idx] = model.toJSON();
+ model.backendConfig.data.rows[idx] = model.toJSON();
}
});
dfd.resolve(model);
@@ -82,7 +68,7 @@ this.recline.Model = this.recline.Model || {};
} else if (method === 'delete') {
var dfd = $.Deferred();
if (model.__type__ == 'Document') {
- this._datasetAsData.data.rows = _.reject(this._datasetAsData.data.rows, function(row) {
+ model.backendConfig.data.rows = _.reject(model.backendConfig.data.rows, function(row) {
return (row.id === model.id);
});
dfd.resolve(model);
@@ -92,7 +78,7 @@ this.recline.Model = this.recline.Model || {};
alert('Not supported: sync on BackendMemory with method ' + method + ' and model ' + model);
}
},
- getDocuments: function(datasetId, numRows, start) {
+ getDocuments: function(model, numRows, start) {
if (start === undefined) {
start = 0;
}
@@ -100,36 +86,32 @@ this.recline.Model = this.recline.Model || {};
numRows = 10;
}
var dfd = $.Deferred();
- rows = this._datasetAsData.data.rows;
+ rows = model.backendConfig.data.rows;
var results = rows.slice(start, start+numRows);
dfd.resolve(results);
return dfd.promise();
}
});
+ my.backends['memory'] = new my.BackendMemory();
- // ## Webstore Backend for connecting to the Webstore
+ // ## BackendWebstore
//
- // Initializing model argument must contain a url attribute pointing to
- // relevant Webstore table.
+ // Connecting to [Webstores](http://github.com/okfn/webstore)
//
- // Designed to only attach to one dataset and one dataset only ...
- // Could generalize to support attaching to different datasets
+ // To use this backend set backendConfig on your Dataset as:
+ //
+ //
+ // {
+ // 'type': 'webstore',
+ // 'url': url to relevant Webstore table
+ // }
+ //
my.BackendWebstore = Backbone.Model.extend({
- getDataset: function(id) {
- var dataset = new my.Dataset({
- id: id
- });
- dataset.backend = this;
- return dataset;
- },
sync: function(method, model, options) {
if (method === "read") {
- // this switching on object type is rather horrible
- // think may make more sense to do work in individual objects rather than in central Backbone.sync
- if (this.__type__ == 'Dataset') {
- var dataset = this;
- // get the schema and return
- var base = this.backend.get('url');
+ if (model.__type__ == 'Dataset') {
+ var dataset = model;
+ var base = dataset.backendConfig.url;
var schemaUrl = base + '/schema.json';
var jqxhr = $.ajax({
url: schemaUrl,
@@ -151,14 +133,14 @@ this.recline.Model = this.recline.Model || {};
}
}
},
- getDocuments: function(datasetId, numRows, start) {
+ getDocuments: function(model, numRows, start) {
if (start === undefined) {
start = 0;
}
if (numRows === undefined) {
numRows = 10;
}
- var base = this.get('url');
+ var base = model.backendConfig.url;
var jqxhr = $.ajax({
url: base + '.json?_limit=' + numRows,
dataType: 'jsonp',
@@ -172,44 +154,40 @@ this.recline.Model = this.recline.Model || {};
return dfd.promise();
}
});
+ my.backends['webstore'] = new my.BackendWebstore();
- // ## DataProxy Backend for connecting to the DataProxy
+ // ## BackendDataProxy
+ //
+ // For connecting to [DataProxy-s](http://github.com/okfn/dataproxy).
//
- // Example initialization:
+ // Set a Dataset to use this backend:
//
- // BackendDataProxy({
- // model: {
- // url: {url-of-data-to-proxy},
- // type: xls || csv,
- // format: json || jsonp # return format (defaults to jsonp)
- // dataproxy: {url-to-proxy} # defaults to http://jsonpdataproxy.appspot.com
- // }
- // })
+ // dataset.backendConfig = {
+ // // required
+ // url: {url-of-data-to-proxy},
+ // format: csv | xls,
+ // }
+ //
+ // When initializing the DataProxy backend you can set the following attributes:
+ //
+ // * dataproxy: {url-to-proxy} (optional). Defaults to http://jsonpdataproxy.appspot.com
+ //
+ // Note that this is a **read-only** backend.
my.BackendDataProxy = Backbone.Model.extend({
defaults: {
dataproxy: 'http://jsonpdataproxy.appspot.com'
- , type: 'csv'
- , format: 'jsonp'
- },
- getDataset: function(id) {
- var dataset = new my.Dataset({
- id: id
- });
- dataset.backend = this;
- return dataset;
},
sync: function(method, model, options) {
if (method === "read") {
- // this switching on object type is rather horrible
- // think may make more sense to do work in individual objects rather than in central Backbone.sync
- if (this.__type__ == 'Dataset') {
- var dataset = this;
- // get the schema and return
- var base = this.backend.get('dataproxy');
- var data = this.backend.toJSON();
- delete data['dataproxy'];
+ if (model.__type__ == 'Dataset') {
+ var dataset = model;
+ var base = my.backends['dataproxy'].get('dataproxy');
// TODO: should we cache for extra efficiency
- data['max-results'] = 1;
+ var data = {
+ url: dataset.backendConfig.url
+ , 'max-results': 1
+ , type: dataset.backendConfig.format
+ };
var jqxhr = $.ajax({
url: base
, data: data
@@ -228,17 +206,19 @@ this.recline.Model = this.recline.Model || {};
alert('This backend only supports read operations');
}
},
- getDocuments: function(datasetId, numRows, start) {
+ getDocuments: function(dataset, numRows, start) {
if (start === undefined) {
start = 0;
}
if (numRows === undefined) {
numRows = 10;
}
- var base = this.get('dataproxy');
- var data = this.toJSON();
- delete data['dataproxy'];
- data['max-results'] = numRows;
+ var base = my.backends['dataproxy'].get('dataproxy');
+ var data = {
+ url: dataset.backendConfig.url
+ , 'max-results': numRows
+ , type: dataset.backendConfig.format
+ };
var jqxhr = $.ajax({
url: base
, data: data
@@ -258,42 +238,35 @@ this.recline.Model = this.recline.Model || {};
return dfd.promise();
}
});
+ my.backends['dataproxy'] = new my.BackendDataProxy();
// ## Google spreadsheet backend
//
- //
+ // Connect to Google Docs spreadsheet. For write operations
my.BackendGDoc = Backbone.Model.extend({
- getDataset: function(id) {
- var dataset = new my.Dataset({
- id: id
- });
- dataset.backend = this;
- return dataset;
- },
sync: function(method, model, options) {
if (method === "read") {
- console.log('fetching data from url', model.backend.get('url'));
var dfd = $.Deferred();
- var dataset = this;
+ var dataset = model;
- $.getJSON(model.backend.get('url'), function(d) {
- result = model.backend.gdocsToJavascript(d);
+ $.getJSON(model.backendConfig.url, function(d) {
+ result = my.backends['gdocs'].gdocsToJavascript(d);
model.set({'headers': result.header});
- model.backend.set({'data': result.data, 'headers': result.header});
+ // cache data onto dataset (we have loaded whole gdoc it seems!)
+ model._dataCache = result.data;
dfd.resolve(model);
})
-
return dfd.promise(); }
},
- getDocuments: function(datasetId, start, numRows) {
+ getDocuments: function(dataset, start, numRows) {
var dfd = $.Deferred();
- var fields = this.get('headers');
+ var fields = dataset.get('headers');
// zip the field headers with the data rows to produce js objs
// TODO: factor this out as a common method with other backends
- var objs = _.map(this.get('data'), function (d) {
+ var objs = _.map(dataset._dataCache, function (d) {
var obj = {};
_.each(_.zip(fields, d), function (x) { obj[x[0]] = x[1]; })
return obj;
@@ -361,6 +334,7 @@ this.recline.Model = this.recline.Model || {};
});
return results;
}
-
});
+ my.backends['gdocs'] = new my.BackendGDoc();
+
}(jQuery, this.recline.Model));
diff --git a/src/model.js b/src/model.js
index 367d5b66..9eccc2f1 100644
--- a/src/model.js
+++ b/src/model.js
@@ -11,9 +11,10 @@ this.recline.Model = this.recline.Model || {};
// * docCount: total number of documents in this dataset (obtained on a fetch for this Dataset)
my.Dataset = Backbone.Model.extend({
__type__: 'Dataset',
- initialize: function() {
+ initialize: function(options) {
this.currentDocuments = new my.DocumentList();
this.docCount = null;
+ this.backend = null;
},
// ### getDocuments
@@ -30,10 +31,14 @@ this.recline.Model = this.recline.Model || {};
// This also illustrates the limitations of separating the Dataset and the Backend
getDocuments: function(numRows, start) {
var self = this;
+ var backend = my.backends[this.backendConfig.type];
var dfd = $.Deferred();
- this.backend.getDocuments(this.id, numRows, start).then(function(rows) {
+ backend.getDocuments(this, numRows, start).then(function(rows) {
var docs = _.map(rows, function(row) {
- return new my.Document(row);
+ var _doc = new my.Document(row);
+ _doc.backendConfig = self.backendConfig;
+ _doc.backend = backend;
+ return _doc;
});
self.currentDocuments.reset(docs);
dfd.resolve(self.currentDocuments);
diff --git a/test/model.test.js b/test/model.test.js
index 186c4d16..2e5f4415 100644
--- a/test/model.test.js
+++ b/test/model.test.js
@@ -12,21 +12,20 @@
var indata = {
headers: ['x', 'y', 'z']
, rows: [
- {id: 0, x: 1, y: 2, z: 3}
- , {id: 1, x: 2, y: 4, z: 6}
- , {id: 2, x: 3, y: 6, z: 9}
- , {id: 3, x: 4, y: 8, z: 12}
- , {id: 4, x: 5, y: 10, z: 15}
- , {id: 5, x: 6, y: 12, z: 18}
- ]
+ {id: 0, x: 1, y: 2, z: 3}
+ , {id: 1, x: 2, y: 4, z: 6}
+ , {id: 2, x: 3, y: 6, z: 9}
+ , {id: 3, x: 4, y: 8, z: 12}
+ , {id: 4, x: 5, y: 10, z: 15}
+ , {id: 5, x: 6, y: 12, z: 18}
+ ]
+ };
+ var dataset = new recline.Model.Dataset(metadata);
+ dataset.backendConfig = {
+ type: 'memory'
+ // deep copy so we do not touch original data ...
+ , data: $.extend(true, {}, indata)
};
- // this is all rather artificial here but would make more sense with more complex backend
- backend = new recline.Model.BackendMemory({
- metadata: metadata,
- data: indata
- });
- recline.Model.setBackend(backend);
- var dataset = backend.getDataset(datasetId);
expect(9);
dataset.fetch().then(function(dataset) {
equal(dataset.get('name'), metadata.name);
@@ -45,13 +44,13 @@
var newVal = 10;
doc1.set({x: newVal});
doc1.save().then(function() {
- equal(backend._datasetAsData.data.rows[0].x, newVal);
+ equal(dataset.backendConfig.data.rows[0].x, newVal);
})
// Test Delete
doc1.destroy().then(function() {
- equal(backend._datasetAsData.data.rows.length, 5);
- equal(backend._datasetAsData.data.rows[0].x, indata.rows[1].x);
+ equal(dataset.backendConfig.data.rows.length, 5);
+ equal(dataset.backendConfig.data.rows[0].x, indata.rows[1].x);
});
});
});
@@ -134,11 +133,11 @@
};
test('Webstore Backend', function() {
- var backend = new recline.Model.BackendWebstore({
+ var dataset = new recline.Model.Dataset();
+ dataset.backendConfig = {
+ type: 'webstore',
url: 'http://webstore.test.ckan.org/rufuspollock/demo/data'
- });
- recline.Model.setBackend(backend);
- dataset = backend.getDataset();
+ };
var stub = sinon.stub($, 'ajax', function(options) {
if (options.url.indexOf('schema.json') != -1) {
@@ -234,11 +233,11 @@
test('DataProxy Backend', function() {
// needed only if not stubbing
// stop();
- var backend = new recline.Model.BackendDataProxy({
+ var dataset = new recline.Model.Dataset();
+ dataset.backendConfig = {
+ type: 'dataproxy',
url: 'http://webstore.thedatahub.org/rufuspollock/gold_prices/data.csv'
- });
- recline.Model.setBackend(backend);
- dataset = backend.getDataset();
+ };
var stub = sinon.stub($, 'ajax', function(options) {
var partialUrl = 'jsonpdataproxy.appspot.com';
@@ -435,10 +434,11 @@
test("GDoc Backend", function() {
- var backend = new recline.Model.BackendGDoc({url: 'https://spreadsheets.google.com/feeds/list/0Aon3JiuouxLUdDQwZE1JdV94cUd6NWtuZ0IyWTBjLWc/od6/public/values?alt=json'
- });
- recline.Model.setBackend(backend);
- dataset = backend.getDataset();
+ var dataset = new recline.Model.Dataset();
+ dataset.backendConfig = {
+ type: 'gdocs',
+ url: 'https://spreadsheets.google.com/feeds/list/0Aon3JiuouxLUdDQwZE1JdV94cUd6NWtuZ0IyWTBjLWc/od6/public/values?alt=json'
+ };
console.log('got gdoc dataset', dataset);