02 Feb 2015
I haven’t had many jobs in my career, so changing jobs feels pretty momentous: two weeks ago I had my last day at Boundless, and today will be my first at CartoDB.
I started with Boundless back in 2009 when it was OpenGeo and still a part of the Open Planning Project, a weird non-profit arm of a New York hedge fund millionaire’s corporate archipelago. (The hedgie, Mark Gorton, is still going strong, despite the brief set-back he endured when LimeWire was sued by RIAA.)
For that six year run, I was fortunate to have a lead role in articulating what it meant to “do open source” in the geospatial world, and to help to build OpenGeo into a self-supporting open source enterprise. We grew, spun out of the non-profit, gained lots of institutional customers, and I got to meet and work with lots of quality folks.
After six years though, I feel like I need a change, an opportunity to learn some new things and meet some new people: To move from the enterprise space, to the consumer space.
So I was very lucky when a new opportunity came along: to work for a company that is reimagining what it means to be a spatial database in a software-as-a-service world. Under the covers, CartoDB uses my favorite open source spatial database, PostGIS, to run their platform, and working for CartoDB gives me a chance to talk about and to work on something I like almost as much as (more than?) open source: spatial SQL!
The team at CartoDB have done a great job with their platform, providing a simple entry-point into map making, while still leaving the power of SQL exposed and available, so that users can transition from beginner, to explorer, to power user.
As someone who currently only knows a portion of their technology (the SQL bit), I’m looking forward to experiencing the rest of their platform as a beginner. I also know the platform folks will have lots of good questions for me on PostGIS internals, and we’ll have many interesting conversations about how to keep pushing PostgreSQL and PostGIS to the limits.
My two week between-jobs break was refreshing, but sometimes a change is as good as rest too. I enjoyed my last six years with Boundless and I’m looking forward to the future with CartoDB.
08 Jan 2015
Some very interesting news out of the USA today, as the GSA (the biggest, baddest procurement agency in the world) has released a Request For Information (RFI) on agile technology delivery.
To shift the software procurement paradigm, GSA’s 18F Team and the Office of Integrated Technology Services (ITS) is collaborating on the establishment of a BPA that will feature vendors who specialize in Agile Delivery Services. The goal of the proposed BPA is to decrease software acquisition cycles to less than four weeks (from solicitation to contract) and expedite the delivery of a minimum viable product (MVP) within three months or less.
In a wonderful “eat your own dogfood” move, the team working on building this new procurement vehicle are themselves adopting agile practices in their own process. Starting small with a pilot, working directly with the vendors who will be trying the new vehicle, etc. If the hidebound old GSA can develop a workable framework for agile procurement, then nobody else has an excuse.
(The reason procurement agencies have found it hard to specify “agile” is that agile by design does not define precise deliverables in advance, so it is damnably hard to fit into a “fixed cost bid” structure. In places where time-and-materials vehicles are already in place, lots of government organizations are already working with vendors in an agile way, but for the kinds of big, boondoggle-prone capital investment projects I write about, the waterfall model still predominates.)
30 Dec 2014
Brian Timoney called out this Tom Macwright quote, and he’s right, it deserves a little more discussion:
…the client side will eat more of the server side stack.
To understand what “more” there is left to eat, it’s worth enumerating what’s already been eaten (or, which is being consumed right now, as we watch):
- Interaction: OK, so this was always on the client side, but it’s worth noting that the impulse towards using a heavy-weight plug-in for interaction is now pretty much dead. The detritus of plug-in based solutions will be around for a long while, inching towards end-of-life, but not many new ones are being built. (I bet some are, though, in the bowels of organizations where IE remains an unbreakable corporate standard.
- Single-Layer Rendering: Go back almost 10 years and you’ll find OpenLayers doing client-side rendering, though using some pretty gnarly hacks at the time. Given the restrictions in rendering performance, a common way to break down an app was a static, tiled base map with a single vector layer of interest on top. (Or, for the truly performance oriented, a single raster layer on top, only switching to vector for editing purposes.) With modern browser technology, and good implementations, rendering very large numbers of features on the client has become commonplace, to the extent that the new bottleneck is no longer the CPU, it’s the network.
- All-the-layers Rendering: Already shown in-principle by Google Maps, tiled vector rendering is moving over the last 12 months rapidly from wow-wizzy-demo to oh-no-not-that-again status. Rather than rendering to raster on the server side, send a simplified version to the client for rendering there. For base maps there’s not a lot of benefit over pre-rendered raster, but there’s some: dynamic labelling means orientation is completely flexible, and also allows for multiple options for labelling; also, simplified vector tiles can serve a wider range of zoom levels while remaining attractively rendered, so the all-important network bandwidth issues can be addressed for mobile devices.
- “GIS” operations: While large scale analysis is not going to happen on a web page, a lot of visual effects that were otherwise hard to achieve can now be pushed to the client. Some of the GIS operations are actually in support of getting attractive client-side rendering: voronoi diagrams can be a great aid to label placement; buffers come in handy for cartography all the time.
- Persistence: Not really designed for long-term storage, but since any mobile application on a modern platform now has access to a storage area of pretty large size, there’s nothing stopping these new “client” applications from wandering far and completely untethered from the server/cloud for long periods of time.
Uh, what’s left?
- Long term storage and coordination remain. If people are going to work together on data, they need a common place to store and access their information from.
- Large scale data handling and analysis remain, thanks to those pesky narrow network pipes, for now.
Have I missed any?
03 Dec 2014
I haven’t had to build CUnit myself for a while, because most of the systems I work with have it in their packaged software repositories, but for Solaris it’s not there, and it turns out, it’s quite painful to build!
- First, the latest version on sourceforge, 2.1.3, is incorrectly bundled, with files (config.h.in) missing, so you have to use 2.1.2.
- Second, the directions in the README don’t include all the details of what autotools commands need to be run.
Here’s the commands I finally used to get a build. Note that you do need to run
libtoolize to get some missing support scripts installed, and that you need to also run automake in “add missing” mode to get let more support scripts. Then and only then do you get a build.
tar xvfj CUnit-2.1-2.tar.bz2
libtoolize -f -c -i \
&& aclocal \
&& autoconf \
&& automake --gnu --add-missing \
&& ./configure --prefix=/usr/local \
&& make \
&& make install
bootstrap file does provide the required autotools flags.
02 Dec 2014
Most users of PostGIS are safely ensconsed in the world of Linux, and their build/deploy environments are pretty similar to the ones used by the developers, so any problems they might experience are quickly found and removed early in development.
Some users are on Windows, but they are our most numerous user base, so we at least test that platform preemptively before release and make sure it is as good as we can make it.
And then there’s the rest. We’ve had a passel of FreeBSD bugs lately, and I’ve found myself doing Solaris builds for customers, and don’t get me started on the poor buggers running AIX. One of the annoyances of trying to fix a problem for a “weird platform” user is just getting the platform setup and running in the first place.
So, having recently learned a bit about vagrant, and seeing that some of the “weird” platforms have boxes already, I thought I would whip off a couple vagrant configurations so it’s easy in the future to throw up a Solaris or FreeBSD box, or even a quick Centos box for debugging purposes.
I’ve just been setting up my Solaris Vagrantfile and using my favourite Solaris crutch: the OpenCSW software repository. But as I use it, I’m not just adding the “things I need”, I’m implicitly choosing an environment:
libxml2 is from OpenCSV
- so is my
gcc, which is version 4, not version 3
- so is my
This is convenient for me, but what are the chances that it’ll be the environment used by someone on Solaris having problems? They might be compiling against libraries from
/usr/sfw/bin, or using the Solaris
gcc-3 package, or any number of other variants. At the end of the day, when testing on such a Solaris environment, will I be testing against a real situation, or a fantasyland of my own making?
For platforms like Ubuntu (apt) or Red Hat (yum) or FreeBSD (port) where there is One True Way to get software, the difficulties are less, but even then there is no easy way to get a “standard environment”, or to quickly replicate the combinations of versions a user might have run into that is causing problems (
libjson is a current source of pain). DLL hell has never really gone away, it has just found new ways to express itself.
(I will send a psychic wedgie to anyone who says “docker”, I’m not kidding.)