Scrape Away From KimonoLabs
Written by Nikos Vaggalis   
Friday, 11 March 2016

A recent startup, KimonoLabs, established in 2014,  set out to change the scraping landscape. Its recently announced takeover by Palantir will shut down its data scraping service. Kimono has produced a desktop app into which devs can import its APIs before the end of March.

Picking up on the growing trend, driven in part by the growing IoT applications Kimono aimed to build web based RESTful APIs around them that would be accessible by developers and non developers alike.

Its early successes included integration with the Pebble smart watch for results of the 2014 World Cup.

Kimono's API's promised to take the burden off the user, who otherwise would had to write code and use tools such as Scrapy for his scraping tasks instead. In short Kimono did the scraping for you and provided the results via a standard API. 

But delay in establishing a sustainable revenue plan was probably the major reason for its takeover by Palantir.

When the platform began operating, signing up was free and access was unlimited to everyone. The plan was to charge users for having private APIs or charge them depending on the quantity of APIs they had deployed. However, there was no premium revenue collection scheme in place, charging business for using the platform, for example.

Pratap Ranade one of Kimono's co-founders explains:


"When we started two years ago, we had a clear mission to help people everywhere structure and extract value from data. Since then, we've realized that continuing to work in isolation on a general data collection tool simply won’t allow us to make the impact we want. When we met the team at Palantir, we were instantly excited by the potential — the incredible talent and access to the world's most important data problems – even if it meant no longer working on the kimono product."


However, there was no comment as to the whys of the takeover. One can speculate that the lack of a revenue stream was a major factor.

Palantir, although absorbing Kimono as an entity, will shut down its data scraping service.

This of course causes a lot of grief and inconvenience to the developers already having deployed software based on Kimono's API's.


To lessen the pain, Kimono has released a desktop application that developers can use to import their APIs. The importing can take place until the end of March. The app will be usable even after that date, but without official support or updates.

Kimono state that the app has similar functionality to the online web based one, but unfortunately it lacks a lot.

Previously we benefitted from a web based endpoint able to scrape data at predefined intervals, emits them classified and in JSON format. Data could be consumed by 'client' applications residing in a VPServer, that performed post processing and imported the resulting data into a database automatically and without user intervention.

Compare that to the desktop based Windows or Mac (there's no Linux version) application, where the user has to manually start the application, whenever that is possible; kick off the scraping task, grab the exported JSON data files and transfer them from local PC to the VPS by FTPing or other clever but involved means. 

Even worse the desktop app would not even run on my Windows 7 machine, dying with an exception instead. 

Update:

Ryan Rowe (co-founder of Kimonolabs) has pointed out that you can have a web based endpoint with the desktop app by using its Firebase integration. This will automatically put the scraped data to the cloud and provide JSON endpoints for VPS or other device based apps to access. Firebase has a generous free tier with 1GByte of storage and 10GBytes of data transfer up and 100GBytes down. So as long as you stay within these limits things aren't bad. 

However basing an application on a desktop app that isn't supported after March obviously doesn't provide any long term development prospects.

Thus alternatives, new or already well established ones, are looking forward to attracting Kimono's developer client base into their platforms.

One of those services is Scrapinhub's Portia that is offering a Kimono2Portia migration service where you just have to fill in your Kimono credentials and let it import your Kimono API's into their platform

Import.io which taken the initiative of setting up a  page Great alternatives to every feature you'll miss from Kimono Labs. While it does cover a wide range of options it does point to the its own offerings of an entry level automatic extraction as an online API, and its more 'advanced' point and click one as a desktop  app, which is therefore prone to the same difficulties as the KimonoLabs app. 

Other alternatives include ParsehubDiffbot or Apifier.  All offer more or less the same functionality, something that makes choosing the right one a task decided on the points. In this arena however, Import's free Udemy course on How to extract data from the web using import.io does gve it a competitive edge.

At least there are plenty of alternatives that will hopefully render Kimono's absence a bit more bearable.

More Information

KimonoLabs

Kimono2Portia

import.io

Related Articles

Doing Data Science 

Google Restricts Autocomplete API

 

To be informed about new articles on I Programmer, sign up for our weekly newsletter,subscribe to the RSS feed and follow us on, Twitter, FacebookGoogle+ or Linkedin

 

Banner


Google Opensources Privacy Library
08/11/2024

Google is making a new differential privacy library available as open source. PipelineDP4J is a Java-based library that can be used to analyse data sets while preserving privacy.



Gender Differences In Coding Style
13/11/2024

A novel investigation into the gender gap between men and women regarding coding ability was undertaken by Dr Siân Brooke. Her conclusion? There is a difference in the Python code [ ... ]


More News

 

espbook

 

Comments




or email your comment to: comments@i-programmer.info

  

Last Updated ( Friday, 11 March 2016 )