
It’s 2006 and the founding team of Michael Berthold, Thomas Gabriel, Peter Ohl and Bernd Wiswedel took a moment here for this photo in between discussing who really did write the first line of code.
The very first version of KNIME Analytics Platform 1.0 was released on July 28, 2006 (and the first release t-shirt was born, sported here by Thorsten Meinl).
Author: Frank Dullweber, Böhringer Ingelheim / April 2016
Modern biological research deals with approaches for modifying genetic information (genotype), i.e. the genetic makeup of a cell - and therefore of an organism. It is this information that determines the characteristics of that cell. In other words the change of the genotype can lead to an observable change of a cell or whole organism (phenotype).
Human diseases are based on different causes - from a clearly identified (mono) genetic disorder to influences of the environment as well as life style or a mixture of both. In pharmaceutical drug development it is becoming more and more important to understand the molecular basis of a disease in order to be able to modify this disease in an efficient manner. Using the vocabulary introduced in the first paragraph, we could say that the researcher who observes a specific phenotype would like to understand the underlying genotype to find the best medication for a patient.
A technique to study such relationships is gene editing, It can be seen as the molecular scissors to insert, delete or replace genomic information in the genome of a cell. The CRISPR-Cas technology is a relatively new, simple technology in the field of gene editing and was selected by the Science Magazine as the "2015 Breakthrough of the Year". This technique was initially described by Jennifer Doudna and Emmanuelle Charpentier. Several scientists call it the biggest biotechnological advance since the polymerase chain reaction (PCR). Its inventor Kary Mullis was honored with the Nobel Prize in Chemistry in the year 1993.
The CRISPR-Cas technology is of high interest in the scientific community. Therefore my goal was to identify the most important authors in this area by more objective indicators and guide colleagues to identify scientific conferences with speakers of high scientific expertise.
Author: Julien Grossmann, Political and Violent Risk Data Analyst, HIS Economic and Country Risk
I’ve been using KNIME Analytics Platform for a year and a half, and in this time, KNIME has become a vital part of my work. As a political and violent risks data scientist, I am often confronted with incomplete or badly structured data. But with KNIME, I can always find ways to efficiently clean up, organize and analyze my data, and this, despite a total lack of programming or coding knowledge.
One of the most frustrating aspects of working with unstructured or semi structured data, is that they are rarely ready made for your needs. Therefore, they always require extensive manual clean up.
Worse, if your needs are constantly shifting, the data restructuration and clean up are too.
For instance, the typical data I work with involve datasets of events (such as civil unrest or terrorism attacks). Most data sets have basic meta data like date, location, and a short description of the incident, but there is only so much you can conclude from such basic dataset. Often I need to drill down, and look for specific groups, or specific actions or targets.
In the old days, we would use basic search functions in excel, or for the less geeky of us, we would do it manually.
So I decided to create a little workflow using the KNIME Textprocessing extension. The idea was to be able to mine large datasets rapidly, using a customizable list of keywords.
Interactive visualizations on the web have become very popular recently. The JavaScript framework D3 has developed into one of the most used libraries by many websites and publishers for rich graphics and visualizations.
In this blog post I’d like to show you how easy it is to harness D3’s capabilities and bring them into KNIME. To do so I am taking an existing example from the D3 website and will create an interactive view using the Generic JavaScript view in just 10 minutes.
So let’s get started.
The KNIME Streaming Executor is an extension that currently "hides" inside KNIME Labs. Not many of our users are aware it exists, so let's find out what it can (and can't) do and why you should be interested.
If you are used to KNIME's standard execution model you will know that connected nodes in a workflow are executed one after the other. The obvious advantage is that you can easily understand what is currently executing and what the (intermediate) results are, produced by each of the nodes in your workflow. That significantly simplifies debugging, because you can see immediately if some intermediate step isn't producing the expected result. You can also reset/continue with the execution at any point in the workflow without re-running the whole workflow -- saving on computation time. Another benefit of the standard execution model is that you can use KNIME's Hiliting to inspect the results and also explore your data using some of the view nodes.
However, the standard execution model also has some drawbacks. Each node needs to cache its results, which requires additional temporary space. (Note that KNIME doesn't duplicate the data from node to node but saves only what has changed between subsequent nodes ... so this isn't as bad as it might sound.) Additionally, as a node doesn't start processing the data until its upstream node has completed, it will need to read the dataset from start to end. Depending on whether those data live in main memory or on hard disc (which is KNIME's choice and depends on data size + available memory) this may require additional I/O operations, which can slow things down. You will notice that if you have a long chain of simple preprocessing nodes, most of the time is spent on just reading the data and caching the intermediate results.
Note. In KNIME Analytics Platform 3.3 or higher, a number of standard examples are in the folder called Example Workflows in the workspace. You’ll find a topic detector for social media, a recommendation engine to be used in retail, some classic examples for customer intelligence (churn prediction, credit scoring, and customer segmentation), and a few additional basics examples including data blending, reporting, and a simple predictive model training.
Supposing you have already downloaded the KNIME® Analytics Platform, here are 7 steps to make your learning phase more practical, more application oriented, and ultimately faster.
If you are a KNIME user you are probably familiar with the mechanism that lets you install additional extensions and update an existing installation with later versions. If you are a KNIME developer you have probably wondered what kind of magic is involved to make this possible: getting from hundreds of line of Java code to an online update that allows users to install and update extensions. In this week's blog post we will reveal some (if not all) of this magic.
The wizard we are using in KNIME is called Buckminster (not be confused with Richard Buckminster Fuller). Buckminster's product sheet describes it as "a set of frameworks and tools for automating build, assemble & deploy (BA&D) development processes in complex or distributed component-based development". The nice thing about Buckminster is that you can use it in your SDK but there is also a headless application which is suitable for fully automating the process.
Although Buckminster is quite well documented (the BuckyBook), initially it can be overwhelming. Therefore we would like to take a closer look at the essentials that are required to build an update site from scratch.

At the KNIME Spring Summit 2016 – Berlin, KNIME invited KNIME users and enthusiasts to meet together for a week of training courses, presentations and workshops. We were really happy to be able to welcome so many people to Berlin.
The Summit opened with a talk by Michael Berthold looking back at how the KNIME Spring Summit has grown from a small user group meeting to a much larger summit – and looking forward to how KNIME plans to advance to ensure even more flexibility, openness and ease of integration in KNIME Analytics Platform for 2016.
![]() |
![]() |
![]() |
![]() |
![]() |
Thanks to Bob Muenchen (muenchen.bob@gmail.com) for permission to share his post from r4stats.com blog post of February 22, 2016
http://r4stats.com/2016/02/22/vcf/
That post refers to two graphs here labelled "Figure1: Gartner Magic Quadrant for 2016. What’s missing?" and "Figure2: Figure 2. Gartner Magic Quadrant for 2015". Since we do not have permission to show those graphs on our site, you might want to quickly click on the link to view them before returning here to read more.
The IT research firm, Gartner, Inc. has released its February 2016 report, Magic Quadrant for Advanced Analytics Platforms. The report’s main graph shows the completeness of each company’s vision plotted against its ability to achieve that vision (Figure 1.) I include this plot each year in my continuously updated article, The Popularity of Data Analysis Software, along with a brief summary of its major points.

In surveys about the most-used tool for data analysis Excel always comes in as one of the most commonly used tools . It is taught in schools and used by countless companies. What you may not know, however, is that anything you can do with Excel you can also do using the nodes in KNIME Analytics Platform! This post is dedicated to getting you started if you already use Excel and want to migrate to KNIME Analytics Platform.
We will be using historical weather data from Berlin. These measurements can be downloaded from the Deutsche Wetterdienst. To be more specific, we took most recent measurements from the "Berlin-Tempelhof" weather station, from September 2014 to January 2016.