Skip to main content

Setting up Spring Cloud Data Flow Server

Spring cloud data flow server can be used to setup the pipeline of tasks or streams. Tasks and streams can be registered with data flow server and can be executed or monitored either using cloud data flow shell or data flow server UI console/ dashboard.

Setting up Data flow server

Spring cloud data flow server can be setup using two ways, either use spring provided spring boot application or develop your own spring cloud data flow server as spring boot application.

Using Spring provided Data flow server application

There are many ways to setup data flow server as given below.
  • Local Machine
  • Cloud Foundry
  • Kubernetes
Please refer below documentation for more details for above installation types.

Create Spring boot application as Data flow server

Here I will show you how to create your own spring boot application as spring cloud data flow server which you can run anywhere locally or in cloud. It gives you more freedom with application to setup the other dependencies like database or security features.

Maven Dependencies

We have used Spring Boot version 2.1.6.RELEASE which is supported by the current version of data flow server (2.2.1.RELEASE). Below are the dependencies required in pom.xml

In this properties file we are just creating an in-memory database using H2. However it is not required as it will automatically setup one if we don't configure it. We can use these properties to use any other database.


This is our main class which need to be annotated  with @EnableDataFlowServer annotation. Since we are going to run this application in our local machine, we need to exclude some auto configurations like cloud foundry, kubernetes etc. until we intend to actually configure them. When required we can remove them from exclusions and configure necessary properties for auto configuration. See the below code.
        exclude = {SessionAutoConfiguration.class,
public class MainDataFlowServerApplication {

    public static void main(String[] args) {, args);


Below GIT URL contains complete source code.

Running data flow server application

Once you done with all required code and configuration setup, you can build the application and use below maven command to start your data flow server.
mvn spring-boot:run
Once application is started, hit the below URL in browser and you will see the given screen as dashboard of data flow server application where you can see all registered task and streams.\
URL: http://localhost:9393/dashboard/


Check my another post on registering and executing Spring Batch application with Spring cloud data flow server.


  1. Thanks for Sharing..Keep Update…

    Web Hosting is the process of buying space for a website on the World Wide Web.
    Web Hosting Explained For Beginners
    Web Hosting


Post a Comment

Popular Posts

Setting up kerberos in Mac OS X

Kerberos in MAC OS X Kerberos authentication allows the computers in same domain network to authenticate certain services with prompting the user for credentials. MAC OS X comes with Heimdal Kerberos which is an alternate implementation of the kerberos and uses LDAP as identity management database. Here we are going to learn how to setup a kerberos on MAC OS X which we will configure latter in our application. Installing Kerberos In MAC we can use Homebrew for installing any software package. Homebrew makes it very easy to install the kerberos by just executing a simple command as given below. brew install krb5 Once installation is complete, we need to set the below export commands in user's profile which will make the kerberos utility commands and compiler available to execute from anywhere. Open user's bash profile: vi ~/.bash_profile Add below lines: export PATH=/usr/local/opt/krb5/bin:$PATH export PATH=/usr/local/opt/krb5/sbin:$PATH export LDFLAGS=&

SpringBoot - @ConditionalOnProperty example for conditional bean initialization

@ConditionalOnProperty annotation is used to check if specified property available in the environment or it matches some specific value so it can control the execution of some part of code like bean creation. It may be useful in many cases for example enable/disable service if specific property is available. Below are the attributes which can be used for property check. havingValue - Provide the value which need to check against specified property otherwise it will check that value should not be false. matchIfMissing - If true it will match the condition and execute the annotated code when property itself is not available in environment. name - Name of the property to be tested. If you want to test single property then you can directly put the property name as string like "" and if you have multiple properties to test then you can put the names like {"prop.name1","prop.name2"} prefix - It can be use when you want to apply some prefix to

Spring Batch tutorial with example

Spring batch is used to create and process the batch jobs. It provides various features like logging, job statistics, transaction management, restarting jobs. It is very helpful in processing of large dataset but with finite volume of data. In this tutorial we will learn how to create and execute the spring batch job. In our example we will create a job which will import all the words from a text file to database and then at last it will print the total number of words available in the database. Below is the project structure. Creating batch job Sample text file to import Below is the contents of text file which we use for importing the words. The list below gives you the 1000 most frequently used English words in alphabetical order. Once you've mastered the shorter vocabulary lists, this is the next step. It would take time to learn the entire list from scratch, but you are probably already familiar with some of these words. Feel free to copy this list into

Multiple data source with Spring boot, batch and cloud task

Here we will see how we can configure different datasource for application and batch. By default, Spring batch stores the job details and execution details in database. If separate data source is not configured for spring batch then it will use the available data source in your application if configured and create batch related tables there. Which may be the unwanted burden on application database and we would like to configure separate database for spring batch. To overcome this situation we will configure the different datasource for spring batch using in-memory database, since we don't want to store batch job details permanently. Other thing is the configuration of  spring cloud task in case of multiple datasource and it must point to the same data source which is pointed by spring batch. In below sections, we will se how to configure application, batch and cloud task related data sources. Application Data Source Define the data source in application properties or yml con