Skip to main content

Entity to DTO conversion in Java using Jackson

It's very common to have the DTO class for a given entity in any application. When persisting data, we use entity objects and when we need to provide the data to end user/application we use DTO class. Due to this we may need to have similar properties on DTO class as we have in our Entity class and to share the data we populate DTO objects using entity objects. To do this we may need to call getter on entity and then setter on DTO for the same data which increases number of code line. Also if number of DTOs are high then we need to write lot of code to just get and set the values or vice-versa.
To overcome this problem we are going to use Jackson API and will see how to do it with minimal code only.

Maven dependency


Entity class

Below is our entity class.
@Table(name = "EMPLOYEE")
public class Employee{
    @Id @GeneratedValue
    @Column(name = "id")
    private long id;
    @Column(name = "name")
    private String name;
    @Column(name = "age")
    private Integer age;
    public Employee() {}
    public Employee(String name, Integer age) {
        super(); = name;
        this.age = age;
    //getter methods
    //setter methods

DTO Class

Below is our DTO class which contains @JsonIgnoreProperties annotation to ignore the Employee instance properties during conversion which are not available in EmpDto class. Also EmptDto can have extra fields which are not available in entity class. I have not create the property for "age" in DTO which is there in entity class to showcase this scenario. Also it has "dept" property which is not there in entity class.
@JsonIgnoreProperties(ignoreUnknown = true)
class EmpDto {
    private String name;
    private long id;
    private String dept = "IT";

    public String getName() {
        return name;

    public void setName(String name) { = name;
    //getter methods
    //setter methods
    public String toString() {
        return "ID: " + id + ", Dept: " + dept + ", Name: " + name;

Conversion code

Below generic method is created using Jackson API which can be used for any type of class conversion.
public  T convertObjToXXX(Object o, TypeReference ref) {
    ObjectMapper mapper = new ObjectMapper();
    return mapper.convertValue(o, ref);

Entity class conversion and execution

Now we will create the object of Employee class (Entity) and convert it to multiple types like Map, Properties and DTO class (can be any POJO class).
//Create entity object to be converted
Employee emp = new Employee("Emp-1", 30);
emp =;//saving to populate the id field

//Convert to Map<String, String> class
Map<String, String> map = convertObjToXXX(emp, new TypeReference<Map<String, String>>(){});

System.out.println("Convert to Map<String, String> :");

//Convert to Properties class
Properties props = convertObjToXXX(emp, new TypeReference<Properties>(){});
System.out.println("\nConvert to Properties :");

//Convert to DTO class
EmpDto dto = convertObjToXXX(emp, new TypeReference<EmpDto>(){});
System.out.println("\nConvert to DTO :");


Below is the output of execution.
Convert to Map :
{id=1, name=Emp-1, age=30}

Convert to Properties :
{age=30, name=Emp-1, id=1}

Convert to DTO :
ID: 1, Dept: IT, Name: Emp-1

You may check my another post where I have explained how to use annotation processor to generate the DTO classes using Entity class at compile time. Link is given below.


Popular Posts

Setting up kerberos in Mac OS X

Kerberos in MAC OS X Kerberos authentication allows the computers in same domain network to authenticate certain services with prompting the user for credentials. MAC OS X comes with Heimdal Kerberos which is an alternate implementation of the kerberos and uses LDAP as identity management database. Here we are going to learn how to setup a kerberos on MAC OS X which we will configure latter in our application. Installing Kerberos In MAC we can use Homebrew for installing any software package. Homebrew makes it very easy to install the kerberos by just executing a simple command as given below. brew install krb5 Once installation is complete, we need to set the below export commands in user's profile which will make the kerberos utility commands and compiler available to execute from anywhere. Open user's bash profile: vi ~/.bash_profile Add below lines: export PATH=/usr/local/opt/krb5/bin:$PATH export PATH=/usr/local/opt/krb5/sbin:$PATH export LDFLAGS=&

SpringBoot - @ConditionalOnProperty example for conditional bean initialization

@ConditionalOnProperty annotation is used to check if specified property available in the environment or it matches some specific value so it can control the execution of some part of code like bean creation. It may be useful in many cases for example enable/disable service if specific property is available. Below are the attributes which can be used for property check. havingValue - Provide the value which need to check against specified property otherwise it will check that value should not be false. matchIfMissing - If true it will match the condition and execute the annotated code when property itself is not available in environment. name - Name of the property to be tested. If you want to test single property then you can directly put the property name as string like "" and if you have multiple properties to test then you can put the names like {"prop.name1","prop.name2"} prefix - It can be use when you want to apply some prefix to

Multiple data source with Spring boot, batch and cloud task

Here we will see how we can configure different datasource for application and batch. By default, Spring batch stores the job details and execution details in database. If separate data source is not configured for spring batch then it will use the available data source in your application if configured and create batch related tables there. Which may be the unwanted burden on application database and we would like to configure separate database for spring batch. To overcome this situation we will configure the different datasource for spring batch using in-memory database, since we don't want to store batch job details permanently. Other thing is the configuration of  spring cloud task in case of multiple datasource and it must point to the same data source which is pointed by spring batch. In below sections, we will se how to configure application, batch and cloud task related data sources. Application Data Source Define the data source in application properties or yml con