Git Product home page Git Product logo

jedaitoolkit's Introduction

Please check pyJedAI for an implementation of JedAI in native Python.

Please check our paper for a detailed description of version 3.0.

The code for running JedAI on Apache Spark is available here.

The Web Application for running JedAI is available here. A video explaining how to use it is available here.

JedAI is also available as a Docker image here. See below for more details.

The latest version of JedAI-gui is available here.

Java gEneric DAta Integration (JedAI) Toolkit

JedAI constitutes an open source, high scalability toolkit that offers out-of-the-box solutions for any data integration task, e.g., Record Linkage, Entity Resolution and Link Discovery. At its core lies a set of domain-independent, state-of-the-art techniques that apply to both RDF and relational data. These techniques rely on an approximate, schema-agnostic functionality based on (meta-)blocking for high scalability.

JedAI can be used in three different ways:

  1. As an open source library that implements numerous state-of-the-art methods for all steps of the end-to-end ER work presented in the figure below.
  2. As a desktop application with an intuitive Graphical User Interface that can be used by both expert and lay users.
  3. As a workbench that compares the relative performance of different (configurations of) ER workflows.

This repository contains the code (in Java 8) of JedAI's open source library. The code of JedAI's desktop application and workbench is available in this repository.

Several datasets already converted into the serialized data type of JedAI can be found here.

You can find a short presentation of JedAI Toolkit here.

Citation

If you use JedAI, please cite the following paper:

George Papadakis, Leonidas Tsekouras, Emmanouil Thanos, George Giannakopoulos, Themis Palpanas and Manolis Koubarakis: "The return of JedAI: End-to-End Entity Resolution for Structured and Semi-Structured Data", in VLDB 2018 (pdf).

Consortium

JEDAI is a collaboration project involving the following partners:

JedAI Workflow

JedAI supports 3 workflows, as shown in the following images:

Below, we explain in more detail the purpose and the functionality of every step.

Data Reading

It transforms the input data into a list of entity profiles. An entity is a uniquely identified set of name-value pairs (e.g., an RDF resource with its URI as identifier and its set of predicates and objects as name-value pairs).

The following formats are currently supported:

  1. CSV
  2. RDF in any format, i.e., XML, OWL, HDT, JSON
  3. Relational Databases (mySQL, PostgreSQL)
  4. SPARQL endpoints
  5. Java serialized objects

Schema Clustering

This is an optional step, suitable for highly heterogeneous datasets with a schema comprising a large diversity of attribute names. To this end, it groups together attributes that are syntactically similar, but are not necessarily semantically equivalent.

The following methods are currently supported:

  1. Attribute Name Clustering
  2. Attribute Value Clustering
  3. Holistic Attribute Clustering

For more details on the functionality of these methods, see here.

Block Building

It clusters entities into overlapping blocks in a lazy manner that relies on unsupervised blocking keys: every token in an attribute value forms a key. Blocks are then extracted, possibly using a transformation, based on its equality or on its similarity with other keys.

The following methods are currently supported:

  1. Standard/Token Blocking
  2. Sorted Neighborhood
  3. Extended Sorted Neighborhood
  4. Q-Grams Blocking
  5. Extended Q-Grams Blocking
  6. Suffix Arrays Blocking
  7. Extended Suffix Arrays Blocking
  8. LSH MinHash Blocking
  9. LSH SuperBit Blocking

For more details on the functionality of these methods, see here.

Block Cleaning

Its goal is to clean a set of overlapping blocks from unnecessary comparisons, which can be either redundant (i.e., repeated comparisons that have already been executed in a previously examined block) or superfluous (i.e., comparisons that involve non-matching entities). Its methods operate on the coarse level of individual blocks or entities.

The following methods are currently supported:

  1. Size-based Block Purging
  2. Cardinality-based Block Purging
  3. Block Filtering
  4. Block Clustering

All methods are optional, but complementary with each other and can be used in combination. For more details on the functionality of these methods, see here.

Comparison Cleaning

Similar to Block Cleaning, this step aims to clean a set of blocks from both redundant and superfluous comparisons. Unlike Block Cleaning, its methods operate on the finer granularity of individual comparisons.

The following methods are currently supported:

  1. Comparison Propagation
  2. Cardinality Edge Pruning (CEP)
  3. Cardinality Node Pruning (CNP)
  4. Weighed Edge Pruning (WEP)
  5. Weighed Node Pruning (WNP)
  6. Reciprocal Cardinality Node Pruning (ReCNP)
  7. Reciprocal Weighed Node Pruning (ReWNP)
  8. BLAST
  9. Canopy Clusetring
  10. Extended Canopy Clustering

Most of these methods are Meta-blocking techniques. All methods are optional, but competive, in the sense that only one of them can part of an ER workflow. For more details on the functionality of these methods, see here. They can be combined with one of the following weighting schemes:

  1. Aggregate Reciprocal Comparisons Scheme (ARCS)
  2. Common Blocks Scheme (CBS)
  3. Enhanced Common Blocks Scheme (ECBS)
  4. Jaccard Scheme (JS)
  5. Enhanced Jaccard Scheme (EJS)
  6. Pearson chi-squared test

Entity Matching

It compares pairs of entity profiles, associating every pair with a similarity in [0,1]. Its output comprises the similarity graph, i.e., an undirected, weighted graph where the nodes correspond to entities and the edges connect pairs of compared entities.

The following schema-agnostic methods are currently supported:

  1. Group Linkage,
  2. Profile Matcher, which aggregates all attributes values in an individual entity into a textual representation.

Both methods can be combined with the following representation models.

  1. character n-grams (n=2, 3 or 4)
  2. character n-gram graphs (n=2, 3 or 4)
  3. token n-grams (n=1, 2 or 3)
  4. token n-gram graphs (n=1, 2 or 3)

For more details on the functionality of these bag and graph models, see here.

The bag models can be combined with the following similarity measures, using both TF and TF-IDF weights:

  1. ARCS similarity
  2. Cosine similarity
  3. Jaccard similarity
  4. Generalized Jaccard similarity
  5. Enhanced Jaccard similarity

The graph models can be combined with the following graph similarity measures:

  1. Containment similarity
  2. Normalized Value similarity
  3. Value similarity
  4. Overall Graph similarity

Any word or character-level pre-trained embeddings are also supported in combination with cosine similarity or Euclidean distance.

Entity Clustering

It takes as input the similarity graph produced by Entity Matching and partitions it into a set of equivalence clusters, with every cluster corresponding to a distinct real-world object.

The following domain-independent methods are currently supported for Dirty ER:

  1. Connected Components Clustering
  2. Center Clustering
  3. Merge-Center Clustering
  4. Ricochet SR Clustering
  5. Correlation Clustering
  6. Markov Clustering
  7. Cut Clustering

For more details on the functionality of these methods, see here.

For Clean-Clean ER, the following methods are supported:

  1. Unique Mapping Clustering
  2. Row-Column Clustering
  3. Best Assignment Clustering

For more details on the functionality of the first method, see here. The 2nd algorithm implements an efficient approximation of the Hungarian Algorithm, while the 3rd one implements an efficient, heuristic solution to the assignment problem in unbalanced bipartite graphs.

Similarity Join

Similarity Join conveys the state-of-the-art algorithms for accelerating the computation of a specific character- or token-based similarity measure in combination with a user-determined similarity threshold.

The following token-based similarity jon algorithms are supported:

  1. AllPairs
  2. PPJoin
  3. SilkMoth

The following character-based similarity jon algorithms are also supported:

  1. FastSS
  2. PassJoin
  3. PartEnum
  4. EdJoin
  5. AllPairs

Comparison Prioritization

Comparison Prioritization associates all comparisons in a block collection with a weight that is proportional to the likelihood that they involve duplicates and then, it emits them iteratively, in decreasing weight.

The following methods are currently supported:

  1. Local Progressive Sorted Neighborhood
  2. Global Progressive Sorted Neighborhood
  3. Progressive Block Scheduling
  4. Progressive Entity Scheduling
  5. Progressive Global Top Comparisons
  6. Progressive Local Top Comparisons

For more details on the functionality of these methods, see here.

How to add JedAI as a dependency to your project

Visit https://search.maven.org/artifact/org.scify/jedai-core

How to run JedAI as a Docker image

After installing Docker on your machine, type the following commands:

docker pull gmandi/jedai-webapp

docker run -p 8080:8080 gmandi/jedai-webapp

Then, open your browser and go to localhost:8080. JedAI should be running on your browser!

How to use JedAI with Python

You can combine JedAI with Python through PyJNIus (https://github.com/kivy/pyjnius).

Preparation Steps:

  1. Install python3 and PyJNIus (https://github.com/kivy/pyjnius).
  2. Install java 8 openjdk and openjfx for java 8 and configure it as the default java.
  3. Create a directory or a jar file with jedai-core and its dependencies. One approach is to use the maven-assembly-plugin (https://maven.apache.org/plugins/maven-assembly-plugin/usage.html), which will package everything to a single jar file: jedai-core-3.0-jar-with-dependencies.jar

In the following code block a simple example is presented in python 3. The code reads the ACM.csv file found at (JedAIToolkit/data/cleanCleanErDatasets/DBLP-ACM) and prints the entities found:

import jnius_config;
jnius_config.add_classpath('jedai-core-3.0-jar-with-dependencies.jar')

from jnius import autoclass

filePath = 'path_to/ACM.csv'
CsvReader = autoclass('org.scify.jedai.datareader.entityreader.EntityCSVReader')
List = autoclass('java.util.List')
EntityProfile = autoclass('org.scify.jedai.datamodel.EntityProfile')
Attribute = autoclass('org.scify.jedai.datamodel.Attribute')
csvReader = CsvReader(filePath)
csvReader.setAttributeNamesInFirstRow(True);
csvReader.setSeparator(",");
csvReader.setIdIndex(0);
profiles = csvReader.getEntityProfiles()
profilesIterator = profiles.iterator()
while profilesIterator.hasNext() :
    profile = profilesIterator.next()
    print("\n\n" + profile.getEntityUrl())
    attributesIterator = profile.getAttributes().iterator()
    while attributesIterator.hasNext() :
        print(attributesIterator.next().toString())

jedaitoolkit's People

Contributors

gpapadis avatar mthanos avatar eni-veres avatar lwj5 avatar xstrtok avatar cyber-cypher avatar alexjoom avatar gabrielepisciotta avatar swamikevala avatar ggianna avatar murray1991 avatar cvedetect avatar zikani03 avatar gioargyr avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.