graphdb
TypeScript icon, indicating that this package has built-in type declarations

3.1.0-newjenkins30 • Public • Published

graphdb.js

Build Status Coverage install size npm version

A GraphDB and RDF4J data access library written in JavaScript to be used in Node.js.

Installation

Prerequisites

  • Node >= 8, tested up to v18
  • NPM (npm)
npm install graphdb

Development

The library is written in ES2016. During the build process source files src/ are transpiled to ES2015 and copied to lib/ directory.

A typescript definition file index.d.ts is generated in the lib/ as well.

Documentation is generated in docs/ from the JSDoc annotations in the source code. Source code documentation can be found here

Project Structure

  • src : The source code of the library.
  • test : Unit and component level tests written with Jest.
  • lib : Transpiled but files are built here when the library is published to npm.
  • docs : Documentation generated with JSDoc when the library is published to npm.
  • src/index.js : The library external API exporting main functional classes. This is also present in lib.
  • lib/index.d.ts : The typescript definitions generated when the library is published to npm.
  • scripts : Service scripts related to building, publishing and so on.

Setup Environment

  • Make sure prerequisites are covered: node.js and npm must be present and versions should be supported.
  • Clone the project from GitHub.
  • In project root directory and execute npm install to install all dependencies. Do the same in test-e2e directory.

Running Tests

The library has two types of tests: unit tests and e2e tests. Both are written with Jest.

Unit Tests

After any change the tests should be run and check if any existing functionality is not broken in result. To do so, run npm run test. This will run all tests in the test directory.

It is sometimes more convenient to run the tests in watch mode in which the tests are run every time a file is changed. To do so, run npm run test:watch.

e2e tests

All e2e tests are run against a local GraphDB instance. The tests are located in the test-e2e directory and can be run with the following command from the project root npm run e2e:run. This command will pack the project locally, install it and run all e2e tests against it. A graphdb instance must be running on http://localhost:7200. Using the docker-compose.yml file, a GraphDB instance can be started with the following command docker-compose up -d --build.

During development, it might be useful to run only a single test file. This can be done with the following command npm run e2e:run -t '{test_file_name.spec.js} or the npm run test:single command from the test-e2e package.

For convenience, there is a command that can be used to do a dry build and run all tests in a docker container. This can be done with the following command npm run e2e:test:docker. The command will build the project, start an instance of GraphDB and run all e2e tests against it the same way as the library is tested in the CI pipeline.

Checking the codestyle

The library uses Google style in conjunction with ESLint's recommended ruleset. The codestyle can be checked with the following command npm run lint

Testing the packaging

Library is managed by NPM package manager. During publishing npm consults the .gitignore, .npmignore and package.json#files property in order to decide which resources should be published. If any change in the project structure, .gitignore or .npmignore is made, then publishing must be verified in order to be guaranteed that needed sources will be properly published. The packaging could be verified using the npm pack command which generates a *.tgz archive in the project root. The archive should contain only the needed resources. Furture the archive could be used as a source for npm install where the path pointing the archive is provided.

Publishing

The library is published to npmjs.com. The publishing process is automated and the version is increased automatically.

Below are steps for manual publishing (usually this should not be needed):

  • Increase the package version.
npm version patch|minor|major
  • Login in npm.
npm login
  • Publish package in npm.
npm publish

Usage

GraphDB supports two REST APIs for working with RDF data: the RDF4J REST API and the GraphDB API. Two classes are used to interact with these APIs: ServerClient and GraphDBServerClient.

GraphDB allows connections to remote locations and data sources from a GraphDB instance. Locations represent individual servers or endpoints that store data. These can be other GraphDB instances or remote endpoints and can be attached, edited, and detached as needed.

ServerClient

The ServerClient implements the RDF4J REST API and handles server-level operations, such as retrieving a list of available repositories, accessing a specific repository, and deleting repositories. It works only with a local instance of GraphDB.

To use the ServerClient, it must first be configured with ServerClientConfig.

  • Configure ServerClient
const {ServerClient, ServerClientConfig} = require('graphdb').server;
const {RDFMimeType} = require('graphdb').http;

const serverConfig = new ServerClientConfig('http://rdf4j-compliant-server/')
    .setTimeout(5000)
    .setHeaders({
        'Accept': RDFMimeType.SPARQL_RESULTS_JSON
    })
    .setKeepAlive(true);

const server = new ServerClient(serverConfig);

When created, configurations receive the following default parameters:

    /**
    * The Server client configuration constructor
    * sets configuration default value to 
    * timeout = 10000,
    * keepAlive = true
    */
  • Fetch repository ids
server.getRepositoryIDs().then(ids => {
    // work with ids
}).catch(err => console.log(err));
  • Check if repository with given name exists
server.hasRepository('repository-name').then(exists => {
    if (exists) {
        // repository exists -> delete it for example
    }
}).catch(err => console.log(err));
  • Delete repository with given name
server.deleteRepository('repository-name').then(() => {
    // successfully deleted
}).catch(err => console.log(err));
  • Although a repository instance can be created using a constructor which can be seen in the examples below a client could obtain an instance of RDFRepositoryClient through the server client
server.getRepository('repository-name').then(repository => {
    // repository is a configured RDFRepositoryClient instance
}).catch(err => console.log(err));

GraphDBServerClient

The GraphDBServerClient extends the functionality of ServerClient by enabling operations with remote locations.

It is used to automate security user management, including adding, editing, and removing users. Additionally, it allows adding, editing, or removing repositories from any attached location.

  • Setup client
// Import all classes needed for work
const {GraphDBServerClient, ServerClientConfig} = require('graphdb').server; 
// Instance the server configuration
const serverConfig = new ServerClientConfig('http://GraphDB-server/')
    .useGdbTokenAuthentication('admin', 'root');
// Instance the server client
const serverClient = new GraphDBServerClient(serverConfig);
  • Create, read, update, delete user
 // create 
 serverClient.createUser('test_user', '123456');
 // update
 serverClient.updateUser('test_user', '111222');
 // read
 serverClient.getUser('test_user');
 //delete user
 serverClient.deleteUser('test_user'); 
  • Update user application settings data
 // Import application settings class
  const {AppSettings} = require('graphdb').server;
 // Use with extreme caution, as the changes that are made to the
 // application settings may possibly change the behavior of the
 // GraphDB Workbench for the logged-in user or for all users
 // if logged in as admin.
  const newAppSettings = new AppSettings(true, true, true, false);
  return serverClient.updateUserData('test_user', '111222', newAppSettings);
  • Fetch repository ids

    • From a local GraphDB instance:
        serverClient.getRepositoryIDs().then((ids) => {
          console.log(ids);
          // [ 'Test_repo' ]
        });
    • From a remote GraphDB instance:
        serverClient.getRepositoryIDs('http://remote-GraphDB-server').then((ids) => {
          console.log(ids);
          // [ 'Test_repo' ]
        });
  • Check if a repository with the given name exists

    • From a local GraphDB instance:
        server.hasRepository('Test_repo').then(exists => {
          if (exists) {
              // Repository exists -> delete it, for example
          }
        }).catch(err => console.log(err));
    • From a remote GraphDB instance:
        server.hasRepository('Test_repo', 'http://remote-GraphDB-server').then(exists => {
          if (exists) {
              // Repository exists -> delete it, for example
          }
        }).catch(err => console.log(err));
  • Get repo type default config

 serverClient.getDefaultConfig(RepositoryType.GRAPHDB).then((response) => {
    console.log(response);
 });
  • Get concrete repo configuration

    • From a local GraphDB instance:
     serverClient.getRepositoryConfig('Test_repo').then((response) => {
     console.log(response);
     });
    • From a remote GraphDB instance:
     serverClient.getRepositoryConfig('Test_repo', 'http://remote-GraphDB-server').then((response) => {
     console.log(response);
     });
  • Get concrete repo configuration as stream in turtle format.

    • From a local GraphDB instance:
       serverClient.downloadRepositoryConfig('Test_repo').then((stream) => {
       stream.on('data', (data) => {
            // data contains requested configuration in turtle format
         });
       });
    • From a remote GraphDB instance:
      serverClient.downloadRepositoryConfig('Test_repo', 'http://remote-GraphDB-server').then((stream) => {
      stream.on('data', (data) => {
           // data contains requested configuration in turtle format
        });
      });
  • Create repository

    • On a local GraphDB instance:
        // Import repository configuration class
        const {RepositoryConfig} = require('graphdb').repository;  
        // Create repository configuration
        const config = new RepositoryConfig('repo_id', '', new Map(), '',  'Repo title', RepositoryType.GRAPHDB);
        // Use the configuration to create new repository
        serverClient.createRepository(config)
            .then(() => {
              // do work
        });
    • On a remote GraphDB instance:
       // Import repository configuration class
       const {RepositoryConfig} = require('graphdb').repository;  
       // Create repository configuration
       // The second argument is the location of the remote GraphDB instnace where the repository will be created.
       const config = new RepositoryConfig('repo_id', 'http://remote-GraphDB-server', new Map(), '',  'Repo title', RepositoryType.GRAPHDB);
       // Use the configuration to create new repository
       serverClient.createRepository(config)
           .then(() => {
             // do work
       });
  • Delete repository

    • From a local GraphDB instance:
        serverClient.deleteRepository('new_repo').then(() => {
          // do work  
        });
    • From a remote GraphDB instance:
        serverClient.deleteRepository('new_repo', 'http://remote-GraphDB-server').then(() => {
          // do work  
        });
  • Checks if GraphDB security is enabled

 serverClient.isSecurityEnabled().then((response) => {
   console.log(response.response.data)
 });
  • Toggle GraphDB security
 // turn security off
 serverClient.toggleSecurity(false);
  • Check free access state
 serverClient.getFreeAccess().then((response) => {
         console.log(response.response.data.enabled);
 });
  • Update free access. Use with extreme caution, as the changes that are made to the application settings may possibly change the behavior of the GraphDB Workbench for all users.
 const authorities = [
      'WRITE_REPO_Test_repo',
      'READ_REPO_Test_repo'
    ];
 const appSettings = new AppSettings(true, true, false, true);
 
 serverClient.updateFreeAccess(true, authorities, appSettings);

RDFRepositoryClient

  • Instantiating repository client
const endpoint = 'http://GDB';
const readTimeout = 30000;
const writeTimeout = 30000;
const config = new RepositoryClientConfig(endpoint)
    .setEndpoints(['http://GDB/repositories/my-repo'])
    .setHeaders({
      'Accept': RDFMimeType.TURTLE
    })
    .setReadTimeout(readTimeout)
    .setWriteTimeout(writeTimeout);
const repository = new RDFRepositoryClient(config);

When created, configurations receive the following default parameters:

    /**
    * The Repository client configuration constructor
    * sets configuration default value to 
    * defaultRDFMimeType = 'application/sparql-results+json',
    * keepAlive = true,
    * readTimeout = 10000,
    * writeTimeout = 10000
    */
  • Obtaining repository client instance through a ServerClient
const {ServerClient, ServerClientConfig} = require('graphdb').server;
const {RepositoryClientConfig} = require('graphdb').repository;

const endpoint = 'http://GDB';
const config = new ServerClientConfig(endpoint);
const server = new ServerClient(config);

const readTimeout = 30000;
const writeTimeout = 30000;

const repositoryClientConfig = new RepositoryClientConfig(endpoint)
    .setEndpoints(['http://GDB/repositories/my-repo'])
    .setReadTimeout(readTimeout)
    .setWriteTimeout(writeTimeout);
return server.getRepository('automotive', repositoryClientConfig)
    .then((rdfRepositoryClient) => {
    // rdfRepositoryClient is a configured instance of RDFRepositoryClient
});

Reading

Statements could be fetched using the RDFRepositoryClient.get, RDFRepositoryClient.query, RDFRepositoryClient.download.

Every reading method can get the response parsed to data objects according to RDFJS data model specification (see Response Parsers).

  • Reading statements
const payload = new GetStatementsPayload()
	.setResponseType(RDFMimeType.RDF_JSON)
	.setSubject('<http://eunis.eea.europa.eu/countries/AZ>')
	.setPredicate('<http://eunis.eea.europa.eu/rdf/schema.rdf#population>')
	.setObject('"7931000"^^http://www.w3.org/2001/XMLSchema#integer')
	.setContext('<http://example.org/graph3>')
	.setInference(true);

return repository.get(payload).then((data) => {
	// data contains requested staments in rdf json format
});
  • Downloading data from repository by consuming a WritableStream
const dest = __dirname + '/statements.ttl';
const output = fs.createWriteStream(dest);
const payload = new GetStatementsPayload()
    .setResponseType(RDFMimeType.TURTLE)
    .get();
repository.download(payload).then((response) => {
    response.on('data', (chunk) => {
        output.write(new Buffer(chunk));
    });
    response.on('end', () => {
        output.end();
    });
});
  • Query evaluation against a sparql endpoint

  • SELECT query returning data objects

repository.registerParser(new SparqlXmlResultParser());

const payload = new GetQueryPayload()
  .setQuery('select * where {?s ?p ?o}')
  .setQueryType(QueryType.SELECT)
  .setResponseType(RDFMimeType.SPARQL_RESULTS_XML)
  .setLimit(100);

return repository.query(payload).then((stream) => {
  stream.on('data', (bindings) => {
    // the bindings stream converted to data objects with the registered parser
  });
  stream.on('end', () => {
    // handle end of the stream
  });
});
  • ASK query returning a boolean result
const payload = new GetQueryPayload()
  .setQuery('ask {?s ?p ?o}')
  .setQueryType(QueryType.ASK)
  .setResponseType(RDFMimeType.BOOLEAN_RESULT);

repository.registerParser(new SparqlJsonResultParser());

return repository.query(payload).then((data) => {
  // data => true|false
});

Writing

  • Uploading data in repository (PUT) using ReadStream
const contentType = RDFMimeType.TURTLE;
const turtleFile = __dirname + '/statements.ttl';
fs.readFile(turtleFile, (err, stream) => {
    repository.upload(stream, contentType).catch((e) => console.log(e));
});
  • Overwriting data in repository (PUT) using ReadStream
const contentType = RDFMimeType.TURTLE;
const file = __dirname + '/statements-overwrite.ttl';
fs.readFile(file, (err, stream) => {
    repository.overwrite(stream, contentType).catch((e) => console.log(e));
});
  • Executing a sparql update query
const payload = new UpdateQueryPayload()
  .setQuery('INSERT {?s ?p ?o} WHERE {?s ?p ?o}')
  .setContentType(QueryContentType.X_WWW_FORM_URLENCODED)
  .setInference(true)
  .setTimeout(5);

return repository.update(payload).then(() => {
    // repository should have been updated at this point
});

Deleting

  • Delete statement from given context
repository.deleteStatements(subj, pred, obj, contexts).then(() => {
    // do work
});

Value bindings

Value bindings can be added using the addBinding() method.

The binding variable must be an N-Triple (RDF triple, URI, blank node or literal).

  • For literals a language tag or data type can be specified.
const payload = new GetQueryPayload()
  .setQuery(query)
  .setQueryType(QueryType.SELECT)
  // Simple string literal
  .addBinding('name', "24")
  // Literal with specified data type
  .addBinding('age', "\"24\"^^<http:\/\/www.w3.org\/2001\/XMLSchema#int>")
  // Literal with specified language tag according to RFC 3066
  .addBinding('name', "\"abc\"@en-gb");

Transactions

Repository operations can be executed in transaction. In order to work with transactions the TransactionalRepositoryClient is used.

Starting a transaction

RDFRepositoryClient can initiate a transaction via beginTransaction() which produces an instance of TransactionalRepositoryClient.

Each started transaction allows to be committed or rolled back by using respectively commit() and rollback()

The following is a short use example of a transaction:

const turtlePath = __dirname + '/statements.ttl';

let transactionClient;
return repository.beginTransaction().then((transaction) => {
  transactionClient = transaction;
  return transactionClient.addFile(turtlePath);
}).then(() => {
  // File upload was successful, commit the changes
  return transactionClient.commit();
}).catch((e) => {
  console.log(e);
  if (transactionClient) {
    // Couldn't upload the file, abort the transaction
    return transactionClient.rollback();
  }
  return Promise.reject(e);
});

For specific isolation level use TransactionIsolationLevel

return repository.beginTransaction(TransactionIsolationLevel.READ_UNCOMMITTED);

The default isolation level is specific for each store implementation.

Important: After commit or rollback, a transaction cannot be reused, any attempts will result in an error. If you are not sure what is the state of the transaction, you can use transaction.isActive()

Working with a transaction

Almost all of the transaction methods for reading & modifying data have the same syntax and parameters as those in RDFRepositoryClient.

Reading

TransactionalRepositoryClient supports the following methods for reading data, including any changes that are not yet committed:

  • getSize()
  • get()
  • download()
  • query()
Writing
  • add()
  • addQuads()
  • upload()
  • addFile()

Deleting

Deleting data during a transaction is different than the one in RDFRepositoryClient, it expects RDF data document instead of statements filter parameters.

Currently, it supports only Turtle or TriG formatted RDF data:

const turtlePath = __dirname + '/statements.ttl';
const turtleData = fs.readFileSync(turtlePath, 'utf8');
return transaction.deleteData(turtleData);

Namespaces

  • Retrieving all available namespace declarations. The resolved value is an array of Namespace instances.
return repository.getNamespaces().then((namespaces) => {
  namespaces.forEach((namespace) => {
    console.log(namespace.getPrefix() + ' -> ' + namespace.getNamespace());
   });
})
  • Retrieving specific namespace declaration
return repository.getNamespace('rdf').then((namespace) => {
  console.log(namespace);
})
  • Setting the namespace declaration. This can act as create or update:
return repository.saveNamespace('rdf', 'http://www.w3.org/1999/02/22-rdf-syntax-ns#');
  • Deleting specific namespace declaration
return repository.deleteNamespace('rdf');
  • Deleting all namespaces declarations
return repository.deleteNamespaces();

Repository management

Repository operations like create, edit, delete, shutdown are not supported by the library at the moment. Supporting these is planned for next versions. Follow the issue.

Authentication

GDB token

If the library is going to be used against a secured server, then all API calls must be authenticated by sending an http authorization header with a token which is obtained after a call to rest/login/user_name with a password provided as a specific header.

In case the server requires that requests should be authenticated, then in the ServerClientConfig and RepositoryClientConfig must be configured the username and password which to be used for the authentication. If those are provided, then the client assumes that authentication is mandatory and the login with the provided credentials is performed authomatically before the first API call. After a successful login, user details which are received and the JWT auth token are stored in the AuthenticationService. From that moment on, with every API call is sent also an authorization header with the GDB token as value.

ServerClient
 const headers = {'Accept': 'text/plain'};
 const config = new ServerClientConfig('/endpoint')
    .setTimeout(5000)
    .setHeaders(headers)
    .useGdbTokenAuthentication('user', 'root');
 const client = new ServerClient(config);
RepositoryClient
const endpoint = 'http://host/';
const endpoints = ['http://host/repositories/repo1'];
const headers = {};
const contentType = '';
const readTimeout = 1000;
const writeTimeout = 1000;

const config = new RepositoryClientConfig(endpoint)
  .setEndpoints(endpoints)
  .setHeaders(headers)
  .setDefaultRDFMimeType(contentType)
  .setReadTimeout(readTimeout)
  .setWriteTimeout(writeTimeout)
  .useGdbTokenAuthentication('testuser', 'pass123');
const repository = new RDFRepositoryClient(config);
const httpRequest = repository.httpClients[0].request;

If the GDB token expires, then the first API call will be rejected with an http error with status 401. The client handles this automatically by re-login the user with the same credentials, updates the stored token and retries the API call. This behavior is the default and can be changed if the ServerClientConfig or RepositoryClientConfig are configured with keepAlive=false.

Note:
GDB token is serialized as “Authorization: GDB” header in every request, so it is vulnerable to a man-in-the-middle attack. Everyone who intercepts the GDB token can reuse the session. To prevent this, we recommend to always enable encryption in transit.

Basic Authentication

Instead of using GDB token, users can access secured GraphDB by passing valid base-64 encoded username:password combinations as a header. In case Basic authentication will be used, then the headers in the ServerClientConfig and RepositoryClientConfig must be configured to send the username and password which to be used for the authentication. From this moment on, with every API call is sent also an authorization header with the encoded credentials as value.

config.useBasicAuthentication('admin', 'root');

Note:
Basic Authentication is even more vulnerable to man-in-the-middle attacks than GDB token! Anyone who intercepts your requests will be able to reuse your credentials indefinitely until you change them. Since the credentials are merely base-64 encoded, they will also get your username and password. This is why it is very important to always use encryption in transit.

Disable authentication

If necessary, authentication can be disabled in the configuration.

config.disableAuthentication();

Response Parsers

Read responses of different content types might be parsed to data objects with parsers registered in the repository instance.

The library provides a way parsers to be implemented and registered with given repository instance which in turn will use them to parse the response before returning it to the client.

Implementing a custom parser

A parser could be implemented by extending the ContentParser and implementing the parse and getSupportedType methods.

class RdfAsJsonParser extends ContentParser {
  getSupportedType() {
    return 'application/rdf+json';
  }

  parse(content) {
    // parse and return the content
    return parsedContent;
  }
}

The getSupportedType method must return one of the supported RDF and SPARQL MIME types this way defining that the parser is responsible for converting from that type.

Registering parser in the repository

Parsers should be registered in the repository before executing any request.

// Import any of the predefined parsers
const {NTriplesParser} = require('graphdb').parser;
// And register it in the repository
repository.registerParser(new NTriplesParser());

Multiple parsers could be registered for different response types.

Registering a second parser for same content type results in overriding the previously registerted parser!

Predefined parsers

The library provides convenient parser wrappers for the rdf formats using third party libraries:

  • text/turtle: TurtleParser (N3)
  • text/rdf+n3: N3parser (N3)
  • text/x-nquads: NQuadsParser (N3)
  • text/plain (N-Triples): NTriplesParser (N3)
  • application/x-trig: TrigParser (N3)
  • application/ld-json: JsonLDParser (jsonld-streaming-parser)
  • application/rdf+xml: RDFXmlParser (rdfxml-streaming-parser)

For SELECT query results in json and xml formats as well as boolean results from ASK queries following parsers are wrapped and exposed:

  • application/sparql-results+xml, text/boolean: SparqlXmlResultParser (sparqlxml-parse)
  • application/sparql-results+json, text/boolean: SparqlJsonResultParser (sparqljson-parse)

rdf*/sparql*

The library provides basic support of extend RDF with a notion of nested triples, also known as reification. Parsers for RDFStar triples are planned for next versions.

When used against server with RDFStar support, for SELECT queries the following Mime-Types are used:

  • application/x-sparqlstar-results+json
  • application/x-sparqlstar-results+tsv

For DESCRIBE and CONSTRUCT queries, the following Mime-Types can be used:

  • application/x-turtlestar
  • application/x-trigstar
const payload = new GetQueryPayload()
    .setQuery('describe <<<http://www.wikidata.org/entity/Q472> <http://www.wikidata.org/prop/direct/P1889> <http://www.wikidata.org/entity/Q202904>>>')
    .setQueryType(QueryType.DESCRIBE)
    .setResponseType(RDFMimeType.TRIG_STAR)
    .setLimit(100);

return repository.query(payload).then((stream) => {
    stream.on('data', (data) => {
      	// data contains requested statements in trig star format
    });
});

When RDFStart triple is requested with non supportive Mime-Types, it resolves to an encoded Base64url string. It can be decoded using TermConverter util class.

const payload = new GetQueryPayload()
    .setQuery('describe <<<http://www.wikidata.org/entity/Q472> <http://www.wikidata.org/prop/direct/P1889> <http://www.wikidata.org/entity/Q202904>>>')
    .setQueryType(QueryType.DESCRIBE)
    .setResponseType(RDFMimeType.RDF_XML)
    .setLimit(100);

repository.registerParser(new RDFXmlParser());
return repository.query(payload).then((stream) => {
    stream.on('data', (data) => {
       console.log(data.subject.value);
      	// urn:rdf4j:triple:PDw8aHR0cDovL3d3dy53aWtpZGF0YS5vcmcvZW50aXR5L1E0NzI-IDxodHRwOi8vd3d3Lndpa2lkYXRhLm9yZy9wcm9wL2RpcmVjdC9QMTg4OT4gPGh0dHA6Ly93d3cud2lraWRhdGEub3JnL2VudGl0eS9RMjAyOTA0Pj4-

       console.log(TermConverter.fromBase64RdfStarTriple(data.subject.value));
       // <<<http://www.wikidata.org/entity/Q472> <http://www.wikidata.org/prop/direct/P1889> <http://www.wikidata.org/entity/Q202904>>>
    });
});

GraphDB and graphdb.js Version Compatibility

graphdb.js GraphDB Version
1+ 8+
1.1+ 9+
2+ 9.6+
3+ 10+

License

LICENSE

Package Sidebar

Install

npm i graphdb

Weekly Downloads

1,098

Version

3.1.0-newjenkins30

License

Apache-2.0

Unpacked Size

550 kB

Total Files

143

Last publish

Collaborators

  • ontotext-user
  • teodossi.dossev