Documentation

Documentation

  • Home
  • Blog
  • API
  • Contact

›Overview

Overview

  • Lentiq introduction
  • Lentiq architecture
  • What is a data pool?
  • What is a project?
  • Migrating from Hadoop

Getting started

  • Deploying applications and processing clusters
  • Connecting to Spark from a notebook
  • Uploading data to Lentiq
  • Creating a data pool
  • Deploying on GCP
  • Deploying on AWS

User Guide

    Managing applications

    • Working with applications
    • Managing compute resources

    Managing data

    • Working with data and metadata
    • Sharing data between data pools
    • Querying data with SQL (DataGrip)
    • Connecting Tableau to Lentiq

    Managing models

    • Working with models
    • Publishing notebooks
    • Training and serializing a model
    • Managing model servers

    Managing workflows

    • Working with workflows
    • Creating a reusable code block from a notebook
    • Creating a docker image based reusable code block
  • Glossary
  • API

Tutorials

  • End-to-end Machine Learning Tutorial

What is a data pool?

A data pool is a fundamental building block of Lentiq. It is to a degree a "mini data lake" designed to be used by a small team of people to share files, tables, notebooks, models etc. The team can also choose it's own stack of applications to use such as Spark or Ray.

lentiq_high_level_architecture

It uses docker containers to run notebooks, additional applications, reusable code blocks that run as part of a workflow etc. The compute layer (the Kubernetes cluster behind the data pool) is separated from the storage layer so both can scale independently as needed.

← Lentiq architectureWhat is a project? →
Copyright © 2019 Lentiq