top of page

Glossary

Platforms & Applications

 

Adaptive Insights – A cloud-based planning and budgeting platform used for financial forecasting, workforce planning, and modeling.

 

ADP – A global leading payroll and human capital management solution.

 

CentralReach – A specialized EHR platform used in ABA practices.

 

KIPU – A specialized EHR platform used in addiction treatment and behavioral healthcare.

 

Lightning Step – A specialized EHR platform used in addiction treatment and behavioral

healthcare.

​

Microsoft Dynamics – An enterprise suite including CRM and ERP for operations, billing, and workflows.

​

SAGE Intacct – A cloud-based financial management solution used for accounting, budgeting, and reporting.

​

Salesforce – A leading CRM platform used for managing sales, marketing, and customer data.

 

Databricks – A unified cloud data platform built on Apache Spark that supports engineering, machine learning, and analytics.​

AI & Analytics

 

AI/BI Dashboards – Visual tools combining artificial intelligence and business intelligence for insight delivery and metric monitoring.

​

AI/BI Genie – A Databricks-native tool for generating dashboards, KPIs, and automated reporting using natural language.

​

AI Agents – Machine learning-powered systems that support validation, anomaly detection, and user interaction tasks.

​

AI-Ready Data – Clean, structured, and governed data prepared for use in AI and machine learning applications.

​

Hallucinations (AI) – Incorrect outputs from generative AI models that are presented as facts.

​

Metrics Views – redefined, reusable logical views in Databricks that standardize business metrics and KPIs. Metrics views ensure consistent calculations across reports and dashboards, simplify querying, and support semantic layers for BI tools.

​​

​Architecture & Engineering

​

Architecture (Data Architecture) – The structural design of data systems covering ingestion, storage, transformation, and analytics.

​

Modern Architecture – A cloud-native, modular approach using layers like raw→silver→gold, with real-time governance and automation.

​

Legacy Architecture – Traditional, static systems characterized by manual processes, poor scalability, and siloed data tools.

​

Config (Configuration Layer) – Metadata-driven logic that governs pipeline behavior and transformation processes.

​​Data Infrastructure

​

Data Lake – A centralized repository for raw, semi-structured, and unstructured data supporting flexible analytics.

 

Data Warehouse – A structured environment for storing transformed data optimized for querying and reporting.

​

Catalog (Unity Catalog) – A Databricks feature for managing data assets, schemas, access, and lineage.

 

Raw → Silver → Gold – A layered architecture for progressing data from raw ingestion to clean, analytics-ready models:

  • Raw: Unprocessed, ingested data

  • Silver: Cleaned and joined data

  • Gold: Final, curated datasets for reporting

 

DLT (Delta Live Tables) – A Databricks tool to build reliable ETL pipelines with quality checks and real-time observability.

 

ETL (Extract, Transform, Load) – A data process involving extraction from sources, transformation, and loading into storage.

​

ELT (Extract, Load, Transform) – A modern data integration process where raw data is first extracted from source systems, then loaded into a centralized repository (like a data lake or warehouse), and finally transformed within the destination system. ELT is optimized for cloud environments and supports scalable, in-place transformations.

 

Pipelines – Automated workflows that manage data flow and transformation between systems or layers.

 

Compute – The processing power used to perform data operations, typically scaled dynamically in cloud environments.

​Governance & Quality

​

Governance – The standards and policies that ensure data quality, security, and traceability across users and systems.

 

Unity Catalog – Databricks’ native governance framework for managing access, schema consistency, and lineage tracking.

 

Mapping Tables – Reference tables used to align and standardize data values across multiple systems.

 

Validation (Data Validation) – Techniques and checks that ensure data accuracy, completeness, and conformance to business rules.

bottom of page