Architecture
232 entities found
Ollama
DataLens uses Ollama (ollama/qwen3-coder-next) on elin for local LLM inference, including embeddings and query synthesis. Integrated with TextToSQLService and vector search. Requires Ollama to run on elin (176.9.90.154) with port 11434 accessible. Deployed on GPU for high-performance AI tasks. Critical for Text-to-SQL and document embedding pipelines, and part of the Arctic-Text2SQL setup. IronClaw Gateway uses Ollama with the Qwen3-coder-next model as fallback for local inference when Anthropic Claude not available.
Ollama (GPU inference on elin)
Runs local LLM models like qwen3-coder-next on elin's GPU for inference, integral for autonomous cataloging and text-to-SQL. Ollama and Arctic LLM inference engines are used by the Data Discovery system for intelligent table discovery and query processing. Qdrant Service uses Ollama Embedding Service to create vector embeddings.
Ollama Arctic Model
DataLens integrates with the Ollama Arctic Model for Text-to-SQL queries in production.
Ollama GPU qwen3-coder-next 80B model
The Ollama GPU qwen3-coder-next 80B model is used by the async embedding queue to generate GPU embeddings for text chunks.
ollama/qwen3-coder-next
The version, license, and approval status of ollama/qwen3-coder-next are not specified; no risk or scan status available. WrenAI integrates with multiple LLM providers including OpenAI, Anthropic, Ollama, and others to support text-to-SQL generation. DataLens uses the Ollama model at localhost:11434 on elin for model inference and forbids downloading models locally.
OpenAI
OpenAI is referenced as a third-party component involved in various AI services, although specific interaction details within DataLens are not provided.
OpenAPI Auto-Generation
DataLens Development implements OpenAPI Auto-Generation to produce TypeScript and Python API clients for frontend and data scientist usability.
OpenClaw config
OpenClaw gateway service
OpenClaw ringfenced integration
The decision is to implement Option A which involves reverting to the last known good commit and deploying it to resolve current issues with DataLens OpenClaw Integration. The stable state of DataLens OpenClaw Integration depends on the last known good commit before timeout/skill loading changes caused regressions. DataLens OpenClaw Integration is preparing deployment by reverting code and pushing to master branch, with deployment expected to follow automatically with Coolify. Claude successfully responded in Danish with a streaming response about the budget database during a previous successful session of DataLens OpenClaw Integration on March 23, 2026, 20:36.
OpenClaw Skill API
IronClaw Gateway depends on OpenClaw Skill API on the agent server to run ringfenced executor skills safely. theo Backend uses the OpenClaw Skill API via HTTP on elin to execute ringfenced database queries safely for the agent. OpenClaw Skill API queries DuckDB which contains 473 extracted budget tables for analytical data. OpenClaw Skill API references PostgreSQL database for metadata of the 132 budget files. DataLens Skill development involves creating an OpenClaw skill following best practices.
openpyxl
Pandas uses openpyxl for Excel file exporting and processing.
pandas
The Extraction pipeline in the DataLens Platform uses pandas for data manipulation and loading extracted data. Pandas uses openpyxl for Excel file exporting and processing. Lux-api depends on pandas for auto-visualization features.
PandasAI
DataLens supports embedded analytics with DuckDB and SQL which covers capabilities similar to PandasAI. PandasAI and DataLens both provide capabilities to query and analyze data, but DataLens uses DuckDB and SQL for more robust analysis.
passlib
passlib is used by the Auth system in the DataLens Platform for hashing passwords securely. Passlib depends on bcrypt for password hashing.
pdfplumber
Pdfplumber uses python-magic for MIME type detection during PDF table extraction.
PgDataService
PgDataService manages PostgreSQL data storage for DataLens, replacing DuckDBService. It handles schema creation, table registry, data insertion, and querying, supporting project-specific schemas, and facilitates text chunk and table metadata storage, with type conversions guiding data loading and query execution in PostgreSQL. PgDataService is an alternative service managing PostgreSQL connections, complementing DuckDBService for DuckDB database management. DataLensPostgresRunner wraps PgDataService to enforce project-scoped schema isolation for SQL execution.
PHASE2_UNIFIED_STRATEGY.md
PHASE2_UNIFIED_STRATEGY.md defines the pipeline design and tool justifications for processing Phase 2 file types Opus 4.6 created the PHASE2_UNIFIED_STRATEGY.md which contains pipeline design, tool justifications, and question-to-data mapping. The PHASE2_UNIFIED_STRATEGY.md and PHASE2_IMPLEMENTATION_PLAN.md documents contain complementary research outputs informing Phase 2 implementation decisions.
Planner
The DataLens DS-STAR Implementation Plan includes the Planner component for creating extraction plans. Planner produces multi-step extraction plans for data processing.
PlannerAgent
DS-STAR Intelligence includes the PlannerAgent component. The plan contains the Planner Agent that creates multi-step extraction plans based on the data catalog. PlannerAgent is a part of the DS-STAR pipeline. DS-STAR Orchestrator has PlannerAgent as a step in its workflow DSStarOrchestrator uses the planner agent in its extraction strategy creation DS-STAR Orchestrator has PlannerAgent as a step in its workflow DSStarOrchestrator uses the planner agent in its extraction strategy creation DS-STAR Intelligence capability encompasses the PlannerAgent component for document analysis and strategy creation. The DS-STAR pipeline includes the PlannerAgent component. The Planner Agent is part of the DataLens DS-STAR Implementation Plan to generate multi-step extraction plans. The DS-STAR Intelligence Layer includes the PlannerAgent component. DS-STAR Intelligence includes the PlannerAgent that analyzes files and creates extraction strategies using Ollama. Planner Agent uses the data catalog generated by FileAnalyzer as input.
Playwright
Playwright is used for end-to-end testing to validate User interactions and the platform's functionality. The npm dev dependency @playwright/test uses the bits-ui component as part of the frontend dependency set in the project.
Playwright testing framework
Playwright testing framework is used to end-to-end test and validate the functionalities of the Datalens SVGV Budget Platform. The E2E Test Suite comprises Playwright tests to validate the Data Discovery workflow.
Plotly
Plotly is used for generating interactive data visualizations in DataLens, supporting various chart types and rendering in findings and reports.
Port 11434
Port 6333
postcss
npm dependency: postcss@^8.5.6, used in frontend, version 8.5.6, license type unknown, no approval or risk assessment noted.
postgres_data volume
PostgreSQL Data Volume
PPTX extractor
The batch upload pipeline depends on new extractors including the PPTX extractor. The PPTX extractor uses the python-pptx third-party component. The PPTX extractor uses python-pptx to extract slide-based chunks during DataLens Phase 2. The PPTX extractor implements slide-based chunking, with potential sub-slide splits for dense content. The PPTX extractor uses python-pptx to extract slide-based chunks during DataLens Phase 2. The PPTX extractor implements slide-based chunking, with potential sub-slide splits for dense content. The PPTX extractor operates via background workers to perform extraction tasks. The PPTX extractor capability is validated by the test_extractors test case.