Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

dltoimport #422

Merged
merged 4 commits into from
Jan 30, 2025
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 4 additions & 0 deletions src/sempy_labs/migration/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,6 +18,9 @@
from sempy_labs.migration._refresh_calc_tables import (
refresh_calc_tables,
)
from sempy_labs.migration._direct_lake_to_import import (
migrate_direct_lake_to_import,
)

__all__ = [
"create_pqt_file",
Expand All @@ -28,4 +31,5 @@
"migrate_tables_columns_to_semantic_model",
"migration_validation",
"refresh_calc_tables",
"migrate_direct_lake_to_import",
]
66 changes: 66 additions & 0 deletions src/sempy_labs/migration/_direct_lake_to_import.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,66 @@
import sempy
from uuid import UUID
import sempy_labs._icons as icons


def migrate_direct_lake_to_import(dataset: str | UUID, workspace: str | UUID):
"""
Migrates a semantic model from Direct Lake mode to import mode. After running this function, you must go to the semantic model settings and update the cloud connection. Not doing so will result in an inablity to refresh/use the semantic model.

Parameters
----------
dataset : str | uuid.UUID
Name or ID of the semantic model.
workspace : str | uuid.UUID, default=None
The Fabric workspace name or ID.
Defaults to None which resolves to the workspace of the attached lakehouse
or if no lakehouse attached, resolves to the workspace of the notebook.
"""

sempy.fabric._client._utils._init_analysis_services()
import Microsoft.AnalysisServices.Tabular as TOM
from sempy_labs.tom import connect_semantic_model

with connect_semantic_model(
dataset=dataset, workspace=workspace, readonly=False
) as tom:

if not tom.is_direct_lake():
print(
f"{icons.warning} The '{dataset}' semantic model within the '{workspace}' workspace is not in Direct Lake mode."
)
return

for t in tom.model.Tables:
table_name = t.Name
if t.Partitions.Count == 1 and all(
p.Mode == TOM.ModeType.DirectLake for p in t.Partitions
):
p = next(p for p in t.Partitions)
partition_name = p.Name
entity_name = p.Source.EntityName
schema_name = p.Source.SchemaName or "dbo"
# Rename Direct Lake partition
t.Partitions[partition_name].Name = f"{partition_name}_remove"

# Generate M expression for import partition
expression = f"""let\n\tSource = DatabaseQuery,\n\tData = Source{{[Schema="{schema_name}",Item="{entity_name}"]}}[Data]\nin\n\tData"""

# Generate M partition
tom.add_m_partition(
table_name=table_name,
partition_name=partition_name,
expression=expression,
mode="Import",
)
# Remove Direct Lake partition
tom.remove_object(object=p)

tom.model.Model.DefaultMode = TOM.ModeType.Import

# Check
# for t in tom.model.Tables:
# if t.Partitions.Count == 1 and all(p.Mode == TOM.ModeType.Import for p in t.Partitions) and t.CalculationGroup is None:
# p = next(p for p in t.Partitions)
# print(p.Name)
# print(p.Source.Expression)