Note
Access to this page requires authorization. You can try signing in or changing directories.
Access to this page requires authorization. You can try changing directories.
Power Automate Process Mining gives you the option to store and read event log data directly from Fabric OneLake through Fabric Lakehouse. This feature simplifies extract, transform, load (ETL) management by connecting directly to your OneLake.
This feature currently supports the ingestion of the following:
Tables
- Delta tables
- Single Delta table in Fabric Lakehouse.
Files and folders
- CSV
- Single CSV file.
- Folder with multiple CSV files that have the same structure. All files are ingested.
- Parquet
- Single parquet file.
- Folder with multiple parquet files that have the same structure. All files are ingested.
- Delta-parquet
- Folder that contains a delta-parquet structure.
Prerequisites
Ensure that Users can access data stored in OneLake with apps external to the Fabric option is switched on in the OneLake tenant settings.
A Fabric workspace that's different from the default My workspace.
The Admin role must be attributed to the user performing the initial workspace setup for the environment for the other users in the same environment.
The Fabric workspace needs to be shared to the Process Insight Prod service principal with an Admin role. To register the Process Insights Prod service principal, follow the steps in Install Azure tools.
A Fabric Lakehouse must be created in this workspace with the data in supported formats.
Important
The following item isn't currently supported: Fabric Lakehouses with Schema support enabled.
CSV data in your Lakehouse should meet the following CSV file format requirements:
- Compression type: None
- Column delimiter: Comma (,)
- Row delimiter: Default and encoding. For example, Default (\r,\n, or \r\n)
All data must be in final event log format and meet the requirements listed in Data requirements. Data should be ready to be mapped to the process mining schema. No data transformation is available post ingestion.
The size (width) of the header row is currently limited to 1 MB.
Important
Ensure that the time stamp represented in your CSV file follows the ISO 8601 standard format (for example, YYYY-MM-DD HH:MM:SS.sss or YYYY-MM-DDTHH:MM:SS.sss).
Connect to Fabric OneLake
On the navigation pane to the left, select Process mining > Start here.
In the Process name field, enter a name for your process.
Under the Data source heading, select Import data > OneLake > Continue.
Select an optional Power BI workspace or select Skip.
On the Connection setup screen, select your Fabric Workspace from the dropdown menu. This populates the Lakehouse dropdown menu. From the menu, select the lakehouse that contains your data and select Next.
Browse the Lakehouse folder structure and select the Delta table, file, or folder containing the event log data.
You can either select a single Delta table, a single file, or a folder with multiple files. All files must have the same headers and format.
Select Next.
On the Map your data screen, map your data to the required schema.
Complete the connection by selecting Save and Analyze.
Define incremental data refresh settings
You can refresh a process ingested from Fabric OneLake on a schedule, either through a full or incremental refresh. Though there are no retention policies, you can ingest data incrementally using one of the following methods:
If you selected a single Delta table in the previous section, append more data to the selected table.
If you selected a single file in the previous section, append more data to the selected file.
If you selected a folder in the previous section, add incremental files to the selected folder.
Important
When you add incremental files to a selected folder or subfolder, make sure you indicate the increment order by naming files with dates such as YYYMMDD.csv or YYYYMMDDHHMMSS.csv.
To refresh a process:
Go to the Details page of the process.
Select Refresh Settings.
On the Schedule refresh screen, complete the following steps:
- Turn on the Keep data up to date toggle switch.
- In the Refresh data every dropdown lists, select the frequency of the refresh.
- In the Start at fields, select the date and time of the refresh.
- Turn on the Incremental refresh toggle switch.