-
Notifications
You must be signed in to change notification settings - Fork 1
Fabric Workshop ‐ Toronto
Welcome to the Fabric Workshop wiki page. This page contains all of the links you will need for today's content, plus any changes/workarounds that may be needed. We hope you enjoy the day!
- This page: https://aka.ms/fabric-toronto
- Eval Form for End of Day: Fabric Event Eval
Portals:
- Azure Portal: https://portal.azure.com
- Fabric Portal: https://app.fabric.microsoft.com
Content Origin (for reference, outside of lab):
- Original GitHub repo with Labs: https://aka.ms/fabricrealtimelab
- Resources Zip File (notebooks): https://aka.ms/fabricrealtimelab-resources
Lab Access and Information:
- Registering for labs: https://cloudweeks.learnondemand.net
- Key: 75ACD06FC7824114
Partner Resources:
- Microsoft Fabric Partner Community
- Learn Fabric
- Fabric Bootcamp
- Fabric Partner Resources
- Industry Dream Demos
- Analytics Campaign in a Box
- Deliver a x-IAD - In a Day Workshop
- Azure Innovate
- Fabric Featured Partner
- How to become a Microsoft Featured Partner
Highlights from Build:
- Microsoft Fabric: What's new and what's next
- Reimagine Real-Time Intelligence with Microsoft Fabric
- The mechanics of Real-Time Intelligence in Microsoft Fabric
- Integrating Azure AI and Microsoft Fabric for Next-Gen AI Solutions
Note: in today's workshop, we are skipping Lab 03: Building a DW using Pipelines.
There are several changes in the labs that are covered in these notes below.
After launching the lab, retrieve your username and password from the Resources tab and keep them handy (such as in a local notepad window). The Azure promo code is used to activate an Azure trial environment. You may use the VM provided in the lab and/or log in using your own browser:
The data generator app we've built uses a container and Event Hub that is deployed in Azure. If you'd like to deploy this yourself, you can do so either using your own Azure subscription (~$1.70/day), or redeem an Azure Pass free trial using the promo code on the Resources tab. However, to help streamline the day, we've deployed the generator for the group, so you will not need Azure access if you'd like to use the group Event Hub. Connection details are included here -- each attendee is assigned a Consumer Group.
If you are using the group Event Hub, you only need to do these Tasks in the module:
- Task 1: Sign in to Power BI account and sign up for the free Microsoft Fabric trial
- Task 2: Start the Microsoft Fabric trial
- Skip Task 3: Redeem Azure Pass
- Skip Task 4: Log Analytics Contributor role assignment
- Skip Task 5: Create Storage Account
-
Task 6: Create Fabric Workspace
- When naming your workspace, you will likely need to make the name globally unique. Instead of "RealTimeWorkspace", add your initials, firstname, or change the name as you see fit. All items from the workshop today will go into this workspace.
- Skip Task 7: Deploy the app via Azure Container Instance
- Task 8: Get data with Eventstream
If you'd like to deploy the stock generator app yourself, follow these steps:
- Task 1: Sign in to Power BI account and sign up for the free Microsoft Fabric trial
- Task 2: Start the Microsoft Fabric trial
- Task 3: Redeem Azure Pass
- Skip Task 4: Log Analytics Contributor role assignment
- Skip Task 5: Create Storage Account
-
Task 6: Create Fabric Workspace
- When naming your workspace, you will likely need to make the name globally unique. Instead of "RealTimeWorkspace", add your initials, firstname, or change the name as you see fit. All items from the workshop today will go into this workspace.
- Task 7: Deploy the app via Azure Container Instance
- Task 8: Get data with Eventstream
In Task 6, give the Workspace a unique name, such as "Workshop-initials" or similar:
In Task 8, be sure to do the following:
- Use the following Event Hub connection details:
Setting | Value |
---|---|
Event Hub namespace | ehns-toronto-fabricworkshop |
Event Hub | stockeventhub |
SAS Name | stockeventhub_sas |
SAS Key | 9TR249DduSgW0hgHucZEwNFVCQ+GOhK1H+AEhHxJpUE= |
Consumer Group | Individually assigned, see handout |
- Delete the default Connection Name and provide a unique name in the textbox, such as EventHub-Connection-{initials} or similar:
As part of the new Real-Time Intelligence features, all KQL databases are part of an EventHouse. Some screenshots may not have been updated.
No additional notes at this time.
Please skip this lab for today's workshop. (Of course, you are free to work on this if time permits or outside of the lab!)
When working with the notebook Lakehouse 2: Build Aggregation Tables, many steps focus on using Data Activator to help shape and aggregate the data. Data Wrangler is a visual tool that outputs Python code into a new notebook cell. If you'd like to save time, each of the 3 Data Wrangler steps has been completed and is commented out, and should look like the below image:
If you'd prefer to avoid going through the Data Wrangler steps or run into any challenges, you can use the code in the commented cell. Highlight all of the code in the cell (CTRL-A) and then uncomment all of the code (CTRL-/). This can be done for any or all of the Data Wrangler steps.
There are 3 notebooks used in this lab:
Notebook | Purpose |
---|---|
DS 1 - Build Model | Builds an ML model using Prophet, examines data, and stores model in MLflow |
DS 2 - Predict Stock Prices | Consume MLflow from a notebook |
DS 3 - Forecast All | Build predictions for all stock symbols and store in a Delta table |
DS 1 and DS 2 are largely examples/theory of how to work with data, evaluate the results, and store the model in MLflow. DS 3 is the main notebook that generates predictions for all stocks and stores the results in the stocks_prediction table.
To save time: if you're primarily interested in data analysis and setting up experiments and comparing runs in MLflow, focus on DS 1 and DS 2. If you're more interested in getting the predictions and building a semantic model and report, consider jumping right to DS 3 (which is Exercise 3: Solution in practice in the lab).