en-us/about_AutomatedLabBasics.help.txt
TOPIC
about_automatedlabbasics SHORT DESCRIPTION Generic help about the basics of AutomatedLab LONG DESCRIPTION Getting started PLACING ISO FILES Of course AutomatedLab (AL) cannot install an operating system without actually having the bits. Hence you need to download an ISO file from MSDN, TechNet Evaluation Center or somewhere else. These files need to go to your folder "ISOs" located in the "LabSoures" folder. The "ISOs" folder contains only one file after the installation of AL: "_Put all ISO images in here.txt". I have downloaded Windows Server 2016 from the TechNet Evaluation Center and put the file like shown below. ! ISOs TESTING THE ISO FILES To make sure that AL can read the file, try to get a list of available operating systems. Open an elevated PowerShell ISE and call the following command (make sure you point to the right location for the LabSources folder: powershell Get-LabAvailableOperatingSystem -Path E:\LabSources This returns a list of all operating system images found on the ISO file (of course this works also if there are a bunch of different OS ISOS in the folder). ! OSList INSTALL THE FIRST LAB Please open an elevated PowerShell ISE and create a new empty document (CTRL+N) if not already open. Copy and paste the following lines into the ISE: * powershell New-LabDefinition -Name GettingStarted -DefaultVirtualizationEngine HyperV Add-LabMachineDefinition -Name FirstServer -OperatingSystem 'Windows Server 2016 SERVERSTANDARD' Install-Lab Show-LabDeploymentSummary * The just press the run button or hit F5 to start the deployment. This is what is going to happen. Many things happen automatically but can be customized: * AutomatedLab starts a new lab named "GettingStarted". The lab defininition will be stored in C:\ProgramData\AutomatedLab\Labs\GettingStarted. The location can be customized "Advanced AutomatedLab Configurations" with the setting LabAppDataRoot. * AL will update download the SysInternals tools and put them into the LabSources folder. * AL looks for an ISO file that contains the specified OS. If the ISO file cannot be found, the deployment stops. * AL adds the one and only machine to the lab and recognizes that no network was defined. In this case, AL creates a virtual switch automatically and uses an free IP range. * The AL measures the disk speed and chooses the fastet drive for the lab, as no location is defined in the call to "New-LabDefinition". In my case, it chooses D. Measuring is done only once and the result is cached. * Then the actual deployment starts. AL creates 1. The virtual switch 2. Then it creates the a base image for the operating system that is shared among all machines with the same OS. 3. Afterwards the VM is created and started 4. AL waits for the machine to become ready and shows the overall installation time. REMOVING A LAB If you want to get rid of the lab, just call Remove-Lab. The cmdlet removes the VMs including the disks and the virtual switches, and leaves the base disks for the next deployment. If you have closed the ISE in the meantime, either specify the lab name or import it first. ! Remove1 SUMMARY With AutomatedLab it is extremely easy to create various kinds of labs. The more you define your lab by code, the easier it is to re-deploy it and the less time you invest in the long term. If you like what you have seen, take a look at the folder "LabSources\Sample Scripts\Introduction" . These scripts demo how to create domains, internet facing labs, PKI, etc. Please provide feedback if something does not work as expected. If you are missing a feature or have some great ideas, please open an issue . NEXT STEPS Now that you have deployed your first lab, what comes next? Would you like to connect to the machines and run remote commands without you knowing the password? Then start with the docs on lab management . Wondering how to transfer data to your new lab? Then start with the docs on data exchange . If you - like us - like to tinker around with things, check out the possible settings . Generally speaking AutomatedLab takes care of everything for you when deploying your labs on Azure. Since additional authentication is required it is possible that you need to login to your Azure account before using AutomatedLab. AutomatedLab works with Azure Resource Manager, so you can either execute the cmdlet `Connect-AzAccount` before deploying your lab or save your Azure Resource Manager profile. If you choose to login to your Azure account before a lab deployment your profile is being saved for you to be able to import the lab at a later stage. Since it is possible that your profile expires you might see an error message indicating your profile expiration. In that case, simply login to your Azure account again. New-LabDefinition -Name 'MyLab' -DefaultVirtualizationEngine Azure ## Optional to set e.g. your preferred location Add-LabAzureSubscription -DefaultLocation 'West Europe' This will enable AutomatedLab to create a lab sources resource group for you as well as a separate resource group for each lab you deploy. Your lab resource group will contain the entire lab deployment and will be removed when you call `Remove-Lab`. Adding machines to a lab ranges from extremely easy to fairly complicated. Both is achieved with one cmdlet though, Add-LabMachineDefinition. ## Add the root domain controller for the domain contoso.com ## Network, memory, ... will all be handled automatically Add-LabMachineDefinition -Name DC1 -Roles RootDC -OperatingSystem 'Windows Server 2019 Datacenter' -Domain contoso.com When you are using the automatic machine definitions, the memory requirements will be calculated based on the machines' roles. DEPLOYING A LAB WITH LINUX VMS With AutomatedLab Linux VMs can be deployed just as easily as Windows VMs. The current implementation should take care of the following distributions: - RHEL 7 (*) - CentOS 7 - Fedora 27 - SLES 12.3 (*) - OpenSuSE At the moment the machines do not support any of AutomatedLab's roles since our roles are Windows-based. However, your VMs should come up domain-joined and capable of receiving WinRM requests. AutomatedLab uses kickstart (RHEL-based) or AutoYAST (SLES-based) to configure everything that would be configured in the unattend file of a Windows machine. WinRM support is spotty with distributions like OpenSuSE and Fedora. Fedora being too cutting-edge does not allow PowerShell and omi-psrp-server to be installed, while OpenSuSE currently has the wrong package dependencies and fails installing PowerShell as well. SIMPLE LINUX LAB You can find the Linux lab here: AL Loves Linux As you can see, integrating Linux clients is very simple in general: Add-LabMachineDefinition -Name LINCN2 -OperatingSystem 'CentOS 7.4' If you want to add additional package groups to be installed during setup, you can do so by specifying them: Add-LabMachineDefinition -RhelPackage domain-client -Name LINCN1 -OperatingSystem 'CentOS 7.4' -DomainName contoso.com You can find all available packages with -LabAvailableOperatingSystem | Select-Object -Expand LinuxPackageGroup . However, the basics should be fine for most cases. At them moment, your Linux-based labs need an internet connection (i.e. a routing VM) so that the PowerShell and omi-psrp-server can be downloaded during setup. Without omid running on the Linux machines, your lab will run into a timeout during installation. While this will not break things, it will certainly cause a long wait. AZURE-SPECIFIC PROPERTIES There are several properties that can be used with the `AzureProperties` parameter of the `Add-LabMachineDefinition` cmdlet. - ResourceGroupName - Resource group this machine is deployed into, if it should be different - UseAllRoleSizes - Use a random role size of the available role sizes - RoleSize - Use specific role size like Standard_D2_v2 - LoadBalancerRdpPort - Use a different port for the inbound NAT rule. Needs to be unique in your lab! - LoadBalancerWinRmHttpPort - Use a different port for the inbound NAT rule. Needs to be unique in your lab! - LoadBalancerWinRmHttpsPort - Use a different port for the inbound NAT rule. Needs to be unique in your lab! - LoadBalancerAllowedIp - A comma-separated string (NOT an array) containing IP addresses allowed to connect, e.g. "$(Get-PublicIpAddress), 1.2.3.4" - SubnetName - The subnet name this machine is deployed into - UseByolImage - Boolean as string indicating that BYOL licensing is used - AutoshutdownTime - The timespan as string when the machines shut be shut down - AutoshutdownTimezoneId - The time zone ID as string for the auto shutdown time - StorageSku - The storage SKU for additional disks. OS disks are managed disks. Either 'Standard_LRS', 'Premium_LRS' or 'StandardSSD_LRS' Creating your lab starts with New-LabDefinition. This cmdlet creates a container that holds all the lab items like networks and machines. It is mandatory to define the name of the lab and the virtualization engine. So far AL supports Hyper-V and Azure. powershell New-LabDefinition -Name $labName -DefaultVirtualizationEngine HyperV -Path D:\AutomatedLabs -VmPath D:\AutomatedLab-VMs - New-LabDefinition Parameters - Name : The name if the lab to create. The name must be unique. - DefaultVirtualizationEngine : HyperV for local deployments or Azure to deploy into the cloud. - Path : This is the path were the lab files will go (screenshot 1). If not defined the labs will be stored in "C:\Users<username>\Documents\AutomatedLab-Labs". - VmPath : This is where AL creates the virtual machines. If this path is not defined AL will choose the fastest drive by trying not to use the system drive and create a folder there names "AutomatedLab-VMs". During the lab deployment, AL exports your definitions into XML files. By default these are stored in 'C:\Users<username>\Documents\AutomatedLab-Labs'. If you close the PowerShell session after the lab deployment, you can import the lab again using the Import-Lab cmdlet. !https://cloud.githubusercontent.com/assets/11280760/20555110/6f12f9a8-b160-11e6-863a-bac119eac71b.png! The virtual machines are stored in a different folder. AL tries to determine the fastest drive and uses the cmdlet Get-DiskSpeed for this. AL tries to avoid using the system drive. The speed test runs only once and results are cached. However if you plug in a new drive, AL takes it into consideration and measure its speed. !https://cloud.githubusercontent.com/assets/11280760/20642671/252e839a-b415-11e6-8c80-307f7662d64a.JPG! Starting with version 5.23, each lab deployment will not only be validated before the deployment but also when the installation is finished. To enable the deployment validation, you need to install Pester 5.0.1+. We did not mark Pester as a dependency, as these tests are optional. LAB BUILDING WITH THE REST API With the custom role 'LabBuilder' it is now possible to create a simple REST API powered by Polaris on any virtualization host. The REST API provides the following functionality: CREATE $request = @{ LabScript = Get-Content "$labsources\Sample Scripts\Introduction\01 Single Win10 Client.ps1" -Raw } | ConvertTo-Json ## Queue new job, use GUID to request status update with GET method $guid = Invoke-RestMethod -Method Post -Uri http://NestedBuilder/Lab -Body $request -ContentType application/json READ ## All labs (Get-Lab -List) Invoke-RestMethod -Method Get -Uri http://NestedBuilder/Labs ## Specific lab Invoke-RestMethod -Method Get -Uri http://NestedBuilder/Lab?Name=Win10 ## Lab creation job (monitoring long running jobs) Invoke-RestMethod -Method Get -Uri http://NestedBuilder/Lab?Id=5b89babb-7402-4a7a-9c16-86e1d52613fa DELETE ## As query parameter Invoke-RestMethod -Method Delete -Uri http://NestedBuilder/Lab?Name=Win10 ## Or as JSON body $request = @{Name = 'Win10'} | ConvertTo-Json Invoke-RestMethod -Method Delete -Uri http://NestedBuilder/Lab -Body $request -ContentType application/json LAB SCENARIO INCLUDING THE BUILDER The following lab scenario configures a simple VM to test-drive this feature. Be aware that when using this on Azure your lab sources need to be uploaded first. Otherwise you will need to manually copy all necessary ISO files to your Azure VM. <# Build a lab with the help of nested virtualization. Adjust the machine memory if necessary. The build worker will use Polaris as a simple REST endpoint that takes your lab data to deploy. AutomatedLab will be copied to the machine. Lab sources will be mirrored to the machine as well, so that lab deployments can start immediately ##> param ( [string] $LabName = 'LabAsAService', [ValidateSet('yes','no')] $TelemetryOptOut = 'no' # Opt out of telemetry for build worker by saying yes here ) New-LabDefinition -Name $labName -DefaultVirtualizationEngine HyperV Add-LabVirtualNetworkDefinition -Name $labName -HyperVProperties @{ SwitchType = 'External'; AdapterName = 'Ethernet' } $role = Get-LabPostInstallationActivity -CustomRole LabBuilder -Properties @{TelemetryOptOut = $TelemetryOptOut} $machineParameters = @{ Name = 'NestedBuilder' PostInstallationActivity = $role OperatingSystem = 'Windows Server 2016 Datacenter (Desktop Experience)' Memory = 16GB Network = $labName DiskName = 'vmDisk' } $estimatedSize = [Math]::Round(((Get-ChildItem $labsources -File -Recurse | Measure-Object -Property Length -Sum).Sum / 1GB + 20), 0) $disk = Add-LabDiskDefinition -Name vmDisk -DiskSizeInGb $estimatedSize -PassThru Add-LabMachineDefinition @machineParameters Install-Lab SUMMARY One basic requirement is sending files to a lab machine as well as receiving files. Both is supported by AutomatedLab, for files and also directory trees. This feature works with Hyper-V VMs as well as Azure VMs. HOW TO USE IT €" SEND To send a file or directory to a lab machine, only the local source path and the machine name is required. There is no difference whether the machine is on Hyper-V or Azure. Note: Copy-LabFileItem wraps the cmdlets Send-File and Send-Directory, both define in the PSFileTransfer module. If the target path is not defined, the file or folder will be put in C:\. The parameter ComputerName is an array and data can be copied to a single or multiple machines with a single command. The following command copies a file to the specified machine trying SMB first and falling back to WinRM. PowerShell Copy-LabFileItem -Path 'D:\170630 AL' -ComputerName wServer1 -DestinationFolderPath C:\Temp The following samples are Using the PSFileTransfer cmdlets directly. These do not support SMB and send the files over the PSSession right away. The Force switch creates the target folder if not already existing. Note: It is recommended to use Copy-LabFileItem as SMB is much faster then sending a byte array over WinRM. PowerShell $s = New-LabPSSession LabVM1 Send-File -SourceFilePath D:\Untitled1.ps1 -DestinationFolderPath C:\Temp\Untitled1.ps1 -Session $s -Force Send-Directory -SourceFolderPath D:\Test -DestinationFolderPath C:\Windows -Session $s HOW TO USE IT €" RECEIVE This also works the other way around. If some process creates files on a lab machine that are required for another process on another lab machine, you can retrieve and re-send the file to another machine. For example, this process is used in the AutomatedLab’s certificate functions like Get-LabCAInstallCertificates. PowerShell Receive-File -SourceFilePath C:\Unattend.xml -DestinationFilePath D:\Unattend.xml -Session $s Receive-Directory -SourceFolderPath 'C:\Program Files\Wireshark' -DestinationFolderPath D:\Wireshark -Session $s INTERNALS Hyper-V machines are usually reachable from the host by SMB. However, it is not sure or pretty unlikely if the other way works as well, accessing the host machines from the VM. Copy-LabFileItem always tries SMB first. If this does not work, it uses the PowerShell session and transfers the file as a Byte[]. Reading and writing a file as a byte array, serializing the by byte array and sending it over the network is much slower than SMB but totally sufficient for most scenarios. If you transfer large files, make sure SMB works from the host to the VMs. Note: All the functions of the module PSFileTransfer do not rely on other modules of AutomatedLab and can be used separately. Offline lab scenario To deploy complex lab scenarios offline you will at some point run into issues due to additional dependencies being downloaded. This article aims to give you all the necessary information to prepare for a fully offline environment. BASICS First of all, install AutomatedLab from the MSI installer and get some ISO files - like you would for an online scenario. Place your ISO files in LabSources\ISOs. The MSI installer contains all static files that we need like ProductKeys.xml, WinSAT and so on. DSC PULL SERVER To deploy the pull server role, we download the Access Database Engine from `Get-LabConfigurationItem -Name AccessDatabaseEngine2016x86` to LabSources/SoftwarePackages without renaming the file. OFFICE To deploy Office Server, the Office deployment toolkit is needed at LabSources/SoftwarePackages/OfficeDeploymentTool.exe. You can download the tool from `Get-LabConfigurationItem -Name OfficeDeploymentTool`. SHAREPOINT SharePoint requires a whole slew of prerequisites. First of all, you should download all CPP redistributables from 2012 up to 2017, 32 and 64 bit, following the naming pattern vcredist_BITS_YEAR e.g. vcredist_64_2012. The links can be easily accessed using `Get-LabConfigurationItem -Name cppredist `. The more descriptive version is `Get-PSFConfig -Module AutomatedLab -Name cppredist ` which shows you the version as well. All packages need to be downloaded to LabSources/SoftwarePackages The rest of the requirements is version-dependent. The links are best accessed using `Get-LabConfigurationItem -Name SharePoint2019Prerequisites`. Please make sure that for SharePoint 2013 the download with an URI containing 1CAA41C7 needs to be renamed to WcfDataServices56.exe. SQL SQL also requires a bunch of prerequisites. First of all, you should download all CPP redistributables from 2015 up to 2017, 32 and 64 bit, following the naming pattern vcredist_xARCHITECTURE_YEAR e.g. vcredist_x86_2015.exe. The links can be easily accessed using `Get-LabConfigurationItem -Name cppredist `. The more descriptive version is `Get-PSFConfig -Module AutomatedLab -Name cppredist ` which shows you the version as well. All packages need to be downloaded to LabSources/SoftwarePackages. Additionally, .NET 4.8 will be required on SQL 2017 and newer when reporting services are deployed, which can be downloaded from `Get-LabConfigurationItem -Name dotnet48DownloadLink`. The download should be stored in LabSources/SoftwarePackages. While not strictly required, you might want to download the following packages as well, thereby reducing some error messages: - SQL Server Reporting Services: `Get-LabConfigurationItem -Name Sql$($server.SqlVersion)SSRS` to `LabSources/SoftwarePackages/Sql$($server.SqlVersion)\SQLServerReportingServices.exe` - SQL Server Report Builder: `Get-LabConfigurationItem -Name SqlServerReportBuilder` to `LabSources\SoftwarePackages\ReportBuilder.msi` - SQL Server Management Studio: `Get-LabConfigurationItem -Name Sql$($server.SqlVersion)ManagementStudio` to `LabSources/SoftwarePackages/Sql$($server.SqlVersion)/SSMS-Setup-ENU.exe` - Sample Databases: Download the desired sample databases as .bak files to LabSources\SoftwarePackages\SqlSampleDbs\SqlServerXXXX. The files should be called as the directory, SqlServerXXXX.bak, where XXXX is the version of SQL. TEAM FOUNDATION SERVER AND AZURE DEVOPS A recent TFS/Azure DevOps agent is required which can be downloaded from `Get-LabConfigurationItem -Name BuildAgentUri` and needs to be stored as `LabSources\Tools\TfsBuildWorker.zip`. Getting started PLACING ISO FILES Of course AutomatedLab (AL) cannot install an operating system without actually having the bits. Hence you need to download an ISO file from MSDN, TechNet Evaluation Center or somewhere else. These files need to go to your folder "ISOs" located in the "LabSoures" folder. The "ISOs" folder contains only one file after the installation of AL: "_Put all ISO images in here.txt". I have downloaded Windows Server 2016 from the TechNet Evaluation Center and put the file like shown below. ! ISOs TESTING THE ISO FILES To make sure that AL can read the file, try to get a list of available operating systems. Open an elevated PowerShell ISE and call the following command (make sure you point to the right location for the LabSources folder: powershell Get-LabAvailableOperatingSystem -Path E:\LabSources This returns a list of all operating system images found on the ISO file (of course this works also if there are a bunch of different OS ISOS in the folder). ! OSList INSTALL THE FIRST LAB Please open an elevated PowerShell ISE and create a new empty document (CTRL+N) if not already open. Copy and paste the following lines into the ISE: * powershell New-LabDefinition -Name GettingStarted -DefaultVirtualizationEngine HyperV Add-LabMachineDefinition -Name FirstServer -OperatingSystem 'Windows Server 2016 SERVERSTANDARD' Install-Lab Show-LabDeploymentSummary * The just press the run button or hit F5 to start the deployment. This is what is going to happen. Many things happen automatically but can be customized: * AutomatedLab starts a new lab named "GettingStarted". The lab defininition will be stored in C:\ProgramData\AutomatedLab\Labs\GettingStarted. The location can be customized "Advanced AutomatedLab Configurations" with the setting LabAppDataRoot. * AL will update download the SysInternals tools and put them into the LabSources folder. * AL looks for an ISO file that contains the specified OS. If the ISO file cannot be found, the deployment stops. * AL adds the one and only machine to the lab and recognizes that no network was defined. In this case, AL creates a virtual switch automatically and uses an free IP range. * The AL measures the disk speed and chooses the fastet drive for the lab, as no location is defined in the call to "New-LabDefinition". In my case, it chooses D. Measuring is done only once and the result is cached. * Then the actual deployment starts. AL creates 1. The virtual switch 2. Then it creates the a base image for the operating system that is shared among all machines with the same OS. 3. Afterwards the VM is created and started 4. AL waits for the machine to become ready and shows the overall installation time. REMOVING A LAB If you want to get rid of the lab, just call Remove-Lab. The cmdlet removes the VMs including the disks and the virtual switches, and leaves the base disks for the next deployment. If you have closed the ISE in the meantime, either specify the lab name or import it first. ! Remove1 SUMMARY With AutomatedLab it is extremely easy to create various kinds of labs. The more you define your lab by code, the easier it is to re-deploy it and the less time you invest in the long term. If you like what you have seen, take a look at the folder "LabSources\Sample Scripts\Introduction" . These scripts demo how to create domains, internet facing labs, PKI, etc. Please provide feedback if something does not work as expected. If you are missing a feature or have some great ideas, please open an issue . NEXT STEPS Now that you have deployed your first lab, what comes next? Would you like to connect to the machines and run remote commands without you knowing the password? Then start with the docs on lab management . Wondering how to transfer data to your new lab? Then start with the docs on data exchange . If you - like us - like to tinker around with things, check out the possible settings . Installation There are two options installing AutomatedLab: - You can use the MSI installer published on GitHub. - Or you install from the PowerShell Gallery using the cmdlet Install-Module. Please note that this is the ONLY way to install AutomatedLab and its dependencies in PowerShell Core/PowerShell 7 on both Windows and Linux/Azure Cloud Shell FROM GALLERY Install-PackageProvider Nuget -Force Install-Module AutomatedLab -AllowClobber ## If you are on Linux and are not starting pwsh with sudo ## This needs to executed only once per user - adjust according to your needs! Set-PSFConfig -Module AutomatedLab -Name LabAppDataRoot -Value /home/youruser/.alConfig -PassThru | Register-PSFConfig ## Prepare sample content - modify to your needs ## Windows New-LabSourcesFolder -Drive C ## Linux Set-PSFConfig -Module AutomatedLab -Name LabSourcesLocation -Value /home/youruser/labsources -PassThru | Register-PSFConfig New-LabSourcesFolder # Linux FROM MSI AutomatedLab (AL) is a bunch of PowerShell modules. To make the installation process easier, it is provided as an MSI. Download Link: https://github.com/AutomatedLab/AutomatedLab/releases There are not many choices when installing AL. ! Install1 The options Typical and Complete are actually doing the same and install AL to the default locations. The PowerShell modules go to "C:\Program Files\WindowsPowerShell\Modules", the rest to "C:\LabSources". As LabSources can grow quite big, you should go for a custom installation and put this component on a disk with enough free space to store the ISO files. This disk does not have to be an SSD. Do not change the location of the modules unless you really know what you are doing. ! Install2 Very important to AL is the LabSources folder that should look like this: ! Install3 If all that worked you are ready to go for Getting Started . SUMMARY Installing software on lab machines is almost as easy as installing <TODO Windows Features>. And there are similar challenges. But additionally, there to the challenges some installers have an extra demand, like the .net 4.5.2 package that cannot be installed remotely. Like for installing Windows Features AutomatedLab tries to provide a solution that makes this task as simple as possible. It does not matter if the installation file is already on the VM or on your host machine . And you do not have to care whether it is a .exe, .msi or .msu file, AL will handle the complexity for you. And there is also a very convenient way to work install stuff from ISOs mounted to the lab VMs. INSTALL-LABSOFTWAREPACKAGE INTRODUCTION The cmdlet that handles the installation is Install-LabSoftwarePackage. It is designed to work in various scenarios that will be discussed in this article. Internally, it uses Invoke-LabCommand again to connect to the lab VMs by using the credentials known to the lab. The files are copied to the lab VMs using Copy-LabFileItem. INSTALL-LABSOFTWAREPACKAGE USAGE INSTALLING AN PACKAGE THAT IS ON THE HOST MACHINE A very simple demonstration of how the cmdlet can help, is installing Notepad++ on a lab VM. Note that $labSources always points to the LabSources folder that is created by AL, so you do not have to provide the full path. This also works on Azure. A lab VM hosted on Azure accesses the share hosted on Azure by default. The Azure LabSources share can be synced with the local one. The next command copes the Notepad++.exe file to the VM and invokes the installer. The package must support a silent installation. For Notepad++, you can switch to the silent mode by providing the argument “/S†(case-sensitive). PowerShell Install-LabSoftwarePackage -ComputerName Server1 -Path $labSources\SoftwarePackages\Notepad++.exe -CommandLine /S If you want to install the same package to all machines in a lab, regardless if they run on Hyper-V or Azure, just change the ComputerName parameter: PowerShell Install-LabSoftwarePackage -ComputerName (Get-LabVM) -Path $labSources\SoftwarePackages\Notepad++.exe -CommandLine /S INSTALLING A PACKAGE THAT IS ON THE HOST MACHINE AS SCHEDULED JOB Some software packages are creating headaches, like the .net Framework 4.5.2. The package calls the Windows Update Service which checks if it is called from remote by checking the token for the NETWORK RID, and cancels the installation if that’s the case. Install-LabSoftwarePackage offers the switch AsScheduledJob to be able to start the installation remotely which then actually runs locally. The next example shows how to install .net 4.5.2 on a VM, restart the VM and then install the Windows Management Framework 5.1. PowerShell Install-LabSoftwarePackage -Path $labSources\SoftwarePackages\NDP452-KB2901907-x86-x64-AllOS-ENU.exe -CommandLine '/q /log c:\dotnet452.txt' -ComputerName Client7 -AsScheduledJob -UseShellExecute Restart-LabVM -ComputerName Client7 -Wait Install-LabSoftwarePackage -Path $labSources\V5\Win7AndW2K8R2-KB3134760-x64.msu -ComputerName Client7 Restart-LabVM -ComputerName Client7 -Wait INSTALLING A PACKAGE THAT IS ALREADY ON THE VM Installing software packages that are already on the VM is also possible. In this case, you can use the LocalPath parameter instead of the Path parameter. The following command installs the Redgate Reflector on all machines with the role visual studio, expecting the package already to be at C:\. The AsJob switch puts the work into the background. PowerShell Install-LabSoftwarePackage -LocalPath C:\ReflectorInstaller.exe -CommandLine '/qn /IAgreeToTheEula' -ComputerName (Get-LabMachine -Role VisualStudio2015) -AsJob INSTALLING FILES FROM ISO FILES MOUNTED TO LAB VMS. The cmdlet Install-LabSoftwarePackage is even more effective when combined with other cmdlets AutomatedLab provides. When mounting an ISO into a VM, it is not guaranteed that the ISO will get always the same drive letter. If a lab VM has only one drive, the ISO will be pretty likely drive D:, but if the lab VM has more than one disk attached, things are getting more complex. Mount-LabIsoImage returns the drive an ISO file actually has inside the VM. This information can be used to construct the path to call. This is demoed in the next example. PowerShell $drive = Mount-LabIsoImage -ComputerName Web1 -IsoPath $labSources\ISOs\SkypeForBusiness2015.iso -PassThru Install-LabSoftwarePackage -ComputerName Web1 -LocalPath "$($drive.DriveLetter)\Setup.exe" -CommandLine /BootStrap Dismount-LabIsoImage -ComputerName Web1 For further help and details on the available parameters please call Get-Help Install-LabSoftwarePackage. SUMMARY A quite common task in a lab is to add Windows features to one or some machines. This can be easily done using the cmdlet Install-WindowsFeature that is available since Windows Server 2008 R2. It is also quite easy to the cmdlet on another machine by using Install-WindowsFeature inside a scriptblock that you invoke on a remote machine using Invoke-Command. However, this requires authentication, creating the scriptblock and other things. The cmdlet Install-LabWindowsFeature does this all for you. INSTALLING WINDOWS FEATURES USING INSTALL-LABWINDOWSFEATURE Install-LabWindowsFeature uses Invoke-LabCommand that is discussed in <TODO>. As Invoke-LabCommand handles obstacles like name resolution and authentication with the standard lab account, which makes this task even easier. The following command installs the Remote Sever Administration tools (RSAT) on the machines Server1 and Server2: PowerShell Install-LabWindowsFeature -FeatureName RSAT -ComputerName 'server1', 'server2' -IncludeAllSubFeature It is also possible to push the activity into the background using the AsJob switch. You can start the installation on multiple machines and then wait for the jobs to complete. If AsJob is used along with PassThru, Install-LabWindowsFeature returns the job objects. If just PassThru is used, the cmdlet return the result of the feature installation. PowerShell $jobs = Install-LabWindowsFeature -FeatureName RSAT -ComputerName 'server1', 'server2' -IncludeAllSubFeature -AsJob -PassThru $result = Wait-LWLabJob -Job $jobs -ProgressIndicator 10 -NoDisplay -PassThru PowerShell offers a great way to run code on remote machines by means of the cmdlet Invoke-Command. This cmdlet just needs to get a script block to run and computer names. However, connecting to a lab machine that is not in the same security context requires authentication. Also, double-hop authentication scenarios (CredSsp) need some configuration to work. Another challenge is making local variables and local functions available to remote PowerShell sessions. In this regards AL as a framework gives you some nice tools that are making PowerShell Remoting even more effective. Lab environments usually do not share the same namespace or DNS servers. Hence AL comes up with its own name resolution system. It lets you connect to the machines with the short names regardless whether they are hosted on Hyper-V or Azure. INVOKE-LABCOMMAND Invoke-LabCommand is like a proxy. It works pretty similar to Invoke-Command but adds a lot of functionality that makes working in a lab even easier. The next sections are describing each of these additional features with some samples. NAME RESOLUTION AL needs to support two different ways of name resolution as it supports labs on the loal Hyper-V and on Windows Azure. For Hyper-V scenarios AL creates an entry in the LMHosts file. Connecting to Hyper-V VMs created by AL works with the standard naming resolution. A bit more work is done behind the scenes for VMs hosted on Azure. AL reads the public IP address of the load balancer each machine is connected to and also discovers the public port for WinRM. This information is added to each machine (see AzureConnectionInfo property) and used inside Invoke-LabCommand. Actually the hard work is done inside the cmdlet New-LabPSSession which also caches connections. The cmdlets Enter-LabPSSession and Connect-LabVM are implementing the AL name resolution as well. Running a command on a specific VM, regardless if it is hosted on Azure or Hyper-V, works like this (Without the PassThru switch, Invoke-LabCommand does not return any data): PowerShell Invoke-LabCommand -ScriptBlock { Get-Date } -ComputerName Client1 -PassThru Running a command on all lab VMs, regardless if they are hosted on Azure or Hyper-V, works like this: PowerShell Invoke-LabCommand -ScriptBlock { Get-Date } -ComputerName (Get-LabVM) MAKING LOCAL VARIABLES AND FUNCTIONS AVAILABLE REMOTELY If you have functions and variables defined locally, they are not available in the remote session. With the using scope (PowerShell 3+) you can access local variables in the remote session but it requires changing the code by adding “$using:†to each variable you want to retrieve from the local scope. Hence this code will no longer work locally which makes debugging your code locally on the VM much harder. Invoke-LabCommand provides the two parameters “Variable†and “Functionâ€. All the variables and functions given here are pushed to the remove session making them available there. The following sample demonstrates this by having a local function that just returns the value of a local variable. Invoke-LabCommand runs the same code on a remote machine without changing anything. The code you may run locally: PowerShell function Foo { "The calue of '`$someVar' is $someVar" } $someVar = 123 Foo To run the same code remotely PowerShell function Foo { "The calue of '`$someVar' is $someVar" } $someVar = 123 Invoke-LabCommand -ComputerName Client1 -ScriptBlock { Foo } -Variable (Get-Variable -Name someVar) -Function (Get-Command -Name Foo) DOUBLE-HOP AUTHENTICATION AND CREDSSP Every machine deployed with AL has CredSsp enabled as a CredSsp server. This is like running “Enable-WSManCredSSP -Role Serverâ€. Invoke-LabCommand always tries to make a connection with CredSsp. If this does not work you will see a warning and a connection without CredSsp is tried. This is definitely not a best practice for a production environment but makes life much easier in a lab. Using CredSsp you can create a remote session from a remote session which is extremely helpful when installing or setting up a lab. For example, reading an AD user account from a remote session does not work without CredSsp as reading data from Active Directory requires and authentication from the remote machine to a domain controller. Inside an AL lab the following code works out of the box. PowerShell Invoke-LabCommand -ComputerName Web1 -ScriptBlock { Get-ADUser -Identity John } TRANSFERRING MODULES TO MACHINES Invoke-LabCommand is a comfortable way to run scripts and ScriptsBlocks defined on your host machine in any lab machine. This cmdlet takes care of the authentication and allows you to send variables and functions to the remote machine. But how can you use cmdlets defined in a PowerShell module that exists on your host machine in a lab VM? With Send-ModuleToPSSession we have provided a way to send any PowerShell module available locally to a lab VM. Note: This cmdlet makes use of other functions provided within the AutomatedLab framework to copy files to Hyper-V or Azure machines like Send-Directory and what is provided by the PSFileTransfer module. Send-ModuleToPSSession will try to send the module using SMB first and if this does not work, using the PSSession. The module will be copied either to the Program Files\WindowsPowerShell\Modules or to the module path of the user used for opening the PSSession, depending of the scope you set: AllUsers, CurrentUser. There is also a switch to look for all module dependencies and copy them as well. DEMO (PROBLEM) The following code block demos the problem. You have a module locally on your machine and have successfully tested some code. Now you want to invoke that code on all lab machines (Azure or Hyper-V) but the module is not there yet. PowerShell Import-Lab -Name POSH -NoValidation ##works locally Get-NTFSAccess -Path C:\ ##does not work remotely unless the module has been copied to the machines $vms = Get-LabVM -Role ADDS Invoke-LabCommand -ScriptBlock { Get-NTFSAccess -Path C:\ } -ComputerName $vms The error is The term 'Get-NTFSAccess' is not recognized as the name of a cmdlet, function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again. + CategoryInfo : ObjectNotFound: (Get-NTFSAccess:String) [], CommandNotFoundException + FullyQualifiedErrorId : CommandNotFoundException + PSComputerName : POSHDC1 The term 'Get-NTFSAccess' is not recognized as the name of a cmdlet, function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again. + CategoryInfo : ObjectNotFound: (Get-NTFSAccess:String) [], CommandNotFoundException + FullyQualifiedErrorId : CommandNotFoundException DEMO (SOLUTION) It requires just one additional line and you can use your local module on the remote machines and this is, how it works: 1. You deploy your lab or import an import an existing one 2. You create a new PSSession to one or more machines 3. You send the module to these sessions 4. Then you can use the module on the lab VM After using Send-ModuleToPSSession, everything works as desired. PowerShell Import-Lab -Name POSH -NoValidation ##works locally Get-NTFSAccess -Path C:\ $vms = Get-LabVM -Role ADDS ##just one line is required to copy the module to the VMs Send-ModuleToPSSession -Module (Get-Module -Name NTFSSecurity -ListAvailable) -Session (New-LabPSSession -ComputerName $vms) ##now works remotely Invoke-LabCommand -ScriptBlock { Get-NTFSAccess -Path C:\ } -ComputerName $vms -PassThru Join Lab VMs to an existing domain AutomatedLab can deploy in your existing infrastructure with very little modifications necessary. At the moment, this is supported for existing domains only, but may be extended in the future. The `Add-LabMachineDefinition` cmdlet can make use of the `SkipDeployment` parameter in order to add an existing domain controller to your lab. You could for example locate an existing DC to join to: $DomainName = 'janhendrikpeters.de' $ctx = [System.DirectoryServices.ActiveDirectory.DirectoryContext]::new('Domain', $DomainName) $domain = [System.DirectoryServices.ActiveDirectory.Domain]::GetDomain($ctx) $dc = $domain.FindDomainController([System.DirectoryServices.ActiveDirectory.LocatorOptions]::WriteableRequired).Name $dcName = $dc.Replace(".$DomainName", '') $dcIp = [System.Net.Dns]::GetHostAddresses($dc).IpAddressToString Now in order to add your domain controller, simply supply the host name and IP address of the system, as well as the domain join credentials: $DomainJoinCredential = Get-Credential Set-LabInstallationCredential -Username ($DomainJoinCredential.UserName -split '\\')[-1] -Password $DomainJoinCredential.GetNetworkCredential().Password ## We are not using a domain admin as we skip the deployment of the DC. Nevertheless, this credential is used for domain joins. Add-LabDomainDefinition -Name $DomainName -AdminUser ($DomainJoinCredential.UserName -split '\\')[-1] -AdminPassword $DomainJoinCredential.GetNetworkCredential().Password $PSDefaultParameterValues = @{ 'Add-LabMachineDefinition:DomainName' = $DomainName 'Add-LabMachineDefinition:OperatingSystem' = 'Windows Server Datacenter' 'Add-LabMachineDefinition:Memory' = 512MB } Add-LabMachineDefinition -Name $dcName -Roles RootDc -SkipDeployment -IpAddress $dcIp Add-LabMachineDefinition -Name POSHFS01 Add-LabMachineDefinition -Name POSHWEB01 Install-Lab The full sample script can be found in the SampleScripts folder . Stay tuned! Mounting ISO files on Azure is as simple as calling the existing cmdlet ``Mount-LabIsoImage``. The only difference is, that the ISO image file path lies on Azure. To find out which images are accessible in your AutomatedLab storage account you can do the following: `` Login-AzureRmAccount (Get-LabAzureLabSourcesContent -RegexFilter \.iso).FullName `` When mounting the ISO file, be sure to specify -PassThru to be able to use the drive letter that was used to mount the ISO in later commands: $mountedVolume = Mount-LabIsoImage -IsoPath https://some/azure/path.iso -ComputerName DC1 -PassThru Invoke-LabCommand DC1 -ScriptBlock { param ( $DriveLetter ) Start-Process (Join-Path $DriveLetter "Path\To\My\Setup.exe") } -ArgumentList $mountedVolume.DriveLetter AutomatedLab can handle everything about networking and IP addresses automatically for you. But, if you prefer, you can define everything by your own. AL supported also working with multiple subnets and internet connected labs. It takes care about the routing if a machine with the role 'Routing' is defined in the network. FULLY AUTOMATED AutomatedLab tries to make defining and deploying labs as easy as possible. Quite often you need some machines that can talk to each other but you are not interested in the IP configuration. In this case you do not have to care about networking and IP addresses at all. AL defines the virtual switch for you and also finds a free subnet for you. If you take a look at the introduction script [04 Single domain-joined server.ps1 you will not find anything that defines a network. When deploying the lab AL tells you the subnet that was chosen. AL tries to find new subnets by simply increasing 192.168.1.0 until a free network is found. Free means that there is no virtual network switch with an IP address in the range of the subnet and the subnet is not routable. If these conditions are not met, the subnet is incremented again. SIMPLE NETWORK DEFINITION To manually create network definitions in order to have specific network address ranges available, the cmdlet ``Add-LabVirtualNetworkDefinition`` can be used for both HyperV- and Azure-based labs. # Adds a virtual network with no special options Add-LabVirtualNetworkDefinition -Name 'MySimpleNetwork' -AddressSpace 10.1.0.0/16 Additionally, the parameters AzureProperties and HyperVProperties can be used to pass parameter hashtables. The following are valid keys that can be used as properties for AzureProperties: * SubnetName: The name of the subnet to create in the virtual network. AL does not segment virtual networks into subnets. One virtual network per network definition will be created * SubnetAddressPrefix: The address prefix (e.g. 24) of the subnet that will be created. * LocationName: The Azure location name (e.g. westeurope). Bear in mind that this should not differ from your lab's default location. * DnsServers: Comma-separated DNS servers for the network. * ConnectToVnets: Connections to other VNETs by leveraging VNET peering. When connecting two or more lab networks through this parameter, please also specifiy this for all additional network definitions For HyperV, the following properties are valid: * SwitchType: Internal or External. Defaults to internal. If External is specified, AdapterName needs to be set as well * AdapterName: The network adapter of the OS that will bridge the connection to the external network SIMPLE INTERNET CONNECTED Reviewing the HyperV and Azure properties for ``Add-LabVirtualNetworkDefinition`` we can see that internet-connected virtual networks can easily be created. To enable HyperV labs to have an internet connection, the following syntax can be used: Add-LabVirtualNetworkDefinition -Name 'MySimpleConnectedNetwork' -AddressSpace 10.1.0.0/16 -HyperVProperties @{SwitchType = 'External'; AdapterName = 'Ethernet'} This is all it takes for machines to connect to the internet. Azure networks by default are connected to the internet through a load balancer that initially only gives access to WinRM and RDP. Through NAT rules random external ports are mapped to the ports 5985, 5986 and 3389 on each machine. The random ports are stored for each machine, so that all connection-related cmdlets like ``Enter-LabPSSession`` or ``Connect-LabVm`` work transparently. Additionally, Azure RDP files can be created by using ``Get-AzureRmRemoteDesktopFile``. COMPLEX INTERNET CONNECTED here be dragons. While creating an Azure-based lab is very easy and only requires to set the default virtualization engine to Azure instead of HyperV not all features are readily available from the start. To fully make use of Azure, you should also synch your lab sources with Azure. Doing so only requires a single line of code: ` Sync-LabAzureLabSources ` This by default will synchronize your entire local lab sources folder with a pre-created Azure file share called labsources in the resource group AutomatedLabSources and the lab's randomly-named storage account. This file share will automatically be mapped on each virtual machine. Operating system images are automatically skipped in the synchronization process as they are readily available on Azure and are not being used. During the lab initialization the builtin variable $labsources will be automatically updated to point to your Azure lab sources location if you are deploying an Azure-based lab. There is no need to declare the variable in any script, as it is dynamically calculated to match the default virtualization engine. To skip large files or entire ISOs for certain products like e.g. Exchange or Skype for Business the following parameters are available: * SkipIsos: This switch parameter indicates that all ISO files should be skipped. * MaxFileSizeInMb: This parameter takes an integer value indicating the maximum size of the files to synch. Now you can simply use the other cmdlets provided by AutomatedLab to access files from the file share: * `$postInstallActivity = Get-LabPostInstallationActivity -ScriptFileName PrepareRootDomain.ps1 -DependencyFolder $labSources\PostInstallationActivities\PrepareRootDomain` * `Install-LabSoftwarePackage -ComputerName $machines -Path $labSources\SoftwarePackages\Notepad++.exe -CommandLine /S -AsJob` Using Azure Bastion Hosts Azure Bastion Hosts are in preview as of 2020-05-19 and allow you to connect to your lab machines via the Remote Desktop Web Client from an Azure Bastion. Starting with release 5.21 AutomatedLab supports these as well. When opting to deploy a bastion, note that this will increase the time your lab deployment takes. This is due to: - The resource provider feature AllowBastionHost takes time to activate, if it is not registered. ( *This step can take up to 20 minutes! *) - The resource group deployment eventually takes up to five minutes longer USAGE The easiest way to add a Bastion to your lab is by using the new AllowBastionHost parameter: New-LabDefinition -Name Bastion1 -DefaultVirtualizationEngine Azure Add-LabAzureSubscription -SubscriptionName JHPaaS -DefaultLocationName 'West Europe' -AllowBastionHost Add-LabMachineDefinition -Name DC1 -Memory 1GB -OperatingSystem 'Windows Server 2016 Datacenter (Desktop Experience)' -Roles RootDC -DomainName contoso.com Install-Lab Show-LabDeploymentSummary -Detailed Specifying this parameter will mean that your virtual network is extended to accomodate the bastion host subnet. Additionally, the lab network security group will receive additional rules. MORE CONTROL WITH CUSTOM SUBNETS If you do not want to extend the virtual network you defined in your lab, you can simply add a bastion subnet called AzureBastionSubnet - AutomatedLab will take care of the rest. New-LabDefinition -Name Bastion2 -DefaultVirtualizationEngine Azure Add-LabAzureSubscription -SubscriptionName JHPaaS -AllowBastionHost -DefaultLocationName 'West Europe' Add-LabVirtualNetworkDefinition -Name Lab2Vnet -AddressSpace 192.168.10.0/23 -AzureProperties @{ Subnets = @{ 'default' = '192.168.10.0/24' 'AzureBastionSubnet' = '192.168.11.0/24' }} Add-LabMachineDefinition -Name DC1 -Memory 1GB -OperatingSystem 'Windows Server 2016 Datacenter (Desktop Experience)' -Roles RootDC -DomainName contoso.com -Network Bastion2Vnet -IpAddress 192.168.10.11 Install-Lab Show-LabDeploymentSummary -Detailed CONNECTING VIA THE BASTION At the moment, you need to use the Azure portal to connect to your bastion host. In the future, we might be able to use Connect-LabVm instead. |