Untitled Document
  Home
Speakers
Sessions
  Schedule
Sponsors
Exhibitors
  Media Sponsors
  Association Sponsors
Untitled Document
2012 East Diamond Sponsor

Untitled Document
2012 East Platinum Plus Sponsors

Untitled Document
2012 East Platinum Sponsors

Untitled Document
2012 East Gold Sponsors

Untitled Document
2012 East Silver Sponsors

Untitled Document
2012 East Bronze Sponsors

Untitled Document
2012 East Exhibitors

Untitled Document
2012 East Big Data Pavilion Sponsor

Untitled Document
2012 East Day 2 & 3 Morning Break Sponsor

Untitled Document
2012 East ODCA Partner Pavilion

Untitled Document
2012 East Association Sponsors

Untitled Document
2012 East Media Sponsors

Untitled Document
2011 West Diamond Sponsor

Untitled Document
2011 West Platinum Sponsors

Untitled Document
2011 West Platinum Plus Sponsor

Untitled Document
2011 West Gold Sponsors

Untitled Document
2011 West Silver Sponsors

Untitled Document
2011 West Bronze Sponsor


Untitled Document
2011 West Exhibitors























































Untitled Document
2011 West Wireless Network Sponsor

Untitled Document
2011 West Break Sponsor

Untitled Document
2011 West Day 3 Coffee Break Sponsor

Untitled Document
2011 West Day 2 Lunch Sponsor

Untitled Document
2011 West Day 3 Lunch Sponsor

Untitled Document
2011 Media Sponsors

Step-by-Step: Build a SharePoint 2013 Lab in the Cloud on Windows Azure
Leverage Windows Azure as your Lab in the Cloud!

My good friend and colleague, Tommy Patterson, recently blogged about leveraging the Windows Azure Infrastructure as a Service (IaaS) preview offering to build a FREE lab environment in the Cloud.  You can read Tommy’s step-by-step article at:

What about a SharePoint 2013 Lab in the Cloud?
Now that SharePoint Server 2013 has been released, I frequently get asked about ways in which a SharePoint 2013 lab environment can be easily built for studying, testing and/or performing a proof-of-concept.  You could certainly build this lab environment on your own hardware, but due to the level of SharePoint 2013 hardware requirements, a lot of us may not have sufficient spare hardware to implement an on-premise lab environment.

This makes a great scenario for leveraging our Windows Azure FREE 90-day Trial Offer to build a free lab environment for SharePoint 2013 in the cloud.  Using the process outlined in this article, you’ll be able to build a basic functional farm environment for SharePoint 2013 that will be accessible for approximately 105 hours of compute usage each month at no cost to you under the 90-day Trial Offer.

After the 90-day trial period is up, you can choose if you’d like to convert to a full paid subscription.  If you choose to convert to a paid subscription, this lab environment will cost approximately $0.56 USD per hour of compute usage ( that’s right – just 56 cents per hour ) plus associated storage and networking costs ( which can typically be less than $10 USD per month for a lab of this nature ). These estimated costs are based on published Pay-As-You-Go pricing for Windows Azure that is current as of this article’s date.

Note: If you are testing advanced SharePoint 2013 scenarios and need more resources than available in the lab configuration below, you can certainly scale-up or scale-out elastically by provisioning larger VMs or additional SharePoint web and application server VMs.  To determine the specific costs associated with higher resource levels, please visit the Windows Azure Pricing Calculator for Virtual Machines.

SharePoint 2013 Lab Scenario
To deliver a functional and expandable lab environment, I’ll be walking through the approach of provisioning SharePoint Server 2013 on Windows Azure VMs as depicted in the following configuration diagram that will require three (3) VMs on a common Windows Azure Virtual Network.

SP2013onAzureScenario

Lab Scenario: SharePoint 2013 on Windows Azure VM

In this lab, we’ll be using a naming convention of XXXlabYYY01, where XXX will be replaced with your unique initials and YYY will be replaced with an abbreviation representing the function of a virtual machine or Windows Azure configuration component (ie., ad, db or app).

Note: This study lab configuration is suitable for study, functional testing and basic proof-of-concept usage.  This configuration is not currently supported for pilot or production SharePoint 2013 farm environments.

Prerequisites

The following is required to complete this step-by-step guide:

  • A Windows Azure subscription with the Virtual Machines Preview enabled.

    DO IT: Sign up for a FREE Trial of Windows Azure

    NOTE: When activating your FREE Trial for Windows Azure, you will be prompted for credit card information.  This information is used only to validate your identity and your credit card will not be charged, unless you explicitly convert your FREE Trial account to a paid subscription at a later point in time. 
  • Completion of the Getting Started tasks in the following article:

    DO IT: Getting Started with Servers in the Cloud
  • This step-by-step guide assumes that the reader is already familiar with configuring Windows Server Active Directory, SQL Server and SharePoint Server 2013 in an on-premise installation. This guide focuses on the unique aspects associated with configuring these components on the Windows Azure cloud platform.

Let’s Get Started!
In this step-by-step guide, you will learn how to:

  • Register a DNS Server in Windows Azure
  • Define a Virtual Network in Windows Azure
  • Configure Windows Server Active Directory in a Windows Azure VM
  • Configure SQL Server 2012 in a Windows Azure VM
  • Configure SharePoint Server 2013 in a Windows Azure VM
  • Export / Import Lab Environment via PowerShell

Exercise 1: Register a DNS Server in Windows Azure

Register the internal IP address that our domain controller VM will be using for Active Directory-integrated Dynamic DNS services by performing the following steps:

  1. Sign in at the Windows Azure Management Portal with the logon credentials used when you signed up for your Free 90-Day Windows Azure Trial.
  2. Select Networks located on the side navigation panel on the Windows Azure Management Portal page.
  3. Click the +NEW button located on the bottom navigation bar and select Networks | Virtual Network | Register DNS Server.
  4. Complete the DNS Server fields as follows:

    - NAME: XXXlabdns01
    - DNS Server IP Address: 10.0.0.4
  5. Click the REGISTER DNS SERVER button.

Exercise 2: Define a Virtual Network in Windows Azure

Define a common virtual network in Windows Azure for running Active Directory, Database and SharePoint virtual machines by performing the following steps:

  1. Sign in at the Windows Azure Management Portal with the logon credentials used when you signed up for your Free 90-Day Windows Azure Trial.
  2. Select Networks located on the side navigation panel on the Windows Azure Management Portal page.
  3. Click the +NEW button located on the bottom navigation bar and select Networks | Virtual Network | Quick Create.
  4. Complete the Virtual Network fields as follows:

    - NAME: XXXlabnet01
    - Address Space: 10.---.---.---
    - Maximum VM Count: 4096 [CIDR: /20]
    - Affinity Group: Select the Affinity Group defined in the Getting Started steps from the Prerequisites section above.
    - Connect to Existing DNS: Select XXXlabdns01 – the DNS Server registered in Exercise 1 above.
  5. Click the CREATE A VIRTUAL NETWORK button.

Exercise 3: Configure Windows Server Active Directory in a Windows Azure VM

Provision a new Windows Azure VM to run a Windows Server Active Directory domain controller in a new Active Directory forest by performing the following steps:

  1. Sign in at the Windows Azure Management Portal with the logon credentials used when you signed up for your Free 90-Day Windows Azure Trial.
  2. Select Virtual Machines located on the side navigation panel on the Windows Azure Management Portal page.
  3. Click the +NEW button located on the bottom navigation bar and select Compute | Virtual Machines | From Gallery.
  4. In the Virtual Machine Operating System Selection list, select Windows Server 2012, December 2012 and click the Next button.
  5. On the Virtual Machine Configuration page, complete the fields as follows:

    - Virtual Machine Name: XXXlabad01
    - New Password and Confirm Password fields: Choose and confirm a new local Administrator password.
    - Size: Small (1 core, 1.75GB Memory)

    Click the Next button to continue.

    Note: It is suggested to use secure passwords for Administrator users and service accounts, as Windows Azure virtual machines could be accessible from the Internet knowing just their DNS.  You can also read this document on the Microsoft Security website that will help you select a secure password: http://www.microsoft.com/security/online-privacy/passwords-create.aspx.
  6. On the Virtual Machine Mode page, complete the fields as follows:

    - Standalone Virtual Machine: Selected
    - DNS Name: XXXlabad01.cloudapp.net
    - Storage Account: Select the Storage Account defined in the Getting Started steps from the Prerequisites section above.
    - Region/Affinity Group/Virtual Network: Select XXXlabnet01 – the Virtual Network defined in Exercise 2 above.
    - Virtual Network Subnets: Select Subnet-1 (10.0.0.0/23)

    Click the Next button to continue.
  7. On the Virtual Machine Options page, click the Checkmark button to begin provisioning the new virtual machine.

    As the new virtual machine is being provisioned, you will see the Status column on the Virtual Machines page of the Windows Azure Management Portal cycle through several values including Stopped, Stopped (Provisioning), and Running (Provisioning).  When provisioning for this new Virtual Machine is completed, the Status column will display a value of Running and you may continue with the next step in this guide.
  8. After the new virtual machine has finished provisioning, click on the name ( XXXlabad01 ) of the new Virtual Machine displayed on the Virtual Machines page of the Windows Azure Management Portal.
  9. On the virtual machine details page for XXXlabad01, make note of the Internal IP Address displayed on this page.  This IP address should be listed as 10.0.0.4

    If a different internal IP address is displayed, the virtual network and/or virtual machine configuration was not completed correctly.  In this case, click the DELETE button located on the bottom toolbar of the virtual machine details page for XXXlabad01, and go back to Exercise 2 and Exercise 3 to confirm that all steps were completed correctly.
  10. On the virtual machine details page for XXXlabad01, click the Attach button located on the bottom navigation toolbar and select Attach Empty Disk.  Complete the following fields on the Attach an empty disk to the virtual machine form:

    - Name: XXXlabad01-data01
    - Size: 10 GB
    - Host Cache Preference: None

    Click the Checkmark button to create and attach the a new virtual hard disk to virtual machine XXXlabad01.
  11. On the virtual machine details page for XXXlabad01, click the Connect button located on the bottom navigation toolbar and click the Open button to launch a Remote Desktop Connection to the console of this virtual machine.  Logon at the console of your virtual machine with the local Administrator credentials defined in Step 5 above.
  12. From the Remote Desktop console of XXXlabad01, create a new partition on the additional data disk attached above in Step 10 and format this partition as a new F: NTFS volume.  This volume will be used for NTDS DIT database, log and SYSVOL folder locations.

    If you need additional guidance to complete this step, feel free to leverage the following study guide for assistance: Windows Server 2012 “Early Experts” Challenge – Configure Local Storage
  13. Using the Server Manager tool, install Active Directory Domain Services and promote this server to a domain controller in a new forest with the following parameters:

    - Active Directory Forest name: contoso.com
    - Volume Location for NTDS database, log and SYSVOL folders: F:

    If you need additional guidance to complete this step, feel free to leverage the following study guide for assistance: Windows Server 2012 “Early Experts” Challenge – Install and Administer Active Directory
  14. After Active Directory has been installed, create the following user accounts that will be used when installing and configuring SharePoint Server 2013 later in this step-by-step guide:

    - CONTOSO\sp_farm – SharePoint Farm Data Access Account
    - CONTOSO\sp_serviceapps – SharePoint Farm Service Applications Account

    If you need additional guidance to complete this step, feel free to leverage the following study guide for assistance: Windows Server 2012 “Early Experts” Challenge – Install and Administer Active Directory

The configuration for this virtual machine is now complete, and you may continue with the next exercise in this step-by-step guide.

Exercise 4: Configure SQL Server 2012 in a Windows Azure VM

Provision a new Windows Azure VM to run SQL Server 2012 by performing the following steps:

  1. Sign in at the Windows Azure Management Portal with the logon credentials used when you signed up for your Free 90-Day Windows Azure Trial.
  2. Select Virtual Machines located on the side navigation panel on the Windows Azure Management Portal page.
  3. Click the +NEW button located on the bottom navigation bar and select Compute | Virtual Machines | From Gallery.
  4. In the Virtual Machine Operating System Selection list, select SQL Server 2012 Evaluation Edition and click the Next button.
  5. On the Virtual Machine Configuration page, complete the fields as follows:

    - Virtual Machine Name: XXXlabdb01
    - New Password and Confirm Password fields: Choose and confirm a new local Administrator password.
    - Size: Medium (2 cores, 3.5GB Memory)

    Click the Next button to continue.
  6. On the Virtual Machine Mode page, complete the fields as follows:

    - Standalone Virtual Machine: Selected
    - DNS Name: XXXlabdb01.cloudapp.net
    - Storage Account: Select the Storage Account defined in the Getting Started steps from the Prerequisites section above.
    - Region/Affinity Group/Virtual Network: Select XXXlabnet01 – the Virtual Network defined in Exercise 2 above.
    - Virtual Network Subnets: Select Subnet-1 (10.0.0.0/23)

    Click the Next button to continue.
  7. On the Virtual Machine Options page, click the Checkmark button to begin provisioning the new virtual machine.

    As the new virtual machine is being provisioned, you will see the Status column on the Virtual Machines page of the Windows Azure Management Portal cycle through several values including Stopped, Stopped (Provisioning), and Running (Provisioning).  When provisioning for this new Virtual Machine is completed, the Status column will display a value of Running and you may continue with the next step in this guide.
  8. After the new virtual machine has finished provisioning, click on the name ( XXXlabdb01 ) of the new Virtual Machine displayed on the Virtual Machines page of the Windows Azure Management Portal.
  9. On the virtual machine details page for XXXlabdb01, make note of the Internal IP Address displayed on this page.  This IP address should be listed as 10.0.0.5

    If a different internal IP address is displayed, the virtual network and/or virtual machine configuration was not completed correctly.  In this case, click the DELETE button located on the bottom toolbar of the virtual machine details page for XXXlabdb01, and go back to Exercise 2 and Exercise 3 to confirm that all steps were completed correctly.
  10. On the virtual machine details page for XXXlabdb01, click the Attach button located on the bottom navigation toolbar and select Attach Empty Disk.  Complete the following fields on the Attach an empty disk to the virtual machine form:

    - Name: XXXlabdb01-data01
    - Size: 50 GB
    - Host Cache Preference: None

    Click the Checkmark button to create and attach the a new virtual hard disk to virtual machine XXXlabdb01.
  11. On the virtual machine details page for XXXlabdb01, click the Connect button located on the bottom navigation toolbar and click the Open button to launch a Remote Desktop Connection to the console of this virtual machine.  Logon at the console of your virtual machine with the local Administrator credentials defined in Step 5 above.
  12. From the Remote Desktop console of XXXlabdb01, create a new partition on the additional data disk attached above in Step 10 and format this partition as a new F: NTFS volume.
  13. Open SQL Server Management Studio from Start | All Programs | Microsoft SQL Server 2012 | SQL Server Management Studio and update default folder locations to the F: volume.

    1. Connect to the SQL Server 2012 default instance using your Windows Account.
    2. Now, you will update the database's default locations for DATA, LOGS and BACKUP folders. To do this, right click on your SQL Server instance and select Properties.
    3. Select Database Settings from the left side pane.
    4. Locate the Database default locations section and update the default values for each path to point to the F: volume you previously formatted.
    5. Close SQL Server Management Studio.
  14. In order to allow SharePoint to connect to the SQL Server, you will need to add an Inbound Rule for the SQL Server requests in the Windows Firewall. To do this, open Windows Firewall with Advanced Security from Start | All Programs | Administrative Tools.

    1. Select Inbound Rules node, right-click it and select New Rule to open the New Inbound Rule Wizard.

    2. In the Rule Type page, select Port and click Next.

    3. In Protocols and Ports page, leave TCP selected, select Specific local ports, and set its value to 1433. Click Next to continue.

    4. In the Action page, make sure that Allow the connection is selected and click Next.

    5. In the Profile page, leave the default values and click Next.

    6. In the Name page, set the Inbound Rule's Name to SQLServerRule and click Finish

    7. Close Windows Firewall with Advanced Security window.

  15. Using the Server Manager tool, join this server to the contoso.com domain and restart the server to complete the domain join operation.
  16. After the server restarts, connect again via Remote Desktop to the server’s console and login with the local Administrator credentials defined above in Step 5.

  17. Open SQL Server Management Studio from Start | All Programs | Microsoft SQL Server 2012 | SQL Server Management Studio and add the CONTOSO\Administrator user to SQL Server with the Sysadmin server role selected.

    1. Expand Security folder within the SQL Server instance. Right-click Logins folder and select New Login.

    2. In the General section, set the Login name to CONTOSO\Administrator, and select the Windows Authentication option.

    3. Click Server Roles on the left pane.  Select the checkbox for the Sysadmin server role.

    4. Click the OK button and close SQL Server Management Studio.

The configuration for this virtual machine is now complete, and you may continue with the next exercise in this step-by-step guide.

Exercise 5: Configure SharePoint Server 2013 in a Windows Azure VM

Provision a new Windows Azure VM to run SharePoint Server 2013 by performing the following steps:

  1. Sign in at the Windows Azure Management Portal with the logon credentials used when you signed up for your Free 90-Day Windows Azure Trial.
  2. Select Virtual Machines located on the side navigation panel on the Windows Azure Management Portal page.
  3. Click the +NEW button located on the bottom navigation bar and select Compute | Virtual Machines | From Gallery.
  4. In the Virtual Machine Operating System Selection list, select Windows Server 2012, December 2012 and click the Next button.
  5. On the Virtual Machine Configuration page, complete the fields as follows:

    - Virtual Machine Name: XXXlabapp01
    - New Password and Confirm Password fields: Choose and confirm a new local Administrator password.
    - Size: Large (4 cores, 7GB Memory)

    Click the Next button to continue.
  6. On the Virtual Machine Mode page, complete the fields as follows:

    - Standalone Virtual Machine: Selected
    - DNS Name: XXXlabapp01.cloudapp.net
    - Storage Account: Select the Storage Account defined in the Getting Started steps from the Prerequisites section above.
    - Region/Affinity Group/Virtual Network: Select XXXlabnet01 – the Virtual Network defined in Exercise 2 above.
    - Virtual Network Subnets: Select Subnet-1 (10.0.0.0/23)

    Click the Next button to continue.
  7. On the Virtual Machine Options page, click the Checkmark button to begin provisioning the new virtual machine.

    As the new virtual machine is being provisioned, you will see the Status column on the Virtual Machines page of the Windows Azure Management Portal cycle through several values including Stopped, Stopped (Provisioning), and Running (Provisioning).  When provisioning for this new Virtual Machine is completed, the Status column will display a value of Running and you may continue with the next step in this guide.
  8. After the new virtual machine has finished provisioning, click on the name ( XXXlabapp01 ) of the new Virtual Machine displayed on the Virtual Machines page of the Windows Azure Management Portal.
  9. On the virtual machine details page for XXXlabapp01, make note of the Internal IP Address displayed on this page.  This IP address should be listed as 10.0.0.6

    If a different internal IP address is displayed, the virtual network and/or virtual machine configuration was not completed correctly.  In this case, click the DELETE button located on the bottom toolbar of the virtual machine details page for XXXlabapp01, and go back to Exercise 2,  Exercise 3 and Exercise 4 to confirm that all steps were completed correctly.
  10. On the virtual machine details page for XXXlabapp01, click the Attach button located on the bottom navigation toolbar and select Attach Empty Disk.  Complete the following fields on the Attach an empty disk to the virtual machine form:

    - Name: XXXlabapp01-data01
    - Size: 50 GB
    - Host Cache Preference: None

    Click the Checkmark button to create and attach the a new virtual hard disk to virtual machine XXXlabapp01.
  11. On the virtual machine details page for XXXlabapp01, click the Connect button located on the bottom navigation toolbar and click the Open button to launch a Remote Desktop Connection to the console of this virtual machine.  Logon at the console of your virtual machine with the local Administrator credentials defined in Step 5 above.
  12. From the Remote Desktop console of XXXlabapp01, create a new partition on the additional data disk attached above in Step 10 and format this partition as a new F: NTFS volume.  
  13. In the Server Manager tool, click on Local Server in the left navigation pane and click on the Workgroup option.  Join this server to the contoso.com domain and restart the server to complete the domain join operation.
  14. After the server restarts, re-establish a Remote Desktop connection to the server and logon with the CONTOSO\Administrator domain user credentials defined earlier in Exercise 3.
  15. In the Server Manager tool, click on Local Server in the left navigation pane and select IE Enhanced Security Configuration. Turn off enhanced security for Administrators and click the OK button.

    Note: Modifying Internet Explorer Enhanced Security configurations is not good practice and is only for the purpose of this particular step-by-step guide. The correct approach should be to download files locally and then copy them to a shared folder or directly to the VM.

  16. Press the Windows key to switch to the Start Screen and launch Internet Explorer.  Download the following files to the F:\INSTALL folder:

    - SharePoint Server 2013 Evaluation Edition

    Make a note of the SharePoint Product Key listed on this page, as you’ll need it for the installation of SharePoint Server 2013.

    - ASP.NET 4.5 hotfix for Windows Server 2012 ( KB2765317 )

  17. Navigate to the F:\INSTALL folder and double-click on the downloaded .IMG file to mount it.  Copy all files and folders from the mounted .IMG file to F:\INSTALL.
  18. Install the SharePoint Server 2013 software prerequisites by running F:\INSTALL\prerequisiteinstaller.exe.  Note that this process may require multiple server restarts to complete.  After all required software is successfully installed, continue with the next step in this step-by-step guide.
  19. Install the ASP.NET 4.5 hotfix downloaded to the F:\INSTALL folder in Step 14 above.
  20. Run F:\INSTALL\setup.exe to launch the SharePoint Server 2013 installation process.
  21. When prompted, on the Server Type tab of the setup program, select the Complete installation option.
  22. On the File Location tab of the setup program, change the data path to use the F: volume formatted in Step 12 above.
  23. At the end of the installation process, ensure the checkbox is selected to Run the SharePoint Products Configuration Wizard Now and click the Close button.
  24. In the SharePoint Products Configuration Wizard, when prompted on the Connect to server farm dialog, select the option to Create a new server farm.
  25. On the Specify Configuration Database Settings, specify the following values for each field:

    - Database Server: XXXlabdb01
    - Username: CONTOSO\sp_farm
    - Password: Type the password specified when the sp_farm domain user account was created earlier in Exercise 3, Step 14.
  26. Click the Next > button and accept all other default values in the SharePoint Products Configuration Wizard.  Click the Finish button when prompted to complete the wizard.
  27. The SharePoint 2013 Central Administration web page should launch automatically.  When prompted, click the Start the Wizard button to begin the Initial Farm Configuration Wizard.
  28. When prompted for Service Account, type the CONTOSO\sp_serviceapps domain username and password specified when this account was created earlier in Exercise 3, Step 14.
  29. Accept all other default values and click the Next > button to continue.
  30. On the Create a Site Collection page, create a new top-level Intranet site collection using the following field values:

    - Title and Description: Enter your preferred Title and Description for the new site collection
    - URL: Select the root URL path – http://XXXlabapp01/
    - Select experience version:
    2013
    - Select a template: Publishing | Publishing Portal

    Click the OK button to provision a new top-level Intranet site collection. 

    After the new top-level Intranet site collection is provisioned, test navigating to the URL for this site collection from within the Remote Desktop session to the server.
  31. On the SharePoint 2013 Central Administration site, configure a Public URL alternate access mapping for accessing the new top-level Intranet site collection from the Internet.
    1. On the Central Administration site home page, click the Configure alternate access mappings link.
    2. On the Alternate Access Mappings page, click the Edit Public URLs link.
    3. On the Edit Public Zone URLs page, select and specify the following values:

      - Alternate Access Mapping Collection: SharePoint - 80
      - Internet: http://XXXlabapp01.cloudapp.net

      Click the Save button to complete the Alternate Access Mapping configuration.
  32. Close the Remote Desktop session to the server.
  33. Sign in at the Windows Azure Management Portal with the logon credentials used when you signed up for your Free 90-Day Windows Azure Trial.
  34. Select Virtual Machines located on the side navigation panel on the Windows Azure Management Portal page.
  35. On the Virtual Machines page, click on the name of the SharePoint virtual machine – XXXlabapp01.
  36. On the XXXlabapp01 virtual machine details page, click on Endpoints in the top navigation area of the page.
  37. Click the +Add Endpoint button in the bottom navigation bar of the page to define a new virtual machine endpoint that will permit HTTP web traffic inbound to the SharePoint virtual machine. 
  38. On the Add an endpoint to a virtual machine form, select the Add Endpoint option and click the Next button to continue.
  39. On the Specify the details of the endpoint form, specify the following field values:

    - Name: Web HTTP
    - Protocol: TCP
    - Public Port: 80
    - Private Port: 80

    Click the Checkmark button to create a new endpoint definition that will permit inbound web traffic to the SharePoint virtual machine.
  40. After the endpoint configuration has been successfully applied, test browsing to the following public URL to confirm that you are able to access the Intranet site collection that is configured on SharePoint:

    - URL: http://XXXlabapp01.cloudapp.net

The configuration for this virtual machine is now complete, and you may continue with the next exercise in this step-by-step guide.

Exercise 6: Export / Import Lab Environment via PowerShell

Our functional SharePoint lab environment is now complete, but if you’re like me, you won’t be using this lab environment 24x7 around-the-clock.  As long as the virtual machines are provisioned, they will continue to accumulate compute hours against your Free 90-Day Windows Azure Trial account regardless of virtual machine state – even in a shutdown state!

To preserve as many of your free compute hours for productive lab work, we can leverage the Windows Azure PowerShell module to de-provision our lab virtual machines when not in use and re-provision our lab virtual machines when we need them again.  Once you’ve configured the PowerShell scripts below, you’ll be able to spin up your SharePoint lab environment when needed in as little as 5-10 minutes!

Note: Prior to beginning this exercise, please ensure that you’ve downloaded, installed and configured the Windows Azure PowerShell module as outlined in the Getting Started article listed in the Prerequisite section of this step-by-step guide.

  1. De-provisioning your lab. Use the PowerShell snippet below to shutdown, export and de-provision your SharePoint lab environment when you’re not using it.  Prior to running this script, be sure to edit the first line to reflect the names of each of your VMs and confirm that the $ExportPath location exists.

    $myVMs = @("XXXlabapp01","XXXlabdb01","XXXlabad01")
    Foreach ( $myVM in $myVMs ) {
    Stop-AzureVM -ServiceName $myVM -Name $myVM
    $ExportPath = "C:\ExportVMs\ExportAzureVM-$myVM.xml"
    Export-AzureVM -ServiceName $myVM -name $myVM -Path $ExportPath
    Remove-AzureVM -ServiceName $myVM -name $myVM
    }

  2. Re-provisioning your lab. Use the PowerShell snippet below to import and re-provision your SharePoint lab environment when you’re ready to use it again.  Prior to running this script, be sure to edit the first two lines to reflect the names of your Virtual Network and VMs.

    $myVNet = "XXXlabnet01"
    $myVMs = @("XXXlabad01","XXXlabdb01","XXXlabapp01")
    Foreach ( $myVM in $myVMs ) {
    $ExportPath = "C:\ExportVMs\ExportAzureVM-$myVM.xml"    
    Import-AzureVM -Path $ExportPath | New-AzureVM -ServiceName $myVM -VNetName $myVNet
    Start-AzureVM -ServiceName $myVM -name $myVM
    }

To ensure safe de-provisioning and re-provisioning of your SharePoint lab environment, note that it is important to preserve the specific order of the VM names listed in both code snippets above to ensure that the dependency order across VMs is properly handled.

What’s Next? Keep Learning!

Now that your SharePoint Server 2013 lab environment is running in the cloud, be sure to explore the resources below to continue your learning:

Build Your Lab! Build Your Lab! Download Windows Server 2012
Build Your Lab in the Cloud! Don’t Have a Lab? Build Your Lab in the Cloud with Windows Azure Virtual Machines
Join our "Early Experts" study group! Want to Get Certified? Join our Windows Server 2012 "Early Experts" Study Group
About Keith Mayer
Keith Mayer is a Technical Evangelist at Microsoft focused on Windows Infrastructure, Data Center Virtualization, Systems Management and Private Cloud. Keith has over 17 years of experience as a technical leader of complex IT projects, in diverse roles, such as Network Engineer, IT Manager, Technical Instructor and Consultant. He has consulted and trained thousands of IT professionals worldwide on the design and implementation of enterprise technology solutions.

Keith is currently certified on several Microsoft technologies, including System Center, Hyper-V, Windows, Windows Server, SharePoint and Exchange. He also holds other industry certifications from IBM, Cisco, Citrix, HP, CheckPoint, CompTIA and Interwoven.

Keith is the author of the IT Pros ROCK! Blog on Microsoft TechNet, voted as one of the Top 50 "Must Read" IT Blogs.

Keith also manages the Windows Server 2012 "Early Experts" Challenge - a FREE online study group for IT Pros interested in studying and preparing for certification on Windows Server 2012. Join us and become the next "Early Expert"!

Untitled Document
Cloud Expo - Cloud Looms Large on SYS-CON.TV


Cloud Expo 2013 East Opening Keynote by IBM
In this Cloud Expo Keynote, Danny Sabbah, CTO & General Manager, Next Generation Platform, will detail the critical architectural considerations and success factors organizations must internalize to successfully implement, optimize and innovate using next generation architectures.
Lisa Larson, Vice President of Enterprise Cloud Solutions of Rackspace Hosting Live From New York City
In the old world of IT, if you didn't have hardware capacity or the budget to buy more, your project was dead in the water. Budget constraints can leave some of the best, most creative and most ingenious innovations on the cutting room floor. It's a true dilemma for developers and innovators – why spend the time creating, when a project could be abandoned in a blink? That was the old world. In the new world of IT, developers rule. They have access to resources they can spin up instantly. A hybrid cloud ignites innovation and empowers developers to focus on what they need. A hybrid cloud blends the best of all worlds, public cloud, private cloud and dedicated servers to fit the needs of developers and offer the ideal environment for each app and workload without the constraints of a one-size-fits-all cloud.

Keynote: Driving Cloud Innovation: SSDs Change Cloud Storage Paradigm
Cloud is a transformational shift in computing that can have a powerful effect on enterprise IT when designed correctly and used to its full potential. Join Citrix in a discussion that centers on building, connecting and empowering users with cloud services and hear examples of how enterprises are solving real-world business challenges with an architecture and solution purpose-built for the cloud.

Go Beyond IaaS to Deliver "Anything As a Service"
Many organizations want to expand upon the IaaS foundation to deliver cloud services in all forms—software, mobility, infrastructure and IT. Understanding the strategy, planning process and tools for this transformation will help catalyze changes in the way the business operates and deliver real value. Join us to learn about the new ITaaS model and how to begin the transformation.


Cloud Expo Breaking News
Cloud Computing is evolving into a Big Three of Amazon Web Services, Google Cloud, and Microsoft Azure. Cloud 360: Multi-Cloud Bootcamp, being held Nov 4–5, 2014, in conjunction with 15th Cloud Expo in Santa Clara, CA, delivers a real-world demonstration of how to deploy and configure a scalable and available web application on all three platforms. The Cloud 360 Bootcamp, led by Janakiram MSV, an analyst with Gigaom Research, is the first bootcamp that introduces the core concepts of Infrastructure as a Service (IaaS) based on the workings of the Big Three platforms – Amazon EC2, Google Compute Engine, and Azure VMs. Bootcamp attendees will get to see the big picture and also receive the knowledge needed to make the best cloud decisions for their business applications and entire enterprise IT organization.
Scott Jenson leads a project called The Physical Web within the Chrome team at Google. Project members are working to take the scalability and openness of the web and use it to talk to the exponentially exploding range of smart devices. Nearly every company today working on the IoT comes up with the same basic solution: use my server and you'll be fine. But if we really believe there will be trillions of these devices, that just can't scale. We need a system that is open a scalable and by using the URL as a basic building block, we open this up and get the same resilience that the web enjoys.
The Internet of Things is a natural complement to the cloud and related technologies such as Big Data, analytics, and mobility. In his session at Internet of @ThingsExpo, Joe Weinman will lay out four generic strategies – digital disciplines – to exploit emerging digital technologies for strategic advantage. Joe Weinman has held executive leadership positions at Bell Labs, AT&T, Hewlett-Packard, and Telx, in areas such as corporate strategy, business development, product management, operations, and R&D.
SYS-CON Events announced today that DevOps.com has been named “Media Sponsor” of SYS-CON's “DevOps Summit at Cloud Expo,” which will take place on June 10–12, 2014, at the Javits Center in New York City, New York. DevOps.com is where the world meets DevOps. It is the largest collection of original content relating to DevOps on the web today Featuring up-to-the-minute news, feature stories, blogs, bylined articles and more, DevOps.com is where the thought leaders of the DevOps movement make their ideas known.
There are 182 billion emails sent every day, generating a lot of data about how recipients and ISPs respond. Many marketers take a more-is-better approach to stats, preferring to have the ability to slice and dice their email lists based numerous arbitrary stats. However, fundamentally what really matters is whether or not sending an email to a particular recipient will generate value. Data Scientists can design high-level insights such as engagement prediction models and content clusters that allow marketers to cut through the noise and design their campaigns around strong, predictive signals, rather than arbitrary statistics. SendGrid sends up to half a billion emails a day for customers such as Pinterest and GitHub. All this email adds up to more text than produced in the entire twitterverse. We track events like clicks, opens and deliveries to help improve deliverability for our customers – adding up to over 50 billion useful events every month. While SendGrid data covers only abo...
SYS-CON Events announced today that the Web Host Industry Review has been named “Media Sponsor” of SYS-CON's 15th International Cloud Expo®, which will take place on November 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA. Since 2000, The Web Host Industry Review has made a name for itself as the foremost authority of the Web hosting industry providing reliable, insightful and comprehensive news, reviews and resources to the hosting community. TheWHIR Blogs provides a community of expert industry perspectives. The Web Host Industry Review Magazine also offers a business-minded, issue-driven perspective of interest to executives and decision-makers. WHIR TV offers on demand web hosting video interviews and web hosting video features of the key persons and events of the web hosting industry. WHIR Events brings together like-minded hosting industry professionals and decision-makers in local communities. TheWHIR is an iNET Interactive property.
SYS-CON Events announced today that O'Reilly Media has been named “Media Sponsor” of SYS-CON's 15th International Cloud Expo®, which will take place on November 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA. O'Reilly Media spreads the knowledge of innovators through its books, online services, magazines, and conferences. Since 1978, O'Reilly Media has been a chronicler and catalyst of cutting-edge development, homing in on the technology trends that really matter and spurring their adoption by amplifying "faint signals" from the alpha geeks who are creating the future. An active participant in the technology community, the company has a long history of advocacy, meme-making, and evangelism.
SYS-CON Events announced today that Verizon has been named “Gold Sponsor” of SYS-CON's 15th International Cloud Expo®, which will take place on November 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA. Verizon Enterprise Solutions creates global connections that generate growth, drive business innovation and move society forward. With industry-specific solutions and a full range of global wholesale offerings provided over the company's secure mobility, cloud, strategic networking and advanced communications platforms, Verizon Enterprise Solutions helps open new opportunities around the world for innovation, investment and business transformation. Visit verizonenterprise.com to learn more.
SYS-CON Events announced today that TMCnet has been named “Media Sponsor” of SYS-CON's 15th International Cloud Expo®, which will take place on November 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA. Technology Marketing Corporation (TMC) is the world's leading business to business and integrated marketing media company, servicing niche markets within the communications and technology industries.
"In my session I spoke about enterprise cloud analytics and how we can leverage analytics as a service," explained Ajay Budhraja, CTO at the Department of Justice, in this SYS-CON.tv interview at the 14th International Cloud Expo®, held June 10-12, 2014, at the Javits Center in New York City. Cloud Expo® 2014 Silicon Valley, November 4–6, at the Santa Clara Convention Center in Santa Clara, CA, will feature technical sessions from a rock star conference faculty and the leading Cloud industry players in the world.
"We are automated capacity control software, which basically looks at all the supply and demand and running a virtual cloud environment and does a deep analysis of that and says where should things go," explained Andrew Hillier, Co-founder & CTO of CiRBA, in this SYS-CON.tv interview at the 14th International Cloud Expo®, held June 10-12, 2014, at the Javits Center in New York City. Cloud Expo® 2014 Silicon Valley, November 4–6, at the Santa Clara Convention Center in Santa Clara, CA, will feature technical sessions from a rock star conference faculty and the leading Cloud industry players in the world.
“We are starting to see people move beyond the commodity cloud and enterprises need to start focusing on additional value added services in order to really drive their adoption," explained Jason Mondanaro, Director of Product Management at MetraTech, in this SYS-CON.tv interview at the 14th International Cloud Expo®, held June 10-12, 2014, at the Javits Center in New York City. Cloud Expo® 2014 Silicon Valley, November 4–6, at the Santa Clara Convention Center in Santa Clara, CA, will feature technical sessions from a rock star conference faculty and the leading Cloud industry players in the world.
Almost everyone sees the potential of Internet of Things but how can businesses truly unlock that potential. The key will be in the ability to discover business insight in the midst of an ocean of Big Data generated from billions of embedded devices via Systems of Discover. Businesses will also need to ensure that they can sustain that insight by leveraging the cloud for global reach, scale and elasticity. In his session at Internet of @ThingsExpo, Mac Devine, Distinguished Engineer at IBM, will discuss bringing these three elements together via Systems of Discover.
The Internet of Things promises to transform businesses (and lives), but navigating the business and technical path to success can be difficult to understand. In his session at 15th Internet of @ThingsExpo, Chad Jones, Vice President, Product Strategy of LogMeIn's Xively IoT Platform, will show you how to approach creating broadly successful connected customer solutions using real world business transformation studies including New England BioLabs and more.
All too many discussions about DevOps conclude that the solution is an all-purpose player: developer and operations guru, complete with pager for round-the-clock duty. For most organizations that is not the way forward. In his session at DevOps Summit, Bernard Golden, Vice President of Strategy at ActiveState, will discuss how to achieve the agility and speed of end-to-end automation without requiring an organization stocked with Supermen and Superwomen.
Top Stories for Cloud Expo 2012 East

In this Big Data Power Panel at the 10th International Cloud Expo, moderated by Cloud Expo Conference Chair Jeremy Geelan, Govind Rangasamy, Director of Product Management at Eucalyptus Systems; Kevin Brown; CEO of Coraid, Inc.; Christos Tryfonas, CTO and Co-Founder of Cetas; and Max Riggsbee, CMO and VP of Products for WhipTail, discussed such topics as: Big Data has existed since the early days of computing; why, then, do you think there is such an industry buzz around it right now? How is Big Data impacting storage and networking architecture in data centers? How about the intersection of Big Data Analytics and Cloud Computing - how big a sector is that and why? What's the difference between Big Data and Fast Data? ... (more)

Best Recent Articles on Cloud Computing & Big Data Topics
As we enter a new year, it is time to look back over the past year and resolve to improve upon it. In 2014, we will see more service providers resolve to add more personalization in enterprise technology. Below are seven predictions about what will drive this trend toward personalization.
IT organizations face a growing demand for faster innovation and new applications to support emerging opportunities in social, mobile, growth markets, Big Data analytics, mergers and acquisitions, strategic partnerships, and more. This is great news because it shows that IT continues to be a key stakeholder in delivering business service innovation. However, it also means that IT must deliver new innovation despite flat budgets, while maintaining existing services that grow more complex every day.
Cloud computing is transforming the way businesses think about and leverage technology. As a result, the general understanding of cloud computing has come a long way in a short time. However, there are still many misconceptions about what cloud computing is and what it can do for businesses that adopt this game-changing computing model. In this exclusive Q&A with Cloud Expo Conference Chair Jeremy Geelan, Rex Wang, Vice President of Product Marketing at Oracle, discusses and dispels some of the common myths about cloud computing that still exist today.
Despite the economy, cloud computing is doing well. Gartner estimates the cloud market will double by 2016 to $206 billion. The time for dabbling in the cloud is over! The 14th International Cloud Expo, co-located with 5th International Big Data Expo and 3rd International SDN Expo, to be held June 10-12, 2014, at the Javits Center in New York City, N.Y. announces that its Call for Papers is now open. Topics include all aspects of providing or using massively scalable IT-related capabilities as a service using Internet technologies (see suggested topics below). Cloud computing helps IT cut infrastructure costs while adding new features and services to grow core businesses. Clouds can help grow margins as costs are cut back but service offerings are expanded. Help plant your flag in the fast-expanding business opportunity that is The Cloud, Big Data and Software-Defined Networking: submit your speaking proposal today!
What do you get when you combine Big Data technologies….like Pig and Hive? A flying pig? No, you get a “Logical Data Warehouse.” In 2012, Infochimps (now CSC) leveraged its early use of stream processing, NoSQLs, and Hadoop to create a design pattern which combined real-time, ad-hoc, and batch analytics. This concept of combining the best-in-breed Big Data technologies will continue to advance across the industry until the entire legacy (and proprietary) data infrastructure stack will be replaced with a new (and open) one.
While unprecedented technological advances have been made in healthcare in areas such as genomics, digital imaging and Health Information Systems, access to this information has been not been easy for both the healthcare provider and the patient themselves. Regulatory compliance and controls, information lock-in in proprietary Electronic Health Record systems and security concerns have made it difficult to share data across health care providers.
Cloud Expo, Inc. has announced today that Vanessa Alvarez has been named conference chair of Cloud Expo® 2014. 14th International Cloud Expo will take place on June 10-12, 2014, at the Javits Center in New York City, New York, and 15th International Cloud Expo® will take place on November 4-6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
12th International Cloud Expo, held on June 10–13, 2013 at the Javits Center in New York City, featured four content-packed days with a rich array of sessions about the business and technical value of cloud computing led by exceptional speakers from every sector of the cloud computing ecosystem. The Cloud Expo series is the fastest-growing Enterprise IT event in the past 10 years, devoted to every aspect of delivering massively scalable enterprise IT as a service.
Ulitzer.com announced "the World's 30 most influential Cloud bloggers," who collectively generated more than 24 million Ulitzer page views. Ulitzer's annual "most influential Cloud bloggers" list was announced at Cloud Expo, which drew more delegates than all other Cloud-related events put together worldwide. "The world's 50 most influential Cloud bloggers 2010" list will be announced at the Cloud Expo 2010 East, which will take place April 19-21, 2010, at the Jacob Javitz Convention Center, in New York City, with more than 5,000 expected to attend.
It's a simple fact that the better sales reps understand their prospects' intentions, preferences and pain points during calls, the more business they'll close. Each day, as your prospects interact with websites and social media platforms, their behavioral data profile is expanding. It's now possible to gain unprecedented insight into prospects' content preferences, product needs and budget. We hear a lot about how valuable Big Data is to sales and marketing teams. But data itself is only valuable when it's part of a bigger story, made visible in the right context.
Cloud Expo, Inc. has announced today that Larry Carvalho has been named Tech Chair of Cloud Expo® 2014. 14th International Cloud Expo will take place on June 10-12, 2014, at the Javits Center in New York City, New York, and 15th International Cloud Expo® will take place on November 4-6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Everyone talks about a cloud-first or mobile-first strategy. It's the trend du jour, and for good reason as these innovative technologies have revolutionized an industry and made savvy companies a lot of money. But consider for a minute what's emerging with the Age of Context and the Internet of Things. Devices, interfaces, everyday objects are becoming endowed with computing smarts. This is creating an unprecedented focus on the Application Programming Interface (API) as developers seek to connect these devices and interfaces to create new supporting services and hybrids. I call this trend the move toward an API-first business model and strategy.
We live in a world that requires us to compete on our differential use of time and information, yet only a fraction of information workers today have access to the analytical capabilities they need to make better decisions. Now, with the advent of a new generation of embedded business intelligence (BI) platforms, cloud developers are disrupting the world of analytics. They are using these new BI platforms to inject more intelligence into the applications business people use every day. As a result, data-driven decision-making is finally on track to become the rule, not the exception.
Register and Save!
Save $500
on your “Golden Pass”!
Call 201.802.3020
or click here to Register
Early Bird Expires June 10th.


Silicon Valley Call For Papers Now OPEN
Submit
Call for Papers for the
upcoming Cloud Expo in
Santa Clara, CA!
[November 5-8, 2012]


Sponsorship Opportunities
Please Call
201.802.3021
events (at) sys-con.com
SYS-CON's Cloud Expo, held each year in California, New York, Prague, Tokyo, and Hong Kong is the world’s leading Cloud event in its 5th year, larger than all other Cloud events put together. For sponsorship, exhibit opportunites and show prospectus, please contact Carmen Gonzalez, carmen (at) sys-con.com.


New York City Expo Floor Plan Revealed
Cloud Expo New York
[June 11-14, 2012]

Floor Plan Revealed


Introducing Big Data Expo
Introducing
There is little doubt that Big Data solutions will have an increasing role in the Enterprise IT mainstream over time. Get a jump on that rapidly evolving trend at Big Data Expo, which we are introducing in June at
Cloud Expo New York.

Follow @CloudExpo New York on Twitter


Testimonials
Cloud Expo was a fantastic event for CSS Corp - we easily exceeded our objectives for engaging with clients and prospects."
AHMAR ABBAS
SVP, Global Infrastructure Management, CSS Corp.
 
With our launch at Cloud Expo, we successfully transformed the company from a relatively unknown European player into the dominant player in the market. Our competitors were taken by surprise and just blown away. We got a huge number of really high quality leads..."
PETE MALCOLM
CEO, Abiquo
 
We were extremely pleased with Cloud Expo this year - I’d say it exceeded expectations all around. This is the same info we got from partners who attended as well. Nice job!"
MARY BASS
Director of Marketing, UnivaUD
 
Cloud Expo helps focus the debate on the critical issues at hand in effect connecting main street with the next frontier."

GREG O’CONNOR
President & CEO, Appzero


Who Should Attend?
Senior Technologists including CIOs, CTOs, VPs of technology, IT directors and managers, network and storage managers, network engineers, enterprise architects, communications and networking specialists, directors of infrastructure Business Executives including CEOs, CMOs, CIOs, presidents, VPs, directors, business development; product and purchasing managers.


Join Us as a Media Partner - Together We Can Rock the IT World!
SYS-CON Media has a flourishing Media Partner program in which mutually beneficial promotion and benefits are arranged between our own leading Enterprise IT portals and events and those of our partners.

If you would like to participate, please provide us with details of your website/s and event/s or your organization and please include basic audience demographics as well as relevant metrics such as ave. page views per month.

To get involved, email Marilyn Moux at marilyn@sys-con.com.

Lastest Blog Posts
I love receiving gifts in the form of new insights! It doesn't matter if others received the same gift years ago and I am just getting it now. If it is new to me, I get excited. It is like waking up in the morning and discovering a new room in your house. I read an article by Thomas Friedman in the New York Times this weekend titled, "And Now for a Bit of Good News." The subject of the article was the new "sharing economy," think Uber, Lyft, Airbnb, etc. In the article, Friedman calls Airbnb a "Trust Platform." To me, this weekend, this term was a gift. He is so right. I have used Airbnb many times when traveling with my family, and to date have been very pleased with our experiences. Often the transactions are sizeable as I am reserving a home in a desirable location for a week. I am engaging in a transaction of some size with a person I don't know, in a home I have never visited, most often in a foreign country using different currency, involving different laws and customs. Why did I risk it? I trusted the platform.
A key development for the Internet of Things will be the evolution and emergence of the ‘Cloud Name System’, a directory system for Cloud applications in the same way DNS (Domain Name System) works for the web and email. Lori MacVittie wrote a while back about the need for an ‘SNS’ – a Service Name System, a DNS type directory approach but for Cloud Services so that they can be entirely loosely coupled from their IT infrastructure. Tim Berners-Lee, the inventor of the web, himself described a scenario of ‘Socially Aware Cloud Storage‘ that applies this same ideal of abstraction to our personal data across all the social networks we use. This refers to a distributed (Cloud) storage service that is used to store personal user data for social networks, rather than the social sites holding it themselves.
I like this word re-imagination from Mary Meeker’s Internet Trends presentation. We are seeing so many aspects of our life being transformed by the internet. Take for example, ordering a cab to go somewhere. Either we phone for a yellow cab here in California or if you are in New York city, then you stand and wave for an incoming yellow cab to stop. The new game-changer is Uber. All you do is touch your smartphone screen for UberX or Black car and you get an instant message about the car coming in less than 5 minutes time with the driver and car info. It is cheaper and you pay by card (pre-registered in your Uber account). This is re-imagining the transport sector. Uber, a San Francisco company is worth about $17B and is operating in 70 cities around the world. Quite a disruptive force!
When people talk about Big Data, the emphasis is usually on the Big. Certainly, Big Data applications are distributed largely because the size of the data on which computations are executed warrants more than a typical application can handle. But scaling the network that provides connectivity between Big Data nodes is not just about creating massive interconnects. In fact, the size of the network might be the least interesting aspect of scaling Big Data fabrics.
Do you avoid stores that have had a credit card breach? You are not alone. About 52% of people avoid merchants who have had a data breach according to a recent Lowcards survey. They surveyed over 400 random consumers to better understand the impact of identity theft on consumer behavior. 17% said they or a family member was a victim of identity theft over the last year with half the cases being credit card theft. 94% said they are more concerned or equally concerned about ID theft. They estimate that there were 13.5 million cases of credit card identity theft in the United States over the last 12 months.
If you've had to test one of today's composite applications, you know that "access" has become one of the most daunting barriers to SDLC acceleration. Whether we're talking about access to dev/test environments or access to dependent applications, the ability to pull all the required pieces together in order to test thoroughly is equivalent to herding feral cats. If you haven't experienced this fun firsthand, consider this: our recent research revealed that systems under test have an average of 30 dependencies, but team members have consistent access to only 6. The time available to access test environments is extremely limited (4 hour slots) and 30% of that limited time was consumed by configuration/setup tasks. Ultimately, testers had time to execute only 50% of the available test plan. Service virtualization is a revolutionary new technology that helps you break through these constraints by providing ubiquitous, global access to complete dev/test environments. Organizations leveraging service virtualization are able to conduct testing earlier, faster, more thoroughly, and more accurately—significantly reducing the risk of application failure. In case you're just starti...
Much has been published about the Open Compute Project. Initiated by Facebook, it has become an industry effort focused on standardization of many parts and components in the datacenter. Initially focused on racks, power and server design, it has also added storage and now networking to its fold. Its goal is fairly straightforward: “how can we design the most efficient compute infrastructure possible”, a direct quote from its web site. The focus of OCP has been mostly around hardware designs and specifications. If you look at the networking arm of OCP, you find several Top of Rack (ToR) ethernet switch hardware designs donated by the likes of Broadcom, Mellanox and Intel. By creating open specifications of hardware designs for fairly standard ethernet switches, the industry can standardize on these designs and economics of scale would drive down the cost to create and maintain this hardware. A noble goal and there are many opinions on both sides of this effort. Mostly referred to as “bare metal” and “commodity”, you can easily spend days reading up on many opinions. Mike Bushong yesterday discussed pricing implications for resellers in this blog post.
In a post Snowden world it is clear that for cloud data security, we need strong encryption. When properly implemented, encryption in the cloud reduces risk to levels acceptable for sensitive data. There is no doubt data protection in the cloud computing era is never going to be a ‘one size fits all’ kind of a solution. It requires a 360-degree view of the company with 365-days a year dedication. The best place to start is with a risk analysis so you know what kind of data you have, its levels of sensitivity, who’s using it, where it’s used and stored, and how and where and over what technologies it’s going to ‘commute’. You need to understand your company’s data – in terms of technology and human weaknesses. Data should be unreadable to an attacker. It must be incoherent at all times to anyone other than you and your trusted personnel: while it’s travelling – or you are; whether it’s in transit, storage, stopping, or resting, data is safest in encrypted form.
Some people are never satisfied. These fearless agents of change are everywhere. They're informed, confident and willing to experiment. They seek out the best business technology solution for the job at hand. They act on instinct. Yes, you could say that they're driven. However, they're also at risk of being labeled as "rogue employees" because they ordered a software-as-a-service (SaaS) offering and perhaps expensed it without prior approval. Sometimes they're the champion of progressive projects that are referred to as Shadow IT -- intentionally bypassing their company's formal evaluation and procurement process. How can this happen? Is it just because their activities are tolerated, or are they being encouraged? If so, by whom? Why would any business leader applaud a team member that breaks the rules? Maybe, the simple answer is that staying within the confines of the status-quo won't enable a top-performer to fully apply their talent, achieving their absolute best.
The GovCloud Media Network features agency specific video playlist for registered members. Please enjoy this feature on the Army IT. Please visit the new GovCloud Network Media Library for more video content. For membership information please send request to info@govcloudnetwork.com.
One of the primary principles of object-oriented programming (OOP) is encapsulation. Encapsulation is the way in which the state of an object is protected from manipulation in a way that is not consistent with the way the variable is intended to be manipulated. The variable (state) is made private, that is to say only the object itself can change it directly. Think of it as the difference between an automatic transmission and a standard (stick). With the latter, I can change gears whenever I see fit. The problem is that when I see fit may not be appropriate to the way in which the gears should be shifted. Which is how engines end up being redlined.
Cloud computing has finally come into its own. While we’ve been hearing for 8 years or more that cloud computing would one day take over the enterprise, the fact of the matter is that it’s been slow going. While the spread of cloud computing solutions hasn’t been as rapid as many early proponents predicted it would be, we are now to a place where cloud solutions are seen as viable for most organizations, and are being utilized regularly.
It's an application world; a world that is rapidly expanding. With new opportunities and markets arising driven by mobility and the Internet of Things, it is only going to keep expanding as applications are deployed to provision, license, and manage the growing sensors and devices in the hands of consumers. Applications are not isolated containers of functionality. No application winds up in production without a robust stack of resources and services to support it. Storage and compute, of course, are required, but so are the networking - both stateless and stateful - services that provide for scale, security and performance.
We're launching the "Tap In Open Forum": a new series of blog posts where we detail a current topic being discussed in the software development and testing community and invite you to participate in an open discussion with our best and brightest here at Skytap. Feel free to share your thoughts and we'll do the same! InformationWeek and IBM Mobile Technical Sales Leader Dustin Amrhein presented a webinar last week as part of their “DevOps Rado” series. The Frictionless Enterprise: Built for Business, looked at DevOps as a viable solution for removing the friction that prevents enterprises from innovating at not just the speed they would like, but at the speed absolutely required to stay relevant, and hopefully ahead of their competition.
Since the birth of Hadoop in 2005-06, the way we think about storing and processing information has evolved considerably. The term “Big Data” has become synonymous with this evolution. But still, many of our customers continue to ask, “What is Big Data?”, “What are its use cases?”, and “What is its business value?”. The Internet is overloaded with definitions, characteristics, and benefits; however, few discussions synthesize all three of these topics in one place. This paper answers these questions, and proposes a total cost calculation framework for CTOs and CIOs that are evaluating solutions for their organization’s use case(s). In the text below, I examine an on-premise Hadoop ecosystem as a general purpose Big Data solution in relation to alternative commercial purpose-built storage technologies (-e.g. Oracle, Teradata, IBM, SAP, Microsoft, EMC, etc). It may be difficult to determine the exact point at which you should leverage one over the other. It is my contention that when the total cost of using all your data exceeds what you are able to spend using purpose-built technologies, it is time to consider using a general purpose solution like Hadoop for process offloading....
Untitled Document
Past SYS-CON Events
    Cloud Expo West
cloudcomputingexpo
2011west.sys-con.com

 
    Cloud Expo East
cloudcexpo
2011east.sys-con.com

 
    Cloud Expo West
cloudcomputingexpo
2010west.sys-con.com

 
    Virtualization Expo West
virtualization
2010west.sys-con.com
    Cloud Expo Europe
cloudexpoeurope2010.
sys-con.com

 
    Cloud Expo East
cloudcomputingexpo
2010east.sys-con.com

 
    Virtualization Expo East
virtualizationconference
2010east.sys-con.com
    Cloud Expo West
cloudcomputingexpo
2009west.sys-con.com

 
    Virtualization Expo West
virtualizationconference
2009west.sys-con.com
    GovIT Expo
govitexpo.com
 
    Cloud Expo Europe
cloudexpoeurope2009.sys-con.com
 

Cloud Expo 2011 Allstar Conference Faculty

S.F.S.
Dell

Singer
NRO

Pereyra
Oracle

Ryan
OpSource

Butte
PwC

Leone
Oracle

Riley
AWS

Varia
AWS

Lye
Oracle

O'Connor
AppZero

Crandell
RightScale

Nucci
Dell Boomi

Hillier
CiRBA

Morrison
Layer 7 Tech

Robbins
NYT

Schwarz
Oracle

What The Enterprise IT World Says About Cloud Expo
 
"We had extremely positive feedback from both customers and prospects that attended the show and saw live demos of NaviSite's enterprise cloud based services."
  –William Toll
Sr. Director, Marketing & Strategic Alliances
Navisite
 


 
"More and better leads than ever expected! I have 4-6 follow ups personally."
  –Richard Wellner
Chief Scientist
Univa UD
 


 
"Good crowd, good questions. The event looked very successful."
  –Simon Crosby
CTO
Citrix Systems
 


 
"It's the largest cloud computing conference I've ever seen."
  –David Linthicum
CTO
Brick Group