Dell Community

Blog Group Posts
Application Performance Monitoring Blog Foglight APM 105
Blueprint for HPC - Blog Blueprint for High Performance Computing 0
Custom Solutions Engineering Blog Custom Solutions Engineering 9
Data Security Data Security 8
Dell Big Data - Blog Dell Big Data 68
Dell Cloud Blog Cloud 42
Dell Cloud OpenStack Solutions - Blog Dell Cloud OpenStack Solutions 0
Dell Lifecycle Controller Integration for SCVMM - Blog Dell Lifecycle Controller Integration for SCVMM 0
Dell Premier - Blog Dell Premier 3
Dell TechCenter TechCenter 1,861
Desktop Authority Desktop Authority 25
Featured Content - Blog Featured Content 0
Foglight for Databases Foglight for Databases 35
Foglight for Virtualization and Storage Management Virtualization Infrastructure Management 256
General HPC High Performance Computing 229
High Performance Computing - Blog High Performance Computing 35
Hotfixes vWorkspace 66
HPC Community Blogs High Performance Computing 27
HPC GPU Computing High Performance Computing 18
HPC Power and Cooling High Performance Computing 4
HPC Storage and File Systems High Performance Computing 21
Information Management Welcome to the Dell Software Information Management blog! Our top experts discuss big data, predictive analytics, database management, data replication, and more. Information Management 229
KACE Blog KACE 143
Life Sciences High Performance Computing 12
On Demand Services Dell On-Demand 3
Open Networking: The Whale that swallowed SDN TechCenter 0
Product Releases vWorkspace 13
Security - Blog Security 3
SharePoint for All SharePoint for All 388
Statistica Statistica 24
Systems Developed by and for Developers Dell Big Data 1
TechCenter News TechCenter Extras 47
The NFV Cloud Community Blog The NFV Cloud Community 0
Thought Leadership Service Provider Solutions 0
vWorkspace - Blog vWorkspace 512
Windows 10 IoT Enterprise (WIE10) - Blog Wyse Thin Clients running Windows 10 IoT Enterprise Windows 10 IoT Enterprise (WIE10) 6
Latest Blog Posts
  • SharePoint for All

    Bill Baer and Michael Lotter to keynote at TEC2012 - SharePoint in San Diego!

    By now I'm sure you've heard about The Experts Conferece - SharePoint...It includes three days of training (April 29-May 2). Just in case you've been too busy controlling your SharePoint Chaos, let me re-cap:

    1. Top SharePoint experts – MVPs, MCTs, MCMs and more!
    2. Incredible attendee-to-speaker ratio – Smaller sessions mean more contact with the experts, so you can be sure to get answers to all of your SharePoint administration questions!
    3. Not a “sales and marketing”-fest – There will be no product demos or advertising allowed in the sessions. You want to learn more about a vendor? You’ll have to visit the solutions lab.
    4. Never-before-presented content – Every presentation will be brand new and focused on three key themes: 1) SharePoint governance; 2) SharePoint’s future; 3) SharePoint business intelligence – featuring SQL Server 2012 “Denali”.
    5. Chris McNulty – TEC welcomes SharePoint expert and author Chris McNulty as the 2012 conference facilitator
    6. Beautiful San Diego! The Marriott Marquis and Marina is a GORGEOUS property. Take a look here>
    7. Red Carpet - Well, I truly don't know what color the carpet is, but as far as SharePoint events go, this conference pulls out all the stops. Receptions, cocktail parties, breakfast and lunch are included in the conference pass. Not to mention the conference staff - Quest's SharePoint team is on site to make sure you have a first class experience!

    And - the part I am most excited about - we have TWO (not just one!) amazing keynotes lined up!

    • Day One – “Always On: The Next Generation of High Availability and Disaster Recovery with SharePoint” with Bill Baer, MCM and senior product expert at Microsoft
    • Day Two Why the SharePoint Community Matters: The Future of SharePoint” with Michael Lotter,MCTS, SharePoint solutions architect at CFA Institute, and founder of SharePoint Saturday

    If you are faced with the pains and pleasures of SharePoint, you should NOT miss this conference.

    This is going to be the HOTTEST TEC event yet! If you'd like to find out more, you can visit or go to and interact directly with our experts!

    Register now!

    And if you plan to regsiter by April 6, please email me for a discount code to register for $1475 which is $200 off the current price!

  • Dell TechCenter

    Wrap Up: First EMEA Cloud Chat at Dell TechCenter

    Last week, we welcomed more than 40 people in our first EMEA Cloud Chat from various geographies, amongst them Dell engineers from our US offices as well as sales and marketing people, trainers and customers from Germany, France, Italy, Spain and Sweden. Thank you all for taking time and engaging with us!

    As announced earlier, we will follow up with further cloud chats, diving deeply into specific solution areas. We will host upcoming cloud chats every third Thursday of a month, 3.00 pm CET. Our next chat is scheduled for April 19th and we will talk about Dell’s OpenStack distribution and Dell Crowbar,the leading open source installation tool for OpenStack and Hadoop. A pre-chat blog post is already in the work, providing a detailed outlook. Stay tuned!

    In case you couldn’t attend (or if you’d like to learn more about the solution areas discussed in the chat) please find below a summarized transcript. We cleaned it up and reordered some fragments, making the transcript more digestible (the entire transcript is almost 20 pages long):

    Dell Openstack & Crowbar

    Stephen Spector: As a starting point, we want to highlight the announcement yesterday (March 21st 2012) on the Openstack (Crowbar) coming to Europe and Asia. The Dell team has a complete Openstack "distro" available globally and the Crowbar project allows you to leverage the concept of Devops to install a complete install of Openstack on bare metal. Dell’s Barton Geore has a nice blog post but we have Rob Hirschfeld here with us. Rob Hirschfeld is the main architect behind Crowbar and the technical lead on the Dell Openstack distro. Rob, can you provide a bigger overview of Crowbar to help people understand how that helps Openstack?

    Rob Hirschfeld: Sure. Crowbar is an open source project that we started to make sure that it was fast and repeatable to install Openstack (and other cloud software). It was very important to us that we could get to production. What we did was start from the Openstack Chef scripts. So Crowbar uses Chef as a foundation, but we needed to have hardware deploy & orchestration. We've been getting some great interaction from EMEA community members on the Crowbar list. So, we've been seeing activity about both Diablo and Essex (recently added).

    Stephen Spector: Is Dell releasing Essex soon or are we staying with Diablo?

    Rob Hirschfeld: Our Diablo distro include Dashboard & Keystone. The current release is Dialbo because that's what's released. But we felt like it was critical to include Keystone & Dashboard in that.

    Stephen Spector: When is Essex set to release?

    Rob Hirschfeld: We don't have an official date yet - we're evaluating it and need to see the final bits.

    Stephen Spector: Can you provide some links on where people can get more info on Crowbar and Openstack from Dell?

    Rob Hirschfeld: and Crowbar is open source, so the best way to start learning it in detail is from the github wiki:

    Dell Boomi

    Stephen Spector: Please welcome Ralph Hibbs, Marketing Director from the Dell Boomi Application Integration solution team. Dell Boomi is an innovative cloud solution allowing Enterprise customers to leverage the latest in SaaS solutions while continuing to leverage their existing enterprise software investments. Ralph, can you tell us more about Dell Boomi?

    Ralph Hibbs: Dell Boomi is an application integration platform that resides in the cloud. It can connect any combination of applications, regardless of where they reside: public cloud, private cloud or on-premise.

    Stephen Spector: Can you provide an example of app integration?

    Ralph Hibbs: Earlier this week, we announced that we are working with several EMEA customers in helping the Oneworld Airline Alliance deploy an IT hub in the cloud to exchange frequent flyer information. This has us working with such EMEA companies as Air Berlin and British Airways. The IT hub between airlines is one example of application integration. Another example, inside a corporation would be connecting to a financial application, such as SAP or Oracle to exchange information about closed orders and invoicing.

    Stephen Spector: I know there was a blog post on that, do you have the link for everyone?

    Ralph Hibbs: Here's a link to the Oneworld press release.

    DVS Simplified Appliance

    Stephen Spector: Please welcome Brent Doncaster, Product Marketing for virtualization solutions. Can you provide an intro to the newly released Dell solution for Virtualized Desktops?

    Brent Doncaster: Dell recently announced a new addition to our DVS portfolio of desktop virtualization solutions - DVS Simplified, an integrated VDI appliance is now available in the USA, and it will available in EMEA in 2 weeks. DVS Simplified Appliance is in addition to our enterprise solution and our Dell Cloud based "as-a-service" offer for virtualizing desktop environments.

    Stephen Spector: Is the DVS Simplified Appliance based on Citrix or another technology?

    Brent Doncaster: DVS Simplified Appliance integrates and factory installs Citrix VDI-in-a-box software with a factory pre-configured Dell sever. It’s an "appliance" is the easy to think of it - we have done all the mechanical software installation and pre-configuration. The cool thing about DVS Simplified Appliance - from out of the box to being up and running can take as little as an afternoon.

    Dell Public VMware vCloud

    Stephen Spector: Please welcome Adam Dawson, Program Manager on Dell Security Solutions, to discuss the new security features added to Dell vCloud. Can you briefly talk about the Dell - Trend Micro relationship for our public VMware vCloud?

    Adam Dawson: This is a really cool solution. Security is every CIO's primary concern in moving to the cloud. So we built a relationship with Trend Micro to offer cloud-based encryption key management for your Dell vCloud infrastructure data. You can encrypt the data in Dell's cloud, and you will be the only one with access to it. Dell won't have access to the keys or the data, but you only pay for the keys you use and the billing will be integrated into your monthly vCloud bill.

    Stephen Spector: Where are the keys stored?

    Adam Dawson: The encryption keys are stored in Trend Micro's data center in Germany. The VM with encrypted data requests the key at startup and if the VM passes integrity checks, the key is dispatched to the VM which allows access to the encrypted data.

    Ralph Hibbs: How is the Dell & Trend Micro solution delivered? A cloud service, appliance?

    Adam Dawson: The Trend Micro service is delivered as a SaaS solution in the cloud. You can order from Dell, and Dell will provision your account through Trend Micro's software portal online.

    Stephen Spector: What gets encrypted on the VM? Just data or the VM itself?

    Adam Dawson: You encrypt individual data volumes (drive letters in Windows, mount points in Linux). So you can pick and choose to encrypt only sensitive data.

    Stephen Spector: Quick note on Dell's Public Cloud - vCloud Datacenter Service currently available in the US. We are working very hard to bring to the European market (stay tuned for more info on this).

    Florian Klaffenbach: The work on the German cloud data center will start next week.

    Stephen Spector: Please welcome Matt Domsch, Solutions Architect in the Office of the CTO, to discuss Dell’s new Disaster Recovery and Backup Solutions.  I want to ask you Matt about Disaster Recovery and how public clouds are changing the way customers view this important deliverable.

    Matt Domsch: A lot of companies I get to speak with daily have DR plans. They've been building second DR data centers and watching their costs skyrocket. Cloud gives companies a second, often less expensive and less hassle method to do DR.

    Stephen Spector: Are companies using public clouds for the backup?

    Matt Domsch: The concept is fairly simple - you create standby application instances in a cloud, and you regularly back up your critical data to the Dr site storage. Having both storage and ability to run VMs in a remote public cloud, gives you the ability to "flip over" fairly quickly. Public cloud for backup is popular, at a consumer level, and increasingly, at a business level.

    The great thing about backups is they are usually "write once, read never", which make them the perfect thing to place in a cloud, where performance variations day-to-day don't really impact your running business. For customers, Dell is working actively on new backup and disaster recovery solutions for the public cloud so stay tuned for more details on those solutions.

    Stephen Spector: Customers - as you can see Dell has a great deal of Cloud solutions for application integration (Dell  Boomi), public and private clouds both open source and VMware based, security (dell Secureworks & partnership with Trend Micro), VDI solutions that are the simplest installs in the marketplace, as well as special cloud hardware to meet customer demands. Feel free to visit for more info.

    Stephen Spector: I have a good blog post on the various Dell Cloud solutions that I wrote last week. This blog post has all the Dell Cloud solutions and what they can do. The trick is to meet the customer's needs!

    Thank you to everyone for attending the 1st Emea Cloud Chat today and I am pleased that we were able to have a variety of Dell cloud experts provide some basic info on the various cloud solutions that we offer. At future chats, we will go in more detail on those solutions to offer you more insight into the products and how they can drive your business forward.

    Chat Hosts

    Stephen Spector, Cloud Evangelist at Dell (Twitter: @SpectorAtDell)
    Florian Klaffenbach, Community Technologist at Dell (Twitter: @FloKlaffenbach)
    Rafael Knuth, Social Media Manager at Dell (Twitter: @RafaelKnuth)

  • Dell TechCenter

    Red Hat Enterprise Linux 5.8 Support on Dell PowerEdge servers

    Posted on behalf of Gireesha US from Dell Linux Engineering Team.

    Red Hat has announced the availability of Red Hat Enterprise Linux (RHEL) 5.8 Operating System on Feb 2012, which includes driver updates, bug fixes and support for new hardware. For detailed information, please refer to RHEL 5.8 release notes

    Red Hat Enterprise Linux 5.8 has been extensively validated on supported Dell PowerEdge Servers. Please refer to the important information guide for a complete list of issues fixed in RHEL 5.8 on Dell PowerEdge servers.

    What’s new :

    Listed are the details of new features including hardware enablement.

    • Support of Catapult-II based CNA/network controllers. Currently shipping servers support Brocade Catapult based controllers - BR1020 and BR1741M-k. The bfa/bna drivers has been upgraded to include the support for new hardware and Bug fixes.
    • Native support for Intel Spring Fountain II (Intel 82599 Dual Port 10Gb KR Mezz) 10G controllers. These controllers are shipping with Dell 12th generation servers.
    • Native support for Broadcom BCM 578xx (Broadcom 57800S/57810S) 10Gb CNA controllers and these are offered with Dell 12th generation platforms.

    Driver updates for the storage & network controllers, these drivers has been upgraded to include the Bug fixes.

    • Updated Intel igb, igbvf, ixgbe and ixgbevf, Driver for Intel 1G and 10GB Network cards.
    • The igb .ko version 3.0.6-k2-2 in Red Hat Enterprise Linux 5.8 drives the Intel i350 1GB controller.
    • The ixgbe driver  version 3.4.8-k  supports  the Intel x540  10Gb controller which is an  NDC option  the Rack servers , Intel x520 10Gb controller which is an NDC option on the M620 and the Intel add-on 10Gb controller.
    • Updated Broadcom tg3, bnx2, bnx2x and  bnx2i drivers.
    • The BCM5720 and BCM5719 are supported by the tg3 driver version 3.119
    • BCM 57810 and 57800 which are the Broadcom 10Gb iscsi capable 10Gb controllers are supported by the bnx2x version 1.70.00-0
    • Updated the mpt2sas (H200) and megaraid_sas (PERC H310, H710, H710P, H810 and PERC H700, H800, PERC6/ir storage controller) driver to upstream version containing bug fixes and performance enhancements to support on 11th and 12th generation servers.


  • Foglight for Virtualization and Storage Management

    Thin Provisioning

    I get asked by a lot of admins about thin provisioning and how that data can be easily understood. Admins are looking to understand the risk in over provisioning datastores, because if you run out of space you can cause outages for several VMs that no longer can read and write to a disk. Unfortunately, there is no metric for a datastore that can show that, but this is where a derived metric can really help. With vFoglight, I can create a metric based on other metrics or anything really. So, in the case of datastores, I want to take all of the VMDKs, thin or thick, add up their committed/uncommitted space and create a datastore version of that. With this new set of metrics, as well as the existing VM level metrics, we can craft some really simple to understand, but powerful views.

    Well, I've taken the liberty of creating a sample Thin Provisioning cartridge you can use to add these derived metrics to your environment and it also has several nice views around thin provisioning in general. You can check it out here:

    Example views:

    Drag out the View in Thin_Provisioning and it will give you a list of all VMs that currently have thin provisioned disks.

    Drag out a datastore to get either the Graph & Table or just the Table. This shows the total allocation that is currently in use and data that could be written (think over provisioning)

    Drag out all datastores to see all Datastores thin provisioning information

  • HPC Storage and File Systems

    Dell HPC NFS Storage Solution with High Availability -- Large Capacity Configuration

    We have posted two blogs (1) (2) to discuss DELL NFS Storage Solution with High Availability (NSS-HA) in the past. This article introduces a new configuration of DELL NSS-HA solution which is able to support larger storage capacities (> 100 TB) compared to the previous configurations of NSS-HA.

    Dell gets the support from Red Hat for XFS capacities greater than a 100 Terabytes . Details on the work that was done to get this exception for Dell is in the blog post: Dell support for XFS greater than 100 TB.


    As the design principles and goals for this configuration remain the same as previous Dell NSS-HA configurations, we will only describe the difference between this configuration and the previous configurations. For complete details, please refer to our white papers titled “Dell HPC NFS Storage Solution High Availability Configurations, Version 1.1.” and “Dell HPC NFS Storage Solution High Availability Configurations with Large Capacities, Version 2.1.”

    Storage density

    In previous configurations of NSS-HA (1), each storage enclosure was equipped with 12 3.5” 2TB NL-SAS disk drives. The larger capacity 3TB disk drives are a new component in the current configuration. The storage arrays in the solution, Dell PowerVault MD3200 and PowerVault MD1200 expansion arrays are the same as in the previous version of the solution but with updated firmware. The higher capacity 3TB disks now allow higher storage densities in the same rack space. Table 1 provides information on new capacity configurations possible with the 3TB drives. This table is not a complete list of options; intermediate capacities are available as well.

    Storage configuration

    In previous configurations of NSS-HA, the file system had a maximum of four virtual disks. A Linux physical volume was created on each virtual disk. The physical volumes (PV) were grouped together into a Linux volume group and a Linux logical volume was created on the volume group. The XFS file system was created on this logical volume.

    With this configuration, if more than four virtual disks are deployed, the Linux logical volume (LV) is extended, in groups of four, to include the additional PVs. In other words, groups of four virtual disks are concatenated together to create the file system. Data is striped across each set of four virtual disks. However it is possible to create users and directories such that different data streams go to different parts of the array and thus ensure that the entire storage array is utilized at the same time. The configuration is shown in Figure 1 for a 144TB configuration and a 288TB configuration.

    Red Hat High Availability Add-On

    Red Hat High Availability Add-On is a key component for constructing a HA cluster. In previous configurations of NSS-HA, the add-on used is distributed with RHEL 5.5. With this release, the version distributed with RHEL6.1 is adopted. There are significant changes in the HA design between the previous RHEL 5.5 release and the new RHEL 6.1 release. New and updated instructions to configure the HA cluster with RHEL 6.1 are listed in the appendix A of our white paper “Dell HPC NFS Storage Solution High Availability Configurations with Large Capacities, Version 2.1.

    Red Hat Scalable File System package

    In previous configurations of NSS-HA, the version of XFS is 2.10.2-7 which is distributed with RHEL 5.5. In the current version of NSS-HA, the version of XFS used is 3.1.1-4 and is distributed with RHEL 6.1. The most important feature of the current XFS for users is that it is able to support greater than a 100 Terabytes of storage capacity.

    Summary of changes

    Table 2 lists the similarities and difference in storage components. Table 3 lists the similarities and differences in the NFS servers.

    Dell NSS-HA solution provides a high availability and high performance storage service to high performance computing clusters via an InfiniBand or 10Gigabit Ethernet network. Performance characterization of this version on the solution is described in “Dell HPC NFS Storage Solution High Availability Configurations with Large Capacities, Version 2.1.” Additionally, in our next few blogs, we will discuss the performance of the random and metadata tests on 10GbE and other performance related topics.

    By Xin Chen and Garima Kochhar 


    1.  Dell NFS Storage Solution with High Availability - an overview

    2.  Dell NFS Storage Solution with High Availability – XL configuration

    3.  Red Hat Enterprise Linux 6 Cluster Administration -- Configuring and Managing the High Availability Add-On.

    4.  Dell HPC NFS Storage Solution High Availability Configurations, Version 1.1

  • HPC Storage and File Systems

    Dell support for XFS greater than 100 TB

    Enterprise storage needs demand solutions that can scale up and scale out in terms of capacity and performance. This is especially true in HPC environments where the additional constraint of cost is paramount. Dell has responded with cost effective solutions for the HPC storage needs in three different spaces (as illustrated in the figure below):

    In particular, NSS is a very cost-efficient solution that can deliver high performance at moderate capacities. However, previous versions of the NSS were restricted to a ceiling of 100 Terabytes (100 * 2^40 or 100 TiB) due to the Red Hat support limit for XFS.

    Given industry current demands for storage solutions with bigger capacity in this space, Dell has been working extensively with Red Hat on XFS testing and validation to expand the support beyond the 100 TiB barrier and meet the current business needs of our customers.

    As a result of this effort, Red Hat has granted Dell support for XFS up to 288 TB (raw disk space) on NFS Storage Solutions with a single namespace, and even bigger capacities on custom design solutions. This is a very important milestone for Dell’s quest towards providing Petabyte storage solutions.

    For details about the performance characteristics and different capacities of our new version of the NSS, please take a look at the NSS white paper.

    Written by Garima Kochhar, Jose Mario Gallegos, and Xin Chen

  • Dell TechCenter

    Managing Enterprise Quality: What are we doing for our next generation products?

    Managing Enterprise Quality: What are we doing for our next generation products?

    You already know the mission of the Enterprise Customer Loyalty Team: It’s about improving the health of our customer relationships and driving actionable customer insights into our product development cycles. Today I want to share a little more context about the larger organization we represent –- Enterprise Quality -- what our Quality Engineers do to manage quality in general, and what we’re doing specifically to support a seamless transition to our next generation server products.

    Enterprise Quality Vision

    Our CULTURE of Quality incorporates the customer in all we do. Built on a foundation of solid quality management, it started with a basic hardware focus followed closely by software. Now our Quality JOURNEY has matured to providing solutions being crucial to everyone’s success. This top-down driven transformation was evident at our recent launch of new Enterprise Products and Solutions, where Michael Dell talked about our continued transformation and evolution from a PC manufacturer to a company providing end-to-end IT solutions.

    For this launch of next generation products, our Quality goal was to provide our customers a seamless transition. This required our support from early product development through post-launch:

    Early product development:

    • Lessons Learned reviewed to ensure that issues/failures identified in previous generation products have been properly addressed to improve our new product offerings
    • NUDDs (New, Unique, Different, Difficult) identified to highlight focus areas throughout development of the product, including enhancements or changes in product features, materials or usability
    • Goal planning and focus on best-ever time-to-quality  or response and resolution times for potential product issue

    Throughout process:

    • Lots of reviews of protos, mechanical samples, raw material, documents, guides, plans, readiness to ensure design and build quality

    At Launch:

    • Rigorous pre-launch review process (e.g., defect reviews, approvals in changes to baseline documents or process assets) to ensure the customer voice is analyzed and prioritized by Quality Engineers on every issue that arises in the days and weeks before launch
    • Pilot Build in-factory ICE (Initial Customer Experienced) audits over several weeks across all systems to audit final build process

    Post Launch:

    • Rapid Response Team formed from day one, including all stakeholders (Core Team, Quality, IPS, Marketing, Operations) to resolve any emerging issues
    • Daily review of customer calls and dispatches as well as multiple COSIPs (Customer On-Site Inspection Programs) planned & executed to discover potential product quality issues
    • Consistent and committed engagement across three global design centers on every issue and potential issue                  

    We’ve already received valuable, timely feedback from customers who are rolling out 12th Generation installations.

     So tell us what you think! Are we missing anything? Are there other activities besides what we have described?  Are you hearing any buzz about our 12th Generation products as they enter the market?

    We’re working hard to improve your customer experience, and we want to hear from YOU.

  • TechCenter News

    Dell TechCenter Tech Tuesday Chat: Open Mic and New Chat Platform Evaluation

    Please join us today for an "Open Mic" chat with the Dell TechCenter crew at 3pm Central.  We'd also like your help in evaluating a new chat platform for the community.  You can login as a guest at the link below.

    Join us here:

    Calendar event here:

    To unsubscribe from the Dell TechCenter distribution, please click "Email unsubscribe to this blog" on the TechCenter News Blog page or email "unsubscribe" to

  • Dell TechCenter

    Deploying Dell Boomi in the Cloud

    As I mentioned a few weeks back in “Digging Deep into Dell Boomi” I planned to come back and write about the various topologies available to deploy Dell Boomi Atoms, or integration run-times.  For more details on the Dell Boomi solution you can read my initial blog post or visit the website

    One of the secret sauces to Dell Boomi is how integration execution is managed. Once your integration processes have been designed and tested in the visual designer, they are loaded into a lightweight, dynamic run-time engine called an “Atom” for execution.  The Atom is a pretty cool technology element—we’ve even applied for a patent on it.

    The cool thing about Atoms is they can be deployed anywhere: the AtomSphere (Boomi’s Atom Cloud), another public cloud (such as Amazon Web Services or Dell’s public vCloud), or safely behind your firewall for on-premise application integration.

    The Atoms control the actual movement of data. If Atoms are deployed behind the firewall, no data flows through the Dell Boomi Atom Cloud. It is extremely secure and requires no holes in your firewall. The Atom stays connected with Dell Boomi for integration updates, Atom updates and sending status to the centralized management console.

    The Dell Boomi Atoms support the following deployment scenarios:

    • Cloud Deployment
    • Third Party Datacenter
    • Behind the Firewall Deployment

    Cloud Deployment

    Full Cloud Deployment

    In this scenario, the Atom is located in Dell Boomi Atom Cloud environment managed and operated by Dell. All data is sent to and from one SaaS application, such as NetSuite to another SaaS application, such as through the Dell Boomi Atom located in the Atom Cloud (see image below).

    Cloud Deployment of Both Cloud and On-Premise Apps

    In this scenario, the customer has a public SaaS solution outside their firewall but also an on-premise legacy application such as SAP or Oracle which requires the data to flow between the Dell Boomi Atom and the two applications; one in a public environment and one behind the firewall.  An image of this type of solution is shown below in the Third Party Datacenter Deployment except the Dell Boomi Atom is hosted and managed by Dell.

    Third Party Datacenter Deployment

    In this scenario, the Dell Boomi Atom is hosted by a third party vendor on a public cloud such as Dell vCloud or in a datacenter managed by a vendor other than Dell. This option requires the Dell Boomi customer to setup and configure their Atom on the third party hardware but it operates similar to the way that the Dell Boomi SaaS setup is shown above.

    Behind the Firewall Deployment

    In this model, the Dell Boomi Atom is deployed within the corporate datacenter providing an on-premise data movement for the customer. The image below demonstrates this solution:


    The Dell Boomi Atom is placed in the datacenter “between” the Oracle and SAP applications where data is transferred following the rules setup before deployment for Oracle/SAP to Salesforce conversion.  Dell does monitor the integration process and log files & status notifications are sent to the Dell Boomi cloud for tracking purposes only. If Atoms are deployed behind the firewall, no data flows through the Dell Boomi Atom Cloud. It is extremely secure and requires no holes in your firewall. The Atom stays connected with Dell Boomi for integration updates, Atom updates and sending status to the centralized management console.

    But wait…there more

    Did I say Atoms are cool.  They can be combined together to form…wait a minute…remember chemistry and you’ll guess it. Yes, Atoms can be combined into Molecules for load balancing, High availability, Disaster recovery and parallel processing.

    To learn more about this unique cloud technology from Dell visit or call -800-732-3602 for sales.


  • High Performance Computing - Blog

    High Performance Computing (HPC) Speeds Scientific Research for Top Chinese Universities

    It's exciting to see Dell HPC making a difference in scientific research across the world. Recently the top publication in China focused on education, called China Informationization Education, posted an article about the role high performance computing (HPC) is playing in advancing research at some of China's top research institutions. Focused on helping to improve the efficiency of scientific research, with an emphasis on biology, HPC is having a large impact.
    The article features Tsinghua University and also cites information provided by Beijing Normal University, and the University of Science and Technology of China. Below is a link to the article that has been translated into English as well as the original version in Chinese.

    High Performance Computing (HPC) Speeds Scientific Research for Top Universities

    As always, we welcome your feedback and comments.

    Christine Fronczak