Recent Blog Posts

Make Your Data Centre Think for Itself

Two-People-Walking-In-A-Data-Center.pngWouldn’t it be nice if your data centre could think for itself and save you some headaches? In my last post, I outlined the principle of the orchestration layer in the software-defined infrastructure (SDI), and how it’s like the brain controlling your data centre organism. Today, I’m digging into this idea in a bit more detail, looking at the neurons that pass information from the hands and feet to the brain, as it were. In data centre terms, this means the telemetry that connects your resources to the orchestration layer.

 

Even the most carefully designed orchestration layer will only be effective if it can get constant and up-to-date, contextual information about the resources it is controlling: How are they performing? How much power are they using? What are their utilisation levels and are there any bottlenecks due to latency issues? And so on and so forth. Telemetry provides this real-time visibility by tracking resources’ physical attributes and sending the intelligence back to the orchestration software.

 

Let me give you an example of this in practice. I call it the ‘noisy neighbour’ scenario. Imagine we have four virtual machines (VMs) running on one server, but one of them is hogging a lot of the resource and this is impacting the performance of the other three. Intel’s cache monitoring telemetry on the server can report this right back to the orchestration layer, which will then migrate the noisy VM to a new server, leaving the others in peace. This is real-time situational feedback informing how the whole organism works. In other words, it’s the Watch, Decide, Act, Learn cycle that I described in my previous blog post – doesn’t it all fit together nicely?

 

Lessons from History…

 

Of course, successful telemetry relies on having the right hardware to transmit it. Just think about another data centre game changer of the recent past – virtualisation. Back in the early 2000s, demand for this technology was growing fast, but the software-only solutions available put tremendous overheard demand on the hardware behind them – not an efficient way to go about it. So, we at Intel helped build in more efficiencies with solutions like Intel® Virtualization Technology, more memory, addressability and huge performance gains. Today, we’re applying that same logic to remove SDI bottlenecks. Another example is Intel® Intelligent Power Node Manager, a hardware engine that works with management software to monitor and control power usage at the server, rack and row level, allowing you to set the usage policies for each.

However, we’re not just adding telemetry capabilities at the chip level and boosting hardware performance, but also investing in high-bandwidth networking and storage technologies.

 

….Applied to Today’s Data Centres

 

With technologies already in the market to enable telemetry within the SDI, there are a number of real-life use cases we can look to for examples of how it can help drive time, cost and labour out of the data centre. Here are some examples of how end-user organizations are using Intelligent Power Node Manager to do this:

 

 

Orchestration-Layer-Explanation-For-Data-Center.png

Another potential use case for the technology is to reduce usage of intelligent power strips, or you could throttle back on server performance and extend the life of your uninterruptable power supply (UPS) in the advent of a power outage, helping lower the risk of a service down time – something no business can afford.

 

So, once you’ve got your data centre functioning like a highly evolved neural network, what’s next? Well, as data centre technologies continue to develop, the extent to which you can build agility into your infrastructure is growing all the time. In my next blog, I’m going to look into the future a bit and explore how silicon photonics can help you create composable architectures that will enable you to build and reconfigure resources on the fly.

 

To pass the time until then, I’d love to hear from any of you that have already started using telemetry to inform your orchestration layer. What impact has it had for you, and can you share any tips for those just starting out?

 

Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations, and functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products. For more information go to http://www.intel.com/performance

 

You can find my first and second blogs on data centres here:   

Is Your Data Centre Ready for the IoT Age?

Have You Got Your Blueprints Ready?

Are You Smarter than a Data Centre?

 

To continue the conversation on Twitter, please follow us at @IntelITCenter or use #ITCenter.

 

*Other names and brands may be claimed as the property of others.

Read more >

Server Refresh Can Reduce Total Cost of Ownership

Snackable-MoreBangforYourBuck.pngMore bang for your buck. Essentially that is the driving force behind my team in Intel IT. Our IT department is on a tight budget, just like most enterprise IT departments. Therefore, return on investment and total cost of ownership are important considerations for deciding when to upgrade the servers that run our silicon design workloads. As a principal engineer in infrastructure engineering, I direct the comparison of the various models of each new generation of Intel® CPU to those of previous generations of processors. (We may re-evaluate the TCO of particular models between generations, if price points significantly change.) We evaluate all the Intel® Xeon® processor families – Intel Xeon processor E3 family, Intel Xeon processor E5 family, and Intel Xeon processor E7 family – each of which have different niches in Intel’s silicon design efforts.

 

We use industry benchmarks and actual electronic design automation (EDA) workloads in our evaluations, which go beyond performance to address TCO – we include throughput, form factor (density), energy efficiency, cost, software licensing costs, and other factors. In many cases over the years, one of the models might turn out better in terms of price/watt, but performance is slower, or the software licensing fees are triple those for a different model.

 

In silicon design, back-end design jobs are time critical and require servers with considerable processing power, large memory capacity, and memory bandwidth. For these types of jobs, the bottleneck has historically been memory, not CPU cycles; with more memory, we can run more jobs in parallel. The Intel Xeon processor E7-8800 v3 product family offers new features that can increase EDA throughput, including up to 20% more cores than the previous generation and DDR4 memory support for higher memory bandwidth. A server based on the Intel Xeon processor E7-8800 v3 can take either DDR3 (thereby protecting existing investment) or DDR4 DIMMs – and supports memory capacity up to 6 TB per 4-socket server (with 64 GB DIMMs) to deliver fast turnaround time for large silicon design jobs.

 

We recently completed an evaluation of the Intel Xeon processor E7-8800 v3 product family, as documented in our recent brief. According to our test results, the Intel Xeon processor E7 v3-based server delivers excellent gains in performance and supports larger models, faster iterations, and greater throughput than was possible with the previous generation of the processor. These improvements can accelerate long-running silicon design jobs and shorten the time required to bring new silicon design to market. These improvements can also reduce data center footprint and help control operational and software licensing costs by achieving greater throughput using fewer systems than were necessary with previous generations of processors.

 

Our tests used a large multi-threaded EDA application operating on current Intel® silicon design data sets. The result shows an Intel Xeon processor E7-8890 v3-based server completed a complex silicon design workload 1.18x faster than the previous-generation Intel Xeon processor E7-4890 v2-based server and 17.04x faster than a server based on Intel® Xeon® processor 7100 series (Intel Xeon processor 7140M).

 

The Intel Xeon processor E7-8800 v3 product family also supports the Intel® Advanced Vector Extensions 2 (Intel® AVX2) instruction set. Benefits of Intel AVX2 include doubling the number of FLOPS (floating-point operations per second) per clock cycle, 256-bit integer instructions, floating-point fused multiply-add instructions, and gather operations. While our silicon design jobs do not currently use AVX2 – mostly because the design cycles can take over a year to complete and during that time we cannot modify the plan of record (POR) EDA tools and infrastructure for those servers – we anticipate that Intel AVX2 can provide a performance boost for many technical applications.

 

I’d like to hear from other IT professionals – are you considering refreshing? If you’ve already refreshed, can you share your observed benefits and concrete ROI? What best-known methods have you developed and what are some remaining pain points? If you have any questions, I’d be happy to answer them and pass on our own best practices in deploying these servers. Please share your thoughts and insights with me – and your other IT colleagues – by leaving a comment below. Join our conversation here in the IT Peer Network.

Read more >

Chromium Command-Line Options for Crosswalk Builds with the Intel XDK

If you are an HTML5 web app developer, something you probably take for granted with that Chrome browser on your desktop are the handy Chromium Command-Line Switches that can be used to enable experimental features in the browser or control useful debugging and … Read more >

The post Chromium Command-Line Options for Crosswalk Builds with the Intel XDK appeared first on Intel Software and Services.

Read more >

Looking to the Future: Smart Healthcare with Big Data – HLRS

HLRS-Healthcare-Techonolgy.png

The NHS is under unbelievable pressure to do more with less; according to the latest research from the Kings Fund, the NHS budget increased by an average of just 0.8 per cent per year in real terms since 2010.

 

Clearly, there is a need for smart ways to improve healthcare around the world. One team of researchers at The University of Stuttgart is using cutting-edge technology and big data to simulate the longest and strongest bone in the human body — the femur — to improve implants.

 

People-Looking-At-Images-On-Screens.png

Medical Marvels

 

For three years, this research team has been running simulations of the types of realistic forces that the thigh bone undergoes on a daily basis for different body types and at a variety of activity levels to try and inform doctors what is needed to create much better implants for patients with severe fractures or hip deterioration. Successful implants can make a significant impact on the wearer’s quality of life, so the lighter and more durable they are the better.

 

Femoral fractures are pretty common, with around 70,000 hip fractures taking place in the UK each year, which is estimated to cost up to £2 billion. So, ascertaining better materials for bone implants that can allow longer wear and better mobility will solve a real need.

 

However, trying to achieve simulation results for a fractured bone requires a huge amount of data. Bone is not a compact structure, but is like a calcified sponge. Such a non-homogenous non-uniform material behaves in different ways under different stresses for different people. This means that the team must collect hundreds of thousands of infinitesimally small scans from genuine bone samples to learn how different femurs are structured. The incredible detail and high resolution provided by high-performance machines powered by the Intel® Xeon® Processor E5-2680 v3 enables them to replicate full femur simulations with this exact material data.

 

Such a level of intricacy cannot be done on a normal cluster. In the University of Stuttgart research team’s experience, one tiny part of the femoral head — a cube of only 0.6mm2 — generates approximately 90,000 samples and each of these samples requires at least six Finite-Element simulations to get the field of anisotropic material data needed to cover the full femoral head. To carry out this large number of simulations they definitely need the super computer! To do this in a commercial way you’d need thousands of patients, but with one supercomputer this team can predict average bone remodelling and develop reliable material models for accurate implant simulations. This is real innovation.

 

High-Performance Computing for Healthcare

 

The High Performance Computing Center in Stuttgart (HLRS) is the institution that makes all this possible. One of just three large supercomputer ‘tier 0’ sites in Germany, it recently upgraded to the Intel Xeon Processor E5-2680 v3, which according to internal tests delivers four times the performance of its previous supercomputer ii. This is great for Stuttgart University, as its computing center now has four times the storage space. ii Research like this requires intensive data processing and accurate analysis, so significant computing capacity is crucial.

 

Person-Using-Interactive-Computing.png

This new system enables breakthroughs that would be otherwise impossible. For more on HLRS’s cutting edge supercomputing offering, click here.

 

I’d love to get your thoughts on the healthcare innovation enabled by the Intel HPC technologies being used by HLRS and its researchers, so please leave a comment below — I’m happy to answer questions you may have too.

 

To continue this conversation on Twitter, please follow me at @Jane_Cooltech.

 

Join the debate in the Intel Health and Life Sciences Community, and look for Thomas Kellerer’s HLRS blog next week!

 

For other blogs on how Intel technology is used in the healthcare space, check out these blogs.

 

i ’Storage and Indexing of Fine Grain, Large Scale Data Sets’ by Ralf Schneider, in Michael M. Resch et. al, Sustained Simulation Performance 2013, Springer International Publishing, 2013, S. 89–104, isbn: 978-3-319-01438-8. doi: 10.1007/978-3-319-01439-5_7

 

Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations, and functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products. For more information go to http://www.intel.com/performance

 

*Other names and brands may be claimed as the property of others.

Photo Credit: copyright Boris Lehner for HLRS

Read more >

Evaluating an OS update? Take a look at eDrive….

As users of Windows 7 consider moving to Windows 8.1 or Windows 10, a new BitLocker feature is available that should be considered.  Nicknamed “eDrive,” “Encrypted Hard Drive,” or “Encrypted Drive,” the feature provides the ability for BitLocker to take advantage of the hardware encryption capabilities of compatible drives, instead of using software encryption.   Hardware encryption provides benefits over software encryption in that encryption activation is near-immediate, and real-time performance isn’t impacted.

 

eDrive is Microsoft’s implementation of managed hardware-based encryption built on the TCG Opal framework and IEEE-1667 protocols.  It is implemented a bit differently from how third-party Independent Software Vendors (ISVs) implement and manage Opal-compatible drives.  It is important to understand the differences as you evaluate your data protection strategy and solution.

 

eDrive information on the internet is relatively sparse currently.  Here are a couple of resources from Intel that will help get you started:

 

And here are a couple of tools from Intel that will be useful when working with the Intel® SSD Pro 2500 Series:

 

If you’re going to do research on the internet, I’ve found that “Opal IEEE 1667 BitLocker” are good search terms to get you started.

 

A special note to those who want to evaluate eDrive with the Intel® SSD Pro 2500 Series: the Intel-provided tool to enable eDrive support only works on “channel SKUs.”  Intel provides SSDs through the retail market (channel) and directly to OEMs (the maker/seller of your laptop).  Support for eDrive on OEM SKUs must be provided by the OEM.  Channel SKUs can be verified by looking at the firmware version on the SSD label, or with the Intel® SSD Toolbox or Intel® SSD Pro Administrator Tool.  Firmware in the format of TG## (TG20, TG21, TG26, etc…)  confirms a channel SKU, and the ability to enable eDrive support on the Intel® SSD Pro 2500 Series.

 

Take a look at eDrive, or managed hardware-based encryption solutions from ISVs such as McAfee, WinMagic, Wave, and others.

 

As always, I look forward to your input on topics you would like covered.

 

Thanks for your time!

 

Doug
intel.com/ssd

Read more >

1, 2, 3…It’s Time to Migrate from Windows Server 2003

Windows 2003 Blog 5.jpgWe’ve told you once. We’ve told you twice. Heck, we’ve told you four times: if you’re still running Windows Server 2003, you need to take action as soon as possible because the EOS is fast approaching on July 14th, 2015.

 

Need a refresher on what lies ahead? Well, good news, we’ve put together all the information you need to stay safe.

 

The upcoming Windows Server 2003 EOS means Microsoft will not be issuing any patches or security updates after the cut off date. While hackers are off rejoicing, this raises major security issues for those still running Windows Server 2003. And that appears to be quite a few of you.

 

According to Softchoice, a company specializing in technology procurement for organizations, 21 percent of all servers are still running Windows Server 2003. More worrisome is that 97 percent of all data centers are still running some form of Windows Sever 2003 within their facilities.

 

But migrating from Windows Server 2003 and ending up with proper security doesn’t have to be a pain. In our previous posts in this series, we’ve highlighted three different options for migration and how to secure the target environment. Let’s recap them here:

 

Option 1: Upgrade to Windows Server 2012

 

Because Windows Server 2008 will be losing support in January 2016, it’s a good idea for organizations to directly upgrade to Windows Server 2012 R2. This will require 64-bit servers and a refreshed application stack for supported configuration.

 

Likely, your organization might be looking to invest in a hybrid cloud infrastructure as part of this upgrade. Depending on what a server is used for, you’ll need optimized security solutions to secure your private virtual machines.

 

Intel Security has you covered. No matter what you’re running, you should at least employ either McAfee Server Security Suite Essentials or McAfee Server Security Suite Advanced.

 

If you’re running an email, sharepoint, or database server, consider McAfee Security for Email Servers, McAfee Security for Microsoft SharePoint or McAfee Data Center Security Suite for Databases, depending on your needs.

 

Option 2: Secure the public cloud

 

As the cloud ecosystem matures, the public cloud is becoming a reasonable alternative for many infrastructure needs. However, one issue remains: while public cloud solutions secure the underlying infrastructure, each company is responsible for securing their virtual servers from the Guest OS and up. Meaning, you’ll need a security solution built for the cloud.

 

Luckily, we’ve a solution that will help you break through the haze and gain complete control over workloads running within an Infrastructure- as-a-Service environment: McAfee Public Cloud Server Security Suite.

 

McAfee Public Cloud Server Security Suite gives you comprehensive cloud security, broad visibility into server instances in the public cloud and dynamic management of cloud environments.

 

Option 3: Protecting the servers you can’t migrate right now

 

For the 1.6 million of you that are behind schedule on Windows Server upgrades and won’t be able to migrate by the EOS date, you have a tough challenge ahead. Hackers know full well that Microsoft won’t be patching any newly discovered security issues, and as such, your servers might be a target.

 

But it’s not all doom and gloom – Intel Security can tie you over and keep you protected, until you’ve finished migrating.

 

With McAfee Application Control for Servers, you can command a centrally managed dynamic whitelisting solution. This solution will help you to protect your unsupported servers from malware and advanced threats, by blocking unauthorized applications, automatically categorizing threats and lowering manual input through a dynamic trust model.

 

Make your migration easy and get started today.

 

Be sure to follow along with @IntelSec_Biz on Twitter for real-time security updates. Stay safe out there!

 

 

Windows Server 2003 EOS

This blog post is episode #5 of 5 in the Tech Innovation Blog Series

View all episodes  >

Read more >

Stuck on Windows Server 2003? Migration Option #3

Windows 2003 Blog 4.pngPop quiz: what’s happening on July 14th, 2015?

 

If you’ve been following along with this series, you’ll know it’s the impending End of Support (EOS) for Windows Server 2003.

 

So far, we’ve covered two of the three migration options available to those still running Windows Server 2003: migrating to Windows Server 2012 or moving to the public cloud. Since migrating to a new server environment takes 200 days on average, and 300 for migrating applications, making the move by the mid-July end-of-life date may not be realistic.

 

This brings us to migration option #3: implementing additional protection for servers that cannot be immediately migrated by the EOS date.

 

Since Windows 2003 servers will no longer be compliant and will be vulnerable to new malware created after mid-July, you’ll need to take additional steps to keep them secure. Especially since hackers are patiently waiting for July 15th, knowing that Microsoft will no longer issue any security updates to these servers.

 

What’s even more concerning is that there are 23.8 million instances of Windows Server 2003 still running, making this a huge and potentially very lucrative target for hackers.

 

Fortunately, Intel Security can provide the security you need to keep your Windows Server 2003 instances safe from hackers’ targeted attacks.

 

If you have workloads that you cannot migrate off Windows Server 2003 by mid-July, be sure to install whitelisting protection to stay secure. McAfee Application Control for Servers is a centrally managed dynamic whitelisting solution which protects from malware and advanced threats. In more details, it:

  • Blocks unauthorized applications, therefore protecting from advanced persistent threats without requiring signature updates or labor-intensive lists to manage
  • Helps lower costs by eliminating manual support requirements associated with other whitelisting technologies
  • And best of all, it does all of that without requiring signature updates!

 

Next week, we’ll wrap up this series with a summary of the migration options for Windows Server 2003 and will highlight how to properly secure each of those paths.

 

Want to learn more about the migration options available to those running Windows Server 2003? Register for an educational webcast on “The 1-2-3 Security Approach for Windows Server 2003 EOS,” by visiting this page.

 

In the meantime, follow @IntelSec_Biz on Twitter for real-time security updates. We’ll see you back here next week!

 

 

 

 

Windows Server 2003 EOS

This blog post is episode #4 of 5 in the Tech Innovation Blog Series

View all episodes  >

Read more >

To the Public Cloud – and Beyond! Migrating from Windows Server 2003

Windows 2003 Blog 3.jpgAs you may remember from the last blog in this series, End of Support (EOS) for Windows Server 2003 goes into effect on July 15th, 2015. The clock is winding down, and we’re highlighting migration options available to those still running Windows Server 2003 in an effort to beat the buzzer.

 

Last week, we highlighted the first of three migration options available: upgrading to a newer version of Windows Server. Now, we will discuss the second migration path: moving to the public cloud.

 

Moving workloads to the public cloud has brought with it some real advantages in that it can help to reduce costs, provide greater agility, and enhance scalability. These benefits will most likely be realized as you move Windows Server 2003 based workloads into a public cloud environment.

 

However, a top concern when workloads are moved into a public cloud environment is security. As protection from a traditional network perimeter is not always available, a very common question is: what are the security controls that are in place for workloads that have moved into public cloud environments?

 

To answer this question, let’s take a look at the importance of security for your public cloud deployments starting with Infrastructure as a Service (IaaS).

 

Many enterprises mistakenly assume that public cloud providers for IaaS will protect their operating systems, applications, and data running on the server instances. While public cloud providers will secure the underlying infrastructure – it’s up to you to secure the virtual servers that run in that infrastructure. In other words, security is a shared responsibility in the cloud.

 

Luckily, we’ve come up with a solution that will help you break through the haze and gain complete control over workloads running within an IaaS environment: McAfee Public Cloud Server Security Suite.

 

This suite of security solutions is ideal for the public cloud, and uses hourly pricing so you only pay for security for the hours that your virtual servers are running in the cloud. The protection provided in McAfee Public Cloud Server Security Suite includes:

  • Comprehensive cloud security to extend and manage security policies for virtual servers in Microsoft Azure, Amazon AWS, and other public clouds such as those based on OpenStack.
  • Broader visibility into server instances in the public cloud via a powerful combination of blacklisting and whitelisting technologies.
  • Dynamic management of cloud environments in concert with on-premise endpoints and physical servers with McAfee ePolicy Orchestrator (ePO) software.

 

What about moving Windows Server 2003 applications such as email and collaboration into a Software as a Service (SaaS) public cloud environment?

 

If you’re planning on migrating to an environment such as Microsoft Office 365, McAfee Email Protection has you covered. This solution provides enterprise-grade security against targeted phishing threats, includes integrated data loss prevention technology, and protects Office 365 Exchange Online, Microsoft’s public cloud email solution.

 

Want to learn more about the security questions you need to ask about protecting email servers? Find all the information you need, here.

 

If moving to the public cloud doesn’t suit you or your business needs, fear not! Next week, we’ll take a deep dive into the third option available to those running Windows Server 2003: implementing additional protection for servers that cannot immediately be migrated by the EOS date.

 

In the meantime, follow @IntelSec_Biz on Twitter for the latest security updates. We’ll see you next week!

 

 

 

Windows Server 2003 EOS

This blog post is episode #3 of 5 in the Tech Innovation Blog Series

View all episodes  >

Read more >

Migrating from Windows Server 2003: Option #1

Windows 2003 Blog 2.jpgIf you read last week’s blog, you are well aware that when Windows Server 2003 goes off support on July 14th, 2015, it has the potential to result in the biggest vulnerability of the year. This End of Support (EOS) will effectively leave customers vulnerable and at risk of no longer being compliant.

 

Luckily, there are three options available to those still running Windows Server 2003. In this blog, we will highlight the first migration path: upgrading to a newer version of Windows Server.

 

You are currently faced with two options. The first, migrating to Windows Server 2008, we do not recommend. This will be taken off support in January 2016, setting you right back to square one. Instead, users might consider migrating to Windows Server 2012 R2, the newest available version.

 

Seems like a piece of cake, right? Not exactly – and here’s why.

 

As with any big change, there are a few challenges associated with this migration path. Here’s what lies ahead:

  • You will most likely need new hardware to install Windows Server 2012 R2, and these need to be 64-bit servers. You will also need to refresh your application stack so that you are in running supported application versions for Windows Server 2012.
  • At the same time, you will likely want to run your applications in virtual machines. In other words: in the private cloud and probably also in the public cloud.

Herein lies the security challenge: how can you best protect this hybrid-computing environment?

 

Intel Security offers a variety of server protection solutions to help you secure all types of servers, whether physical or even virtual servers in the cloud. No matter what you run on these servers, we strongly recommend one of the following options:

If your server is a specialized server such as Email, SharePoint, or Database server, you’ll need some further protection in addition to the above:

 

In next week’s Windows Server 2003 EOS blog, we will discuss the second migration path available to you: moving workloads to the public cloud instead of to Windows Server 2012. Trust us, it’s not as scary as it sounds!

 

Want to learn more about the migration options available to those running Windows Server 2003? Register for an educational webcast on “The 1-2-3 Security Approach for Windows Server 2003 EOS,” by visiting this page.

 

In the meantime, follow @IntelSec_Biz on Twitter and like us on Facebook for real-time security updates. See you back here next week!

 

 

Windows Server 2003 EOS

This blog post is episode #2 of 5 in the Tech Innovation Blog Series

View all episodes  >

Read more >

Bye, Bye Windows Server 2003 – Hello, Hackers

Still running a Windows Server 2003? 

You’ll want to read this.


Windows 2003 Blog 1.jpgAs you may or may not have heard, the end of life for Windows Server 2003 is now set for July 14th, 2015 – meaning after that date, it will no longer be supported by Microsoft.

 

So, what does this mean for you? It means that, along with other support, Microsoft will no longer be providing any security updates to Windows 2003 servers, leaving them exposed to malware and open for attack.

 

Security pros and IT specialists around the globe have already begun preparing as there are still 23.8 million Windows 2003 Servers out there. Some are even going as far to call the end of life for this product the “biggest security threat of 2015.”

 

You can bet that most hackers have this date marked on their calendars and are eagerly awaiting its arrival.

 

Lucky for you, with this blog (the first of a series of five intended to outline your options and solutions) we’ll help you to ensure you’re adequately protected when the date rolls around. Right now, there are three paths you can choose to go down in order to prepare:

  1. Upgrade to a newer version of Windows such as Windows 2012 R2
  2. Migrate Windows 2003 workloads to the public cloud
  3. Stay on Windows 2003 for now

However, each of these three paths comes with their own associated challenges, namely:

  • How to secure a hybrid compute environment and protect virtualized servers.
  • How to secure virtual servers in the public cloud.
  • How to secure Windows 2003 servers after July 14th, 2015.

 

Ultimately, the “do nothing” model is not an option as customers run the risk of no longer being compliant and leaving themselves vulnerable to malware.

 

At Intel Security we’re preparing materials to help you transition as seamlessly as possible, and overcome these challenges by securing your server environment no matter which of the three paths you decide to take.

 

Stay tuned for our next blog where we will outline in more depth each of the three migration options available to you and how you can overcome the challenges inherent with each. And, be sure to follow @IntelSec_Biz on Twitter for the latest security updates.

 

 

Windows Server 2003 EOS

This blog post is episode #1 of 5 in the Tech Innovation Blog Series

View all episodes  >

Read more >

How End-To-End Network Transformation Fuels the Digital Service Economy

To see the challenge facing the network infrastructure industry, I have to look no farther than the Apple Watch I wear on my wrist.

 

That new device is a symbol of the change that is challenging the telecommunications industry. This wearable technology is an example of the leading edge of the next phase of the digital service economy, where information technology becomes the basis of innovation, services and new business models.

 

I had the opportunity to share a view on the end-to-end network transformation needed to support the digital service economy recently with an audience of communications and cloud service providers during my keynote speech at the Big Telecom Event.

 

These service providers are seeking to transform their network infrastructure to meet customer demand for information that can help grow their businesses, enhance productivity and enrich their day-to-day lives.  Compelling new services are being innovated at cloud pace, and the underlying network infrastructure must be agile, scalable, and dynamic to support these new services.

 

The operator’s challenge is that the current network architecture is anchored in purpose-built, fixed function equipment that is not able to be utilized for anything other than the function for which it was originally designed.  The dynamic nature of the telecommunications industry means that the infrastructure must be more responsive to changing market needs. The challenge of continuing to build out network capacity to meet customer requirements in a way that is more flexible and cost-effective is what is driving the commitment by service providers and the industry to transform these networks to a different architectural paradigm anchored in innovation from the data center industry.

 

Network operators have worked with Intel to find ways to leverage server, cloud, and virtualization technologies to build networks that cost less to deploy, giving consumers and business users a great experience, while easing and lowering their cost of deployment and operation.

 

Transformation starts with reimagining the network

 

This transformation starts with reimagining what the network can do and how it can be redesigned for new devices and applications, even including those that have not yet been invented. Intel is working with the industry to reimagine the network using Network Functions Virtualization (NFV) and Software Defined Networking (SDN).

 

For example, the evolution of the wireless access network from macro basestations to a heterogeneous network or “HetNet”, using a mix of macro cell and small cell base-stations, and the addition of mobile edge computing (MEC) will dramatically improve network efficiency by providing more efficient use of spectrum and new radio-aware service capabilities.  This transformation will intelligently couple mobile devices to the access network for greater innovation and improved ability to scale capacity and improve coverage.

 

In wireline access, virtual customer premises equipment moves service provisioning intelligence from the home or business to the provider edge to accelerate delivery of new services and to optimize operating expenses. And NFV and SDN are also being deployed in the wireless core and in cloud and enterprise data center networks.

 

This network transformation also makes possible new Internet of Things (IoT) services and revenue streams. As virtualized compute capabilities are added to every network node, operators have the opportunity to add sensing points throughout the network and tiered analytics to dynamically meet the needs of any IoT application.

 

One example of IoT innovation is safety cameras in “smart city” applications. With IoT, cities can deploy surveillance video cameras to collect video and process it at the edge to detect patterns that would indicate a security issue. When an issue occurs, the edge node can signal the camera to switch to high-resolution mode, flag an alert and divert the video stream to a central command center in the cloud. With smart cities, safety personnel efficiency and citizen safety are improved, all enabled by an efficient underlying network infrastructure.

 

NFV and SDN deployment has begun in earnest, but broad-scale deployment will require even more innovation: standardized, commercial-grade solutions must be available; next-generation networks must be architected; and business processes must be transformed to consume this new paradigm. Intel is investing now to lead this transformation and is driving a four-pronged strategy anchored in technology leadership: support of industry consortia, delivery of open reference designs, collaboration on trials and deployments, and building an industry ecosystem.

 

The foundation of this strategy is Intel’s role as a technology innovator. Intel’s continued investment and development in manufacturing leadership, processor architecture, Ethernet controllers and switches, and optimized open source software provide a foundation for our network transformation strategy.

 

Open standards are a critical to robust solutions, and Intel is engaged with all of the key industry consortia in this industry, including the European Telecommunications Standards Institute (ETSI), Open vSwitch, Open Daylight, OpenStack, and others. Most recently, we dedicated significant engineering and lab investments to the Open Platform for NFV’s (OPNFV) release of OPNFV Arno, the first carrier-grade, open source NFV platform.

 

The next step for these open source solutions is to be integrated with operating systems and other software into open reference software to provide an on-ramp for developers into NFV and SDN. That’s what Intel is doing with our Open Network Platform (ONP); a reference architecture that enables software developers to lower their development cost and shorten their time to market.  The innovations in ONP form the basis of many of our contributions back to the open source community. In the future, ONP will be based on OPNFV releases, enhanced by additional optimizations and proofs-of-concept in which we continue to invest.

 

We also are working to bring real-world solutions to market and are active in collaborating on trials and deployments and deeply investing in building an ecosystem that brings companies together to create interoperable solutions.

 

As just one example, my team is working with Cisco Systems on a service chaining proof of concept that demonstrates how Intel Ethernet 40GbE and 100GbE controllers, working with a Cisco UCS network, can provide service chaining using network service header (NSH).  This is one of dozens of PoCs that Intel has participated in in just this year, which collectively demonstrate the early momentum of NFV and SDN and its potential to transform service delivery.

 

A lot of our involvement in PoCs and trials comes from working with our ecosystem partners in the Intel Network Builders. I was very pleased to have had the opportunity to share the stage with Martin Bäckström and announce that Ericsson has joined Network Builders. Ericsson is an industry leader and innovator, and their presence in Network Builders demonstrates a commitment to a shared vision of end-to-end network transformation.

 

The companies in this ecosystem are passionate software and hardware vendors, and also end users, that work together to develop new solutions. There are more than 150 Network Builder members taking advantage of this program and driving forward with a shared vision to accelerate the availability of commercial grade solutions.

 

NFV and SDN are deploying now – but that is just the start of the end-to-end network transformation. There is still a great deal of technology and business innovation required to drive NFV and SDN to scale, and Intel will continue its commitment to drive this transformation.




I invited the BTE audience – and I invite you – to join us in this collaboration to create tomorrow’s user experiences and to lay the foundation for the next phase of the digital services economy.

Read more >