Recent Blog Posts

High-Performance Computing Helping to Make Dreams Come True – HLRS

Audience-Watching-Movie-In-A-Theater.jpgThis year is shaping up to be one of the best years for cinema in a while. Some of Hollywood’s most iconic characters are returning to the big screen in 2015, with new releases from James Bond*, Star Wars* and The Hunger Games* franchises. However, few of us stop to wonder how 007 can plummet through a glass ceiling unscathed or how those X-wings crashing look so realistic… It’s all down to hidden technology.


Technology in the Talkies


Space drama, Gravity*, won a few Oscars at the 2014 Academy Awards, including Best Visual Effects, and it’s not hard to see why. Apparently, around 80 per cent of the scenes in Gravity were animated and computer generated. In many scenes, only Sandra Bullock and George Clooney’s faces existed as anything other than 1s and 0s on computers. Everything else, from space shuttles, jetpacks and space debris, was created by graphic artists using Dell* workstations powered by 4th generation Intel® Core™ i5 and i7 vPro™ processors.


Only last month, we released What Lives Inside, the fourth installment of Intel’s Inside Films series. Directed by two-time Oscar-winner Robert Stromberg, our latest social film stars Colin Hanks, J.K. Simmons and Catherine O’Hara alongside the recently launched Dell* Venue 8 7000 Series super-thin tablet with Intel® RealSenseTM technology and powered by an Intel® Atom™ processor Z3580. The film took eight days to shoot and relied on 200 visual effects artists, which just goes to show what it takes to bring such whimsical worlds to life on the big screen.


HPC Helping the Film Industry


3D movies rely a lot on technology and require significant computing capacity. In a change of pace from the usual manufacturing or university research projects, The High Performance Computing Center in Stuttgart (HLRS) recently supported a local production company by rendering a 3D kids’ movie called Maya, The Bee*. This 3D flick, starring Kodi Smit-McPhee and Miriam Margolyes, is not your typical HPC challenge, but the amount of data behind the 3D visuals presented quite a mountain to climb for the makers, Studio 100.


To ensure the film was ready in time, the project was transferred to HLRS, which had recently upgraded to the Intel® Xeon® Processor E5-2680 v3, enabling it to undertake more projects and better serve critical industry needs like this one because this system delivers four times the performance of its previous supercomputer. 1 Thanks to the HPC capacity available at HLRS, Maya, The Bee was released last month in all its 3D glory.2 “We are addicted to giving best possible service, so it is vital that we run on reliable technology,” said Bastian Koller, Manager of HLRS. For more on the HLRS supercomputer, click here.


Bringing Characters to Life


Intel and Framestore have been working together for almost five years now. However, Paddington* is the first film that Framestore has worked on with a computer-animated animal as the lead character, and the mischievous little bear caused quite a few challenges. Many characters brought to film, such as J. K. Rowling’s Dobby* or Andy Serkis as Gollum*, are shot using motion-capture technology to make them appear more lifelike, but the actor playing Paddington wore a head-mounted camera during the voice recordings so animators could see how a human face moved at every point and try to mimic it in bear form. While this gave audiences an incredible character, animating and rendering a photo-real lead character for every shot required significant processing capacity. It took 350 people across two continents working for three years to bring Paddington the CGI bear to life, but with high-performance, flexible IT solutions based on servers powered by Intel® Xeon® processors, it was a piece of cake (or a marmalade sandwich!).


Intel VP & GM, Gordon Graylish, shared his thoughts on the red carpet at the Paddington premiere, saying: “It is wonderful when [Intel® technology] gets to be used for creativity like this, and this is something that would have been impossible or prohibitively expensive [to make] even two or three years ago. Technology allows you to take the limits off your imagination.”


I’d love to hear what you think about all this hidden HPC tech used to get fantastic blockbusters into our movie theaters, so please leave a comment below or continue this conversation by connecting with me on LinkedIn or Twitter, with #ITCenter – I’m happy to answer questions you may have.


1 Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations, and functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products. For more information go to


2 Released in March 2015 in the US and the UK


*Other names and brands may be claimed as the property of others.

Read more >

The Challenge of the Smart Megacity

People-Talking-By-The-Water.pngOn my recent visit to China, I was struck by the country’s commitment to investing in smart cities. China’s most recent five-year plan set aside $70 billion for smart city technologies, with around 200 cities competing for funding. This is part of a huge project of urbanization, which saw $1 trillion allocated for urban infrastructure under the same plan. Last year, the Chinese government announced its intention to increase its urban population from 53.7 percent to 60 percent by 2020, and there are already 15 megacities in China with more than 10 million people.


With 1.3 million people per week moving to and trying to build lives in cities globally through 2050, it’s no surprise that the impetus to bring “smart” to these locations has risen on the agenda of many of the most prominent cities worldwide.


If done properly, a smarter city environment should have a measurable impact on the economy, the citizens and their lifestyles, business and the environment. There’s certainly no shortage of examples of how technology can be applied to making a city smart, as listed below:



BUT do applications like this amount to the “smart city”? One thing I’ve noticed, from my discussions with Intel customers in China and other countries, is there is no single definition of what the smart city is. Government bodies recognize the opportunities presented by technologies like those I’ve mentioned and it’s clear there’s a healthy degree of friendly competition amongst the cities but where I see many struggle is working out what they should do first or next, and what the smart city really means to them.


While this may be well understood by many of you, the focus areas we see coming up most frequently are:


  • Smart Transport/Mobility
  • Smart Home-Building-Facility
  • Smart Public Infra & Community Driven Services
  • Smart Fixed-Mobile Security Surveillance
  • Analytics and Big-Data Strategy-Planning



Irrespective of which area a city focuses on first, one thing is for sure: with the proliferation of millions of smart connected devices – on the transport network, and in anything from buildings to street lights to manholes – the result is a huge amount of flowing data. To get the best return on investment, it’s essential to plan how that data will be managed, how value can be extracted from it and what you plan to do with it. While almost every customer I talk to acknowledges they need to do something with the data most struggle with what they want to do with it. Without these plans in place, the data simply piles up and creates mountains in minutes. If you haven’t done so already I’d recommend hiring some Data Scientists – typically mathematicians or statisticians who can help you determine what data you need and what you might want to do with it.


On a somewhat related note, many of you will be familiar with SMAC stack (social, mobile, analytics, cloud). This is the digital platform being laid down across industries to underpin transformation. It’s been a core part of the rapid rise we’ve seen in shared economy companies like Uber and AirBnB. It is also fundamental to the smart city. The smart city is not just about adding connectivity to a building or other asset: it’s about the data you gather, the insights you gain, the services you can create and deliver, the accessibility you provide, the economic growth you stimulate and the communities you grow. Clearly, this all needs to be done and delivered in a secure and predictable manner. The point is not to look at or use SMAC and just look at one part of it. The impact comes from the multiplicative effect it has.


In the smart city, as much as anywhere these days, all roads lead to data. The question we need to be asking is: which roads do we want to travel?


What do you think defines the smart city? I’d be interested to read your comments below.


To continue the conversation on Twitter, please follow us at @IntelITCenter or use #ITCenter.

Read more >

Ordering Kiosks Give Hardee’s a Tasty Solution for Satisfying Customers and Growing Sales

Today’s Hardees.jpgconsumers move at breakneck speedwhich is one reason quick-service restaurants like Hardee’s are so popular. And to stay out front, it’s essential for those restaurants to keep finding new ways to delight customers and keep them coming back. Hardee’s did it by installing new quick-service customer ordering kiosks with 24-inch, touch-based screens. Instead of waiting in line, customers can see enticing images of what’s on the menu and then order with a few quick taps on the screen. And since orders go directly to the kitchen, the food is ready sooner. It all means Hardee’s can serve more customers and bring in more revenue.


Based on Industry-Standard Technology


Hardee’s had investigated the idea of ordering kiosks a few years ago, but those available at the time were based on proprietary technology and too expensive to be practical.




The new kiosks Hardee’s chose are based on a Dell OptiPlex all-in-one system equipped with Intel® Core™ i5 vPro™ processors and Windows Pro 8.1.


Using Hardees_Tweet.jpgindustry-standard technology like Windows gives Hardee’s the flexibility to run other applications including software used by employees and managers. It’s also convenient for software developers, who can use familiar programming environments, and for the restaurant’s IT administrators, who can use existing Microsoft systems management tools.


The all-in-one form factor increases deployment flexibility, since Hardee’s can mount the kiosks in a variety of places, depending on the layout of each restaurant.


Controlling Costs in the Future


With the success of the kiosks, Hardee’s is now considering using all-in-one systems to gradually replace point-of-sale (POS) systems at the counter as a tasty solution for delivering an outstanding customer experience and controlling costs.


To learn more, take a look at the Hardee’s solution here or read more about it here. To explore more technology success stories, visit or follow us on Twitter.












Read more >

Big Data is Changing the Football Game

The football authorities have been slow to embrace technology, at times actively resisting it. It’s only been two seasons since some of Europe’s top leagues were authorized to use goal-line technology to answer the relatively simple question of whether or not a goal has been scored, i.e., has the whole ball crossed the goal line.


This is something the games of tennis and cricket have been doing for nearly ten years, but for one of the world’s richest sports, it risked becoming a bit of a joke.  As one seasoned British manager once said, after seeing officials deny his team a perfectly good goal: “We can put a man on the moon, time serves of 100 miles per hour at Wimbledon, yet we cannot place a couple of sensors in a net to show when a goal has been scored.” The authorities eventually relented, of course, their hand forced by increasingly common, high profile and embarrassing slip-ups.


But while the sport’s governing bodies were in the grips of technological inertia, the world’s top clubs have dived in head first in the last ten to fifteen years, turning to big data analytics in search of a new competitive advantage. In turn, this has seen some innovative companies spring up to serve this new ‘industry’, companies like Intelcustomer Scout7.


Taking the Guesswork out of the Beautiful Game


Big data has become important in football in part because it is big business. And for a trend that is only in its second decade, things have moved fast since the days of teams of hundreds of scouts collecting ‘data’ in the form of thousands of written reports in an effort to provide teams with insights into the opposition or potential new signings.


Now, with tools like Scout7’s football database, which is powered by a solution based on the Intel® Xeon® Processor E3 Family solution, they have a fast, sophisticated system that clubs can use to enhance their scouting and analysis operations.


For 138 clubs in 30 leagues, Scout7 makes videos of games from all over the world available for analysis within two hours of the final whistle[1]. At the touch of a button, they can take some of the guess work and ‘instinct’ out of deciding who gets on the pitch, as well as the legwork of keeping tabs on players and prospects from all over the world.



Pass master: Map of one player’s passes and average positions from the Italian Serie A during the 2014-15 season


Using big data analytics to enable smarter player recruitment is among Scout7’s specialties. For young players, without several seasons of experience on which to judge them, this can be especially crucial. How do you make a call on their temperament or readiness to make the step up? How will they handle the pressure? As we enter the busiest recruitment period of the football calendar – the summer transfer window – questions like this are being asked throughout the football world right now.


Delving into the Data


It’s a global game, and Scout7 deals in global data, so we can head to a league less travelled for an example: the Czech First League. The UEFA Under-21 European Championships also took place this summer and, with international tournaments often acting as shop windows for the summer transfer market (which opened on 1st July – a day after tournament’s final), it makes sense to factor this into our analysis.


So, let’s look at the Scout7 player database for players in the Czech First League that are currently Under-21 internationals, to see who has had the most game time and therefore exposure to the rigors of competitive football. We can see that a 22-year-old FC Hradec Králové defender, played every single minute of his team’s league campaign this season – 2,700 minutes in total.


Another player’s on-field time for this season was 97% — valuable experience for a youngster. Having identified two potential first-team ready players, Scout7’s database would allow us to take a closer look at the key moments from these games in high-definition video.


Check out our infographic, detailing a fledgling career of another player in the context of the vast amount of data collection and analysis that takes place within Scout7.


Scout-7-Player-Profile.pngScout7 player profile


“Our customers are embracing this transition to data-driven business decision-making, breaking away from blind faith in the hunches of individuals and pulling insights from the raft of new information sources, including video, to extract value and insights from big data,” explains Lee Jamison, managing director and founder, Scout7.


Scout7’s platform uses Intel® technology to deliver the computing power and video transcoding speed that clubs need to mine and analyze more than 3 million minutes of footage per year, and its database holds 135,000 active player records.


Lonely at the Top


There’s only room at the top of the elite level of sport for one and the margins between success and failure can be centimeters or split seconds. Identifying exactly where to find those winning centimeters and split seconds is where big data analytics really comes into its own.


Read the full case study.


To continue this conversation on Twitter, please follow us at @IntelITCenter or use #ITCenter.

Find me on LinkedIn.

Keep up with me on Twitter.


*Other names and brands may be claimed as the property of others.


[1] Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations and functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products. For more information go to

Intel does not control or audit the design or implementation of third party benchmark data or Web sites referenced in this document. Intel encourages all of its customers to visit the referenced Web sites or others where similar performance benchmark data are reported and confirm whether the referenced benchmark data are accurate and reflect performance of systems available for purchase.

Intel technologies’ features and benefits depend on system configuration and may require enabled hardware, software or service activation. Performance varies depending on system configuration. Check with your system manufacturer or retailer or learn more at

Read more >

10 Mobile BI Strategy Questions: Executive Sponsorship

Man-On-Morning-Comute-Using-Tablet.pngOf the ten mobile BI questions I outlined in my last post, “Do we have an executive sponsor?” is the most important one because the success of a mobile BI journey depends on it more than any other. While the role of an executive sponsor is critical in all tech projects, several aspects of mobile BI technology make it easy for executive management to be involved closely and play a unique role.


Moreover, although the CIO or the CTO plays a critical role in making sure the right technology is acquired or developed, the executive sponsorship from the business side provides the right level of partnership in order to run on all three cylinders of BI: insight into right data for the right role and at the right time.


Why Do We Need an Executive Sponsor?


We need executive sponsorship because, unlike grassroots efforts, business and technology projects require a top-down approach. Whether the strategy is developed as part of a structured project or as a standalone engagement, the executive sponsor delivers three critical ingredients:


  1. The mobile BI strategy is in line with the overall business strategy.
  2. The required resources are made available.
  3. Necessary guidance is provided in order to stay the course.


Is Having an Executive Sponsor Enough?


Having an executive only on paper isn’t enough, however. How much commitment an executive sponsor makes and the leadership he/she provides has a direct impact on the outcome of the strategy. Thus, the ideal executive sponsor of a mobile BI initiative is a champion of the cause, an ardent mobile user, and the most active consumer of its assets.


What Makes an Ideal Executive Sponsor for Mobile BI?


How does the executive champion the mobile BI initiative? First and foremost, he/she leads by example — no more printing paper copies of reports or dashboards. This means that the executive is keen not only to consume the data on mobile devices but also to apply the insight derived from these mobile assets to decisions that matter. Using the technology demonstrates firsthand the mobile mindset that sets an example for the rest of the direct reports and their teams. In addition, by recognizing the information available on these mobile BI assets as the single version of the truth, the executive provides a clear and consistent message for everyone to follow.


Is Mobile BI Easier to Adopt by Executive Sponsors?


Without a doubt, mobile BI, just like mobility, is conducive to a wide range of users, starting with executives. Unlike the PC, which wasn’t mobile at all, and the laptop, which provided limited mobility, tablets and smartphones provide a perfect combination of mobility and convenience. This ease of use makes these devices an ideal candidate in winning over even those executives who may have been initially uneasy to include mobile BI in their arsenals or to use them in their daily decision-making activities.


The mobility and simplicity may give the executives additional incentives to get involved in the development of requirements for the first set of mobile BI assets because they can easily see the benefits of having access to critical information at their fingertips. These benefits include an additional opportunity for sales and marketing to use mobile BI to showcase new products and services to customers (an approach that reflects the innovation inherent in the use of this technology).


Bottom Line: Executive Sponsorship Matters


The most important goal of a mobile BI strategy is to enable faster, better-informed decision making. Executive sponsorship matters because with the right sponsorship, the mobile BI initiative will have the best chance to drive growth and profitability. Without this sponsorship — even with the most advanced technology in place — a strategy will face an uphill battle.


What other aspects of executive sponsorship do you see playing a role in mobile BI strategy?


Stay tuned for my next blog in the Mobile BI Strategy series.


Connect with me on Twitter at @KaanTurnali and LinkedIn.


This story originally appeared on the SAP Analytics Blog.

Read more >

NVM Express: Windows driver support decoded


NVMe Drivers and SSD support in Windows

Microsoft enabled native, support for NVM Express (NVMe) in Windows 8.1 and Windows Server 2012 R2 by way of inbox drivers, and subsequent versions of each OS family are expected to have native support moving forward.  Additionally, native support for NVMe in Windows 7 and Windows Server 2008 R2 was added via product updates. 


Intel also provides an NVMe driver for Microsoft OS’s that releases with each version of our NVMe hardware products internally and using Microsoft’s WHCK.  The list of supported OS’s is the same as those above (for both 32-bit and 64-bit versions), along with Windows 8 and Windows Server 2012 (R2). The Intel NVMe driver supports only Intel SSDs and is required for power users or server administrators who plan to use the Intel® Solid-State Drive Data Center Tool to perform administrative commands on an NVMe SSD (e.g. firmware updates).  The Intel driver is intended to provide the best overall experience in terms of performance and supportability, it is strongly recommended.



Download Links by Operating Systems


NVMe Drivers for Windows

Operating System

Intel Driver Download

Microsoft Driver Download

Windows 7

Windows Server 2008 R2

Windows 8

supported by upgrade to Windows 8.1

Windows Server 2012

supported by upgraded to Windows Server 2012 R2

Windows 8.1

N/A (inbox driver)

Windows Server 2012 R2

N/A (inbox driver)



Other Links of Interest




Intel® Solid-State Drive Data Center Tool

The Intel® Solid-State Drive Data Center Tool (Intel SSD DCT) is a drive management tool for Intel SSD Data Center Family of products.

Intel® SSD Data Center Family Overview

Provides access to more information on Intel’s NVMe PCIe SSDs.

More information on what NVMe is, why you should consider using it, and news/upcoming events.



Other blogs by Operating Systems with NVM Express driver information:

NVM Express: Linux driver support decoded

The Skinny on NVM Express and ESXi

Read more >

Empathizing with Teachers and Learners Leads to the Read With Me App

Teaching is tough work. In one design thinking project that I mentioned in a previous blog post, empathy for teachers and students led to development of the Read With Me app (available now on Chrome and select Android devices) co-developed … Read more >

The post Empathizing with Teachers and Learners Leads to the Read With Me App appeared first on Intel Software and Services.

Read more >

Why Choose the Mini PC? Part 2

Retail and finance industries turn to Mini PCs for high performance, compact computing power



Whether it’s tucked away on a bookshelf, hidden behind a fitting room mirror or mounted on a digital display, Intel technology-based solutions featuring the Mini PC are helping to power industries as varied as the retail and the financial sectors. Thanks to their energy efficiency, compact design and high performance computing power, these tiny form factors bring full-sized PC power to the smallest of spaces. Here are some real-world examples of the endless possibilities with Mini PCs:


Mini PCs as Part of An Overall Technology Solution for Retail


One of my favorite Mini PC success stories is that of Galleria Rizzoli in Milan, Italy. Galleria Rizzoli saw the impact of digital book sales firsthand, and decided to respond with a complete digital overall of its operations.


With the help of Intel technology, Galleria Rizzoli launched a pilot program that gave their store a complete technology makeover. Mini PCs powered new in-store digital signage and seven new in-store customer kiosks. Mini PCs replaced bulky desktop towers, freeing up valuable store space. Thanks to the technology makeover, sales increased 40 percent.


Galleria Rizzoli is a great example of how Mini PCs can enhance the user experience to help drive sales.


Overall, it’s a winning solution for Intel, for Rizzoli, and for consumers who might be looking to quickly find the perfect kids’ book for a boy who likes to play with trucks. Read the full story of how Mini PCs modernized the bookstore.


Embedded Mini PCs Enable Next-Gen Vending Machines


Whether you’re grabbing a quick snack at the office or simply refueling at the gas station, vending machines today are operating on a complex system of motherboards, dispensing tanks, and printing and credit card machines. Many new OEMs are currently working on consolidating all these disparate parts into one Mini PC solution.


Mini PCs in the Fitting Room


Instead of treating the fitting room like a revolving door, imagine being able to tap a screen to request a different size or color. Some retailers are exploring the idea of using the Mini PC to power touch-screen consoles in fitting rooms to provide instant inventory access to customers while also recommending referential products for purchase.



National Grocery Chains Power POS with Mini PCs


The days of the bulky cash register have given way to more compact Mini PC-powered POS systems in grocery stores as well. Not only do Mini PCs leave a smaller footprint in tight cashier stalls, they also provide the high performance computing power necessary to ring up multiple items in quick succession.


Hospitality Industry Welcomes Mini PCs

Look inside many hotel business centers and you’ll likely see a row of monitors with Mini PCs tucked neatly behind them. The Mini PC offers a compact solution that won’t slow guests down. And some hotels are exploring the use of Mini PCs in guest rooms attached to the TVs along with concierge-type software to enhance the in-room guest experience.


Banks Turn to Mini PCs for Increased Efficiency

A growing number of banks are reaching for Mini PCs, not only for their compact size, but for their energy efficiency and speed. For many clients, a visit to the local bank reveals tellers relying on Mini PCs where desktop towers once stood. Mini PCs free up valuable desk space, offer compact security, and integrate with legacy systems.


Day Traders Turn to Mini PCs for Quick Calculations


For day traders, Mini PCs featuring solid-state-drives (SSDs) are the desktop PCs of choice. While traditional hard disk drives in PCs and laptops are fairly inexpensive, they are also slow. SSDs offer greater capacity, are considered more reliable, and enable faster access to data, which is critical to an industry where seconds matter.


Where have you seen the Mini PC in use? Join the conversation using #IntelDesktop or view our other posts in the Desktop World Series and rediscover the desktop.


To read part 1, click here: Why Choose the Mini PC? Part 1

Read more >

Future of IoT: 5 Questions with Technology Futurist Angela Orebaugh

We sat down with technology futurist Angela Orebaugh recently to chat about emerging Internet of Things (IoT) trends. In 2011, Angela was named Booz Allen Hamilton’s first Cybersecurity Fellow, a position reserved for the firm’s most notable experts in their … Read more >

The post Future of IoT: 5 Questions with Technology Futurist Angela Orebaugh appeared first on IoT@Intel.

Read more >

Population Health Management Best Practices for Today and Tomorrow’s Healthcare System

By Justin Barnes and Mason Beard


The transition to value-based care is not an easy one. Organizations will face numerous challenges on their journey towards population health management.


We believe there are five key elements and best practices to consider when transitioning from volume to value-based care:  managing multiple quality programs; supporting both employed and affiliated physicians and effectively managing your network and referrals; managing organizational risk and utilization patterns; implementing care management programs; and ensuring success with value-based reimbursement.


When considering the best way to proactively and concurrently manage multiple quality programs, such as pay for performance, accountable care and/ or patient-centered medical home initiatives, you must rally your organization around a wide variety of outcomes-based programs. This requires a solution that supports quality program automation. Your platform must aggregate data from disparate sources, analyze that data through the lens of a program’s specific measures, and effectively enable the actions required to make improvements. Although this is a highly technical and complicated process, when done well it enables care teams to utilize real-time dashboards to monitor progress and identify focus areas for improving outcomes.


In order to provide support to both employed and affiliated physicians, and effectively manage your network and referrals, an organization must demonstrate its value to healthcare providers. Organizations that do this successfully are best positioned to engage and align with their healthcare providers. This means providing community-wide solutions for value-based care delivery. This must include technology and innovation, transformation services and support, care coordination processes, referral management, and savvy representation with employers and payers based on experience and accurate insight into population health management as well as risk.


To effectively manage organization risk and utilization patterns, it is imperative to optimize episodic and longitudinal risk, which requires the application of vetted algorithms to your patient populations using a high quality data set. In order to understand the difference in risk and utilization patterns you need to aggregate and normalize data from various clinical and administrative sources, and then ensure that the data quality is as high as possible. You must own your data and processes to be successful. And importantly, do not rely entirely on data received from payers.


It is also important to consider the implementation of care management programs to improve individual patient outcomes. More and more organizations are creating care management initiatives for improving outcomes during transitions of care and for complicated, chronically ill patients. These initiatives can be very effective.  It is important to leverage technology, innovation and processes across the continuum of care, while encompassing both primary and specialty care providers and care teams in the workflows. Accurate insight into your risk helps define your areas of focus. A scheduled, trended outcomes report can effectively identify what’s working and where areas of improvement remain.


Finally, your organization can ensure success with value-based reimbursement when the transition is navigated correctly. The shift to value-based reimbursement is a critical and complicated transformation—oftentimes a reinvention—of an organization. Ultimately, it boils down to leadership, experience, technology and commitment. The key to success is working with team members, consultants and vendor partners who understand the myriad details and programs, and who thrive in a culture of communication, collaboration, execution and accountability.


Whether it’s PCMH or PCMH-N, PQRS or GPRO, CIN or ACO, PFP or DSRIP, TCM or CCM, HEDIS or NQF, ACG’s or HCC’s, care management or provider engagement, governance or network tiering, or payer or employer contracting, you can find partners with the right experience to match your organizations unique needs. Because much is at stake, it is necessary to ensure that you partner with the very best to help navigate your transition to value-based care.


Justin Barnes is a corporate, board and policy advisor who regularly appears in journals, magazines and broadcast media outlets relating to national leadership of healthcare and health IT. Barnes is also host of the weekly syndicated radio show, “This Just In.”


Mason Beard is Co-Founder and Chief Product Officer for Wellcentive. Wellcentive delivers population health solutions that enable healthcare organizations to focus on high quality care, while maximizing revenue and transforming to support value-based models.

Read more >