Tagged: Human-Machine Systems

IEEE ETR 2018, Emerging Technologies Reliability Roundtable – Human Factors Session (2)


Following up on my last post about IEEE ERT 2018, here are a couple of charts for my “discussion brief,” which include a Human-Machine-System Capaility Mapping chart (above) and concept illustrations of the Experiential Decision Support System (below.)  The charts’ text conveys context setting remarks, which I am also providing here.


Slide1


The goal of furthering machine intelligence is to make humans more able and smarter: the opposite engineering approach typically becomes a source of self-defeating technical myopia waiting to happen and missed opportunities. This simple mapping exercise can be customized to assess and roadmap capability levels.

The more sophisticated automation becomes, the more obvious the criticality of the human factor in both consumer and enterprise environments… rather than less. And, in any case, customer acceptance and adoption criteria remain Quality’s litmus test for emerging technologies.

Digitalization is fostering (a) XaaS,  (b) Self-Service, (c) the Shared Economy and the (d) Maker Movement. All elevate human involvement and drive the push for opening and democratizing technologies. These make (e) citizen science and citizen developers shape the next generation prosumers at mass market scale.

Digital Transformation initiatives embracing the above allow (f) nimbler enterprise teams to operate at far greater scale, scope and speed, and shift focus from routine operations to dynamic value creation coupled with extreme efficiencies.

This entails (g) interdisciplinary workstyles and collaborative organizational behaviors that include (h) customer co-creation models. In this new context, humans remain (i) the ultimate critical element in system reliability and safety. Left shifting Quality by Design (QbD) prioritizes Human-Centered-Design tools and processes to deliver high performance workforce automation systems.


Slide2


Cost-effective Lean Ops systems intertwine analytics, automation, programmability and flexible systems integration. All optimized for dynamic behaviors given Soft System’s perpetual motion. This means designing “for-ever” rapid and seamless reconfigurability instead of just engineering “day 1” implementations.

Operational Excellence dictates system-wide as well as subsystem level visualization, and a combination of centralized & distributed closed loop controls under user friendly operational modes. Cognitive models involve Situational Awareness (SA,) Sense Making (SM,) Root Cause Analysis (RCA,) Scenario Planning (SP,) and ROA (Real Options Analysis.)

The Experiential element is not just about programming known rules and policies but, most importantly, it grows by assimiliating iterative learning in the context of cyclical automation: routine decisions and manual operations can be streamlined and colapsed, then switching to “exception” based management for that particular event.

Productivity calls for streamlining operations so that (a) waste can be eliminated & prevented, and (b) value based tasks can be performed effortlessly, in less steps, at speed & without error. High performance behaviors and sustainable competitiveness also call for the ability to (c) experiment and create new capabilities, as well as leveraging (d) process mining for customer journeys & value stream mapping (CJM & VSM) to continuously optimize them and guarantee service levels.

Service Operations Centers (SOC) should be equipped with Experiential Decision Support Systems (DSS) featuring (d) collaborative filtering, (e) actionable data stories conveying hindsight, insight & foresight and (f) adaptive cybernetics. Advanced visualization for both (f) intuitive & highly abstracted infographics and (g) scientific views is of the essence.

Quality is best addressed as a human experience, which determines (d) meaning and, therefore, the degree to which a system is lean vs. over-engineered or subpar (both being defective and carrying obvious and hidden costs.) A new take on QbD for Soft Systems, which are inherently fluid by definition, emphasizes acceptance testing probing for: usefulness & utility, usability & affectivity, consumability & serviceability and safety thru use cases and lifecycle events.

 

IEEE ETR 2018, Emerging Technologies Reliability – Human Factors Session

IEEE ETR 2018 on Twitter


ETR turned out to be a very productive undertaking and I would like to thank IEEE’s Spilios Markis, Chi-Ming Chen and Chris Mayer for all the help provided prior and during workshop.

My contribution focusing on addressing the unprecedented flexibility of advanced software defined systems and artificial intelligence. That intersection defines game changing technologies leading to zero-touch automation and, therefore, fostering self-service opportunities at both operational and service consumption levels.

“Zero touch” implies extreme automation to its fullest while self-service reveals that this new order elevates the criticality of HMS (Human Machine Systems.) More touch points surface compared to what legacy technologies allowed given their constraint and restricted nature. That prompts a new take on HCI (Human Computer Interaction) and QbD (Quality by Design) to best deliver service quality throughout: concept exploration and service definition, fulfilment and adaptation, assurance and security… across multi-domain, highly decomposed, re-configurable and exceptionally dynamic end-to-end systems involving integration and service delivery in continuous motion.

These are thought out to (a) dramatically optimize support personnel ratios and (b) shift staff’s attention and efforts to value based activities and innovation. These are small agile teams and new talent tasked with jobs involving (c) far greater scale with (d) a wider interdisciplinary scope, and all to be performed at (e) digital speed. In this next-level productivity and more demanding and challenging context, success relies on new tools embracing Design Thinking’s HCD (Human-Centered-Design.)

That is applied to capability models and subsequent modes of operation for (f) HITL (Human “IN” The Loop) Computing largely devoted to  deep domain expertise supported by Science Visualization, as well as (g) HOTL (Human “ON” the Loop) for system-wide supervisory responsibilities and ease of service creation and onboarding. HOTL draws from highly abstracted Visualization techniques and Low Code Development revealing the behavior of end-to-end systems and subsystems and adequate flow control.

These are coupled with effective Cybernetics gearing up for context aware 360-closed-loop-control, zooming in and out between distributed and central levels. Last but not least, effective and efficient tools that are characterized by ease of use and consumability do attract many more new users from many more different domains to interact with these systems in a self-service fashion and create new business opportunities as a result.

 

Software’s Defining Age


1k Fulton Market


I am on my way to Mobile World Congress and last night I had the opportunity to speak at DevMynd’s “Agile Software in a Hardware World.” That panel discussion featured BMW Technology Corporation (BMW, Mini, Rolls-Royce,) Monsanto’s “The Climate Corporation,” and Nokia Software Group, which I was proud to represent. The venue, 1KFulton, is a century-old and former cold storage building in the Fulton Market neighborhood, home to Google’s Chicago campus.


DEVMYND EVENT


Reflecting on that panel discussion, small group conversations and one-on-one chats before and after the event, I think that it is fair to state the following:

(A) software is undergoing a defining moment while re-shaping industries. “Software defined instruments and systems” have superseded capaibilities of hardware-centric deployments.

In other words, economic value and profitability are migrating from conventional products to software dominated environments that control tools, systems, and processes.


In this new context, (B) collaborative undertakings (co-creation, open source,) platforms, modularization and mashups are paving the way for rapid experimentation and for a wide-range of services to surface.

Back to economics… a venture capital firm operating in the Silicon Valley shared with me that when comparing current investments with equivalent old-school ones, they experienced x3 times time-to-market speed at 1/3 of the investment, which allows them to better diversify risk and fund more start-ups in the process.


Moreover, we are now operating at (C) unprecedented speed, scale and scope. For that reason alone, software should improve our ability to “pivot” and dynamically adapt to changing circumstances.

Most plans don’t survive first contact and many start-ups and emerging technologies don’t survive the so-called “crossing-the-chasm” or “Valley of Death.” So, remaining lean and embracing continuous/iterative improvement are of the essence. That’s a quality mantra rather than an excuse for forgoing best quality practices.

Back to economics again: quality management’s definition of “customer satisfaction” is now table-stakes and compliance in that area drives low-cost commoditization. “Customer delight” is the higher benchmark that commands a premium and the kind of margins enabling us to re-invest to further innovate.


Let’s now state the obvious, “customers” are human beings, aren’t they? Interestingly enough, the more sophistication and diversification, the higher the need for (D) humanizing technology so that we can better create, consume, use and democratize any digital services. In turn, this has fostered (E) Design Thinking as a leading innovation practice that intersects art and science. Design Thinking addresses HMS, Human-Machine-Systems, by prioritizing HCD, Human-Centered-Design.

In terms of economic effectiveness and efficience, that means outcome-oriented system-sizing, rather than over-engineering waste. It also means the definition of meaningful and purposeful requirements: some are designed to meet customer satisfaction metrics, while others are explicetly thought out to exceed that baseline and, hence, to actually deliver the X-Factor prompting customer delight. All key to customer acceptance and adoption growth.

Better yet, one of the event’s participants volunteered the fact that “good design” factoring inuitive interaction, advanced dataviz (data visualization) and effortless controls was proven to shrink the sales cycle by literally half: not only customers perceived and experienced the service’s taginble value early, the sales team was also able to approach more customers in that timeframe. Innovative Human-Computer-Interaction based on information design, value based tasks, streamlined processes, intuitive data visualization, effortless controls and overall UX, User Exeperience, double as compeling demonstration tools.


This is a side note: that has already become a critical success factor in Artificial Intelligence’s new developments, AI being software’s top transformational exponent as DSS, Decision Support Systems for humans and/or machines become quintessential. I will detail that in another post.


One last thought… (F) software’s pervasiveness has also brought along Agile development practices. These include “user stories” borrowing a Design Thinking technique by which application features are defined by sinthesizing human optics (persona/outcome/rationale) to put technical myopia at bay.

After all, we should all be in the business of making tech human. Otherwise, what would negating or ignoring that say about each of us and our collective culture?

Nokia @ Service Design Week 2017


Exploring Other Methods. November 7, 4:00 PM Understanding How Design Thinking, Lean and Agile Play within Service Design.

“Since service design serves as the umbrella discipline for delivering service experiences, there are many sub methods to address different types of problems. For example, Design Thinking is helpful on the front end to empathize and identify customer needs where Agile is helpful in software development and digital experience design. This group explores well-known methods and how they play a role in the service design universe.”


image


image

I’m back in Chicago and I would first like to thank everyone who joined my session about “Exploring Other Methods” for your participation (full house) and encouraging feedback. I hope to cross paths again in the near future. In the meantime, we can take advantage of LinkedIn to stay in touch. I would also like to express my gratitude to Michael DeJager and Tyler Peterson for all of their tireless help.

Here are the links for a couple of the items that I briefly discussed when providing context for Exploring Other Methods: a photo album of where I work, Nokia’s Chicago Technology Center, and the first version of the Human Factors Engineering Manifesto. Regarding requests about the slideware for my talk… I ran an interactive whiteboarding session with my iPad connected to the projector and I did not produce formal slides.


The discussion’s narrative was centered on how to best approach HSM, Human-Machine-Systems, to craft a compelling Service Experience. In that context, “Human” refers to relevant stakeholders and “Machine” to any technology involved. The “Systems” approach prompts a holistic undertaking which includes Front Stage, Back Stage factors and the continuum across the too.

Service Design is about innovation, whether capability-wise that qualifies as incremental, breakthrough and/or disruptive innovation. Today’s Service Design also entails a wide range of low and high-tech at any point in the process. While this is just anecdotal evidence, when I asked everyone about who can do away without any technology, there was an implicit understanding of the rhetorical nature of my question and, therefore, the obvious pervasiveness of digital experiences.

We are a technological society. Good design is concerned with human factors and crafts technological solutions to enable human experiences that contribute to our quality of life and the quality of the work we do. That is Human Factors Engineering (HFE) reason for being, a field pioneered by Nokia Bell Labs in 1947.


From that perspective, it pays to intertwine any relevant practices and tools for the healthy purpose of figuring out what combination works best for any given Service Design project. While process repeatability is a desired outcome, what makes an interdisciplinary team smart is the ability to mix, match and blend what’s needed for each undertaking.

We can think of it as an a-la-carte menu featuring elements from Design Thinking, Agile and Lean methodologies just to name a popular handful to start with. I did not discuss some other such as Concept of Operations, Goal Directed Design or Outcome Driven Innovation, but I do recommend expanding one’s horizons beyond the aforementioned few. Note that while featuring commonalities, each one works with different optics. A holistic approach to Service Design also requires a composite method, leveraging as much (or as little) as needed from any, and with any needed adaptations.


Rather than summarizing what I shared at Service Design Week, I’m taking this chance to further reflect on those insights. So, given that we operate in highly dynamic environments, why wouldn’t designers also apply dynamic methodologies?

I’d like to think twice about cookie-cutter and one-size-fits-all approaches because Service Design typically prompts problems and opportunities where fixed-gear-techniques that might have worked well in the past can end up betraying one’s confidence: they might no longer serve or be the best fit whichever purpose they were originally conceived for. Design typically takes us beyond our comfort level, and that makes it an exciting profession.

Statistically speaking, the more one does the very same thing, the closer one gets to mastering that craft (e.g. deliberate practice model). But, paradoxically, you also get closer and closer to confronting environmental deviations, anomalies and rare events in an ever-changing world with even-growing moving parts and targets (e.g. black swan model). Besides, Service Design practitioners shouldn’t deny themselves the benefits that come with continuous improvement. So, here is a quick recap: innovation in Service Design’s outcomes and method innovation go hand by hand. As Einstein put it:

“Insanity is doing the same thing over and over and expecting a different result.”

“If we knew what it was we were doing, it would not be called research, would it?”


Nokia C3LM @ Service Design Week


“Service Design is big. Being holistic, it includes the researching, envisioning and orchestrating of service experiences that happen over time and across multiple touch points with many stakeholders involved, both frontstage and backstage.”

“At Service Design Week, we seek to strip away any fluff, examining service design methods and processes at their core, and unpack the practical tools and skill-sets, hard and soft, needed for this way of working. Service Design Week will gather service design leaders from various functions and disciplines across all flavors of Service Design. With content for all levels of Service Design maturity, we look forward to drawing both fledging and experienced service designers.” 

 www.servicedesignweekusa.com

image


I am looking forward to joining Service Design Week and I would like to thank Michel DeJager and the team at the International Quality & Productivity Center for their kind invitation. My talk will discuss C3LM, Customer Co-Creation Lifecycle Methodology, in the context of Blended Service Design, which I will take care of defining and demystifying in my talk.

I am proud to share that C3LM is the recipient of a Nokia Innovation Award. My work seeks to interweave a set of known and brand new interdisciplinary practices to best address end-to-end solutions for complex and dynamic environments, also known as soft systems given their organic and morphing nature. And, most importantly, achieving that by optimizing for the delivery of quality experiences while humanizing low and high tech in the process.


image


Widespread digitalization in our everyday activities is not just far reaching, but is also leading to a renaissance in Human Factors disciplines. The delivery of “effective quality services” with “highly efficient end-to-end solutions” is the reason for being and rationale behind creating C3LM. This new brave world entails Blended Services that intersect Data Science, Automation and Programmability, all orchestrated with Human Centered Design in mind.

My talk will also cover how we can best experience Artificial Intelligence and how to make it transparent to Blended Services. That will be a sneak preview in advance to another talk that I’m giving early next year. In case you have already heard what Elon Musk has to say about AI, let me share that Human Factors Engineering has been revisited and redefined to come to the rescue. More on that when we get to meet at Service Design Week : )  



image


Here is the event’s registration page. See you in Boston : )

Pictures courtesy of Service Design Week.