Categories

Tag: AI

The Latest AI Craze: Ambient Scribing

By MATTHEW HOLT

Okay, I can’t do it any longer. As much as I tried to resist, it is time to write about ambient scribing. But I’m going to do it in a slightly odd way

If you have met me, you know that I have a strange English-American accent, and I speak in a garbled manner. Yet I’m using the inbuilt voice recognition that Google supplies to write this story now.

Side note: I dictated this whole thing on my phone while watching my kids water polo game, which has a fair amount of background noise. And I think you’ll be modestly amused about how terrible the original transcript was. But then I put that entire mess of a text  into ChatGPT and told it to fix the mistakes. it did an incredible job and the output required surprisingly little editing.

Now, it’s not perfect, but it’s a lot better than it used to be, and that is due to a couple of things. One is the vast improvement in acoustic recording, and the second is the combination of Natural Language Processing and artificial intelligence.

Which brings us to ambient listening now. It’s very common in all the applications we use in business, like Zoom and others like transcript creation from videos on Youtube. Of course, we have had something similar in the medical business for many years, particularly in terms of radiology and voice recognition. It has only been in the last few years that transcribing the toughest job of all–the clinical encounter–has gotten easier.

The problem is that doctors and other professionals are forced to write up the notes and history of all that has happened with their patients. The introduction of electronic medical records made this a major pain point. Doctors used to take notes mostly in shorthand, leaving the abstraction of these notes for coding and billing purposes to be done by some poor sap in the basement of the hospital.

Alternatively in the past, doctors used to dictate and then send tapes or voice files off to parts unknown, but then would have to get those notes back and put them into the record. Since the 2010s, when most American health care moved towards using  electronic records, most clinicians have had to type their notes. And this was a big problem for many of them. It has led to a lot of grumpy doctors not only typing in the exam room and ignoring their patients, but also having to type up their notes later in the day. And of course, that’s a major contributor to burnout.

To some extent, the issue of having to type has been mitigated by medical scribes–actual human beings wandering around behind doctors pushing a laptop on wheels and typing up everything that was said by doctors and their patients. And there have been other experiments. Augmedix started off using Google Glass, allowing scribes in remote locations like Bangladesh to listen and type directly into the EMR.

But the real breakthrough has been in the last few years. Companies like Suki, Abridge, and the late Robin started to promise doctors that they could capture the ambient conversation and turn it into proper SOAP notes. The biggest splash was made by the biggest dictation company, Nuance, which in the middle of this transformation got bought by one of the tech titans, Microsoft. Six years ago, they had a demonstration at HIMSS showing that ambient scribing technology was viable. I attended it, and I’m pretty sure that it was faked. Five years ago, I also used Abridge’s tool to try to capture a conversation I had with my doctor — at that time, they were offering a consumer-facing tool – and it was pretty dreadful.

Fast forward to today, and there are a bunch of companies with what seem to be really very good products.

Continue reading…

Are AI Clinical Protocols A Dobb-ist Trojan Horse?

By MIKE MAGEE

For most loyalist Americans at the turn of the 19th century, Justice John Marshall Harlan’s decision in Jacobson v. Massachusetts (1905). was a “slam dunk.” In it, he elected to force a reluctant Methodist minister in Massachusetts to undergo Smallpox vaccination during a regional epidemic or pay a fine.

Justice Harlan wrote at the time: “Real liberty for all could not exist under the operation of a principle which recognizes the right of each individual person to use his own, whether in respect of his person or his property, regardless of the injury that may be done to others.”

What could possibly go wrong here? Of course, citizens had not fully considered the “unintended consequences,” let alone the presence of President Wilson and others focused on “strengthening the American stock.”

This involved a two-prong attack on “the enemy without” and “the enemy within.”

The The Immigration Act of 1924, signed by President Calvin Coolidge, was the culmination of an attack on “the enemy without.” Quotas for immigration were set according to the 1890 Census which had the effect of advantaging the selective influx of Anglo-Saxons over Eastern Europeans and Italians. Asians (except Japanese and Filipinos) were banned.

As for “the enemy within,” rooters for the cause of weeding out “undesirable human traits” from the American populace had the firm support of premier academics from almost every elite university across the nation. This came in the form of new departments focused on advancing the “Eugenics Movement,” an excessively discriminatory, quasi-academic approach based on the work of Francis Galton, cousin of Charles Darwin.

Isolationists and Segregationists picked up the thread and ran with it focused on vulnerable members of the community labeled as paupers, mentally disabled, dwarfs, promiscuous or criminal.

In a strategy eerily reminiscent of that employed by Mississippi Pro-Life advocates in Dobbs v. Jackson Women’s Health Organization in 2021, Dr. Albert Priddy, activist director of the Virginia State Colony for Epileptics and Feebleminded, teamed up with radical Virginia state senator Aubrey Strode to hand pick and literally make a “federal case” out of a young institutionalized teen resident named Carrie Buck.

Their goal was to force the nation’s highest courts to sanction state sponsored mandated sterilization.

In a strange twist of fate, the Dobbs name was central to this case as well.

Continue reading…

The 7 Decade History of ChatGPT

By MIKE MAGEE

Over the past year, the general popularization of AI orArtificial Intelligence has captured the world’s imagination. Of course, academicians often emphasize historical context. But entrepreneurs tend to agree with Thomas Jefferson who said, “I like dreams of the future better than the history of the past.”

This particular dream however is all about language, its standing and significance in human society. Throughout history, language has been a species accelerant, a secret power that has allowed us to dominate and rise quickly (for better or worse) to the position of “masters of the universe.”

Well before ChatGPT became a household phrase, there was LDT or the laryngeal descent theory. It professed that humans unique capacity for speech was the result of a voice box, or larynx, that is lower in the throat than other primates. This permitted the “throat shape, and motor control” to produce vowels that are the cornerstone of human speech. Speech – and therefore language arrival – was pegged to anatomical evolutionary changes dated at between 200,000 and 300,000 years ago.

That theory, as it turns out, had very little scientific evidence. And in 2019, a landmark study set about pushing the date of primate vocalization back to at least 3 to 5 million years ago. As scientists summarized it in three points: “First, even among primates, laryngeal descent is not uniquely human. Second, laryngeal descent is not required to produce contrasting formant patterns in vocalizations. Third, living nonhuman primates produce vocalizations with contrasting formant patterns.”

Language and speech in the academic world are complex fields that go beyond paleoanthropology and primatology. If you want to study speech science, you better have a working knowledge of “phonetics, anatomy, acoustics and human development” say the  experts. You could add to this “syntax, lexicon, gesture, phonological representations, syllabic organization, speech perception, and neuromuscular control.”

Professor Paul Pettitt, who makes a living at the University of Oxford interpreting ancient rock paintings in Africa and beyond, sees the birth of civilization in multimodal language terms. He says, “There is now a great deal of support for the notion that symbolic creativity was part of our cognitive repertoire as we began dispersing from Africa.  Google chair, Sundar Pichai, maintains a similarly expansive view when it comes to language. In his December 6, 2023, introduction of their ground breaking LLM (large language model), Gemini (a competitor of ChatGPT), he described the new product as “our largest and most capable AI model with natural image, audio and video understanding and mathematical reasoning.”

Continue reading…

The Optimism of Digital Health

By JONATHON FEIT

Journalists like being salty.  Like many venture investors, we who are no longer “green” have finely tuned BS meters that like to rip off the sheen of a press release to reach the truthiness underneath. We ask, is this thing real? If I write about XYZ, will I be embarrassed next year to learn that it was the next Theranos?

Yet journalists must also be optimistic—a delicate balance: not so jaded that one becomes boooring, not so optimistic that one gets giddy at each flash of potential; and still enamored of the belief that every so often, something great will remake the present paradigm.

This delicately balanced worldview is equally endemic to entrepreneurs that stick around: Intel founder Andy Grove’s famously said “only the paranoid survive,” a view that is inherently nefarious since it points out that failure is always lurking nearby. Nevertheless, to venture is to look past the risk, as in, “Someone has to reach that tall summit someday—it may as well be our team!” Pragmatic entrepreneurs seek to do something else, too: deliver value for one’s clients / customers / partners / users in excess of what they pay—which makes they willing to pay in excess of what the thing or service costs to produce. We call that metric “profit,” and over the past several years, too many young companies, far afield of technology and healthcare, forgot about it.

Once upon a time, not too many years ago, during the very first year that my company (Beyond Lucid Technologies) turned a profit, I presented to a room of investors in San Francisco, and received a stunning reply when told that people were willing to pay us for our work.  “But don’t you want to grow?” the investor asked. 

Continue reading…

Can Generative AI Improve Health Care Relationships?

By MIKE MAGEE

“What exactly does it mean to augment clinical judgement…?”

That’s the question that Stanford Law professor, Michelle Mello, asked in the second paragraph of a May, 2023 article in JAMA exploring the medical legal boundaries of large language model (LLM) generative AI.

This cogent question triggered unease among the nation’s academic and clinical medical leaders who live in constant fear of being financially (and more important, psychically) assaulted for harming patients who have entrusted themselves to their care.

That prescient article came out just one month before news leaked about a revolutionary new generative AI offering from Google called Genesis. And that lit a fire.

Mark Minevich, a “highly regarded and trusted Digital Cognitive Strategist,” writing in a December issue of  Forbes, was knee deep in the issue writing, “Hailed as a potential game-changer across industries, Gemini combines data types like never before to unlock new possibilities in machine learning… Its multimodal nature builds on, yet goes far beyond, predecessors like GPT-3.5 and GPT-4 in its ability to understand our complex world dynamically.”

Health professionals have been negotiating this space (information exchange with their patients) for roughly a half century now. Health consumerism emerged as a force in the late seventies. Within a decade, the patient-physician relationship was rapidly evolving, not just in the United States, but across most democratic societies.

That previous “doctor says – patient does” relationship moved rapidly toward a mutual partnership fueled by health information empowerment. The best patient was now an educated patient. Paternalism must give way to partnership. Teams over individuals, and mutual decision making. Emancipation led to empowerment, which meant information engagement.

In the early days of information exchange, patients literally would appear with clippings from magazines and newspapers (and occasionally the National Inquirer) and present them to their doctors with the open ended question, “What do you think of this?”

But by 2006, when I presented a mega trend analysis to the AMA President’s Forum, the transformative power of the Internet, a globally distributed information system with extraordinary reach and penetration armed now with the capacity to encourage and facilitate personalized research, was fully evident.

Coincident with these new emerging technologies, long hospital length of stays (and with them in-house specialty consults with chart summary reports) were now infrequently-used methods of medical staff continuous education. Instead, “reputable clinical practice guidelines represented evidence-based practice” and these were incorporated into a vast array of “physician-assist” products making smart phones indispensable to the day-to-day provision of care.

At the same time, a several decade struggle to define policy around patient privacy and fund the development of medical records ensued, eventually spawning bureaucratic HIPPA regulations in its wake.

The emergence of generative AI, and new products like Genesis, whose endpoints are remarkably unclear and disputed even among the specialized coding engineers who are unleashing the force, have created a reality where (at best) health professionals are struggling just to keep up with their most motivated (and often mostly complexly ill) patients. Needless to say, the Covid based health crisis and human isolation it provoked, have only made matters worse.

Continue reading…

AI Inside

by KIM BELLARD

Well: 2024. I’m excited about the Paris Olympics, but otherwise I’d be just as happy to sleep through all the nonsense that the November elections will bring. In any event, I might as well start out talking about one of the hottest topics of 2023 that will get even more so in 2024: AI.

In particular, I want to look at what is being billed as the “AI PC.” 

Most of us have come to know about ChatGPT. Google has Bard (plus DeepMind’s Gemini), Microsoft is building AI into Bing and its other products, Meta released an open source AI, and Apple is building its AI framework. There is a plethora of others. You probably have used “AI assistants” like Alexa or Siri.

What most of the large language model (LLM) versions of AI have in common is that they are cloud-based. What AI PCs offer to do is to take AI down to your own hardware, not dissimilar to how PCs took mainframe computing down to your desktop.  

As The Wall Street Journal tech gurus write in their 2024 predictions in their 2024 predictions:

In 2024, every major manufacturer is aiming to give you access to AI on your devices, quickly and easily, even when they’re not connected to the internet, which current technology requires. Welcome to the age of the AI PC. (And, yes, the AI Mac.)

Continue reading…

2024 Prediction: Society Will Arrive at an Inflection Point in AI Advancement

By MIKE MAGEE

For my parents, March, 1965 was a banner month. First, that was the month that NASA launched the Gemini program, unleashing “transformative capabilities and cutting-edge technologies that paved the way for not only Apollo, but the achievements of the space shuttle, building the International Space Station and setting the stage for human exploration of Mars.” It also was the last month that either of them took a puff of their favored cigarette brand – L&M’s.

They are long gone, but the words “Gemini” and the L’s and the M’s have taken on new meaning and relevance now six decades later.

The name Gemini reemerged with great fanfare on December 6, 2023, when Google chair, Sundar Pichai, introduced “Gemini: our largest and most capable AI model.” Embedded in the announcement were the L’s and the M’s as we see here: “From natural image, audio and video understanding to mathematical reasoning, Gemini’s performance exceeds current state-of-the-art results on 30 of the 32 widely-used academic benchmarks used in large language model (LLM) research and development.

Google’s announcement also offered a head to head comparison with GPT-4 (Generative Pretrained Transformer-4.) It is the product of a non-profit initiative, and was released on March 14, 2023. Microsoft’s helpful AI search engine, Bing, helpfully informs that, “OpenAI is a research organization that aims to create artificial general intelligence (AGI) that can benefit all of humanity…They have created models such as Generative Pretrained Transformers (GPT) which can understand and generate text or code, and DALL-E, which can generate and edit images given a text description.”

While “Bing” goes all the way back to a Steve Ballmer announcement on May 28, 2009, it was 14 years into the future, on February 7, 2023, that the company announced a major overhaul that, 1 month later, would allow Microsoft to broadcast that Bing (by leveraging an agreement with OpenAI) now had more than 100 million users.

Which brings us back to the other LLM (large language model) – GPT-4, which the Gemini announcement explores in a head-to-head comparison with its’ new offering. Google embraces text, image, video, and audio comparisons, and declares Gemini superior to GPT-4.

Mark Minevich, a “highly regarded and trusted Digital Cognitive Strategist,” writing this month in Forbes, seems to agree with this, writing, “Google rocked the technology world with the unveiling of Gemini – an artificial intelligence system representing their most significant leap in AI capabilities. Hailed as a potential game-changer across industries, Gemini combines data types like never before to unlock new possibilities in machine learning… Its multimodal nature builds on yet goes far beyond predecessors like GPT-3.5 and GPT-4 in its ability to understand our complex world dynamically.”

Expect to hear the word “multimodality” repeatedly in 2024 and with emphasis.

Continue reading…

AI Could Have “Unimaginable Consequences” For Democratic Societies, Says Expert.

By MIKE MAGEE

His biography states, “He speaks to philosophical questions about the fears and possibilities of new technology and how we can be empowered to shape our future. His work to bridge cultures spans artificial intelligence, cognition, language, music, creativity, ethics, society, and policy.”

He embraces the title “cross-disciplinary,” and yet his PhD thesis at UC Berkeley in 1980 “was one of the first to spur the paradigm shift toward machine learning based natural language processing technologies.” Credited with inventing and building “the world’s first global-scale online language translator that spawned Google Translate, Yahoo Translate, and Microsoft Bing Translator,” he is clearly a “connector” in a world currently consumed by “dividers.” In 2019, Google named De Kai as “one of eight inaugural members of its AI Ethics Council.”

The all encompassing challenge of our day, as he sees it, is relating to each other. As he says, “The biggest fear is fear itself – the way AI amplifies human fear exponentially…turning us upon ourselves through AI powered social media driving misinformation, divisiveness, polarization, hatred and paranoia.” The value system he embraces “stems from a liberal arts perspective emphasizing creativity in both technical and humanistic dimensions.”

Dr. De Kai is feeling especially urgent these days, which is a bit out of character.

Continue reading…

Where is AI in Medicine Going? Ask The Onion.

By MIKE MAGEE

One of the top ten headlines of all time created by the satirical geniuses at The Onion was published 25 years ago this December. It read, “God Answers Prayers Of Paralyzed Little Boy. ‘No,’ Says God.”

The first paragraph of that column introduced us to Timmy Yu, an optimistic 7-year old, who despite the failures of the health system had held on to his “precious dream.” As the article explained, “From the bottom of his heart, he has hoped against hope that God would someday hear his prayer to walk again. Though many thought Timmy’s heavenly plea would never be answered, his dream finally came true Monday, when the Lord personally responded to the wheelchair-bound boy’s prayer with a resounding no.”

But with a faith that rivals the chieftains of today’s American health care system who continue to insist this is “the best health care in the world,” this Timmy remained undeterred. As The Onion recorded the imagined conversation, “‘I knew that if I just prayed hard enough, God would hear me,’ said the joyful Timmy,., as he sat in the wheelchair to which he will be confined for the rest of his life. ‘And now my prayer has been answered. I haven’t been this happy since before the accident, when I could walk and play with the other children like a normal boy.’”

According to the article, the child did mildly protest the decision, but God held the line, suggesting other alternatives. “God strongly suggested that Timmy consider praying to one of the other intercessionary agents of Divine power, like Jesus, Mary or maybe even a top saint,” Timmy’s personal physician, Dr. William Luttrell, said. ‘The Lord stressed to Timmy that it was a long shot, but He said he might have better luck with one of them.’”

It didn’t take a wild leap of faith to be thrust back into the present this week. Transported by a headline to Rochester, Minnesota, the banner read, “Mayo Clinic to spend $5 billion on tech-heavy redesign of its Minnesota campus.” The “reinvention” is intended to “to present a 21st-century vision of clinical care” robust enough to fill 2.4 million square feet of space.

The Mayo Clinic’s faith in this vision is apparently as strong as little “Timmy’s”, and their “God” goes by the initials AI.

Continue reading…

Pin Me, Please

By KIM BELLARD

You had to know I’d write about the new Humane AI Pin, right?

After all, I’d been pleading for the next big thing to take the place of the smartphone, as recently as last month and as long ago as six years, so when a start-up like Humane suggests it is going to do just that, it has my attention.  Even more intriguing, it is billed as an AI device, redefining “how we interact with AI.”  It’s like catnip for me.

For anyone who has missed the hype – and there has been a lot of hype, for several months now – Humane is a Silicon Valley start-up founded by two former Apple employees, Imran Chaudhri and Bethany Bongiorno (who are married).  They left Apple in 2016, had the idea for the AI Pin by 2018, and are ready to launch the actual device early next year.  It is intended to be worn as a pin on the lapel, starts at $699, and requires a monthly $24 subscription (which includes wireless connectivity).  Orders start November 16.

Partners include OpenAI, Microsoft, T-Mobile, Tidal, and Qualcomm.

Mr. Chaudhri told The New York Times that artificial intelligence  “can create an experience that allows the computer to essentially take a back seat.” He also told TechCrunch that the AI Pin represented “a new way of thinking, a new sense of opportunity,” and that it would “productize AI” (hmm, what are all those other people in AI doing?).  

Humane’s press release elaborates:

Ai Pin redefines how we interact with AI. Speak to it naturally, use the intuitive touchpad, hold up objects, use gestures, or interact via the pioneering Laser Ink Display projected onto your palm. The unique, screenless user interface is designed to blend into the background, while bringing the power of AI to you in multi-modal and seamless ways.

Basically, you wear a pin that is connected with an AI, which – upon request – will listen and respond to your requests. It can respond verbally, or it can project a laser display into the palm of your hand, which you can control with a variety of gestures that I am probably too old to learn but which younger people will no doubt pick up quickly.  It can take photos or videos, which the laser display apparently does not, at this point, do a great job projecting. 

Here’s Humane’s introductory video:

Some cool features worth noting:

  • It can summarize your messages/emails;
  • It can make phone calls or send messages;
  • It can search the web for you to answer questions/find information;
  • It can act as a translator;
  • It has trust features that include not always listening and a “Trust Light” that indicates when it is.
Continue reading…