Category: General

Research from Harvard, MIT, and Wharton highlights the risks of depending on junior staff to train AI systems.

As businesses rapidly incorporate artificial intelligence (AI) into their operations, the prevailing belief is that younger, technologically adept employees will spearhead educating their senior managers on effectively harnessing these advanced tools. However, a recent study challenges this notion, especially concerning the use of generative AI technologies.

Study Details and Key Participants

The study was a collaborative effort involving scholars from prestigious institutions such as Harvard Business School, MIT, and Wharton, in partnership with Boston Consulting Group. The research focused on the interactions and experiences of junior employees with generative AI systems, particularly GPT-4, in real-world business scenarios.

Unexpected Findings from Junior Consultants

Contrary to expectations, the study revealed that junior employees, often presumed to be tech-savvy, might not be the best resources for guiding senior professionals in the effective use of emerging technologies like generative AI. The findings showed that the risk mitigation strategies proposed by these junior consultants frequently contradicted expert advice and lacked a deep understanding of AI’s capabilities.

Key Insights from the Study:

  • Junior Consultants’ Tactics: The research highlighted that the strategies suggested by junior employees to alleviate seniors’ concerns about AI risks were often misaligned with expert recommendations. These tactics were more about altering human behavior than enhancing the AI system’s design.
  • Focus on Short-term Solutions: Many recommendations were project-specific rather than aimed at broader organizational or industry-wide applications, suggesting a narrow scope of understanding.

In-depth Analysis of the Research Findings

1. Limited Technical Expertise The study found that junior consultants typically had minimal technical expertise in AI. Their recommendations were based more on general knowledge and less on a technical understanding of AI systems like GPT-4.

2. Risk Mitigation Approaches Junior employees tended to focus on immediate, surface-level solutions rather than systemic changes or in-depth strategies that could be more beneficial in the long run.

Challenges in Adopting Generative AI in Business

The rapid evolution of generative AI technologies presents significant challenges and opportunities for businesses. These AI systems can perform tasks such as engaging in detailed dialogues, responding to follow-up questions, and assisting in writing, analysis, and coding tasks. However, the study underscores the necessity of comprehensive AI governance and the need for expert input at all organizational levels.

Navigating AI Implementation Challenges:

  • Top-down Governance: Effective AI implementation requires informed leadership rather than relying solely on the knowledge of digital natives within the organization.
  • Expert Involvement: Incorporating AI experts into strategic planning and implementation processes is crucial to address potential risks and optimize AI usage.

Moving Forward: Recommendations for Effective AI Adoption

The findings advocate for a structured approach to AI adoption in corporate settings:

  • Upskilling Programs: Develop extensive training programs to enhance the AI competence of employees across all levels.
  • Leadership Roles: Senior professionals should take proactive roles in understanding and integrating new technologies to lead their teams effectively.
  • Future-proofing Strategies: Businesses need to anticipate future technological advancements and their potential impacts on industry and internal operations.

Summary in Bullet Points

  • Study Collaboration: Involvement of top academic institutions and Boston Consulting Group.
  • Key Finding: Junior employees may not be ideal mentors for senior staff in AI adoption.
  • Recommendations: Emphasize top-down governance, expert involvement, and comprehensive training.

This extensive study not only highlights a critical gap in the assumed capabilities of junior employees concerning AI but also sets the stage for rethinking how businesses should approach the integration of these powerful technologies into their workflows. Senior leaders are encouraged to take a more active role in understanding and guiding AI initiatives to ensure that their organizations can fully leverage AI’s capabilities responsibly and effectively.

McKesson and Merck Invest in Atropos Health’s $33M Funding Round to Boost AI-Driven Drug Development

Silicon Valley-based Atropos Health has successfully raised $33 million in a Series B funding round, marking a significant step forward in its mission to integrate AI-powered, personalized real-world evidence into healthcare decision-making.

In a recent announcement, Atropos Health, a pioneer in generating personalized real-world evidence, disclosed a substantial $33 million acquisition in Series B funding. This investment round featured prominent contributions from healthcare behemoths like McKesson, Merck, and Cencora Ventures, indicating a robust industry endorsement of Atropos’ innovative approach to healthcare.

The funds are earmarked for a strategic expansion aimed at enhancing the company’s operational capacity and doubling down on critical initiatives. These include a deeper penetration into the life sciences sector, broadening channel partnerships in value-based care and oncology, and expanding its network of data partners to enrich its evidence base.

Brigham Hyde, PhD, CEO and co-founder of Atropos Health, expressed his enthusiasm in a VentureBeat interview, stating, “We’re on a mission to bring personalized evidence for care to everybody in the world. This funding is a pivotal step in that journey. Specifically, we’ll be focusing on reinforcing our strategic initiatives, continuing our successful launch in life sciences, and enhancing our partnerships, particularly in value-based and specialty care oncology.”

Atropos Health is not just another player in the healthcare field; it is a trailblazer aiming to close the pervasive “evidence gap” in medical decision-making. The company’s flagship technology, Geneva OS, harnesses artificial intelligence (AI) and automation to rapidly generate clinical-grade evidence from real-world data. This platform, which has been developed over nearly a decade of research at Stanford University, powers applications such as the generative AI assistant, ChatRWD.

The technology enables clinicians, researchers, and other healthcare stakeholders to swiftly access reliable clinical evidence, personalized to specific patient populations—a capability often missing in current healthcare practices. Dr. Hyde highlighted a concerning statistic in his interview: “Only about 14% of daily medical decisions have any high-quality evidence behind them. Our goal is to use high-quality data, analyzed correctly, to fill this evidence gap.”

The central mission of Atropos is to provide clinicians with easy access to personalized evidence, thereby enhancing patient outcomes. Dr. Hyde used the example of heart failure patients to illustrate the need for tailored evidence that caters to subpopulations with unique characteristics and comorbidities, which could lead to more effective treatments and cost control.

Atropos’ applications extend beyond clinical decision-making. The company collaborates with pharmaceutical leaders, including Janssen, to expedite drug development by leveraging real-world evidence for clinical trial design, patient recruitment, and more. Dr. Hyde even suggested that the platform could simulate clinical trials, potentially revolutionizing the way pharmaceutical research is conducted by reducing cycle times and de-risking trials.

Despite the excitement surrounding large language models (LLMs) and generative AI, Atropos prioritizes building trust through methodological rigor and transparency. Dr. Hyde expressed concerns about the “hallucination rates” in current AI models and emphasized that Geneva OS ensures clinical-grade quality and transparency, backed by a decade of publications.

Strategic Use of Series B Funding

Initiative Objective Expected Impact
Expansion in Life Sciences Enhance presence and partnerships in life sciences Broaden application of real-world evidence in R&D
Channel Partnerships Growth Focus on value-based care and oncology Improve treatment strategies and patient outcomes
Data Network Expansion Increase the network of data partners Enrich the quality and diversity of clinical evidence

With a fresh influx of capital and a roster of strategic backers, Atropos is poised to bring its vision of personalized, automated clinical evidence to the global healthcare landscape. “Evidence is the currency of value in healthcare,” Dr. Hyde posited. “What if I could give doctors more evidence, more personalized, so they make better decisions? Fundamentally, we’re trying to move the world to a point where all patients and all providers have access to quality, personalized evidence for their decision-making.”

This bold vision by Atropos Health not only promises to transform patient care but also positions the company as a frontrunner in the integration of AI and healthcare. As they continue to bridge the evidence gap, the future of healthcare looks promisingly precise, personalized, and powered by artificial intelligence.

Y Combinator’s Garry Tan Backs AI Rules, Warns of Monopolies

Garry Tan, the influential president and CEO of the startup incubator Y Combinator, recently addressed an audience at The Economic Club of Washington, D.C., emphasizing the need for regulatory frameworks in the rapidly evolving field of artificial intelligence (AI). Tan’s comments come at a critical juncture as AI technologies continue to permeate various aspects of societal and economic activities.

The Argument for Regulation

During a detailed interview with Teresa Carlson, a board member at General Catalyst, Tan shared his views on a multitude of topics, from entry paths into Y Combinator to the broader implications of AI developments. He highlighted the unprecedented opportunities currently available in the technology sector, stating, “There is no better time to be working in technology than right now.”

Tan voiced his support for the efforts by the National Institute of Standards and Technology (NIST) to create a risk mitigation framework for generative AI (GenAI). He believes that the Executive Order (EO) by the Biden Administration aligns well with necessary steps towards responsible AI deployment. The NIST’s framework includes several important guidelines:

  • Compliance with Existing Laws: GenAI must adhere to laws governing data privacy and copyright.
  • Disclosure Requirements: Companies must inform end users about their use of GenAI technologies.
  • Prohibitions on Harmful Content: Regulations should prevent GenAI from generating or distributing harmful materials such as child sexual abuse content.

Further, President Biden’s executive order mandates AI companies to share safety data with governmental bodies and ensures that small developers have equitable access to the technology market.

Concerns Over State Legislation

Despite his general support for federal efforts, Tan expressed concerns about AI-related bills progressing through state legislatures, particularly in California and San Francisco. One controversial bill, introduced by California State Senator Scott Wiener, could potentially allow the state attorney general to sue AI companies if their products cause harm. This bill, among others, has stirred significant debate within the tech community regarding its implications on innovation and business operations.

Key Regulatory Proposals Discussed by Garry Tan

Regulation Aspect Description Potential Impact
NIST Framework Guidelines for risk mitigation in GenAI applications Enhances safety and compliance standards
Biden’s Executive Order Comprehensive directives for AI deployment and oversight Aims for balanced growth and safety
California Legislative Bills Potential legal actions against harmful AI products Raises concerns about innovation stifling

The Balance Between Innovation and Control

Tan highlighted the delicate balance that needs to be maintained between fostering technological innovation and mitigating potential harms. He cited UK AI expert Ian Hogarth’s approach, which is thoughtful about maintaining a balance between limiting the concentration of power within the AI sector and encouraging innovative progress. Hogarth, a former YC entrepreneur, is part of an AI model taskforce in the UK, working towards viable policy solutions.

Y Combinator’s Ethical Stance

Tan shared insights into Y Combinator’s internal decision-making processes regarding AI startups. He emphasized that the incubator only funds startups that align with positive societal impacts. “If we don’t agree with a startup’s mission or its potential effects on society, YC just doesn’t fund it,” Tan explained. This cautious approach has led them to avoid backing several companies after reviewing their potential implications through media reports and internal evaluations.

AI Industry Challenges and the Future Landscape

The discussion also touched upon recent industry controversies, including high-profile issues at OpenAI and Meta. These instances have sparked a broader debate on the ethical responsibilities of AI firms and the transparency required in their operations.

  • OpenAI’s Responsibility Team: Recently, it was reported that OpenAI might be scaling back its AI responsibility team, raising questions about the commitment to ethical AI development.
  • Voice Mimicry Issues: OpenAI faced criticism for using a voice resembling actress Scarlett Johansson’s in demos, without her consent, leading to further scrutiny of its practices.
  • Meta’s AI Advisory Council Composition: Meta’s decision to form an AI advisory council predominantly comprising white men has also drawn criticism for not reflecting diversity.

The Vision for AI’s Future

Looking ahead, Tan is optimistic about the potential for AI to enable a diverse range of consumer choices and empower founders. He envisages a future where AI does not lead to monopolistic practices but instead fosters a vibrant landscape of varied solutions accessible to billions globally.

In conclusion, while Tan acknowledges the potential dangers of AI, his primary concern remains the risk of a monopolistic concentration of power within the industry, which could lead to restrictive practices detrimental to innovation and consumer choice. His vision for AI emphasizes both caution and enthusiasm, aiming for a future where technology serves humanity broadly and equitably.

Fintech lender SoLo Funds faces another lawsuit from the government regarding its lending practices.

The Consumer Financial Protection Bureau (CFPB) has filed a lawsuit against SoLo Funds, a fintech company specializing in peer-to-peer lending. The CFPB accuses the company of employing “digital dark patterns” to mislead borrowers and illegally extracting fees, despite advertising a fee-free service.

Allegations of Deceptive Practices

In a press release on May 17, CFPB Director Rohit Chopra stated, “The CFPB is suing SoLo for using digital trickery to hide interest and fees on its online loans. SoLo has had repeated run-ins with state regulators, and we are putting a stop to their fake tipping scheme.”

The CFPB’s lawsuit details several allegations against SoLo Funds:

  • Misrepresentation of Loan Costs: The company is accused of obscuring the true cost of loans, making it difficult for consumers to understand the financial commitments they were making.
  • Illegal Fee Collection: Despite advertising no fees, SoLo allegedly took fees from borrowers under the guise of tips and donations.
  • False Credit Reporting Threats: SoLo is charged with making unfounded threats related to credit reporting to coerce payments.
  • Lack of Safeguards: The CFPB claims that SoLo’s business model lacks necessary consumer protection measures.

The CFPB contends that these practices not only deceived consumers but also contravened federal consumer financial laws.

Breakdown of Allegations Against SoLo Funds

Allegation Description
Misrepresentation of Loan Costs Obscuring true loan costs, interfering with consumers’ understanding of agreements.
Illegal Fee Collection Collecting fees despite advertising no fees, through lender ‘tips’ and SoLo ‘donations.’
False Credit Reporting Threats Making baseless threats related to credit reporting to compel payments.
Lack of Consumer Safeguards Operating without adequate measures to protect consumers from predatory practices.

Background and Growth of SoLo Funds

SoLo Funds was founded in 2018 by Rodney Williams and Travis Holoway, with a mission to provide financial services to underserved Americans who are frequently targeted by predatory lenders. The company’s innovative approach aimed to create a peer-to-peer lending platform where users could borrow money from each other.

SoLo Funds has garnered significant financial backing, raising approximately $13 million in venture funding. Notable investors include Serena Ventures, founded by tennis star Serena Williams, Endeavor Catalyst, Alumni Ventures, and Techstars. In 2021, TechCrunch highlighted SoLo Funds when it secured $10 million in Series A funding. The platform has seen considerable growth, claiming to have reached one million registered users and over 1.3 million downloads by 2023.

Recent Legal Challenges

The CFPB lawsuit is the latest in a series of legal troubles for SoLo Funds. The company settled multiple lawsuits last year with the District of Columbia and the State of California, addressing accusations of predatory lending practices. Additionally, the Connecticut Department of Banking issued a temporary cease-and-desist order against SoLo in 2022.

In December 2023, the company faced scrutiny from the State of Maryland, adding to its ongoing regulatory challenges.

SoLo Funds’ Response to the Lawsuit

In a statement to TechCrunch, SoLo Funds asserted that it had been working cooperatively with the CFPB for the past 18 months to establish a regulatory framework. The company claimed that an agreement had been reached on May 16, only to be “blindsided” by the lawsuit the following morning.

CEO Travis Holoway expressed his frustration, stating, “Minority innovators were challenged to create new models to address our communities’ financial inequalities. Now that we are doing that, the regulators seem driven by press releases when they should be motivated by true consumer protection and empowering equitable solutions.”

CFPB’s Objectives in the Lawsuit

The CFPB aims to rectify SoLo Funds’ practices and secure refunds for affected customers. The bureau is seeking several remedies, including:

  • Preventing Future Violations: Ensuring that SoLo Funds implements compliant business practices.
  • Monetary Relief: Providing financial restitution to consumers harmed by SoLo’s practices.
  • Disgorgement of Gains: Requiring SoLo to return any ill-gotten gains.
  • Civil Penalties: Imposing additional fines to deter future violations.

The lawsuit seeks to hold SoLo Funds accountable and establish a precedent for other fintech companies, emphasizing the importance of transparent and fair lending practices.

Key Takeaways

  • The CFPB is suing SoLo Funds for allegedly using deceptive digital practices to mislead borrowers.
  • SoLo is accused of misrepresenting loan costs, illegally collecting fees, and making false credit threats.
  • The company has faced multiple legal challenges and settlements with various state regulators.
  • SoLo Funds claims it was working towards regulatory compliance and was surprised by the lawsuit.
  • The CFPB aims to enforce consumer protection laws, secure refunds for customers, and impose financial penalties on SoLo Funds.

This legal action by the CFPB underscores the regulatory scrutiny faced by fintech companies and highlights the ongoing efforts to protect consumers from deceptive financial practices.

Why Microsoft Continues to Lead Google in the Battle for AI Developers

In a cheeky nod to last year’s viral video featuring a montage of every utterance of the word “AI” during the Google I/O keynotes, CEO Sundar Pichai decided to turn the tables this year. As he brought the keynote to a close, Pichai couldn’t resist the temptation to showcase the capabilities of Google’s latest Gemini AI model. With a glimmer of amusement in his eyes, he fed the entire transcript of the keynote into Gemini’s eager algorithms, challenging it to count the number of times “AI” appeared. The result? A staggering 120 mentions, a testament to Google’s unwavering commitment to AI.

Google I/O 2024 Announcements: AI Integration and Competition

Over the past few days, the announcements from Google I/O were nothing short of impressive. Google is going head-to-head with OpenAI and Microsoft, integrating Gemini into a wide range of products. From the introduction of Astra, a multimodal AI tool, to Gemini’s integration across search, photos, and personal assistants, Google is cementing itself as a strong competitor in the AI landscape.

Key Highlights from Google I/O 2024:

  • Astra AI Tool: A new multimodal AI tool.
  • Gemini Integration: Extending across search, photos, and personal assistants.
  • AI Agents and AI Teammate: New AI functionalities aimed at improving user experience.
  • Gemini Models: Range from lightweight Gemini Nano to powerful Gemini 1.5 Pro and Flash.
  • Developer-Focused Tools: Enhancements in Google Colab and other development platforms.

The variety of Gemini’s applications—from the lightweight Gemini Nano to the powerful Gemini 1.5 Pro and Flash—demonstrates Google’s ambitious plans. With innovations like AI Agents, AI Teammate, and the new Gemini app, Google aims to make AI a central part of everyday digital experiences, highlighting its strategy to outpace its rivals.

A Glimpse into the Past: Lessons from Microsoft’s Developer Focus

As I attended the keynote and subsequent breakout sessions, a sense of cognitive dissonance emerged. I was reminded of a distant memory from another tech giant and its eccentric CEO. The year was 2006, and a middle-aged man stood on a brightly illuminated stage, wearing a corporate blue button-down shirt and pleated khaki pants. This man, sweating profusely to the point of absurdity, was Steve Ballmer, the CEO of Microsoft. He pounded his fist into the palm of his hand and began to chant the word “developers” over and over again, whipping the crowd into a frenzy.

Suddenly, I was back in the Google I/O lab session. Developers, a hundred or more, sat in front of me, tuned in. Some struggled to follow along, raising their hands timidly to ask questions but quickly putting them down again as the speaker marched on, leaving no room for stragglers in the AI revolution. It hit me: Google has the wrong mantra. The AI war is being fought for the hearts and minds of developers, but Google is distracted by its consumer core.

Developer Sentiments: Google vs. Microsoft

Despite Google’s impressive technical achievements, developers are the ones who have to build the apps, use the tools, and ultimately decide the victors. Microsoft, in collaboration with OpenAI, remains focused on delighting developers. OpenAI, while known for its consumer-facing products, is actually highly developer-focused, providing comprehensive APIs, extensive documentation, and robust support to make AI integration seamless. Meta, with its open-source Llama 3 models, also positions itself as developer-friendly, allowing anyone to take its AI and build whatever they want, free of charge.

Comparison of AI Developer Focus:

Feature Google Microsoft + OpenAI
Developer Tools Google Colab, Gemini AI Comprehensive APIs, extensive documentation
Integration Consumer products focus Developer-centric, enterprise solutions
Accessibility Technically advanced, but can be intimidating Streamlined, user-friendly
Support Improving, but mixed feedback Robust, highly supportive

This is not the first time Google has fumbled a lead. Developer traction has been the Achilles’ heel of Google Cloud, a marvelous infrastructure technically superior in many respects. Yet, after over a decade and $45 billion invested, Google Cloud still trails behind with an 11% market share, versus Amazon Web Services at 31% and Microsoft Azure at 25%. Google had a three-year head start, but Microsoft now has almost three times the market share. The reason? Developers.

Google’s Renewed Developer Push with AI

Will Google’s AI push finally break Microsoft’s grip on developers? I asked some of them at I/O. As our lab session came to a close, I approached Layla Bouzoubaa, a doctoral candidate at Drexel University, who was particularly attentive during a session on fine-tuning Gemini models.

“I think impressed is a strong word. I’m currently working with large language models (LLMs) in my research, and I am curious to see if I can use Gemini for my specific research topic which involves fine-tuning,” said Bouzoubaa. “I use OpenAI regularly, and it’s been very straightforward. Learning how to use Gemini within Google Colab seemed a little bit more intimidating from a developer’s perspective.”

Despite some challenges, Google has plenty of inroads with developers and significant opportunities. The company even landed OpenAI’s former developer advocate Logan Kilpatrick, which Business Insider called a big win in the war for AI talent.

Bouzoubaa sees fine-tuning smaller language models (SLMs) as something that is quickly becoming essential. “I don’t have the hardware to host my own model. So, before, fine-tuning a large model wasn’t really a possibility.” However, with smaller language models, Bouzoubaa sees the potential for better results with health data. “I know the models aren’t necessarily trained on the health data that I use, so being able to fine-tune is going to enable more AI applications.”

Developer Perspectives: Opportunities and Challenges

A few rows away, another developer was deeply focused on her laptop. “I’m kind of a Google fangirl,” said April Johnson from Extensis, a 30-year-old software company focused on managing licensed fonts for teams. “Microsoft is doing amazing stuff, but I think Google is still ahead with the deeper problems to solve with AI.”

Johnson was enthusiastic about learning with Google tools. “I use Google Colab all the time, and I like the way you can easily set up prototyping. It’s so easy to just prototype on my computer and then move it to the cloud.”

Near the last row of the room, a man with a beard, suggesting he almost certainly has the root password, hunches over a laptop with a thousand stickers from a hundred conferences like this one. He is Nick Bates from Cyber Drive, a startup creating a smartphone specifically for children. “What we’re using ML for right now is to check for patterns of abuse,” Bates says. “We are not creating our own AI models ourselves, but with Gemini being open, we’re looking at ways to leverage it.”

Another developer seated nearby who wished to remain anonymous echoed, “The problem is the big models are really big, so you need a big machine, and you need a lot of users to make it profitable to actually host yourself. Many of the open-source models are so big I don’t have the space to run them on my laptop.”

Microsoft’s Edge with Developers

As I left the Google I/O conference, it was clear that despite Google’s renewed push to win over developers with AI, Microsoft still has the upper hand. The Redmond giant’s deep roots in the enterprise, combined with its strong developer ecosystem and strategic partnership with OpenAI, give it a significant advantage in the battle for the hearts and minds of developers.

Microsoft’s release of the Phi-3 family of small language models, particularly the Phi-3-mini, demonstrates its capability to compete with SLMs. The model’s optimization for various platforms and its support for larger context windows make it an attractive option for developers looking to integrate AI into their projects.

Google, while technically impressive, still has some way to go in terms of making its AI tools more accessible and user-friendly for developers. The feedback from developers at I/O suggests that Google’s AI offerings can be intimidating to approach, especially compared to the more streamlined experience offered by Microsoft and OpenAI.

The Path Forward: Empowering Developers

As the AI race intensifies, it’s becoming increasingly clear that the key to success lies not just in the quality of the AI models themselves, but in the ability to empower developers to harness their potential. Microsoft, with its long-standing developer focus and strategic partnerships, seems poised to maintain its grip on this crucial audience. Google, despite its technical prowess, will need to double down on its efforts to simplify and streamline its AI offerings for developers if it hopes to catch up.

In the end, the winner of the AI race may well be determined by which company can most effectively rally developers to its cause. And right now, it looks like Microsoft is the one still chanting that word.

Scroll to top