Author: yasmeeta

Research from Harvard, MIT, and Wharton highlights the risks of depending on junior staff to train AI systems.

As businesses rapidly incorporate artificial intelligence (AI) into their operations, the prevailing belief is that younger, technologically adept employees will spearhead educating their senior managers on effectively harnessing these advanced tools. However, a recent study challenges this notion, especially concerning the use of generative AI technologies.

Study Details and Key Participants

The study was a collaborative effort involving scholars from prestigious institutions such as Harvard Business School, MIT, and Wharton, in partnership with Boston Consulting Group. The research focused on the interactions and experiences of junior employees with generative AI systems, particularly GPT-4, in real-world business scenarios.

Unexpected Findings from Junior Consultants

Contrary to expectations, the study revealed that junior employees, often presumed to be tech-savvy, might not be the best resources for guiding senior professionals in the effective use of emerging technologies like generative AI. The findings showed that the risk mitigation strategies proposed by these junior consultants frequently contradicted expert advice and lacked a deep understanding of AI’s capabilities.

Key Insights from the Study:

  • Junior Consultants’ Tactics: The research highlighted that the strategies suggested by junior employees to alleviate seniors’ concerns about AI risks were often misaligned with expert recommendations. These tactics were more about altering human behavior than enhancing the AI system’s design.
  • Focus on Short-term Solutions: Many recommendations were project-specific rather than aimed at broader organizational or industry-wide applications, suggesting a narrow scope of understanding.

In-depth Analysis of the Research Findings

1. Limited Technical Expertise The study found that junior consultants typically had minimal technical expertise in AI. Their recommendations were based more on general knowledge and less on a technical understanding of AI systems like GPT-4.

2. Risk Mitigation Approaches Junior employees tended to focus on immediate, surface-level solutions rather than systemic changes or in-depth strategies that could be more beneficial in the long run.

Challenges in Adopting Generative AI in Business

The rapid evolution of generative AI technologies presents significant challenges and opportunities for businesses. These AI systems can perform tasks such as engaging in detailed dialogues, responding to follow-up questions, and assisting in writing, analysis, and coding tasks. However, the study underscores the necessity of comprehensive AI governance and the need for expert input at all organizational levels.

Navigating AI Implementation Challenges:

  • Top-down Governance: Effective AI implementation requires informed leadership rather than relying solely on the knowledge of digital natives within the organization.
  • Expert Involvement: Incorporating AI experts into strategic planning and implementation processes is crucial to address potential risks and optimize AI usage.

Moving Forward: Recommendations for Effective AI Adoption

The findings advocate for a structured approach to AI adoption in corporate settings:

  • Upskilling Programs: Develop extensive training programs to enhance the AI competence of employees across all levels.
  • Leadership Roles: Senior professionals should take proactive roles in understanding and integrating new technologies to lead their teams effectively.
  • Future-proofing Strategies: Businesses need to anticipate future technological advancements and their potential impacts on industry and internal operations.

Summary in Bullet Points

  • Study Collaboration: Involvement of top academic institutions and Boston Consulting Group.
  • Key Finding: Junior employees may not be ideal mentors for senior staff in AI adoption.
  • Recommendations: Emphasize top-down governance, expert involvement, and comprehensive training.

This extensive study not only highlights a critical gap in the assumed capabilities of junior employees concerning AI but also sets the stage for rethinking how businesses should approach the integration of these powerful technologies into their workflows. Senior leaders are encouraged to take a more active role in understanding and guiding AI initiatives to ensure that their organizations can fully leverage AI’s capabilities responsibly and effectively.

Crystal Sonic Emerges Victorious in Lam Capital Venture Competition, Securing $250,000 Investment

At the Enabling Future Semiconductors event held at Lam Research’s headquarters in Fremont, California, Crystal Sonic emerged as the winner of the third Lam Capital Venture Competition, securing a $250,000 investment from Lam Capital. The event, which focused on exploring novel semiconductor technology and manufacturing technologies, highlighted a dozen different startups who were the finalists among 70 applicants in the semiconductor-focused competition.

Winner and Runner-Up
Crystal Sonic, a chip-related startup, won the competition by showcasing its innovative technology that helps semiconductor manufacturers reduce waste and cost by enabling thin device lift-off and substrate re-use. This technology allows for the separation of parts of the substrate and reusing it, thereby reducing the waste of chip-making materials. The runner-up, Lidrotec, makes tools for cutting semiconductor chips.

Lam Capital Venture Competition
The Lam Capital Venture Competition aims to invest in disruptive companies that advance the semiconductor ecosystem through next-generation industrial automation, technology, and product innovation. This is the third annual event for the competition, with the first event taking place before the pandemic in 2019. The competition is a significant platform for Lam Research, a 44-year-old semiconductor equipment manufacturing company, to nurture collaboration with customers and the wider chip ecosystem.

Judges and Applications
The six judges for the competition included Weili Dai, serial entrepreneur and cofounder of Marvell and a frequent investor in semiconductor startups including Silicon Box. Other judges included Rene Do, senior investment director, SK Hynix; Ben Haskell, investment director, Lam Capital; Amir Salek, senior managing director, Cerberus Capital Management; Vera Schroeder, partner, Safar Partners; and Lucas Tsai, senior director, market development and emerging business, TSMC North America.Many of the applicants had a heavy emphasis on AI, particularly as a way to counteract growing costs, increasing technological complexity, and sustainability issues. Lam Research has been investing in chip-related startups for years and has made 20 investments so far, with $1 million to $10 million going into each startup.

Impact on Lam Research
The competition is beneficial for Lam Research, as it enables the company to stay ahead of the curve in terms of innovation and to nurture collaboration with customers and the wider chip ecosystem. As Audrey Charles, senior vice president for corporate strategy at Lam Research, noted, “We can only be successful based on the types of innovation we see today.”

Lam Capital Venture Competition Winners

Year Winner Runner-Up Investment Amount
2019
2022
2024 Crystal Sonic Lidrotec $250,000

Key Points

  • Crystal Sonic: Helps semiconductor manufacturers reduce waste and cost by enabling thin device lift-off and substrate re-use.
  • Lidrotec: Makes tools for cutting semiconductor chips.
  • Lam Capital Venture Competition: Aims to invest in disruptive companies that advance the semiconductor ecosystem through next-generation industrial automation, technology, and product innovation.
  • Judges: Include Weili Dai, Rene Do, Ben Haskell, Amir Salek, Vera Schroeder, and Lucas Tsai.
  • Applications: Many applicants emphasized AI as a way to counteract growing costs, increasing technological complexity, and sustainability issues.
  • Lam Research Investments: Has made 20 investments so far, with $1 million to $10 million going into each startup

Nvidia boosts AI assistants on PCs with new GeForce RTX features.

Nvidia made a splash at Computex 2024, showcasing a suite of new RTX AI technologies designed to revolutionize AI assistants, digital humans, and content creation on laptops. Here’s a breakdown of the key announcements:

Project G-Assist: The AI Assistant of the Future

Project G-Assist is an RTX-powered AI assistant demo that offers context-aware help for PC games and applications. It leverages generative AI to understand players’ needs and provide assistance within the game itself. Here’s how it works:

Feature Description
Contextual Awareness G-Assist analyzes voice or text inputs and game screen information to understand the situation.
Large Language Model (LLM) A powerful LLM linked to a game knowledge database processes the information and generates tailored responses.
Personalized Support G-Assist personalizes its responses based on the player’s current game session.

Benefits of Project G-Assist:

  • Improved Gameplay: Gain insights into creatures, items, lore, objectives, and boss strategies.
  • Streamlined Workflows: Receive assistance with complex creative tasks within applications.
  • Performance Optimization: G-Assist can configure your system for optimal performance and efficiency.

Nvidia ACE Comes to RTX AI PCs

Nvidia is bringing its digital human development platform, Nvidia ACE, to RTX AI laptops and workstations. This allows developers to create lifelike digital humans with capabilities like natural language understanding, speech synthesis, and facial animation.

Key benefits of Nvidia ACE on RTX AI PCs:

  • Reduced Deployment Times: Leverage Nvidia NIM (inference microservices) to streamline development.
  • High-Quality Inference: Enables local processing of complex AI tasks for smooth performance.

Collaboration with Microsoft on Windows Copilot

Microsoft and Nvidia are teaming up to bring new AI capabilities to Windows applications. This collaboration will provide developers with access to GPU-accelerated small language models (SLMs) for tasks like content summarization, generation, and automation.

Faster and Smaller AI Models with RTX AI Toolkit

The RTX AI Toolkit is a suite of tools and resources designed to empower developers to build and deploy AI models specifically for RTX AI PCs. Here’s what it offers:

  • Model Customization: Open-source QLoRa tools allow for tailoring pre-trained models.
  • Model Optimization: Nvidia TensorRT optimizes models for faster performance and reduced memory usage.
  • Deployment Support: The Nvidia AI Inference Manager simplifies AI integration into PC applications.

Benefits of RTX AI Toolkit for Developers:

  • Faster Performance: Achieve up to 4x faster performance compared to pre-trained models.
  • Reduced Model Size: Models can be optimized to consume up to 3x less RAM.
  • Simplified Integration: The AI Inference Manager streamlines the integration process.

Integration with Popular Creative Applications

Several software partners are integrating components of the RTX AI Toolkit into their applications. This will unlock new possibilities for AI-powered content creation.

Examples of RTX AI Toolkit Integration:

  • Adobe Photoshop: Leverages TensorRT for faster performance and AI-powered capabilities.
  • ComfyUI: RTX acceleration delivers a 60% performance improvement over the current version.
  • Automatic1111: RTX acceleration streamlines workflows for Stable Diffusion users.

RTX Remix: A Boon for Modders

Nvidia is expanding the capabilities of RTX Remix, its modding platform for classic games. This update allows modders to create even more stunning remasters:

  • Open-Source Toolkit: More of the platform is becoming open-source, empowering the modding community.
  • Expanded Functionality: New features include streamlined asset replacement, scene relighting, and enhanced AI texture tools.
  • REST API and SDK: Integration with other content creation tools and games is now possible.

Benefits of RTX Remix for Modders:

  • Simplified Workflows: Streamlined asset replacement and scene relighting.
  • Enhanced Capabilities: New AI texture tools and support for more file formats.
  • Greater Integration: Live link with other tools and deployable renderer for broader use.

RTX Video Goes Beyond Browsers

Previously available only in web browsers, Nvidia RTX Video, the AI-powered video upscaling feature, is now available as an SDK for developers. This allows them to integrate AI for upscaling, sharpening, and HDR conversion within their applications.

Future of RTX Video:

  • Video Editing Software: Upscale video to 4K and convert SDR to HDR in DaVinci Resolve and Filmora.
  • VLC Media Player: Soon to offer RTX Video HDR capabilities for existing super-resolution features.

McKesson and Merck Invest in Atropos Health’s $33M Funding Round to Boost AI-Driven Drug Development

Silicon Valley-based Atropos Health has successfully raised $33 million in a Series B funding round, marking a significant step forward in its mission to integrate AI-powered, personalized real-world evidence into healthcare decision-making.

In a recent announcement, Atropos Health, a pioneer in generating personalized real-world evidence, disclosed a substantial $33 million acquisition in Series B funding. This investment round featured prominent contributions from healthcare behemoths like McKesson, Merck, and Cencora Ventures, indicating a robust industry endorsement of Atropos’ innovative approach to healthcare.

The funds are earmarked for a strategic expansion aimed at enhancing the company’s operational capacity and doubling down on critical initiatives. These include a deeper penetration into the life sciences sector, broadening channel partnerships in value-based care and oncology, and expanding its network of data partners to enrich its evidence base.

Brigham Hyde, PhD, CEO and co-founder of Atropos Health, expressed his enthusiasm in a VentureBeat interview, stating, “We’re on a mission to bring personalized evidence for care to everybody in the world. This funding is a pivotal step in that journey. Specifically, we’ll be focusing on reinforcing our strategic initiatives, continuing our successful launch in life sciences, and enhancing our partnerships, particularly in value-based and specialty care oncology.”

Atropos Health is not just another player in the healthcare field; it is a trailblazer aiming to close the pervasive “evidence gap” in medical decision-making. The company’s flagship technology, Geneva OS, harnesses artificial intelligence (AI) and automation to rapidly generate clinical-grade evidence from real-world data. This platform, which has been developed over nearly a decade of research at Stanford University, powers applications such as the generative AI assistant, ChatRWD.

The technology enables clinicians, researchers, and other healthcare stakeholders to swiftly access reliable clinical evidence, personalized to specific patient populations—a capability often missing in current healthcare practices. Dr. Hyde highlighted a concerning statistic in his interview: “Only about 14% of daily medical decisions have any high-quality evidence behind them. Our goal is to use high-quality data, analyzed correctly, to fill this evidence gap.”

The central mission of Atropos is to provide clinicians with easy access to personalized evidence, thereby enhancing patient outcomes. Dr. Hyde used the example of heart failure patients to illustrate the need for tailored evidence that caters to subpopulations with unique characteristics and comorbidities, which could lead to more effective treatments and cost control.

Atropos’ applications extend beyond clinical decision-making. The company collaborates with pharmaceutical leaders, including Janssen, to expedite drug development by leveraging real-world evidence for clinical trial design, patient recruitment, and more. Dr. Hyde even suggested that the platform could simulate clinical trials, potentially revolutionizing the way pharmaceutical research is conducted by reducing cycle times and de-risking trials.

Despite the excitement surrounding large language models (LLMs) and generative AI, Atropos prioritizes building trust through methodological rigor and transparency. Dr. Hyde expressed concerns about the “hallucination rates” in current AI models and emphasized that Geneva OS ensures clinical-grade quality and transparency, backed by a decade of publications.

Strategic Use of Series B Funding

Initiative Objective Expected Impact
Expansion in Life Sciences Enhance presence and partnerships in life sciences Broaden application of real-world evidence in R&D
Channel Partnerships Growth Focus on value-based care and oncology Improve treatment strategies and patient outcomes
Data Network Expansion Increase the network of data partners Enrich the quality and diversity of clinical evidence

With a fresh influx of capital and a roster of strategic backers, Atropos is poised to bring its vision of personalized, automated clinical evidence to the global healthcare landscape. “Evidence is the currency of value in healthcare,” Dr. Hyde posited. “What if I could give doctors more evidence, more personalized, so they make better decisions? Fundamentally, we’re trying to move the world to a point where all patients and all providers have access to quality, personalized evidence for their decision-making.”

This bold vision by Atropos Health not only promises to transform patient care but also positions the company as a frontrunner in the integration of AI and healthcare. As they continue to bridge the evidence gap, the future of healthcare looks promisingly precise, personalized, and powered by artificial intelligence.

Y Combinator’s Garry Tan Backs AI Rules, Warns of Monopolies

Garry Tan, the influential president and CEO of the startup incubator Y Combinator, recently addressed an audience at The Economic Club of Washington, D.C., emphasizing the need for regulatory frameworks in the rapidly evolving field of artificial intelligence (AI). Tan’s comments come at a critical juncture as AI technologies continue to permeate various aspects of societal and economic activities.

The Argument for Regulation

During a detailed interview with Teresa Carlson, a board member at General Catalyst, Tan shared his views on a multitude of topics, from entry paths into Y Combinator to the broader implications of AI developments. He highlighted the unprecedented opportunities currently available in the technology sector, stating, “There is no better time to be working in technology than right now.”

Tan voiced his support for the efforts by the National Institute of Standards and Technology (NIST) to create a risk mitigation framework for generative AI (GenAI). He believes that the Executive Order (EO) by the Biden Administration aligns well with necessary steps towards responsible AI deployment. The NIST’s framework includes several important guidelines:

  • Compliance with Existing Laws: GenAI must adhere to laws governing data privacy and copyright.
  • Disclosure Requirements: Companies must inform end users about their use of GenAI technologies.
  • Prohibitions on Harmful Content: Regulations should prevent GenAI from generating or distributing harmful materials such as child sexual abuse content.

Further, President Biden’s executive order mandates AI companies to share safety data with governmental bodies and ensures that small developers have equitable access to the technology market.

Concerns Over State Legislation

Despite his general support for federal efforts, Tan expressed concerns about AI-related bills progressing through state legislatures, particularly in California and San Francisco. One controversial bill, introduced by California State Senator Scott Wiener, could potentially allow the state attorney general to sue AI companies if their products cause harm. This bill, among others, has stirred significant debate within the tech community regarding its implications on innovation and business operations.

Key Regulatory Proposals Discussed by Garry Tan

Regulation Aspect Description Potential Impact
NIST Framework Guidelines for risk mitigation in GenAI applications Enhances safety and compliance standards
Biden’s Executive Order Comprehensive directives for AI deployment and oversight Aims for balanced growth and safety
California Legislative Bills Potential legal actions against harmful AI products Raises concerns about innovation stifling

The Balance Between Innovation and Control

Tan highlighted the delicate balance that needs to be maintained between fostering technological innovation and mitigating potential harms. He cited UK AI expert Ian Hogarth’s approach, which is thoughtful about maintaining a balance between limiting the concentration of power within the AI sector and encouraging innovative progress. Hogarth, a former YC entrepreneur, is part of an AI model taskforce in the UK, working towards viable policy solutions.

Y Combinator’s Ethical Stance

Tan shared insights into Y Combinator’s internal decision-making processes regarding AI startups. He emphasized that the incubator only funds startups that align with positive societal impacts. “If we don’t agree with a startup’s mission or its potential effects on society, YC just doesn’t fund it,” Tan explained. This cautious approach has led them to avoid backing several companies after reviewing their potential implications through media reports and internal evaluations.

AI Industry Challenges and the Future Landscape

The discussion also touched upon recent industry controversies, including high-profile issues at OpenAI and Meta. These instances have sparked a broader debate on the ethical responsibilities of AI firms and the transparency required in their operations.

  • OpenAI’s Responsibility Team: Recently, it was reported that OpenAI might be scaling back its AI responsibility team, raising questions about the commitment to ethical AI development.
  • Voice Mimicry Issues: OpenAI faced criticism for using a voice resembling actress Scarlett Johansson’s in demos, without her consent, leading to further scrutiny of its practices.
  • Meta’s AI Advisory Council Composition: Meta’s decision to form an AI advisory council predominantly comprising white men has also drawn criticism for not reflecting diversity.

The Vision for AI’s Future

Looking ahead, Tan is optimistic about the potential for AI to enable a diverse range of consumer choices and empower founders. He envisages a future where AI does not lead to monopolistic practices but instead fosters a vibrant landscape of varied solutions accessible to billions globally.

In conclusion, while Tan acknowledges the potential dangers of AI, his primary concern remains the risk of a monopolistic concentration of power within the industry, which could lead to restrictive practices detrimental to innovation and consumer choice. His vision for AI emphasizes both caution and enthusiasm, aiming for a future where technology serves humanity broadly and equitably.

Scroll to top