11x.ai, a startup focused on automating workflows through AI bots, has raised $24 million in a Series A funding round led by Benchmark. The company, which specializes in creating “automated digital workers,” aims to replace repetitive tasks typically handled by human employees, allowing them to focus on more strategic responsibilities. Originally based in London, 11x.ai has now relocated its headquarters to San Francisco.
Founded in 2022, the company gained attention with its AI agent, Alice, designed for sales teams. It has since expanded its product lineup to include Jordan, an AI phone representative capable of holding intelligent conversations for up to 30 minutes. CEO Hasan Sukkar noted that these agents are a key part of 11x.ai’s strategy to create a suite of virtual employees tailored to different job categories. Sukkar also mentioned that the company is nearing $10 million in annual recurring revenue and counts companies such as Brex, DataStax, and Otter among its clients.
The new funding round comes just a year after 11x.ai raised a $2 million seed round, also led by Project A Ventures. With this latest investment, the company plans to accelerate product development and expand its team, currently numbering 27 employees. The funds will also support the continued growth of its AI worker suite, with additional virtual employees set to be launched in the coming months.
11x.ai’s digital workers are designed to automate tasks that traditionally require human labor, with agents like Jordan trained across 25 languages, including Swedish, Italian, German, and Hebrew. While competition in the AI automation space is fierce, with companies like UiPath, ServiceNow, and Salesforce already established, investors remain optimistic about 11x.ai’s potential. Sarah Tavel, general partner at Benchmark and a board member of 11x.ai, expressed confidence in the company’s future, noting that the demand for AI in workplace automation is rapidly increasing.
Sukkar emphasized that the development of AI agents capable of operating autonomously could transform industries on a scale similar to the internet or cloud computing. He envisions a future where AI agents handle highly skilled tasks without human involvement, drastically enhancing productivity.
While the company’s future plans include further expanding its product lineup and team, 11x.ai will maintain an office in London but relocate most of its key staff to San Francisco, positioning itself closer to the heart of tech innovation.
Luma AI has intensified competition in the AI video generation field by launching its Dream Machine API just hours after Runway introduced its own API. The Dream Machine API allows developers and businesses to create custom applications based on Luma’s video model, which has gained popularity since its debut in June 2024. Prior to this API, users could only generate videos on Luma’s website, but the new API expands access to a broader audience, offering features like text-to-video, image-to-video, and keyframe control.
Dream Machine’s API distinguishes itself from Runway’s by being immediately available, unlike Runway’s which requires interested parties to join a waitlist. Developers at Hugging Face have already created a public demo, showcasing the versatility of Luma’s API. Amit Jain, co-founder and CEO of Luma, emphasized the company’s commitment to making AI-driven visual creation tools accessible globally, allowing more people to experiment with video creation without needing advanced technical skills.
The API provides advanced tools for video creation, including text-to-video, image-to-video transformations, keyframe control, and video looping. These capabilities are designed to simplify the video creation process, allowing developers to integrate them into their applications without relying on complex video editing software. The API can also optimize videos for various platforms by supporting different aspect ratios and offering camera motion control, which guides the scene’s perspective.
Dream Machine’s API pricing is set at $0.32 per million pixels generated, equating to $0.35 for a 5-second video at 720p. This makes the technology accessible to smaller developers while also offering a scalable option for enterprises needing higher rate limits and support. However, Runway’s pricing remains undisclosed, making direct comparisons between the two platforms difficult.
Luma AI has incorporated multi-layered moderation into its API, allowing developers to customize settings based on their user base. The company ensures that user inputs and outputs are not used to train their models unless explicit permission is granted, safeguarding intellectual property rights. Despite concerns from human artists about potential copyright violations in AI models, Luma continues to expand its offerings, positioning the Dream Machine API as a tool to drive innovation in AI video creation.
HyperWrite’s Reflection 70B, a new large language model (LLM) based on Meta’s Llama 3.1-70B Instruct, has been introduced by Matt Shumer, CEO of HyperWrite. Shumer claims that Reflection 70B is now the most powerful open-source AI model in the world, outperforming Meta’s Llama series and competing with top commercial models. What sets Reflection 70B apart is its innovative self-correction capability, allowing the model to recognize and fix its own mistakes during inference, enhancing its accuracy in various tasks.
Reflection 70B has been tested on multiple benchmarks, including MMLU and HumanEval, using LMSys’s LLM Decontaminator to ensure clean, reliable results. These benchmarks demonstrate the model’s ability to surpass its competitors in terms of performance. Users can interact with the model through a demo on a dedicated website, but the overwhelming demand has caused traffic issues, with Shumer’s team working to scale up GPU resources to meet the surge.
A key feature of Reflection 70B is “reflection tuning,” which allows the model to detect errors in its reasoning and correct them in real time. The system uses special tokens to mark reasoning steps, enabling more structured interaction and accurate outputs. This makes it particularly useful for tasks requiring precision, such as comparing numbers or counting characters in words, areas where many other models often fail.
Shumer also revealed plans for an even more powerful model, Reflection 405B, set to release soon, which is expected to outpace top closed-source models like OpenAI’s GPT-4. In addition, HyperWrite intends to integrate Reflection 70B into its main AI writing assistant, with details on this integration expected soon.
The rapid development of Reflection 70B was made possible by Glaive, a startup specializing in generating synthetic data for AI model training. Glaive’s technology allowed the HyperWrite team to quickly create high-quality datasets tailored to their needs, reducing the model’s training time to just three weeks. Shumer credited Glaive for this efficiency, highlighting how synthetic data can accelerate AI development.
HyperWrite, originally founded as Otherside AI in 2020, has grown significantly since its inception. Initially launched as a tool for crafting emails, HyperWrite has expanded to provide a range of AI-driven services, with two million users as of late 2023. The company’s cautious approach to accuracy and safety in AI development mirrors the reflection capabilities embedded in its latest models.
Looking forward, Shumer’s ambition is for Reflection models to set new standards for open-source AI, challenging established players in the industry like OpenAI, Anthropic, and Microsoft. With Reflection 405B on the horizon, the balance of power in generative AI might be shifting towards open-source innovations.
Concerns about an impending fourth AI winter are rising as doubts emerge over whether artificial intelligence will deliver enough practical benefits to justify its high costs. Recent reports from Goldman Sachs and other research bodies highlight this skepticism. However, a solution has been present for some time—engineered intelligence, a concept emphasizing the practical application of AI through engineering principles.
Traditionally, scientific breakthroughs in fields like chemistry and physics are first made in laboratories and then transferred to engineers to develop real-world applications. This process ensures that discoveries are turned into practical solutions. However, AI lacks a similar transition mechanism. Instead of a dedicated discipline for applied AI, organizations often hire data scientists—primarily researchers—to work on developing practical AI solutions. This mismatch has contributed to a significant failure rate, with 87% of AI projects not reaching successful outcomes.
Engineered intelligence, or intelligence engineering, is emerging as a new field focusing on applying AI research in practical settings, much like how chemical or mechanical engineers utilize scientific discoveries. This discipline allows experts, scientists, and engineers to develop intelligent solutions without needing to become data scientists. By reestablishing research-to-engineering pipelines and forming partnerships with academic institutions and technology vendors, industrial organizations are setting the stage for engineered intelligence. This approach mirrors how breakthroughs in other scientific areas are handed off to specialized engineers.
With intelligence engineering, AI research can be transformed into breakthrough applications, yielding tangible value and producing outcomes that might not have been identified by data scientists alone. The process facilitates creating value-driven AI solutions that are feasible and safe for production use, contributing to meaningful advancements across various industries.
To introduce intelligence engineering into an organization, practical experience is crucial. Here’s a five-step guide that differs from traditional AI implementation methods:
By adopting intelligence engineering, organizations can expand their capabilities, moving beyond existing expertise to identify new opportunities. This approach enables safe and practical value creation, both within organizations and across broader ecosystems. As more industries and educational institutions develop programs focused on engineered intelligence, the resulting innovations will unlock unrealized economic and societal benefits, paving the way for new job categories and a surge in value creation.
Brian Evergreen, author of Autonomous Transformation: Creating a More Human Future in the Era of Artificial Intelligence, and Kence Anderson, author of Designing Autonomous AI, both emphasize the potential of engineered intelligence to redefine AI’s impact in practical, everyday applications.