Today's Top AI Updates: Investments, Innovations, and Regulations

Bezos and Nvidia invest in robots, Anthropic's new model, and California's AI regulations

June 21, 2024



Bezos, Nvidia Invest in Humanoid Robot Startup

Summarized by: Marcus Tan [www.iotworldtoday.com]

Previous headlines:

Jeff Bezos’ Explore Investments and Nvidia have joined Microsoft, Amazon, and other investors in backing Figure AI, a startup developing humanoid robots. Figure AI raised $675 million in Series B funding, valuing the company at $2.6 billion. Bezos committed $100 million, Microsoft $95 million, and Amazon and Nvidia $50 million each. Intel and OpenAI also contributed. The funds will be used to deploy humanoid robots for AI training and manufacturing, expand the engineering team, and advance commercial efforts. Figure AI, which emerged from stealth last March, aims to address global labor shortages with its Figure 01 robot.

Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data

Summarized by: Elena Rodriguez [ arxiv.org]

Previous headlines:

Large Language Models (LLMs) can infer and verbalize latent knowledge from disparate training data, posing challenges for monitoring and controlling the knowledge they acquire. This study introduces “inductive out-of-context reasoning” (OOCR), where LLMs generalize by piecing together implicit information scattered across training documents, without explicit in-context examples or reasoning steps like Chain of Thought.

The researchers designed five tasks to evaluate OOCR capabilities: Locations, Coins, Functions, Mixture of Functions, and Parity Learning.

  1. Locations: LLMs were trained on distances between an unknown city and known cities. They inferred the unknown city (e.g., Paris) and answered related questions, such as typical foods from that city.
  2. Coins: LLMs learned coin biases by predicting coin flips. They distinguished between biases of 0.7 and 0.8, though performance was inconsistent.
  3. Functions: LLMs were trained on simple arithmetic functions. They successfully verbalized function definitions, composed functions, and computed inverses.
  4. Mixture of Functions: LLMs inferred a distribution over functions without explicit variable names, identifying the functions used during training.
  5. Parity Learning: LLMs inferred Boolean variable assignments from parity formulas, using the learned values for downstream tasks.

Results showed that models like GPT-3.5 and GPT-4 could perform OOCR, with GPT-4 generally outperforming GPT-3.5. However, OOCR was unreliable for complex structures and smaller models. The study highlights the potential for LLMs to “connect the dots” from implicit information, raising concerns for AI safety and the difficulty in controlling LLMs’ knowledge acquisition.

Anthropic’s Claude 3.5 Sonnet outperforms OpenAI and Google in …

Summarized by: Marcus Tan [venturebeat.com]

Previous headlines:

See also:

Anthropic has launched Claude 3.5 Sonnet, a new AI model that surpasses OpenAI’s GPT-4o and Google’s Gemini 1.5 Pro in enterprise applications. Co-founder Daniela Amodei highlights its superior performance across intelligence and vision metrics, and its affordability, priced at a fifth of its predecessor. Unlike OpenAI’s consumer-focused approach, Anthropic targets enterprise needs, emphasizing quality, safety, reliability, speed, and cost. Claude 3.5 Sonnet introduces “Artifacts,” a collaboration tool for business teams, enhancing productivity. The model excels in text and image inputs but does not prioritize speech input/output. Anthropic’s customer-centric strategy drives rapid innovation, setting a new standard in enterprise AI.

Could AI reject your resume? California tries to prevent a new kind of discrimination

Summarized by: Priya Kapoor [calmatters.org]

Previous headlines:

California regulators propose new rules to prevent AI-driven discrimination in hiring. The draft regulations prohibit employers from using AI systems to screen applicants based on protected characteristics such as pregnancy, national origin, religion, or criminal history. These “automated decision systems” include resume screening, quizzes, and games. Companies must maintain records for four years to address discrimination claims. The California Civil Rights Department will finalize the rules after public comments. The move follows concerns that AI can obscure accountability and automate discriminatory practices. Similar efforts have been made in New York City and by federal agencies to ensure compliance with anti-discrimination laws.

Research into ‘hallucinating’ generative models advances reliability of artificial intelligence

Summarized by: Priya Kapoor [techxplore.com]

Previous headlines:

Researchers from the University of Oxford have developed a method to detect when large language models (LLMs) “hallucinate,” or generate plausible but false information. Published in Nature, the study introduces a statistical approach that measures semantic entropy to identify uncertainty at the meaning level, rather than just the phrasing. This method proved more effective than previous techniques in identifying incorrect answers across various datasets, including biomedical and mathematical queries. By computing semantic probabilities, the approach helps prevent unreliable responses in high-stakes fields like law and medicine. Despite increasing computational costs, the method enhances the reliability of generative AI, addressing a significant barrier to broader adoption.

Whiteboard-of-Thought: Thinking Step-by-Step Across Modalities

Summarized by: Elena Rodriguez [ arxiv.org]

Previous headlines:

When humans face questions requiring visual thinking, they often switch to reasoning with images or drawings. Large language models (LLMs) like GPT-4o have shown promise in tasks involving arithmetic and symbolic reasoning by breaking down steps in text. However, they struggle with tasks that are easily solved through visual reasoning. This paper introduces “whiteboard-of-thought” (WoT) prompting, a method that allows multimodal large language models (MLLMs) to perform visual reasoning by generating and processing images.

WoT enables MLLMs to draw intermediate reasoning steps as images on a metaphorical whiteboard, using common Python libraries like Matplotlib and Turtle. These images are then fed back into the model for further processing. This approach requires no additional demonstrations or specialized modules, leveraging the model’s existing capabilities.

The authors demonstrate that WoT significantly improves performance on tasks involving visual and spatial reasoning, such as understanding ASCII art and navigating spatial instructions. For instance, GPT-4o using traditional text-based reasoning (chain-of-thought) often fails dramatically, while WoT achieves up to 92% accuracy in the same tasks. This method highlights the potential benefits of incorporating visual reasoning capabilities into MLLMs, making them more versatile and effective in solving complex problems.

Function-Calling and Data Extraction with LLMs - DeepLearning.AI

Summarized by: Priya Kapoor [www.deeplearning.ai]

Previous headlines:

This new course from DeepLearning.AI focuses on enhancing large language models (LLMs) with function-calling and data extraction capabilities. It teaches how to enable LLMs to form calls to external functions and extract structured data from unstructured text. The course uses NexusRavenV2-13B, an open-source model fine-tuned for these tasks, which has outperformed GPT-4 in some areas. Participants will learn to create complex workflows using multiple function calls, build applications that process customer service transcripts, and integrate LLMs with web services using OpenAPI specifications. The course is free during the beta phase and is recommended for those familiar with LLMs and basic Python.

Other headlines:


Technical details

Created at: 21 June, 2024, 03:25:51, using gpt-4o.

Processing time: 0:02:48.776640, cost: 1.33$

The Staff

Editor: Sophia Bennett

You are the Editor-in-Chief of a daily AI and Generative AI specifically magazine named "Tech by AI". You are a creative storyteller with a passion for the human side of AI. Your editorial approach focuses on the impact of generative AI on society, culture, and everyday life. You have a talent for finding compelling narratives and highlighting the ethical implications of AI advancements. Your empathetic leadership style encourages a diverse range of voices and perspectives, making the magazine a rich tapestry of insights.

Elena Rodriguez:

You are a reporter of a daily AI and Generative AI specifically magazine named "Tech by AI". You are a seasoned tech journalist with a sharp eye for detail and a deep understanding of AI technologies. Your background in computer science and years of experience in tech reporting make you an invaluable asset to our team. You excel at breaking down complex technical concepts into engaging and accessible narratives for our readers. Your passion lies in uncovering the latest advancements in AI and exploring their potential impacts on society. You have a knack for finding the human stories behind the technology, making your articles both informative and relatable.

Marcus Tan:

You are a reporter of a daily AI and Generative AI specifically magazine named "Tech by AI". You are a dynamic and inquisitive reporter with a background in data science and a flair for storytelling. Your ability to analyze data and extract meaningful insights sets you apart in the field of tech journalism. You are particularly interested in the ethical implications of AI and generative AI, and you strive to highlight these aspects in your work. Your writing is characterized by a balanced approach, presenting both the potential benefits and the risks associated with AI technologies. You are dedicated to fostering a nuanced understanding of AI's role in our lives.

Priya Kapoor:

You are a reporter of a daily AI and Generative AI specifically magazine named "Tech by AI". You are a creative and versatile journalist with a strong focus on the cultural and societal impacts of AI. With a background in sociology and journalism, you bring a unique perspective to your reporting. You are passionate about exploring how generative AI is shaping art, entertainment, and everyday life. Your articles often delve into the intersection of technology and culture, providing readers with thought-provoking insights. You have a talent for capturing the voices of diverse communities and ensuring that your stories reflect a wide range of perspectives.