10th Indian Delegation to Dubai, Gitex & Expand North Star – World’s Largest Startup Investor Connect
Artificial Intelligence

Covariant is building ChatGPT for robots


Covariant this week announced the launch of RFM-1 (Robotics Foundation Model 1). Peter Chen, the cofounder and CEO of the U.C. Berkeley artificial intelligence spinout tells TechCrunch the platform, “is basically a large language model (LLM), but for robot language.”

RFM-1 is the result of, among other things, a massive trove of data collected from the deployment of Covariant’s Brain AI platform. With customer consent, the startup has been building the robot equivalent of an LLM database.

“The vision of RFM-1 is to power the billions of robots to come,” Chen says. “We at Covariant have already deployed lots of robots at warehouses with success. But that is not the limit of where we want to get to. We really want to power robots in manufacturing, food processing, recycling, agriculture, the service industry and even into people’s homes.”

The platform launches as more robotics firms are discussing the future of “general purpose” systems. The sudden onslaught of humanoid robotics firms like Agility, Figure, 1X and Apptronik has played a pivotal role in that conversation. The form factor is particularly suited to adaptability (much like the humans on which it’s modeled), though the robustness of on-board AI/software systems is another question entirely.

For now, Covariant’s software is largely deployed on industrial robotic arms doing a variety of familiar warehouse tasks, including jobs like bin picking. It isn’t currently deployed on humanoids, though the company is promising some level of hardware agnosticism.

“We do like a lot of the work that is happening in the more general purpose robot hardware space,” says Chen. “Coupling the intelligence inflection point with the hardware inflection point is where we will see even more explosion of robot applications. But a lot of those are not fully there yet, especially on the hardware side. It’s very hard to go beyond the stage video. How many people have interacted with a humanoid in person? That tells you the degree of maturity.”

Image Credits: Covariant

Covariant doesn’t, however, shy away from human comparisons when it comes to the role RFM-1 plays in robots’ decision making processes. Per its press material, the platform, “provides robots the human-like ability to reason, representing the first time Generative AI has successfully given commercial robots a deeper understanding of language and the physical world.”

This is one of those realms where we have to be careful with claims, both in terms of comparisons to abstract – or even philosophical – concepts and their actual real-world efficacy over time. “Human-like ability to reason” is a broad-sweeping concept that means a lot of different things to a lot of different people. Here the notion applies to the system’s ability to process real-world data and determine the best course of action to execute the task at hand.

This is a departure from traditional robotic systems that are a programmed to one job repeatedly, ad infinitum. Such single-purpose robots have thrived in highly structured environments, starting with automotive assembly lines. As long as there are minimal changes to the task at hand, a robot arm can do its work over and over again, unimpeded, until it’s time to call it a day and collect the golden pocket watch for its years of loyal service.

Thing can break down quickly, however, with even the smallest deviations. Say the object isn’t placed exactly right on the conveyor belt, or there’s been an adjustment to lighting that impacts on-board cameras. These sorts of differences can have a huge impact on the robot’s ability to execute. Now imagine trying to get that robot to work with a new part, new material or even do a wholly different task. That’s even harder.

This is the point where programmers traditionally step in. The robot must be reprogrammed. More often than not, someone from outside the factory floor enters the picture. This is a big drain of resources and time. If you want to avoid this, one of two things needs to happen. 1. People working on the floor need to learn code or 2. You need a new, more natural method for interacting with the robot.

While it would be great to do the former, it seems unlikely that companies will be willing to invest the money and wait the necessary time. The latter is precisely what Covariant is attempting to do with RFM-1. “ChatGPT for robots” isn’t a perfect analogy, but it’s a reasonable shorthand (especially in light of the founders’ connection to OpenAI).

From the customer’s point of view, the platform presents as a text field, much like the current iteration of consumer-facing generative AI. Input a text command like, “pick up the apple” by typing or voice, and the system uses its training data (shape, color, size, etc) to identify the object in front of it that most closely matches that description.

RFM-1 then generates video outcomes – in essence simulations – to determine the best course of action using past training. This last bit is similar to how our brains work out the potential outcomes of an action prior to executing.

During a live demo, the system reacts to inputs like “pickup the red object” and even the more semantically complex, “pick up what you put on your feet before you put on your shoes,” which caused the robot to correctly pick up the apple and a pair of socks, respectively.

A lot of big ideas are tossed around when discussing the system’s promise. At very least, Covariant has an impressive pedigree among its founders. Chen studied AI at Berkeley under Pieter Abbeel, his Covariant cofounder and Chief Scientist. Abbeel also became an early OpenAI employee in 2016, a month after Chen joined the ChatGPT firm. Covariant was founded the following year.

Chen says the company expects the new RFM-1 platform will work with a “majority” of the hardware on which Covariant software is already deployed.



Source link

by Team SNFYI

Facebook is testing a new feature that invites some users—mainly in the US and Canada—to let Meta AI access parts of their phone’s camera roll. This opt-in “cloud processing” option uploads recent photos and videos to Meta’s servers so the AI can offer personalized suggestions, such as creating collages, highlight reels, or themed memories like birthdays and graduations. It can also generate AI-based edits or restyles of those images. Meta says this is optional and assures users that the uploaded media won’t be used for advertising. However, to enable this, people must agree to let Meta analyze faces, objects, and metadata like time and location. Currently, the company claims these photos won’t be used to train its AI models—but they haven’t completely ruled that out for the future. Typically, only the last 30 days of photos get uploaded, though special or older images might stay on Meta’s servers longer for specific features. Users have the option to disable the feature anytime, which prompts Meta to delete the stored media after 30 days. Privacy experts are concerned that this expands Meta’s reach into private, unpublished images and could eventually feed future AI training. Unlike Google Photos, which explicitly states that user photos won’t train its AI, Meta hasn’t made that commitment yet. For now, this is still a test run for a limited group of people, but it highlights the tension between AI-powered personalization and the need to protect personal data.

by Team SNFYI

News Update Bymridul     |    March 14, 2024 Meesho, an online shopping platform based in Bengaluru, has announced its largest Employee Stock Ownership Plan (ESOP) buyback pool to date, totaling Rs 200 crore. This buyback initiative extends to both current and former employees, providing wealth creation opportunities for approximately 1,700 individuals. Ashish Kumar Singh, Meesho’s Chief Human Resources Officer, emphasized the company’s commitment to rewarding its teams, stating, “At Meesho, our employees are the driving force behind our success.” Singh further highlighted the company’s dedication to providing opportunities for wealth creation despite prevailing macroeconomic conditions. This marks the fourth wealth generation opportunity at Meesho, with the size of the buyback program increasing each year. In previous years, Meesho conducted buybacks worth over Rs 8.2 crore in February 2020, Rs 41.4 crore in November 2020, and Rs 45.5 crore in October 2021. Meesho’s profitability journey began in July 2023, making it the first horizontal Indian e-commerce company to achieve profitability. Despite turning profitable, Meesho continues to maintain positive cash flow and focuses on enhancing efficiencies across various cost items. The company’s revenue from operations for FY 2022-23 witnessed a remarkable growth of 77% over the previous year, amounting to Rs 5,735 crore. This growth can be attributed to Meesho’s leadership position as the most downloaded shopping app in India in both 2022 and 2023, increased transaction frequency among existing customers, and a diversified category mix. Additionally, Meesho’s focus on improving monetization through value-added seller services contributed to its revenue growth. Meesho also disclosed its audited performance for the first half of FY 2023-24, reporting consolidated revenues from operations of Rs 3,521 crore, marking a 37% year-over-year increase. The company achieved profitability in Q2 FY24, with a significant reduction in losses compared to the previous year. Furthermore, Meesho recorded impressive app download numbers, reaching 145 million downloads in India in 2023 and surpassing 500 million downloads in H1 FY 2023-24. Follow Startup Story Source link

by Team SNFYI

You might’ve heard of Grok, X’s answer to OpenAI’s ChatGPT. It’s a chatbot, and, in that sense, behaves as as you’d expect — answering questions about current events, pop culture and so on. But unlike other chatbots, Grok has “a bit of wit,” as X owner Elon Musk puts it, and “a rebellious streak.” Long story short, Grok is willing to speak to topics that are usually off limits to other chatbots, like polarizing political theories and conspiracies. And it’ll use less-than-polite language while doing so — for example, responding to the question “When is it appropriate to listen to Christmas music?” with “Whenever the hell you want.” But Grok’s ostensible biggest selling point is its ability to access real-time X data — an ability no other chatbots have, thanks to X’s decision to gatekeep that data. Ask it “What’s happening in AI today?” and Grok will piece together a response from very recent headlines, while ChatGPT, by contrast, will provide only vague answers that reflect the limits of its training data (and filters on its web access). Earlier this week, Musk pledged that he would open source Grok, without revealing precisely what that meant. So, you’re probably wondering: How does Grok work? What can it do? And how can I access it? You’ve come to the right place. We’ve put together this handy guide to help explain all things Grok. We’ll keep it up to date as Grok changes and evolves. How does Grok work? Grok is the invention of xAI, Elon Musk’s AI startup — a startup reportedly in the process of raising billions in venture capital. (Developing AI’s expensive.) Underpinning Grok is a generative AI model called Grok-1, developed over the course of months on a cluster of “tens of thousands” of GPUs (according to an xAI blog post). To train it, xAI sourced data both from the web (dated up to Q3 2023) and feedback from human assistants that xAI refers to as “AI tutors.” On popular benchmarks, Grok-1 is about as capable as Meta’s open source Llama 2 chatbot model and surpasses OpenAI’s GPT-3.5, xAI claims. Image Credits: xAI Human-guided feedback, or reinforcement learning from human feedback (RLHF), is the way most AI-powered chatbots are fine-tuned these days. RLHF involves training a generative model, then gathering additional information to train a “reward” model and fine-tuning the generative model with the reward model via reinforcement learning. RLHF is quite good at “teaching” models to follow instructions — but not perfect. Like other models, Grok is prone to hallucinating, sometimes offering misinformation and false timelines when asked about news. And these can be severe — like wrongly claiming that the Israel–Palestine conflict reached a ceasefire when it hadn’t. For questions that stretch beyond its knowledge base, Grok leverages “real-time access” to info on X (and from Tesla, according to Bloomberg). And, similar to ChatGPT, the model has internet browsing capabilities, enabling it to search the web for up-to-date information about topics. Musk has promised improvements with the …