Transcript
A (0:00)
Foreign. Welcome to Reshaping Workflows with Dell Pro
B (0:07)
Precision and Nvidia, where innovation meets real world impact in high performance computing.
A (0:20)
Welcome back to another exciting episode of Reshaping Workflows with Delpro Precision and Nvidia RTX GPUs. I'm your host Logan Mahler and today you've got rishi mini part 2. But I'm super excited. Like before we even kicked off this episode, Remy, who's my guest, well, I'll let him introduce himself in a second. Like kind of showed me a few things. It's going to be a great conversation. You can tell I'm, I'm super excited. So with that, Remy, thank you for joining us. Can you take a couple of minutes, kind of explain your role, you know, as working kind of at Hugging Face through Pollen Robotics, you know, as kind of a partner, what you do as it relates to Rishi Mini and then just your background in general.
B (1:02)
My name is, I've been a robotics engineer for close to 15 years now. I worked on many, many projects before working on this guy I worked on a full humanoid with arms doing grasping and stuff, some autonomous navigation too. And yeah, the last six months or so I went into this interactional direction. How do you express emotion through motion? How do you make interaction with a non technical user fulfilling? How do you integrate the novelties of AI that changes every week into both the robot and the workflows? We're trying to empower people to build on top of this. And when you see that now we've got AI agents that are capable of coding one shot in entire apps almost. This is to me incredible transformation. So yeah, it's very exciting times. And yeah, I've been, I've been exploring many things, mostly doing software.
A (2:01)
Very cool. Well, I'm excited. So you know, if you're watching the, the video version of this, you can see if you're listening to podcast, we'll describe behind it. But you have kind of a Rishi Mini behind you. So we've talked kind of heavily. One of the things with, with you know, Rishi, if you buy, you know, you buy a Reishi Mini, there's all types of like, I'm going to call them apps but like basically kind of software kits that you can install that, you know, there's cool little ones that are, you know, you can play like red light, green light and like do a lot of really great stuff. But the big one that we want to talk about is kind of the emotions. So first, before we get into anything, can you talk a little bit about the emotions part within Rishi, what it's designed to do, and then we can go into kind of showing a few examples.
B (2:46)
So the emotions is one of the first things that we did on this robot. And the technical elements of it is quite simple. We have an Charlotte at home that's very talented at this. And we tailor operated the robot and recorded sound at the same time. And they've recorded like 80 emotions. And this robot, it has six degrees of freedom for the head, can move forward, left, right, up, up and down, then 3 degrees of rotation, the body rotates and two antennas. So you've got a lot of expressivity. And with the right sound, it's incredible the amount of emotion you can display. This has became like a foundation block that's reusable in any app when you start. An app is generally much better when an emotion plays or when you want to react to something that the user does. So this is the v0 of emotions. These are pre recorded. We have several projects that still in their infancy to generate new motion automatically programmatically. I can show you an example the that one of my students did like last week. Very interesting. And then on top of that, I said, the main thing that we've done is connecting an LLM to it. Think it like ChatGPT when you have it on your phone in the conversation mode. But instead of just being a phone, it can control the robot. It can decide to play an emotion, to wiggle its head as it moves. It can look at you and follow you, and then the other hand will decide which emotion to play at which moment.
