下載App 希平方
攻其不背
App 開放下載中
下載App 希平方
攻其不背
App 開放下載中
IE版本不足
您的瀏覽器停止支援了😢使用最新 Edge 瀏覽器或點選連結下載 Google Chrome 瀏覽器 前往下載

免費註冊
! 這組帳號已經註冊過了
Email 帳號
密碼請填入 6 位數以上密碼
已經有帳號了?
忘記密碼
! 這組帳號已經註冊過了
您的 Email
請輸入您註冊時填寫的 Email,
我們將會寄送設定新密碼的連結給您。
寄信了!請到信箱打開密碼連結信
密碼信已寄至
沒有收到信嗎?
如果您尚未收到信,請前往垃圾郵件查看,謝謝!

恭喜您註冊成功!

查看會員功能

註冊未完成

《HOPE English 希平方》服務條款關於個人資料收集與使用之規定

隱私權政策
上次更新日期:2014-12-30

希平方 為一英文學習平台,我們每天固定上傳優質且豐富的影片內容,讓您不但能以有趣的方式學習英文,還能增加內涵,豐富知識。我們非常注重您的隱私,以下說明為當您使用我們平台時,我們如何收集、使用、揭露、轉移及儲存你的資料。請您花一些時間熟讀我們的隱私權做法,我們歡迎您的任何疑問或意見,提供我們將產品、服務、內容、廣告做得更好。

本政策涵蓋的內容包括:希平方學英文 如何處理蒐集或收到的個人資料。
本隱私權保護政策只適用於: 希平方學英文 平台,不適用於非 希平方學英文 平台所有或控制的公司,也不適用於非 希平方學英文 僱用或管理之人。

個人資料的收集與使用
當您註冊 希平方學英文 平台時,我們會詢問您姓名、電子郵件、出生日期、職位、行業及個人興趣等資料。在您註冊完 希平方學英文 帳號並登入我們的服務後,我們就能辨認您的身分,讓您使用更完整的服務,或參加相關宣傳、優惠及贈獎活動。希平方學英文 也可能從商業夥伴或其他公司處取得您的個人資料,並將這些資料與 希平方學英文 所擁有的您的個人資料相結合。

我們所收集的個人資料, 將用於通知您有關 希平方學英文 最新產品公告、軟體更新,以及即將發生的事件,也可用以協助改進我們的服務。

我們也可能使用個人資料為內部用途。例如:稽核、資料分析、研究等,以改進 希平方公司 產品、服務及客戶溝通。

瀏覽資料的收集與使用
希平方學英文 自動接收並記錄您電腦和瀏覽器上的資料,包括 IP 位址、希平方學英文 cookie 中的資料、軟體和硬體屬性以及您瀏覽的網頁紀錄。

隱私權政策修訂
我們會不定時修正與變更《隱私權政策》,不會在未經您明確同意的情況下,縮減本《隱私權政策》賦予您的權利。隱私權政策變更時一律會在本頁發佈;如果屬於重大變更,我們會提供更明顯的通知 (包括某些服務會以電子郵件通知隱私權政策的變更)。我們還會將本《隱私權政策》的舊版加以封存,方便您回顧。

服務條款
歡迎您加入看 ”希平方學英文”
上次更新日期:2013-09-09

歡迎您加入看 ”希平方學英文”
感謝您使用我們的產品和服務(以下簡稱「本服務」),本服務是由 希平方學英文 所提供。
本服務條款訂立的目的,是為了保護會員以及所有使用者(以下稱會員)的權益,並構成會員與本服務提供者之間的契約,在使用者完成註冊手續前,應詳細閱讀本服務條款之全部條文,一旦您按下「註冊」按鈕,即表示您已知悉、並完全同意本服務條款的所有約定。如您是法律上之無行為能力人或限制行為能力人(如未滿二十歲之未成年人),則您在加入會員前,請將本服務條款交由您的法定代理人(如父母、輔助人或監護人)閱讀,並得到其同意,您才可註冊及使用 希平方學英文 所提供之會員服務。當您開始使用 希平方學英文 所提供之會員服務時,則表示您的法定代理人(如父母、輔助人或監護人)已經閱讀、了解並同意本服務條款。 我們可能會修改本條款或適用於本服務之任何額外條款,以(例如)反映法律之變更或本服務之變動。您應定期查閱本條款內容。這些條款如有修訂,我們會在本網頁發佈通知。變更不會回溯適用,並將於公布變更起十四天或更長時間後方始生效。不過,針對本服務新功能的變更,或基於法律理由而為之變更,將立即生效。如果您不同意本服務之修訂條款,則請停止使用該本服務。

第三人網站的連結 本服務或協力廠商可能會提供連結至其他網站或網路資源的連結。您可能會因此連結至其他業者經營的網站,但不表示希平方學英文與該等業者有任何關係。其他業者經營的網站均由各該業者自行負責,不屬希平方學英文控制及負責範圍之內。

兒童及青少年之保護 兒童及青少年上網已經成為無可避免之趨勢,使用網際網路獲取知識更可以培養子女的成熟度與競爭能力。然而網路上的確存有不適宜兒童及青少年接受的訊息,例如色情與暴力的訊息,兒童及青少年有可能因此受到心靈與肉體上的傷害。因此,為確保兒童及青少年使用網路的安全,並避免隱私權受到侵犯,家長(或監護人)應先檢閱各該網站是否有保護個人資料的「隱私權政策」,再決定是否同意提出相關的個人資料;並應持續叮嚀兒童及青少年不可洩漏自己或家人的任何資料(包括姓名、地址、電話、電子郵件信箱、照片、信用卡號等)給任何人。

為了維護 希平方學英文 網站安全,我們需要您的協助:

您承諾絕不為任何非法目的或以任何非法方式使用本服務,並承諾遵守中華民國相關法規及一切使用網際網路之國際慣例。您若係中華民國以外之使用者,並同意遵守所屬國家或地域之法令。您同意並保證不得利用本服務從事侵害他人權益或違法之行為,包括但不限於:
A. 侵害他人名譽、隱私權、營業秘密、商標權、著作權、專利權、其他智慧財產權及其他權利;
B. 違反依法律或契約所應負之保密義務;
C. 冒用他人名義使用本服務;
D. 上載、張貼、傳輸或散佈任何含有電腦病毒或任何對電腦軟、硬體產生中斷、破壞或限制功能之程式碼之資料;
E. 干擾或中斷本服務或伺服器或連結本服務之網路,或不遵守連結至本服務之相關需求、程序、政策或規則等,包括但不限於:使用任何設備、軟體或刻意規避看 希平方學英文 - 看 YouTube 學英文 之排除自動搜尋之標頭 (robot exclusion headers);

服務中斷或暫停
本公司將以合理之方式及技術,維護會員服務之正常運作,但有時仍會有無法預期的因素導致服務中斷或故障等現象,可能將造成您使用上的不便、資料喪失、錯誤、遭人篡改或其他經濟上損失等情形。建議您於使用本服務時宜自行採取防護措施。 希平方學英文 對於您因使用(或無法使用)本服務而造成的損害,除故意或重大過失外,不負任何賠償責任。

版權宣告
上次更新日期:2013-09-16

希平方學英文 內所有資料之著作權、所有權與智慧財產權,包括翻譯內容、程式與軟體均為 希平方學英文 所有,須經希平方學英文同意合法才得以使用。
希平方學英文歡迎你分享網站連結、單字、片語、佳句,使用時須標明出處,並遵守下列原則:

  • 禁止用於獲取個人或團體利益,或從事未經 希平方學英文 事前授權的商業行為
  • 禁止用於政黨或政治宣傳,或暗示有支持某位候選人
  • 禁止用於非希平方學英文認可的產品或政策建議
  • 禁止公佈或傳送任何誹謗、侮辱、具威脅性、攻擊性、不雅、猥褻、不實、色情、暴力、違反公共秩序或善良風俗或其他不法之文字、圖片或任何形式的檔案
  • 禁止侵害或毀損希平方學英文或他人名譽、隱私權、營業秘密、商標權、著作權、專利權、其他智慧財產權及其他權利、違反法律或契約所應付支保密義務
  • 嚴禁謊稱希平方學英文辦公室、職員、代理人或發言人的言論背書,或作為募款的用途

網站連結
歡迎您分享 希平方學英文 網站連結,與您的朋友一起學習英文。

抱歉傳送失敗!

不明原因問題造成傳送失敗,請儘速與我們聯繫!
希平方 x ICRT

「Leila Takayama:身為機器人是什麼感覺?」- What's It like to Be a Robot?

觀看次數:2285  • 

框選或點兩下字幕可以直接查字典喔!

You only get one chance to make a first impression, and that's true if you're a robot as well as if you're a person. The first time that I met one of these robots was at a place called Willow Garage in 2008. When I went to visit there, my host walked me into the building and we met this little guy. He was rolling into the hallway, came up to me, sat there, stared blankly past me, did nothing for a while, rapidly spun his head around 180 degrees and then ran away.

And that was not a great first impression. The thing that I learned about robots that day is that they kind of do their own thing, and they're not totally aware of us. And I think as we're experimenting with these possible robot futures, we actually end up learning a lot more about ourselves as opposed to just these machines. And what I learned that day was that I had pretty high expectations for this little dude. He was not only supposed to be able to navigate the physical world, but also be able to navigate my social world—he's in my space; it's a personal robot. Why didn't it understand me? My host explained to me, "Well, the robot is trying to get from point A to point B, and you were an obstacle in his way, so he had to replan his path, figure out where to go, and then get there some other way," which was actually not a very efficient thing to do. If that robot had figured out that I was a person, not a chair, and that I was willing to get out of its way if it was trying to get somewhere, then it actually would have been more efficient at getting its job done if it had bothered to notice that I was a human and that I have different affordances than things like chairs and walls do.

You know, we tend to think of these robots as being from outer space and from the future and from science fiction, and while that could be true, I'd actually like to argue that robots are here today, and they live and work amongst us right now. These are two robots that live in my home. They vacuum the floors and they cut the grass every single day, which is more than I would do if I actually had time to do these tasks, and they probably do it better than I would, too. This one actually takes care of my kitty. Every single time he uses the box, it cleans it, which is not something I'm willing to do, and it actually makes his life better as well as mine. And while we call these robot products—it's a "robot vacuum cleaner, it's a robot lawnmower, it's a robot littler box," I think there's actually a bunch of other robots hiding in plain sight that have just become so darn useful and so darn mundane that we call them things like, "dishwasher," right? They get new names. They don't get called robot anymore because they actually serve a purpose in our lives. Similarly, a thermostat, right? I know my robotics friends out there are probably cringing at me calling this a robot, but it has a goal. Its goal is to make my house 66 degrees Fahrenheit, and it senses the world. It knows it's a little bit cold, it makes a plan and then it acts on the physical world. It's robotics. Even if it might not look like Rosie the Robot, it's doing something that's really useful in my life so I don't have to take care of turning the temperature up and down myself.

And I think these systems live and work amongst us now, and not only are these systems living amongst us but you are probably a robot operator, too. When you drive your car, it feels like you are operating machinery. You are also going from point A to point B, but your car probably has power steering, it probably has automatic braking systems, it might have an automatic transmission and maybe even adaptive cruise control. And while it might not be a fully autonomous car, it has bits of autonomy, and they're so useful and they make us drive safer, and we just sort of feel like they're invisible-in-use, right? So when you're driving your car, you should just feel like you're going from one place to another. It doesn't feel like it's this big thing that you have to deal with and operate and use these controls because we spent so long learning how to drive that they've become extensions of ourselves. When you park that car in that tight little garage space, you know where your corners are. And when you drive a rental car that maybe you haven't driven before, it takes some time to get used to your new robot body. And this is also true for people who operate other types of robots, so I'd like to share with you a few stories about that.

Dealing with the problem of remote collaboration. So, at Willow Garage I had a coworker named Dallas, and Dallas looked like this. He worked from his home in Indiana in our company in California. He was a voice in a box on the table in most of our meetings, which was kind of OK except that, you know, if we had a really heated debate and we didn't like what he was saying, we might just hang up on him.

Then we might have a meeting after that meeting and actually make the decisions in the hallway afterwards when he wasn't there anymore. So that wasn't so great for him. And as a robotics company at Willow, we had some extra robot body parts laying around, so Dallas and his buddy Curt put together this thing, which looks kind of like Skype on a stick on wheels, which seems like a techy, silly toy, but really it's probably one of the most powerful tools that I've seen ever made for remote collaboration. So now, if I didn't answer Dallas' email question, he could literally roll into my office, block my doorway and ask me the question again until I answered it. And I'm not going to turn him off, right? That's kind of rude. Not only was it good for these one-on-one communications, but also for just showing up at the company all-hands meeting. Getting your butt in that chair and showing people that you're present and committed to your project is a big deal and can help remote collaboration a ton. We saw this over the period of months and then years, not only at our company but at others, too. The best thing that can happen with these systems is that it starts to feel like you're just there. It's just you, it's just your body, and so people actually start to give these things personal space. So when you're having a stand-up meeting, people will stand around the space just as they would if you were there in person. That's great until there's breakdowns and it's not. People, when they first see these robots, are like, "Wow, where's the components? There must be a camera over there," and they start poking your face. "You're talking too softly, I'm going to turn up your volume," which is like having a coworker walk up to you and say, "You're speaking too softly, I'm going to turn up your face." That's awkward and not OK, and so we end up having to build these new social norms around using these systems.

Similarly, as you start feeling like it's your body, you start noticing things like, "Oh, my robot is kind of short." Dallas would say things to me—he was six-foot tall—and we would take him via robot to cocktail parties and things like that, as you do, and the robot was about five-foot-tall, which is close to my height. And he would tell me, "You know, people are not really looking at me. I feel like I'm just looking at this sea of shoulders, and it's just—we need a taller robot." And I told him, "Um, no. You get to walk in my shoes for today. You get to see what it's like to be on the shorter end of the spectrum." And he actually ended up building a lot of empathy for that experience, which was kind of great. So when he'd come visit in person, he no longer stood over me as he was talking to me, he would sit down and talk to me eye to eye, which was kind of a beautiful thing. So we actually decided to look at this in the laboratory and see what others kinds of differences things like robot height would make. And so half of the people in our study used a shorter robot, half of the people in our study used a taller robot and we actually found that the exact same person who has the exact same body and says the exact same things as someone, is more persuasive and perceived as being more credible if they're in a taller robot form. It makes no rational sense, but that's why we study psychology. And really, you know, the way that Cliff Nass would put this is that we're having to deal with these new technologies despite the fact that we have very old brains. Human psychology is not changing at the same speed that tech is and so we're always playing catch-up, trying to make sense of this world where these autonomous things are running around. Usually, things that talk are people, not machines, right? And so we breathe a lot of meaning into things like just height of a machine, not a person, and attribute that to the person using the system.

You know, this, I think, is really important when you're thinking about robotics. It's not so much about reinventing humans, it's more about figuring out how we extend ourselves, right? And we end up using things in ways that are sort of surprising. So these guys can't play pool because the robots don't have arms, but they can heckle the guys who are playing pool and that can be an important thing for team bonding, which is kind of neat. People who get really good at operating these systems will even do things like make up new games, like robot soccer in the middle of the night, pushing the trash cans around.

But not everyone's good. A lot of people have trouble operating these systems. This is actually a guy who logged into the robot and his eyeball was turned 90 degrees to the left. He didn't know that, so he ended up just bashing around the office, running into people's desks, getting super embarrassed, laughing about it—his volume was way too high. And this guy here in the image is telling me, "We need a robot mute button." And by that what he really meant was we don't want it to be so disruptive. So as a robotics company, we added some obstacle avoidance to the system. It got a little laser range finder that could see the obstacles, and if I as a robot operator try to say, run into a chair, it wouldn't let me, it would just plan a path around, which seems like a good idea. People did hit fewer obstacles using that system, obviously, but actually, for some of the people, it took them a lot longer to get through our obstacle course, and we wanted to know why. It turns out that there's this important human dimension—a personality dimension called locus of control, and people who have a strong internal locus of control, they need to be the masters of their own destiny—really don't like giving up control to an autonomous system—so much so that they will fight the autonomy; "If I want to hit that chair, I'm going to hit that chair." And so they would actually suffer from having that autonomous assistance, which is an important thing for us to know as we're building increasingly autonomous, say, cars, right? How are different people going to grapple with that loss of control? It's going to be different depending on human dimensions. We can't treat humans as if we're just one monolithic thing. We vary by personality, by culture, we even vary by emotional state moment to moment, and being able to design these systems, these human-robot interaction systems, we need to take into account the human dimensions, not just the technological ones. Along with a sense of control also comes a sense of responsibility. And if you were a robot operator using one of these systems, this is what the interface would look like. It looks a little bit like a video game, which can be good because that's very familiar to people, but it can also be bad because it makes people feel like it's a video game. We had a bunch of kids over at Stanford play with the system and drive the robot around our office in Menlo Park, and the kids started saying things like, "10 points if you hit that guy over there. 20 points for that one." And they would chase them down the hallway. I told them, "Um, those are real people. They're actually going to bleed and feel pain if you hit them." And they'd be like, "OK, got it."But five minutes later, they would be like, "20 points for that guy over there, he just looks like he needs to get hit." It's a little bit like "Ender's Game," right? There is a real world on that other side and I think it's our responsibility as people designing these interfaces to help people remember that there's real consequences to their actions and to feel a sense of responsibility when they're operating these increasingly autonomous things.

These are kind of a great example of experimenting with one possible robotic future, and I think it's pretty cool that we can extend ourselves and learn about the ways that we extend ourselves into these machines while at the same time being able to express our humanity and our personality. We also build empathy for others in terms of being shorter, taller, faster, slower, and maybe even armless, which is kind of neat.

We also build empathy for the robots themselves. This is one of my favorite robots. It's called the Tweenbot. And this guy has a little flag that says, "I'm trying to get to this intersection in Manhattan," and it's cute and rolls forward, that's it. It doesn't know how to build a map, it doesn't know how to see the world, it just asks for help. The nice thing about people is that it can actually depend upon the kindness of strangers. It did make it across the park to the other side of Manhattan—which is pretty great—just because people would pick it up and point it in the right direction. And that's great, right?

We're trying to build this human-robot world in which we can coexist and collaborate with one another, and we don't need to be fully autonomous and just do things on our own. We actually do things together. And to make that happen, we actually need help from people like the artists and the designers, the policy makers, the legal scholars, psychologists, sociologists, anthropologists—we need more perspectives in the room if we're going to do the thing that Stu Card says we should do, which is invent the future that we actually want to live in.

And I think we can continue to experiment with these different robotic futures together, and in doing so, we will end up learning a lot more about ourselves. Thank you.

播放本句

登入使用學習功能

使用Email登入

HOPE English 播放器使用小提示

  • 功能簡介

    單句重覆、重複上一句、重複下一句:以句子為單位重覆播放,單句重覆鍵顯示綠色時為重覆播放狀態;顯示白色時為正常播放狀態。按重複上一句、重複下一句時就會自動重覆播放該句。
    收錄佳句:點擊可增減想收藏的句子。

    中、英文字幕開關:中、英文字幕按鍵為綠色為開啟,灰色為關閉。鼓勵大家搞懂每一句的內容以後,關上字幕聽聽看,會發現自己好像在聽中文說故事一樣,會很有成就感喔!
    收錄單字:框選英文單字可以收藏不會的單字。
  • 分享
    如果您有收錄很優秀的句子時,可以分享佳句給大家,一同看佳句學英文!