下載App 希平方
攻其不背
App 開放下載中
下載App 希平方
攻其不背
App 開放下載中
IE版本不足
你的 IE 瀏覽器太舊了 更新 IE 瀏覽器或點選連結下載 Google Chrome 瀏覽器 前往下載

免費註冊
! 這組帳號已經註冊過了
Email 帳號
密碼請填入 6 位數以上密碼
已經有帳號了?
忘記密碼
! 這組帳號已經註冊過了
您的 Email
請輸入您註冊時填寫的 Email,
我們將會寄送設定新密碼的連結給您。
寄信了!請到信箱打開密碼連結信
密碼信已寄至
沒有收到信嗎? 點這裡重寄一次
如果您尚未收到信,請前往垃圾郵件查看,謝謝!

恭喜您註冊成功!

查看會員功能

註冊未完成

《HOPE English 希平方》服務條款關於個人資料收集與使用之規定

隱私權政策
上次更新日期:2014-12-30

希平方 為一英文學習平台,我們每天固定上傳優質且豐富的影片內容,讓您不但能以有趣的方式學習英文,還能增加內涵,豐富知識。我們非常注重您的隱私,以下說明為當您使用我們平台時,我們如何收集、使用、揭露、轉移及儲存你的資料。請您花一些時間熟讀我們的隱私權做法,我們歡迎您的任何疑問或意見,提供我們將產品、服務、內容、廣告做得更好。

本政策涵蓋的內容包括:希平方 如何處理蒐集或收到的個人資料。
本隱私權保護政策只適用於: 希平方 平台,不適用於非 希平方 平台所有或控制的公司,也不適用於非 希平方 僱用或管理之人。

個人資料的收集與使用
當您註冊 希平方 平台時,我們會詢問您姓名、電子郵件、出生日期、職位、行業及個人興趣等資料。在您註冊完 希平方 帳號並登入我們的服務後,我們就能辨認您的身分,讓您使用更完整的服務,或參加相關宣傳、優惠及贈獎活動。希平方 也可能從商業夥伴或其他公司處取得您的個人資料,並將這些資料與 希平方 所擁有的您的個人資料相結合。

我們所收集的個人資料, 將用於通知您有關 希平方 最新產品公告、軟體更新,以及即將發生的事件,也可用以協助改進我們的服務。

我們也可能使用個人資料為內部用途。例如:稽核、資料分析、研究等,以改進 希平方公司 產品、服務及客戶溝通。

瀏覽資料的收集與使用
希平方 自動接收並記錄您電腦和瀏覽器上的資料,包括 IP 位址、希平方 cookie 中的資料、軟體和硬體屬性以及您瀏覽的網頁紀錄。

隱私權政策修訂
我們會不定時修正與變更《隱私權政策》,不會在未經您明確同意的情況下,縮減本《隱私權政策》賦予您的權利。隱私權政策變更時一律會在本頁發佈;如果屬於重大變更,我們會提供更明顯的通知 (包括某些服務會以電子郵件通知隱私權政策的變更)。我們還會將本《隱私權政策》的舊版加以封存,方便您回顧。

服務條款
歡迎您加入看 ”希平方”
上次更新日期:2013-09-09

歡迎您加入看 ”希平方”
感謝您使用我們的產品和服務(以下簡稱「本服務」),本服務是由 希平方 所提供。
本服務條款訂立的目的,是為了保護會員以及所有使用者(以下稱會員)的權益,並構成會員與本服務提供者之間的契約,在使用者完成註冊手續前,應詳細閱讀本服務條款之全部條文,一旦您按下「註冊」按鈕,即表示您已知悉、並完全同意本服務條款的所有約定。如您是法律上之無行為能力人或限制行為能力人(如未滿二十歲之未成年人),則您在加入會員前,請將本服務條款交由您的法定代理人(如父母、輔助人或監護人)閱讀,並得到其同意,您才可註冊及使用 希平方 所提供之會員服務。當您開始使用 希平方 所提供之會員服務時,則表示您的法定代理人(如父母、輔助人或監護人)已經閱讀、了解並同意本服務條款。 我們可能會修改本條款或適用於本服務之任何額外條款,以(例如)反映法律之變更或本服務之變動。您應定期查閱本條款內容。這些條款如有修訂,我們會在本網頁發佈通知。變更不會回溯適用,並將於公布變更起十四天或更長時間後方始生效。不過,針對本服務新功能的變更,或基於法律理由而為之變更,將立即生效。如果您不同意本服務之修訂條款,則請停止使用該本服務。

第三人網站的連結 本服務或協力廠商可能會提供連結至其他網站或網路資源的連結。您可能會因此連結至其他業者經營的網站,但不表示希平方與該等業者有任何關係。其他業者經營的網站均由各該業者自行負責,不屬希平方控制及負責範圍之內。

兒童及青少年之保護 兒童及青少年上網已經成為無可避免之趨勢,使用網際網路獲取知識更可以培養子女的成熟度與競爭能力。然而網路上的確存有不適宜兒童及青少年接受的訊息,例如色情與暴力的訊息,兒童及青少年有可能因此受到心靈與肉體上的傷害。因此,為確保兒童及青少年使用網路的安全,並避免隱私權受到侵犯,家長(或監護人)應先檢閱各該網站是否有保護個人資料的「隱私權政策」,再決定是否同意提出相關的個人資料;並應持續叮嚀兒童及青少年不可洩漏自己或家人的任何資料(包括姓名、地址、電話、電子郵件信箱、照片、信用卡號等)給任何人。

為了維護 希平方 網站安全,我們需要您的協助:

您承諾絕不為任何非法目的或以任何非法方式使用本服務,並承諾遵守中華民國相關法規及一切使用網際網路之國際慣例。您若係中華民國以外之使用者,並同意遵守所屬國家或地域之法令。您同意並保證不得利用本服務從事侵害他人權益或違法之行為,包括但不限於:
A. 侵害他人名譽、隱私權、營業秘密、商標權、著作權、專利權、其他智慧財產權及其他權利;
B. 違反依法律或契約所應負之保密義務;
C. 冒用他人名義使用本服務;
D. 上載、張貼、傳輸或散佈任何含有電腦病毒或任何對電腦軟、硬體產生中斷、破壞或限制功能之程式碼之資料;
E. 干擾或中斷本服務或伺服器或連結本服務之網路,或不遵守連結至本服務之相關需求、程序、政策或規則等,包括但不限於:使用任何設備、軟體或刻意規避看 希平方 - 看 YouTube 學英文 之排除自動搜尋之標頭 (robot exclusion headers);

服務中斷或暫停
本公司將以合理之方式及技術,維護會員服務之正常運作,但有時仍會有無法預期的因素導致服務中斷或故障等現象,可能將造成您使用上的不便、資料喪失、錯誤、遭人篡改或其他經濟上損失等情形。建議您於使用本服務時宜自行採取防護措施。 希平方 對於您因使用(或無法使用)本服務而造成的損害,除故意或重大過失外,不負任何賠償責任。

版權宣告
上次更新日期:2013-09-16

希平方 內所有資料之著作權、所有權與智慧財產權,包括翻譯內容、程式與軟體均為 希平方 所有,須經希平方同意合法才得以使用。
希平方歡迎你分享網站連結、單字、片語、佳句,使用時須標明出處,並遵守下列原則:

  • 禁止用於獲取個人或團體利益,或從事未經 希平方 事前授權的商業行為
  • 禁止用於政黨或政治宣傳,或暗示有支持某位候選人
  • 禁止用於非希平方認可的產品或政策建議
  • 禁止公佈或傳送任何誹謗、侮辱、具威脅性、攻擊性、不雅、猥褻、不實、色情、暴力、違反公共秩序或善良風俗或其他不法之文字、圖片或任何形式的檔案
  • 禁止侵害或毀損希平方或他人名譽、隱私權、營業秘密、商標權、著作權、專利權、其他智慧財產權及其他權利、違反法律或契約所應付支保密義務
  • 嚴禁謊稱希平方辦公室、職員、代理人或發言人的言論背書,或作為募款的用途

網站連結
歡迎您分享 希平方 網站連結,與您的朋友一起學習英文。

抱歉傳送失敗!

不明原因問題造成傳送失敗,請儘速與我們聯繫!
希平方 x ICRT

「Kevin Slavin:掌控世界的演算法」- How Algorithms Shape Our World


框選或點兩下字幕可以直接查字典喔!

This is a photograph by the artist Michael Najjar, and it's real, in the sense that he went there to Argentinato take the photo. But it's also a fiction. There's a lot of work that went into it after that. And what he's done is he's actually reshaped, digitally, all of the contours of the mountains to follow the vicissitudes of the Dow Jones index. So what you see, that precipice, that high precipice with the valley, is the 2008 financial crisis. The photo was made when we were deep in the valley over there. I don't know where we are now. This is the Hang Seng index for Hong Kong. And similar topography. I wonder why.

And this is art. This is metaphor. But I think the point is that this is metaphor with teeth, and it's with those teeth that I want to propose today that we rethink a little bit about the role of contemporary math—not just financial math, but math in general. That its transition from being something that we extract and derive from the world to something that actually starts to shape it—the world around us and the world inside us. And it's specifically algorithms, which are basically the math that computers use to decide stuff. They acquire the sensibility of truth because they repeat over and over again, and they ossify and calcify, and they become real.

And I was thinking about this, of all places, on a transatlantic flight a couple of years ago, because I happened to be seated next to a Hungarian physicist about my age and we were talking about what life was like during the Cold War for physicists in Hungary. And I said, "So what were you doing?" And he said, "Well we were mostly breaking stealth."

And I said, "That's a good job. That's interesting. How does that work?" And to understand that, you have to understand a little bit about how stealth works. And so—this is an over-simplification—but basically, it's not like you can just pass a radar signal right through 156 tons of steel in the sky. It's not just going to disappear. But if you can take this big, massive thing, and you could turn it into a million little things—something like a flock of birds—well then the radar that's looking for that has to be able to see every flock of birds in the sky. And if you're a radar, that's a really bad job.

And he said, "Yeah." He said, "But that's if you're a radar. So we didn't use a radar; we built a black box that was looking for electrical signals, electronic communication. And whenever we saw a flock of birds that had electronic communication, we thought, 'Probably has something to do with the Americans.'"

And I said, "Yeah. That's good. So you've effectively negated 60 years of aeronautic research. What's your act two? What do you do when you grow up?" And he said, "Well, financial services." And I said, "Oh." Because those had been in the news lately. And I said, "How does that work?" And he said, "Well there's 2,000 physicists on Wall Street now, and I'm one of them." And I said, "What's the black box for Wall Street?"

And he said, "It's funny you ask that, because it's actually called black box trading. And it's also sometimes called algo trading, algorithmic trading." And algorithmic trading evolved in part because institutional traders have the same problems that the United States Air Force had, which is that they're moving these positions—whether it's Proctor & Gamble or Accenture, whatever—they're moving a million shares of something through the market. And if they do that all at once, it's like playing poker and going all in right away. You just tip your hand. And so they have to find a way—and they use algorithms to do this—to break up that big thing into a million little transactions. And the magic and the horror of that is that the same math that you use to break up the big thing into a million little things can be used to find a million little things and sew them back together and figure out what's actually happening in the market.

So if you need to have some image of what's happening in the stock market right now, what you can picture is a bunch of algorithms that are basically programmed to hide, and a bunch of algorithms that are programmed to go find them and act. And all of that's great, and it's fine. And that's 70 percent of the United States stock market, 70 percent of the operating system formerly known as your pension, your mortgage.

And what could go wrong? What could go wrong is that a year ago, nine percent of the entire market just disappears in five minutes, and they called it the Flash Crash of 2:45. All of a sudden, nine percent just goes away, and nobody to this day can even agree on what happened because nobody ordered it, nobody asked for it. Nobody had any control over what was actually happening. All they had was just a monitor in front of them that had the numbers on it and just a red button that said, "Stop."

And that's the thing, is that we're writing things, we're writing these things that we can no longer read. And we've rendered something illegible, and we've lost the sense of what's actually happening in this world that we've made. And we're starting to make our way. There's a company in Boston called Nanex, and they use math and magic and I don't know what, and they reach into all the market data and they find, actually sometimes, some of these algorithms. And when they find them they pull them out and they pin them to the wall like butterflies. And they do what we've always done when confronted with huge amounts of data that we don't understand—which is that they give them a name and a story. So this is one that they found, they called the Knife, the Carnival, the Boston Shuffler, Twilight.

And the gag is that, of course, these aren't just running through the market. You can find these kinds of things wherever you look, once you learn how to look for them. You can find it here: this book about flies that you may have been looking at on Amazon. You may have noticed it when its price started at 1.7 million dollars. It's out of print—still ... If you had bought it at 1.7, it would have been a bargain. A few hours later, it had gone up to 23.6 million dollars, plus shipping and handling. And the question is: Nobody was buying or selling anything; what was happening? And you see this behavior on Amazon as surely as you see it on Wall Street. And when you see this kind of behavior, what you see is the evidence of algorithms in conflict, algorithms locked in loops with each other, without any human oversight, without any adult supervision to say, "Actually, 1.7 million is plenty."

And as with Amazon, so it is with Netflix. And so Netflix has gone through several different algorithms over the years. They started with Cinematch, and they've tried a bunch of others—there's Dinosaur Planet; there's Gravity. They're using Pragmatic Chaos now. Pragmatic Chaos is, like all of Netflix algorithms, trying to do the same thing. It's trying to get a grasp on you, on the firmware inside the human skull, so that it can recommend what movie you might want to watch next—which is a very, very difficult problem. But the difficulty of the problem and the fact that we don't really quite have it down, it doesn't take away from the effects Pragmatic Chaos has. Pragmatic Chaos, like all Netflix algorithms, determines, in the end, 60 percent of what movies end up being rented. So one piece of code with one idea about you is responsible for 60 percent of those movies.

But what if you could rate those movies before they get made? Wouldn't that be handy? Well, a few data scientists from the U.K. are in Hollywood, and they have "story algorithms"—a company called Epagogix. And you can run your script through there, and they can tell you, quantifiably, that that's a 30 million dollar movie or a 200 million dollar movie. And the thing is, is that this isn't Google. This isn't information. These aren't financial stats; this is culture. And what you see here, or what you don't really see normally, is that these are the physics of culture. And if these algorithms, like the algorithms on Wall Street, just crashed one day and went awry, how would we know? What would it look like?

And they're in your house. They're in your house. These are two algorithms competing for your living room. These are two different cleaning robots that have very different ideas about what clean means. And you can see it if you slow it down and attach lights to them, and they're sort of like secret architects in your bedroom. And the idea that architecture itself is somehow subject to algorithmic optimization is not far-fetched. It's super-real and it's happening around you.

You feel it most when you're in a sealed metal box, a new-style elevator; they're called destination-control elevators. These are the ones where you have to press what floor you're going to go to before you get in the elevator. And it uses what's called a bin-packing algorithm. So none of this mishegas of letting everybody go into whatever car they want. Everybody who wants to go to the 10th floor goes into car two, and everybody who wants to go to the third floor goes into car five. And the problem with that is that people freak out. People panic. And you see why. You see why. It's because the elevator is missing some important instrumentation, like the buttons. Like the things that people use. All it has is just the number that moves up or down and that red button that says, "Stop." And this is what we're designing for. We're designing for this machine dialect. And how far can you take that? How far can you take it? You can take it really, really far.

So let me take it back to Wall Street. Because the algorithms of Wall Street are dependent on one quality above all else, which is speed. And they operate on milliseconds and microseconds. And just to give you a sense of what microseconds are, it takes you 500,000 microseconds just to click a mouse. But if you're a Wall Street algorithm and you're five microseconds behind, you're a loser. So if you were an algorithm, you'd look for an architect like the one that I met in Frankfurt who was hollowing out a skyscraper—throwing out all the furniture, all the infrastructure for human use, and just running steel on the floors to get ready for the stacks of servers to go in—all so an algorithm could get close to the Internet.

And you think of the Internet as this kind of distributed system. And of course, it is, but it's distributed from places. In New York, this is where it's distributed from: the Carrier Hotel located on Hudson Street. And this is really where the wires come right up into the city. And the reality is that the further away you are from that, you're a few microseconds behind every time. These guys down on Wall Street, Marco Polo and Cherokee Nation, they're eight microseconds behind all these guys going into the empty buildings being hollowed out up around the Carrier Hotel. And that's going to keep happening. We're going to keep hollowing them out, because you, inch for inch and pound for pound and dollar for dollar, none of you could squeeze revenue out of that space like the Boston Shuffler could.

But if you zoom out, if you zoom out, you would see an 825-mile trench between New York City and Chicago that's been built over the last few years by a company called Spread Networks. This is a fiber optic cable that was laid between those two cities to just be able to traffic one signal 37 times faster than you can click a mouse—just for these algorithms, just for the Carnival and the Knife. And when you think about this, that we're running through the United States with dynamite and rock saws so that an algorithm can close the deal three microseconds faster, all for a communications framework that no human will ever know, that's a kind of manifest destiny; and we'll always look for a new frontier.

Unfortunately, we have our work cut out for us. This is just theoretical. This is some mathematicians at MIT. And the truth is I don't really understand a lot of what they're talking about. It involves light cones and quantum entanglement, and I don't really understand any of that. But I can read this map, and what this map says is that, if you're trying to make money on the markets where the red dots are, that's where people are, where the cities are, you're going to have to put the servers where the blue dots are to do that most effectively. And the thing that you might have noticed about those blue dots is that a lot of them are in the middle of the ocean. So that's what we'll do: we'll build bubbles or something, or platforms. We'll actually part the water to pull money out of the air, because it's a bright future if you're an algorithm.

And it's not the money that's so interesting actually. It's what the money motivates, that we're actually terraforming the Earth itself with this kind of algorithmic efficiency. And in that light, you go back and you look at Michael Najjar's photographs, and you realize that they're not metaphor, they're prophecy. They're prophecy for the kind of seismic, terrestrial effects of the math that we're making. And the landscape was always made by this sort of weird, uneasy collaboration between nature and man. But now there's this third co-evolutionary force: algorithms—the Boston Shuffler, the Carnival. And we will have to understand those as nature, and in a way, they are.

Thank you.

播放本句

登入使用學習功能

使用Email登入

HOPE English 播放器使用小提示

  • 功能簡介

    單句重覆、上一句、下一句:顧名思義,以句子為單位重覆播放,單句重覆鍵顯示橘色時為重覆播放狀態;顯示灰色時為正常播放狀態。按上一句鍵、下一句鍵時就會自動重覆播放該句。
    收錄佳句:點擊可增減想收藏的句子。

    中、英文字幕開關:中、英文字幕按鍵為綠色為開啟,灰色為關閉。鼓勵大家搞懂每一句的內容以後,關上字幕聽聽看,會發現自己好像在聽中文說故事一樣,會很有成就感喔!
    收錄單字:用滑鼠框選英文單字可以收藏不會的單字。
  • 分享
    如果您覺得本篇短片很有趣或很喜歡,在短片結束時有分享連結,可以分享給朋友一同欣賞,一起看YouTube學英文!

    或是您有收錄很優秀的句子時,也可以分享佳句給大家,一同看佳句學英文!