Placeholder Image

字幕表 動画を再生する

  • [MUSIC PLAYING]

  • DAVID MALAN: This is CS50.

  • Hello, world.

  • This is the CS50 podcast my name is David Malan,

  • and this is Episode Zero, our very first,

  • and I'm joined here by CS50's own Colt--

  • COLTON OGDEN: Yep.

  • [LAUGHTER]

  • Colton Odgen. This is an interesting new direction that we're going.

  • DAVID MALAN: Yeah, it's one in which we clearly haven't rehearsed.

  • COLTON OGDEN: Yeah.

  • DAVID MALAN: So, but what we thought we'd do with the CS50 podcast is really

  • focus on the week's current events as it relates to technology,

  • use this as an opportunity to talk about the implications

  • of various technologies, and really explain things as it comes up,

  • but really in a non-visual way.

  • And so, perhaps, I think the topics Colton

  • and I'll hit on here will focus on things you, yourself, might have read

  • in the news that maybe didn't register necessarily

  • or maybe you didn't really understand how it pertains to technologies

  • that you, yourself, use.

  • COLTON OGDEN: Yeah, and I think that's ties as well to prior, when

  • we did CS50 live, and this was kind of the same idea.

  • DAVID MALAN: Yeah, absolutely.

  • Whereas CS50 Live, when we did it on video, was much more visual--

  • we prepared slides, we actually looked at sample videos and such--

  • here, we thought we'd really try to focus on ideas.

  • And it'll be up to you to decide if this works well or not well,

  • but we come prepared with a look at some of the past week's news.

  • And why don't we get right into it?

  • COLTON OGDEN: Yeah, absolutely.

  • One of the things I noticed, actually, is--

  • I put together this list of topics, but the one thing

  • that I didn't put in here that you actually found and put in here,

  • today, was something about Facebook passwords.

  • DAVID MALAN: Yeah, so a website named Krebs on Security, the author of this,

  • was contacted apparently by some employee-- presumably

  • and a current employee of Facebook-- who revealed

  • to him that during some recent audit of their security processes,

  • they discovered that for like seven years,

  • since 2012, had one or more processes inside of Facebook

  • been storing passwords-- users' passwords, like yours and mine

  • potentially, in the clear, so to speak, clear text, not

  • cipher text, which means unencrypted-- in some database or some file

  • somewhere.

  • COLTON OGDEN: Typically, people will use some sort

  • of hashing algorithm to store things cryptographically and much more

  • securely?

  • DAVID MALAN: Indeed, even like ROT13, like rotate every character 13 places,

  • would have been arguably more secure.

  • And there's not a huge amount of technical detail out there.

  • If you go to krebsonsecurity.com, you can actually

  • dig up the blog post itself.

  • And then Facebook actually did respond, and I

  • think there's a link in Krebs on Security to the Facebook announcement.

  • But to be honest, the Facebook announcement which is

  • on newsroom.fb.com, pretty, to be honest,

  • it's pretty nondescript and really doesn't--

  • I mean, it's kind of disingenuous.

  • They seem to use this as an opportunity to talk about best practices

  • when it comes to passwords and all of the various other mechanisms

  • that they have in place to help you secure your password.

  • And yet, they really kind of didn't address the topic

  • at hand, which is, well, despite all of those mechanisms,

  • you were storing our passwords in the clear, or at least

  • millions of Facebook users, particularly on Facebook Light,

  • lighter-weight version of the app that's useful in low bandwidth locations

  • or where bandwidth is very expensive or slow.

  • COLTON OGDEN: So this strikes you sort of as an opportunity for them to,

  • what, hand wave over the issue and sort of distract people?

  • Is that sort of how it-- this rubs you?

  • DAVID MALAN: Yeah, maybe.

  • I think they, you know, acknowledged the issue,

  • but then used this as an opportunity to emphasize all

  • of the things that are being done well.

  • And that's fine, but I think the world is done a disservice

  • when companies aren't just candid with their mea culpas

  • and what they got wrong.

  • I think there's learning opportunities and, as I read this,

  • there's really little for me as a technical person

  • or as an aspiring program to really learn from,

  • other than the high order bit which is, encrypt your passwords.

  • But how did this happen?

  • What are the processes that failed?

  • I mean if companies like Facebook can't get this right,

  • how can little old me, an aspiring programmer,

  • get these kinds of details right?

  • I wonder.

  • COLTON OGDEN: So an article more about how they failed

  • and how they could address it, and how other companies could address it,

  • you think that would've been more productive?

  • DAVID MALAN: I think so.

  • I mean, postmortems, as they're called in many contexts, including in tech,

  • and I've always really admired companies that when they do

  • have some significant mistake or human error, where they own up to it

  • and they explain in technical terms exactly what went wrong.

  • They can still have a more layman's explanation

  • of the problem too, where most people might only take

  • an interest in that level of detail.

  • But for the technophiles and for the students

  • and the aspiring technophiles out there, I think it's just appreciated.

  • And these are such teachable moments and all

  • that-- but I would respect the persons, the company all the more

  • if they really just explained what it is they failed so that we can all learn

  • from it and not repeat those mistakes.

  • COLTON OGDEN: If a large company like Facebook is doing something like this,

  • how prevalent do you think this practice is in the real world?

  • DAVID MALAN: Yeah.

  • Oh my God.

  • I mean, probably frighteningly common, and it's

  • just if you have fewer users or fewer eyes on the company,

  • you probably just notice these things less frequently.

  • But I do think things are changing.

  • I mean with laws like GDPR in the EU, the European Union,

  • I think there's increased pressure on companies now, increased

  • legal pressure, on them to disclose when these kinds of things happen,

  • to impose penalties when it does, to therefore

  • discourage this from even happening.

  • And you know, I'm wondering why this audit detected this in 2019, and not

  • in 2012 or 2013 or 2014 and so forth.

  • COLTON OGDEN: GDPR, did that happened back in 2012?

  • Oh no, that was--

  • DAVID MALAN: No, this was recent.

  • COLTON OGDEN: That one Came onto force--

  • OK.

  • DAVID MALAN: Recent months, actually, has this been rolled out.

  • COLTON OGDEN: Was this-- is this related at all

  • to the proliferation, now, of cookie messages that you see on websites?

  • DAVID MALAN: That's US-specific, where I believe it's now being enforced.

  • Because that actually has been around for quite some time in Europe.

  • Anytime you took your laptop abroad, for instance,

  • would you notice that almost every darn site asks you,

  • hey, can we store cookies.

  • And honestly, that's a very annoying and almost silly manifestation of it

  • because the reality is, as you know, I mean

  • the web doesn't work without cookies or at least

  • dynamic applications don't work.

  • And anyone who's taken CS50 or who's and a bit of web programming,

  • really, in any language, know that the only way

  • to maintain state in most HTTP-based applications is with cookies.

  • So, I mean, we've created a culture where people just

  • dismiss yet another message, and I don't think that's a net positive either.

  • COLTON OGDEN: I think I see a lot, too, of the messages that say,

  • by continuing to use this site, you acknowledge

  • that we have access to whatever information, using cookies, and so on.

  • So I almost think that they do it already and sort of legally

  • can get away with it by having this message visible.

  • DAVID MALAN: Yeah, I mean, it's like cigarette ads

  • which, abroad, as well, there was--

  • before the US, there was much more of, I presume,

  • law around having to have very scary warnings on packages.

  • And companies somewhat cleverly, but somewhat tragically, kind of

  • steered into that and really owned that and put the scariest of messages.

  • And it-- you almost become desensitized to it because it's just so silly

  • and it's so over the top, you know, smoking kills.

  • And then, here's the price tag and here's the brand name.

  • Like, you start to look past those kinds of details too,

  • so I'm not sure even that is all that effective.

  • But someone who's looked at this and studied it

  • can perhaps attest quantitatively just how effective it's been.

  • COLTON OGDEN: Yeah, indeed.

  • Well, scary to know that our passwords may

  • have been reflected visibly on somebody's server,

  • a big website like Facebook.

  • Related to that, another of the topics that I sort of dug into a little bit

  • yesterday-- or not yesterday, a few days ago,

  • was Gmail Confidential Mode, a new feature

  • that they're starting to roll out.

  • DAVID MALAN: Yeah.

  • Yeah, I saw that, just in March, one of the articles on Google's blog

  • discussed this.

  • What, so do you understand what the--

  • what they're offering now as a service?

  • COLTON OGDEN: I have to reread back through the article.

  • So from what I understood, though, it was encrypting not P2P emails,

  • but encrypting the emails sort of towards a sort of proxy,

  • towards a center point, and then forwarding

  • that encrypted email to the other person on the receiving end.

  • But I remember reading in the article that P2P encryption wasn't something

  • that they were actually going to start implementing just yet.

  • DAVID MALAN: Yeah, and this is--

  • I mean, this is kind of the illusion of security or confidentiality.

  • In fact, I was just reading, after you sent me

  • this link on the EFF's website, the Electronic Frontier Foundation who

  • are really very progressively-minded, security-conscious, privacy-conscious

  • individuals as a group, they noted how this really isn't confidential.

  • Google, of course, still has access to the plain text of your email.

  • They don't claim to be encrypting it Peer-to-Peer,

  • so they're not being disingenuous.

  • But I think they, too, are sort of creating and implying

  • a property, confidentiality, that isn't really there.

  • And what this does, for those unfamiliar,

  • is when you send an email in Gmail, if you or your company

  • enables this feature--

  • I think it might still be in beta mode or in trial mode.

  • COLTON OGDEN: I don't think it's officially fully deployed yet,

  • but yeah, [INAUDIBLE].

  • DAVID MALAN: Yeah, so you can opt into it if you have a corporate account,

  • for instance.

  • It gives you an additional, like, Lock icon on the Compose window

  • for an email, where you can say that this message expires, sort of James

  • Bond style, after some number of hours.

  • You can add an SMS code to it so the human who is receiving it

  • has to type in a code that they get on their cell phone.

  • And so it also prevents users from forwarding it, for instance,

  • therefore accidentally or intentionally sending it to someone else.

  • But there's the fundamental issue because you

  • start to condition people, potentially, into thinking,

  • oh, this is confidential.

  • No one can see the message that I'm sending or that I've received.

  • And that's just baloney, right?

  • And you or I could take out our cell phone, right

  • now and not screenshot, but photograph anything on our screen.

  • You could certainly highlight and Copy-Paste it into some other email.

  • And so I think these kinds of features are dangerous if users don't really

  • understand what's going on.

  • And honestly, this is going to be a perfect topic

  • in CS50 itself or CS50 for MBAs or for JDs at Harvard's graduate schools.

  • Because if you really push on this, what does confidential mean?

  • Well, not really much.

  • You're just kind of-- it's more of a social contract between two people.

  • Like, OK, OK, I won't forward this.

  • It's just raising the bar.

  • It's not stopping anything.

  • COLTON OGDEN: Part of this kind of reminds me, too, of the point

  • you like to mention in most of the courses

  • that you teach in that security doesn't really mean much just by virtue

  • of seeing something.

  • Somebody sees a padlock icon in their browser in, let's say,

  • bankofamerica.com, that doesn't necessarily mean

  • that anything that they see is secure.

  • DAVID MALAN: Yeah, well that too.

  • I mean, there too, we humans learned, years ago, well maybe we

  • shouldn't be putting padlocks in places that

  • have no technical meaning for exactly that reason.

  • People just assume it means something that it doesn't, so we seem doomed,

  • as humans, to repeat these mistakes.

  • And this isn't to say that I think this is a bad feature.

  • Frankly, I wish that I could somehow signal to recipients of emails

  • I send, sometimes, please don't forward this to someone else

  • because it's not going to reflect well or it's going to sound overly harsh

  • or whatever the email is.

  • You sometimes don't want other people to see it,

  • even if it's not the end of the world if they actually do.

  • But short of writing in all caps, like, do not forward this email,

  • at the very start of your message, most people might not realize.

  • So I think having a software mechanism that says

  • don't, not forwardable, isn't bad.

  • But, you know, it should probably be like, please don't forward, and not

  • imply that this is confidential and no one else is going to see it.

  • COLTON OGDEN: Do you think that there is some sort of risk involved

  • in making these emails self-destructive, in as much as maybe

  • it will bite people sort of in the future when maybe they

  • want to look back on records that are important like this?

  • DAVID MALAN: Could be.

  • I mean, there too, I suspect there are business motivations for this,

  • for retention policies, where there might be laws or policies in place

  • where companies do or don't want to keep information around

  • because it can come back to bite them.

  • And so maybe it's a good thing if emails do expire after some amount of time,

  • so long as that's within the letter of the law.

  • But I presume it's motivated, in part, by that.

  • So this is a software technique that helps with that.

  • And so, in that sense, you know, confidential

  • does have that kind of meaning, but it's not secure

  • and I worry that you put a padlock on it-- that doesn't necessarily

  • mean to people what you think.

  • I mean, so many people, and kids especially,

  • might think or once thought that Snapchat messages are indeed ephemeral

  • and they'll disappear.

  • But ah, I mean, they're still on the servers.

  • COLTON OGDEN: Undoubtedly.

  • DAVID MALAN: They can be on the servers.

  • You can snap-- screenshot them or record them with another device.

  • So I think we do humans a disservice if we're not really upfront as

  • to what a feature means and how it works, and I

  • think we should label things appropriately so as to not oversell

  • them.

  • COLTON OGDEN: And it sort of takes unfortunate advantage

  • of those who are not as technically literate as well, allowing

  • them to sort of-- or at least capitalizing

  • on people taking for granted these things that they assume to be true.

  • DAVID MALAN: Yeah.

  • I mean, we've been doing this, honestly, as humans,

  • for like, what, 20, 30 years with DOS.

  • You might recall that when you format a hard drive, which generally means to--

  • kind of means to erase it and prepare it to have something new installed

  • on it, the command, back then, when you used to delete it or fdisk it

  • or whatever it was, was are you sure you want to proceed?

  • This will erase the entire disk, something like that,

  • and I think it actually was in all caps.

  • But it was false, technically.

  • Right?

  • All it would do is rewrite part of the headers on disk,

  • but it would leave all of your zeros and ones from previous files

  • there in place.

  • And there, too, we said it would delete or erase information, but it doesn't.

  • And so, for years maybe to this day, do people

  • assume that when you delete something from your Mac or PC

  • or empty the Recycle Bin or whatnot, that it's gone?

  • But anyone who's taken CS50 knows that's not the case.

  • I mean, we have students recover data in their forensics homework alone.

  • COLTON OGDEN: You have a background, certainly, in this too.

  • You did this for a few years.

  • DAVID MALAN: Yeah, or a couple, a year or two.

  • Yeah, yeah, in graduate school.

  • COLTON OGDEN: If you were to advise our listeners on the best way

  • to sort of format their hard drive and avoid this fallacy,

  • what would be your suggestion?

  • DAVID MALAN: Drop it in a volcano.

  • [LAUGHTER]

  • COLTON OGDEN: So then, are you insinuating

  • that there is no truly safe way to clean a hard drive?

  • DAVID MALAN: No, no.

  • Well, in software, it's risky.

  • COLTON OGDEN: In software.

  • DAVID MALAN: I think if you really want peace of mind

  • because you have personal documents, financial documents, family documents,

  • whatever it is that you want to destroy, physical destruction is probably

  • the most safe.

  • And there are companies that allow you to physically destroy hard drives.

  • They drill holes in it or they crush it or whatnot,

  • or you can take out a hammer and try to break through the device.

  • But it's difficult, as we've seen in class when

  • we've disassembled things, you and I, for CS50's Introduction to Technology

  • class.

  • It's hard just to get the damn screws open.

  • So that's the most robust way, is physical destruction.

  • You can wipe the disk in software.

  • Frankly, it tends not to be terribly easy.

  • It's easier with mechanical drives, hard disk drives that spin around.

  • But with SSDs, the Solid State Drives that are purely electronic these days,

  • it's even harder.

  • Because those things, in a nutshell, are designed

  • to only have certain parts of them written to a finite number of times.

  • And eventually, the hard drive, after a certain number of writes

  • or after a certain amount of time, will stop using certain parts of the disks.

  • And that does mean you have a slightly less space available, potentially,

  • but it ensures that your data's still intact.

  • That means that even if you try to overwrite that data,

  • it's never going to get written to because the device isn't

  • going to write to it anymore.

  • COLTON OGDEN: Oh, I see.

  • It closes off certain sectors--

  • DAVID MALAN: Exactly.

  • COLTON OGDEN: --that might have data written in.

  • That's interesting.

  • I didn't know that.

  • DAVID MALAN: So you're better off just destroying that disk, at that point,

  • too.

  • So it's wasteful unfortunately, financially,

  • but if you want true peace of mind, you shouldn't just wipe it with software.

  • You shouldn't hand it off to someone and assume that Best Buy

  • or whatever company is doing it for you is going to do it properly as well.

  • You should probably just remove the device if you can,

  • destroy it, and sell the rest of the equipment.

  • COLTON OGDEN: I think this is reflected too,

  • in Mr. Robot, where he microwaves an SD card that he

  • trusts I'll get off of his--

  • DAVID MALAN: Did he?

  • I Don't know if I saw that episode, then.

  • COLTON OGDEN: This was, I think, the second episode.

  • DAVID MALAN: That's probably not the right way to do it.

  • That's probably just very dangerous.

  • COLTON OGDEN: That's probably very-- yeah, I think it exploded in the video,

  • but yeah.

  • DAVID MALAN: You don't put metal thing-- for our CS50 listeners

  • out there, don't put metal things in microwaves.

  • COLTON OGDEN: Yeah, generally not advisable.

  • DAVID MALAN: No, I think never advisable.

  • [LAUGHTER]

  • COLTON OGDEN: Yeah, so off of the-- well,

  • I guess sort of related to the topic of security,

  • there was an article recently published on Gizmodo

  • about how the FCC admitted in court that it can't track who

  • submits fake comments to its database.

  • DAVID MALAN: Yeah, I was reading that.

  • And as best I could tell, it sounded like they had a web-based form

  • to solicit feedback on, what was it, net neutrality or some topic like that,

  • and they claimed that they couldn't trace

  • who it was because apparently there were millions of bogus comments generated

  • by script kiddies or just adversaries who wrote programs

  • to just submit comments again and again and again and again.

  • And as best I could infer, it sounds like they

  • were weren't logging, maybe, who they were coming from,

  • maybe it's IP address.

  • It sounded like maybe they didn't even have a CAPTCHA in place

  • to sort of force a presumed human to answer some challenge like a math

  • problem or what is this blurry text or click

  • all of the icons that have crosswalks in them or something like that.

  • COLTON OGDEN: Right.

  • DAVID MALAN: And so they just don't have much metadata,

  • it seemed, about who the users were.

  • So short of looking at the text that was submitted alone,

  • it sounds like they can't necessarily filter things out.

  • It's a little strange to me because it sounded

  • like they do have IP addresses, at least in the article that I read,

  • and the FCC doesn't want to release that for reasons of privacy.

  • But you could certainly filter out a good amount of the traffic,

  • probably, if it all seems to be coming from the same IP.

  • I'm guessing many of the adversaries weren't as thoughtful

  • as to use hundreds or thousands of different IPs,

  • so that's a little curious too.

  • COLTON OGDEN: Is it all related to onion routing?

  • And this is more of my sort of lack of knowledge of Tor and onion

  • routing, but is this sort of how onion routing works in that you can

  • spoof your IP from a million locations?

  • DAVID MALAN: Not even spoof your IP.

  • You just really send the data through an anonymized network such

  • that it appears to be--

  • that it is coming from someone else that's not you.

  • So yeah, that's an option.

  • I've not used that kind of software in years

  • or looked very closely at how it's advanced, but that's the general idea.

  • Like, you just get together with a large enough group of other people

  • who you presumably don't know, so n is large, so to speak,

  • and all these computers are running the same software.

  • And even though you might originate a message in email or form submission,

  • that information gets routed through n minus 1 other people, or some subset

  • thereof, so that you're kind of covering your tracks.

  • It's like in the movies, right, when they show a map of the world

  • and like the bad guys' data is going from here to here to here

  • and a red line is bouncing all over the world.

  • That's pretty silly, but it's actually that kind of idea.

  • You just don't have software that visualizes it.

  • COLTON OGDEN: And that's why it looks-- that's why they call it onion routing,

  • because it's like layers of an onion, kind of going all around?

  • DAVID MALAN: Oh is it?

  • I never thought about it.

  • COLTON OGDEN: I thought that that was why it was called onion routing.

  • DAVID MALAN: Maybe.

  • That sounds pretty compelling.

  • So sure, yes.

  • COLTON OGDEN: They apparently, per the article,

  • the API logs contain dozens of IP addresses

  • that belong to groups that uploaded millions of comments combined.

  • So to your point, it does sound like that indeed is what happened.

  • DAVID MALAN: Oh, so that's presumably how they know minimally

  • that there were bogus comments?

  • COLTON OGDEN: Yeah.

  • DAVID MALAN: And but hard to distinguish maybe

  • some of the signal from the noise.

  • COLTON OGDEN: Folks concerns were that people

  • were sort of creating these bogus comments that

  • were propaganda, essentially malicious-- maliciously-oriented comments.

  • DAVID MALAN: Yeah, well this is pretty stupid then,

  • sounding then because, honestly, there are

  • so many like available APIs via which you can at least raise

  • the barrier to adversaries.

  • Right, using CAPTCHAs so that, theoretically, you

  • can't just write a program to answer those kinds of challenge questions;

  • a human actually has to do it.

  • So you might get bogus submissions, but hopefully not thousands or millions

  • of them.

  • COLTON OGDEN: Yeah.

  • No, it sounded more like a technical--

  • I might not want to--

  • I don't want to stretch my words here-- but it

  • sounded like there was a little bit of potential technical illiteracy involved

  • at least in this.

  • DAVID MALAN: Could be.

  • COLTON OGDEN: Potentially.

  • DAVID MALAN: It could be.

  • COLTON OGDEN: I want to try to sound as diplomatic as possible.

  • DAVID MALAN: Good thing they're making all these decisions around technology.

  • COLTON OGDEN: Ah, yeah, exactly.

  • Right?

  • And I have a picture--

  • OK, I'm not going to go in that direction, but--

  • DAVID MALAN: I don't think we can show pictures on this podcast, though.

  • COLTON OGDEN: Another topic sort of related

  • to this was-- and John Oliver sort of did

  • a skit on this related to robocalls-- is, well, robocalls.

  • For those that-- do you want to maybe explain

  • what robocalls are for our audience?

  • DAVID MALAN: Yeah.

  • I mean, a robocall is like a call from a robot, so to speak.

  • Really, a piece of software that's pretending to dial the phone,

  • but is doing it all programmatically through software.

  • And it's usually because they want to sell you something or it's

  • an advertisement or it's a survey or they

  • want to trick you into giving your social security number

  • or that you owe taxes.

  • I mean, they can be used for any number of things and, sometimes, good things.

  • You might get a reminder from a robocall from like an airline saying hey,

  • your flight has been delayed an hour.

  • That's useful, and you might invite that.

  • But robocalls have a bad rap because they're often unsolicited

  • and because I have not signed up for someone to call me.

  • And indeed, these have been increasing in frequency for me

  • too, on my cell phone in particular, which theoretically is unlisted.

  • And I added the Do Not Track thing, years ago,

  • but that's really just the honor system.

  • You don't have to honor people who are on those lists.

  • COLTON OGDEN: They just supposedly have to check the list, right,

  • but they can still call you afterwards?

  • DAVID MALAN: Yeah, and I mean certainly if the problem is with bad actors,

  • then, by definition, those people aren't respecting these lists

  • in the first place.

  • So it's the good people who you might want to hear from who you're not

  • because they are honoring the Do Not Call lists.

  • COLTON OGDEN: I've noticed that I've received a great many as well.

  • Most of them from--

  • DAVID MALAN: Oh, yeah, sorry about those.

  • COLTON OGDEN: Most of them from 949, which is where I grew up in California,

  • and that's where the bulk of all the messages are coming from.

  • DAVID MALAN: But well, per-- seem to be coming from.

  • I've noticed this too.

  • I get them from 617, which is Boston's area code, too.

  • They're doing that on purpose.

  • I just read this, and it makes perfect sense, now,

  • in retrospect why I keep seeing the same prefix in these numbers.

  • because they're trying to trick you and me

  • into thinking that, oh, this is from a neighbor or someone

  • I know in my locality.

  • No it's just another obnoxious technique, to be honest.

  • COLTON OGDEN: I live in Massachusetts now,

  • so I know that it's not a neighbor, definitely, if they're 949.

  • DAVID MALAN: Yeah.

  • COLTON OGDEN: Likely.

  • DAVID MALAN: Well, the thing is, I don't know anyone's number, now.

  • So if I don't, it's not in my contacts, I know it's probably a robocall.

  • So no, it's really awful and, I mean, I think

  • one of the primary reasons this is becoming even more of a problem

  • is that making calls is so darn cheap.

  • I mean, you and I have experimented with Twilio,

  • which is a nice service that has, I think, a free tier but a paid tier

  • too where you can automate phone calls, hopefully for good purposes.

  • And I was just reading on their website that they actually

  • deliberately, though this certainly is a business advantage for them

  • too, charge minimally by the minute, not by the second,

  • because they want to charge even potential adversaries at least

  • 60 seconds for the call.

  • Though, of course, this means that if you and I are

  • writing an app that just needs a few seconds of airtime,

  • we're overpaying for it.

  • But it's a hard problem because calls are just so cheap.

  • And this is why spam has so proliferated, right?

  • Because it's close to zero cents to even send a bogus email, these days.

  • And so those two are dominating the internet, too.

  • And thankfully, companies like Google have

  • been pretty good at filtering it out.

  • You know, we don't really have a middleman filtering out our phone

  • calls, and it's unclear if you'd want a middleman-- software

  • picking up the phone figuring out if it's is legit

  • and then connecting them to you.

  • COLTON OGDEN: Right.

  • DAVID MALAN: It feels a little invasive and a time-consuming, too.

  • COLTON OGDEN: And it's kind of alarming just

  • how easy it is for your average person, your average beginning programmer

  • to set up an automated robocall system.

  • This was illustrated, and again, back to the John Oliver segment,

  • this was illustrated on there where they had literally had--

  • they were showing a clip of somebody who wrote a little command line script

  • or something like that.

  • And even John Oliver made light of it in this skit

  • where he said that his tech person took only 15 minutes to sort

  • of bomb the FCC with phone calls.

  • But I mean, the demonstration showed writing a simple script,

  • 20 phones just light up on the table.

  • And this can be scaled and, you know, however large you want to go with it.

  • DAVID MALAN: No.

  • And, fun fact, I actually did this in CS50 once, a few years ago,

  • and have not done it since because this blew up massively on me.

  • Long story short-- and we have video footage of this if you dig through--

  • several years ago, maybe if it's 2018, most recently,

  • it's probably 2014, give or take.

  • In one of the lectures, mid-semester, we were talking about web programming

  • and APIs and I wrote a script in advance to send a message via text--

  • well, technically, via email to text.

  • It was sent through what's called an email

  • to SMS gateway that would send a message to every CS50 student in the room.

  • And at the time, I foolishly thought it would

  • be cute to say something like, where are you?

  • Why aren't you in class?

  • Question mark.

  • And the joke was supposed to be, because if anyone were,

  • you know, cutting class that day and weren't

  • there they'd get this message from CS50's bot thinking,

  • oh my God, they know I'm not there.

  • When, really, everyone else in the classroom was in on it

  • because they saw me running the program and they knew what was going to happen.

  • And it was pretty cool in that, all of the sudden,

  • a whole bunch of people in the room started

  • getting text messages with this, I thought, funny message.

  • But I had a stupid bug in my code, and essentially my loop

  • sent one text message, the first iteration;

  • Then two text messages, the second iteration;

  • then three text messages, the third iteration.

  • Whereby the previous recipients would get another and another and another

  • because I essentially kept appending to an array or to a list of recipients

  • instead of blowing away the previous recipient list.

  • COLTON OGDEN: It's like a factorial operation.

  • DAVID MALAN: Well, a geometric series, technically.

  • COLTON OGDEN: [INAUDIBLE].

  • DAVID MALAN: Or if you did it--

  • I did, I think I did out the math.

  • If I had not hit Control-C pretty quickly

  • to cancel or to interrupt the process, I would have sent 20,000 text messages.

  • And they were going out quickly.

  • And I felt horrible because this was enough years ago where some people were

  • still paying for text messaging plans.

  • It wasn't unlimited, which is pretty common, at least in the US these days,

  • to just have unlimited texts or iMessage or whatever.

  • So, you know, this could have been costing students $0.10 to $0.25

  • or whatever.

  • So we offered to compensate anyone for this,

  • and I did have to single out a $20 bill I think to one student whose phone I

  • had overwhelmed.

  • But, there too, it was also, phones were old enough that they only had finite--

  • well, they always had finite memory.

  • They had terribly little memory, and so when

  • you get a whole bunch of text messages, back in the day,

  • it would push out older text messages and I felt awful about that.

  • COLTON OGDEN: Oh.

  • DAVID MALAN: Kind of overwhelming people's memory.

  • So anyhow, this is only to say that even hopefully

  • good people with good intentions can use robocalls or robotexting accidentally

  • for ill.

  • And if you're trying to do that deliberately,

  • maliciously, it's just so darn easy.

  • COLTON OGDEN: So solutions to this then?

  • DAVID MALAN: Don't let me in front of a keyboard.

  • [LAUGHTER]

  • COLTON OGDEN: Do we-- so there is a little bit of reading I was doing,

  • and it might have been in this same article,

  • but cryptographically signing phone calls,

  • is this something that you think is possible?

  • DAVID MALAN: Yeah.

  • I mean, I don't know terribly much about the phone industry

  • other than it's pretty backwards or dated

  • in terms of how it's all implemented.

  • I mean, I'm sure this is solvable.

  • But the catch is how do you roll it out when you have old-school copper phone

  • lines, when you have all of us using cell phones on different carriers?

  • It just feels like a very hard coordination problem.

  • And honestly, now that data plans are so omnipresent and we decreasingly

  • need to use voice, per se--

  • you can use Voice over IP, so to speak--

  • you know, I wouldn't be surprised if we don't fix the phone industry,

  • but we instead replace it with some equivalent of WhatsApp or Facebook

  • Messenger or Skype or Signal or any number of tools

  • that communicate voice, but over software.

  • And at that point, then yes, you can authenticate.

  • COLTON OGDEN: OK, that makes sense.

  • And especially if this keeps scaling, I feel like this is an eventuality.

  • DAVID MALAN: I imagine, yeah.

  • I mean, even now, right, like I don't get calls via any of those apps--

  • well, some of them, the Facebook ones I do--

  • from people that aren't in your contacts.

  • Sometimes, it just goes to your other folder or whatnot.

  • But I'm pretty sure you can prevent calls

  • from people who weren't on your whitelist on those apps like Signal

  • and WhatsApp that do use end-to-end encryption.

  • COLTON OGDEN: Sure, yeah.

  • That makes sense.

  • That makes sense.

  • DAVID MALAN: So we shall see.

  • COLTON OGDEN: There is a--

  • so away from the, I guess, the security, which

  • has been a major theme of the podcast today,

  • towards something a little bit different actually-- and this is pretty cool

  • and I'm, particularly for me because I'm into games--

  • but Google actually announced a brand-new streaming service

  • that people are really talking about.

  • DAVID MALAN: Yeah, that's really interesting.

  • You probably know more about this world than I do, since I am a fan of the NES,

  • the original.

  • [LAUGHTER]

  • COLTON OGDEN: Well, it's called Stadia, and I've

  • done a little bit of reading on it, not terribly,

  • because there's actually not that much about it, right now.

  • DAVID MALAN: OK.

  • COLTON OGDEN: Because it just was announced maybe two or three days ago

  • and, actually, Karim, one of our team, actually kindly showed it to me

  • because I wasn't aware.

  • This is going on at a live event.

  • DAVID MALAN: OK.

  • COLTON OGDEN: But it's essentially an idea that's been done before.

  • The companies have done this sort of, we process all the games

  • and then we stream the video signal to you and you play,

  • and there's this back and forth.

  • My initial sort of qualm about this is that we're fundamentally

  • dealing with streaming latency.

  • DAVID MALAN: Yeah, of course.

  • COLTON OGDEN: And it's--

  • I find it highly unlikely that we can-- especially for a geographically distant

  • locations amongst servers and amongst consumers--

  • that we can deal with less than 13 milliseconds of latency

  • in between a given frame and the input on someone's machine.

  • DAVID MALAN: Maybe right now, but this seems inevitable.

  • So I kind of give Google credit for being a little bleeding edge here.

  • Like, this probably won't work well for many people.

  • But it feels inevitable, right?

  • Like eventually, we'll have so much bandwidth and so low

  • latency that these kinds of things seem inevitable to me, these applications.

  • So I'm kind of comfortable with it being a bit bleeding edge,

  • especially if it maybe has sort of lower quality graphics, more Nintendo

  • style than Xbox style which-- or at least with the Wii, the original Wii,

  • was like a design decision.

  • I think it could kind of work, and I'm very curious to see how well it works.

  • But, yeah, I mean, even latency, we for CS50's IDE

  • and for the sandbox tool and the lab tool

  • that to support X-based applications, which

  • is the windowing system for Linux, the graphical system,

  • it doesn't work very well for animation.

  • I mean, even you, I think, implemented Breakout for us a while ago.

  • COLTON OGDEN: A while back.

  • DAVID MALAN: And tried it out and, eh, you know, it's OK,

  • but it's not compelling.

  • But I'm sure there are games that would be compelling.

  • COLTON OGDEN: Yeah.

  • I know that in their examples, they were doing things like Assassin's Creed

  • Odyssey, you know, very recent games that are very high graphic quality.

  • I mean, I would like to--

  • I would definitely like to see at work, if possible.

  • DAVID MALAN: Yeah.

  • No, I think that would be pretty cool.

  • One less thing to buy, too, and it hopefully lowers the barrier to entry

  • to people.

  • You don't need the hardware.

  • You don't need to connect something else.

  • You don't need to draw the power for it.

  • I mean, there's some upsides here, I think.

  • COLTON OGDEN: I think especially if they are doing this at scale--

  • and Google already does this, surely-- but, you

  • know, they have a CDN network that's very--

  • and it's very, maybe a US-centric thing at first,

  • and then can scale it out to other countries.

  • Maybe the latency between any given node on their network

  • is either-- or the gap is small enough such that the latency is minimal.

  • DAVID MALAN: Yeah, hopefully.

  • COLTON OGDEN: As long as it's less than 13 milliseconds, though.

  • That's the one in 60-- one's over 60, which is the--

  • typically, the games are 60 frames per second--

  • that's the amount of time it needs to be to process input

  • and feel like it's a native game.

  • DAVID MALAN: Well, to be honest, I mean this

  • is similar to their vision for Chromebooks

  • which, if you're unfamiliar, is a relatively low-cost laptop that

  • is kind of locked down.

  • It pretty much gives you a browser, and that's it,

  • the presumption being that you can use things like Gmail and Google

  • Docs and Google Calendar, even partly offline, if you're on an airplane,

  • so long as you pre-open them in advance and sort of cache some of the code.

  • I mean, that works well so long as you have good internet.

  • But we've chatted with some of our high school students and teachers

  • whose schools use Chromebooks and it's not great

  • when the students need to or want to take the laptops home.

  • Maybe they don't have or can't afford their own internet access at home,

  • so there's certainly some downsides.

  • But I don't know.

  • I feel like within enough years, we'll be at the point

  • where internet access of some sort is more commodity like electricity

  • in the wall and so long as you have that kind of flowing into the house,

  • that it'll be even more omnipresent than it is now.

  • COLTON OGDEN: Sure.

  • Yeah that makes total sense.

  • I would definitely like to see it happen.

  • I hope it does.

  • DAVID MALAN: Yeah, so.

  • COLTON OGDEN: Well, I think that's all the topics

  • that we've sort of had lined up.

  • We covered a nice sort of breadth of them.

  • This was great.

  • I like this format.

  • DAVID MALAN: Yeah.

  • No, hopefully you're still listening because I

  • feel like we should offer a couple bits of advice here.

  • I mean, one, on the Facebook password front,

  • I mean, even I did change my password.

  • I don't know if mine was among the millions that were apparently

  • exposed in the clear, and it's not clear that any humans noticed

  • or used the password in any way, but changing your password's

  • not a bad idea.

  • And as you may recall from CS50, itself, if you've taken the class,

  • you should probably be using a password manager anyway and not

  • just picking something that's pretty easy for you to remember.

  • Better to let software do it instead.

  • And on the robocall front, I, mean there's a couple defenses here.

  • I mean, even on my phone, I block numbers once I realized, wait a minute,

  • I don't want you calling me.

  • But you can also use things like Google Voice,

  • right, where they have a feature which seems a little socially obnoxious where

  • Google will pick up the phone for you and they

  • will ask the human to say who they are.

  • Then you get, on your phone, a little preview of who it is,

  • so it's like your own personal assistant.

  • COLTON OGDEN: That's kind of interesting.

  • I actually didn't realize that was thing.

  • DAVID MALAN: It's an interesting buffer, but it's kind of obnoxious.

  • COLTON OGDEN: Yeah.

  • DAVID MALAN: Right, to have that intermediate.

  • COLTON OGDEN: You could have a whitelist, surely though, that--

  • DAVID MALAN: For sure.

  • COLTON OGDEN: Yeah.

  • DAVID MALAN: No, so for unrecognized calls,

  • but people tried rolling this out for email, though, years ago,

  • and I remember even being put off by it then.

  • If I email you for the first time, we've never met,

  • you could have an automated service bounce back

  • and say, oh, before Colton will reply to this, you need to confirm who you are

  • and click a link, or something like that.

  • And at least for me at the time-- maybe I

  • was being a little, you know, a little presumptuous-- but it just felt like,

  • ugh, this is, why is the burden being put on me to solve this problem.

  • COLTON OGDEN: Also, a little high and mighty, potentially.

  • DAVID MALAN: Yeah.

  • But, I mean, it's an interesting software solution,

  • and that's what Google Voice and there's probably

  • other services that do the same thing.

  • So you can look into things like that.

  • And as for like Stadia, I'll be curious to try this out

  • when it's more available.

  • COLTON OGDEN: Yeah, me too.

  • Me too.

  • DAVID MALAN: Yeah.

  • You know, I think it's worth noting that podcasting is how CS50 itself ended up

  • online, way back when.

  • Long story short, before CS50, as Colton knows,

  • I taught a class at Harvard's Extension School, the Continuing Ed program,

  • called Computer Science E-1, Understanding Computers

  • and the Internet.

  • And we, in 2005, I believe, started podcasting

  • that course, which initially meant just distributing MP3s,

  • which are audio files, of the lectures.

  • And then, I think one year later, when the video iPod, of all things,

  • came out, we started distributing videos in QuickTime format and Flash format,

  • probably.

  • COLTON OGDEN: Definitely MOVs.

  • DAVID MALAN: Yeah, of the course's lectures.

  • And it was really for the convenience of our own students

  • who might be commuting on a train or maybe they're on a treadmill,

  • and it was just kind of trying to make it easier for people

  • to access the course's content.

  • And, long story short, a whole bunch of other people

  • who weren't in the class online took an interest, found the material valuable.

  • And certainly, these days, there's such a proliferation

  • of educational content online.

  • But it was because that course we started

  • podcasting that when I took over CS50 in 2007,

  • it just felt natural, at that point, to make the course's

  • videos available online as well.

  • And even though we've kind of come full circle now and taken away the video

  • and replaced it just with audio, I think it really

  • allows us to focus on the conversation and the ideas

  • without really any distractions of visuals or need to rely on video.

  • So hopefully, this opens up possibilities

  • for folks to listen in, as opposed to having to be rapt

  • attention on a screen.

  • COLTON OGDEN: And what I like is this is a more current events focused talk,

  • too.

  • Yeah, we have so much other content, it's

  • nice to sort of have a discussion on the things that are relevant in the tech

  • world, or otherwise.

  • You know, it fits this format very well.

  • DAVID MALAN: Yeah, absolutely.

  • Well, so this was Episode Zero of CS50's podcast.

  • We hope you'll join us soon for Episode 1, our second podcast.

  • Thanks so much for CS50's own Colton Ogden, whose idea this has been,

  • and thank you for spearheading.

  • COLTON OGDEN: And thanks, David, for sort of leading the way here.

  • DAVID MALAN: Yeah, absolutely.

  • Talk to you all soon.

  • COLTON OGDEN: Bye-bye.

[MUSIC PLAYING]

字幕と単語

ワンタップで英和辞典検索 単語をクリックすると、意味が表示されます

B1 中級

ロボコール、Facebookパスワード - CS50 Podcast, Ep. (Robocalls, Facebook Passwords - CS50 Podcast, Ep. 0)

  • 4 0
    林宜悉 に公開 2021 年 01 月 14 日
動画の中の単語